repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
knative/docs | 539546673 | Title: Page needs improvement
Question:
username_0: ## Expected Behavior
When following the examples from [camel source page](https://knative.dev/docs/eventing/samples/apache-camel-source/) in a completely new k8s minikube environment, they work as the page says.
## Actual Behavior
Executing the example `Run a Timer CamelSource` in Minikube (knative with gloo) does not work and I cannot find any documentation that helps me to find what is happening.
Once finished that example, I expect a new pod created (due to knative serving) and this should be logging the event generated by the camel source.
When I check, the channel is ready, but the camel source does not have any event nor info in READY column.
## Steps to Reproduce the Problem
1. `minikube delete`
2. `minikube start --memory=8192 --cpus=6 \
--vm-driver=hyperkit \
--disk-size=30g \
--extra-config=apiserver.enable-admission-plugins="LimitRanger,NamespaceExists,NamespaceLifecycle,ResourceQuota,ServiceAccount,DefaultStorageClass,MutatingAdmissionWebhook"`
3. `kubectl create ns gloo-system`
4. `glooctl install knative -e -k`
5. `minikube addons enable registry`
6. `kamel install`
7. `kubectl apply -f camel.yaml`
8. `kubectl apply -f display_resource.yaml`
9. `kubectl apply -f source_timer.yaml`
## Additional Info
### Versions
minikube: v1.6.1
k8s: Client v1.17.0, Server v1.17.0
gloo: Client 1.2.1, Server 1.2.1
kamel: 1.0.0-M4
### camel.yaml
```yaml
apiVersion: v1
kind: Namespace
metadata:
labels:
contrib.eventing.knative.dev/release: "v0.11.0"
name: knative-sources
---
apiVersion: v1
kind: ServiceAccount
metadata:
labels:
contrib.eventing.knative.dev/release: "v0.11.0"
name: camel-controller-manager
namespace: knative-sources
---
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
labels:
[Truncated]
source:
flow:
from:
uri: timer:tick
parameters:
period: 2s
steps:
- set-header:
name: Content-Type
constant: text/plain
- set-body:
constant: Hello world!
sink:
ref:
apiVersion: messaging.knative.dev/v1alpha1
kind: InMemoryChannel
name: camel-test
```
Answers:
username_1: @username_0 Ive pinged the Eventing Slack channel for help with your issue. If you don't get a response here, you might try opening your issue in the knative/eventing or eventing-contrib repos (or reaching out directly on Slack too)?
username_2: @username_0 can you provide some more context on the status of your current namespace, that will help investigate the issue.
E.g.
```
kubectl get pod
kubectl get camelsource -o yaml
kubectl get integrationplatform -o yaml
kubectl get it -o yaml
kubectl get ik -o yaml
```
username_2: If there's a pod started from the timer source, it will also help having its logs.
username_0: Hi @username_2, here the additional context information you asked for.
```bash
+ kubectl get pods
NAME READY STATUS RESTARTS AGE
camel-k-cache 0/1 Completed 0 85m
camel-k-operator-5cdb648587-2s9ff 1/1 Running 0 85m
```
```bash
+ kubectl get camelsource -o yaml
apiVersion: v1
items:
- apiVersion: sources.eventing.knative.dev/v1alpha1
kind: CamelSource
metadata:
annotations:
kubectl.kubernetes.io/last-applied-configuration: |
{"apiVersion":"sources.eventing.knative.dev/v1alpha1","kind":"CamelSource","metadata":{"annotations":{},"name":"camel-timer-source","namespace":"default"},"spec":{"sink":{"ref":{"apiVersion":"messaging.knative.dev/v1alpha1","kind":"InMemoryChannel","name":"camel-test"}},"source":{"flow":{"from":{"parameters":{"period":"2s"},"steps":[{"set-header":{"constant":"text/plain","name":"Content-Type"}},{"set-body":{"constant":"Hello world!"}}],"uri":"timer:tick"}}}}}
creationTimestamp: "2019-12-19T08:57:37Z"
generation: 1
name: camel-timer-source
namespace: default
resourceVersion: "88780"
selfLink: /apis/sources.eventing.knative.dev/v1alpha1/namespaces/default/camelsources/camel-timer-source
uid: d87f72c0-364e-4eb6-945b-290d3ebff105
spec:
sink:
ref:
apiVersion: messaging.knative.dev/v1alpha1
kind: InMemoryChannel
name: camel-test
source:
flow:
from:
parameters:
period: 2s
steps:
- set-header:
constant: text/plain
name: Content-Type
- set-body:
constant: Hello world!
uri: timer:tick
kind: List
metadata:
resourceVersion: ""
selfLink: ""
```
```bash
+ kubectl get integrationplatform -o yaml
apiVersion: v1
items:
- apiVersion: camel.apache.org/v1alpha1
kind: IntegrationPlatform
metadata:
creationTimestamp: "2019-12-19T08:57:17Z"
generation: 2
labels:
[Truncated]
items: []
kind: List
metadata:
resourceVersion: ""
selfLink: ""
```
```bash
+ kubectl get ik -o yaml
apiVersion: v1
items: []
kind: List
metadata:
resourceVersion: ""
selfLink: ""
```
I don't see any log related to the timer source, only the camel-k-operator.
In advance, thanks for your help.
username_2: It's strange, there's no status section in the Camel Source. This means probably that the camel-source-controller is not able to change it.
Can you provide the logs of the controller?
```
kubectl logs camel-controller-manager-0 -n knative-sources
```
username_0: Now there is something, here the logs:
```bash
+ kubectl logs camel-controller-manager-0 -n knative-sources
2019/12/19 10:57:01 Camel K is correctly installed in the cluster.
2019/12/19 10:57:02 Registering Components.
2019/12/19 10:57:02 Setting up Controllers.
2019/12/19 10:57:02 Adding the Camel Source controller.
{"level":"panic","ts":1576753022.1587267,"logger":"fallback","caller":"addressable/addressable.go:56","msg":"Unable to fetch knative.dev/pkg/apis/duck.InformerFactory from context.","stacktrace":"knative.dev/eventing-contrib/vendor/knative.dev/pkg/client/injection/ducks/duck/v1/addressable.Get\n\t/home/prow/go/src/knative.dev/eventing-contrib/vendor/knative.dev/pkg/client/injection/ducks/duck/v1/addressable/addressable.go:56\nknative.dev/eventing-contrib/vendor/knative.dev/pkg/resolver.NewURIResolver\n\t/home/prow/go/src/knative.dev/eventing-contrib/vendor/knative.dev/pkg/resolver/addressable_resolver.go:53\nknative.dev/eventing-contrib/camel/source/pkg/reconciler.Add\n\t/home/prow/go/src/knative.dev/eventing-contrib/camel/source/pkg/reconciler/camelsource.go:80\nmain.main\n\t/home/prow/go/src/knative.dev/eventing-contrib/camel/source/cmd/controller/main.go:78\nruntime.main\n\t/usr/local/go/src/runtime/proc.go:200"}
panic: Unable to fetch knative.dev/pkg/apis/duck.InformerFactory from context.
goroutine 1 [running]:
knative.dev/eventing-contrib/vendor/go.uber.org/zap/zapcore.(*CheckedEntry).Write(0xc00021f6b0, 0x0, 0x0, 0x0)
/home/prow/go/src/knative.dev/eventing-contrib/vendor/go.uber.org/zap/zapcore/entry.go:229 +0x546
knative.dev/eventing-contrib/vendor/go.uber.org/zap.(*SugaredLogger).log(0xc000134090, 0x1557204, 0x0, 0x0, 0xc00091da98, 0x1, 0x1, 0x0, 0x0, 0x0)
/home/prow/go/src/knative.dev/eventing-contrib/vendor/go.uber.org/zap/sugar.go:234 +0xf7
knative.dev/eventing-contrib/vendor/go.uber.org/zap.(*SugaredLogger).Panic(...)
/home/prow/go/src/knative.dev/eventing-contrib/vendor/go.uber.org/zap/sugar.go:123
knative.dev/eventing-contrib/vendor/knative.dev/pkg/client/injection/ducks/duck/v1/addressable.Get(0x1a87380, 0xc0006f2f60, 0x6239b5a2c000, 0xc0002e6fc0)
/home/prow/go/src/knative.dev/eventing-contrib/vendor/knative.dev/pkg/client/injection/ducks/duck/v1/addressable/addressable.go:56 +0x150
knative.dev/eventing-contrib/vendor/knative.dev/pkg/resolver.NewURIResolver(0x1a87380, 0xc0006f2f60, 0x17db690, 0x1a87740)
/home/prow/go/src/knative.dev/eventing-contrib/vendor/knative.dev/pkg/resolver/addressable_resolver.go:53 +0xe7
knative.dev/eventing-contrib/camel/source/pkg/reconciler.Add(0x1ab43c0, 0xc000318640, 0xc00091ddf0, 0x0, 0x0)
/home/prow/go/src/knative.dev/eventing-contrib/camel/source/pkg/reconciler/camelsource.go:80 +0x559
main.main()
/home/prow/go/src/knative.dev/eventing-contrib/camel/source/cmd/controller/main.go:78 +0x57f
```
username_2: Thanks @username_0 , it looks like a problem due to a refactoring in pgk that broke it in 0.11.0.
I'm opening an issue on eventing-contrib to track.
username_3: It looks like this issue may have been resolved by a fix in another repo, so I'm closing this for now.
If that's incorrect, please feel free to re-open!
Status: Issue closed
|
yuhiisk/starter-kit | 67062970 | Title: Roadmap of starter-kit
Question:
username_0: Future plans I am thought to be as follows.
- Create a Generator of `Yeoman`.
- Additional template files of `Jade`.
- Add a custom build of `Modernizr` using gulp.
Please wait!!
Answers:
username_1: :blush:
username_2: :sushi:
username_0: The rest is Yeoman. |
beeminder/road | 660924281 | Title: Animated watermarks in tutorial stopped working
Question:
username_0: http://road.beeminder.com/tutorial
## Reseminata
1. Go to http://road.beeminder.com/tutorial
2. Wait for the part where it says "this is how much time time you have until that happens"
## Expectata
See the watermark fade in and out to bring your attention to it
## Resultata
It does that but it turns so faint you can't really see it so it looks more like it's disappearing.
Status: Issue closed
Answers:
username_0: http://beeminder.com/changelog#3478 |
jhelvy/splitKbCompare | 981942331 | Title: Install dependencies incomplete, and issue running still
Question:
username_0: runGitHub('username_1/splitKbCompare')
Downloading https://github.com/username_1/splitKbCompare/archive/master.tar.gz
Rows: 65 Columns: 22
โโ Column specification โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
Delimiter: ","
chr (5): id, name, colStagger, url_source, url_store
dbl (17): nKeysMin, nKeysMax, include, hasNumRow, hasFuncRow, numRows, rowSt...
โน Use `spec()` to retrieve the full column specification for this data.
โน Specify the column types or set `show_col_types = FALSE` to quiet this message.
Listening on http://127.0.0.1:7143
Error in utils::browseURL(appUrl) :
'browser' must be a non-empty character string
```
Answers:
username_0: FWIW, I just duplicated this effort on a fresh R install from binary downloaded at [r-project.org in Win10](https://cloud.r-project.org/bin/windows/base/R-4.1.1-win.exe). It works there. The Downloading line pops up, and then it launches the interface in browser on port 3943 with no more terminal output.
username_1: Yeah this occasionally goes down on shinyapps.io, which is free. I'm never sure why sometimes, and I'm seeing what you mean - it never seems to load. Usually I just have to re-upload the whole app and it runs again.
Good catch on the additional package dependencies. Yes, a user will need to install all of those to run it locally. I'll add that in a comment at the top of the app to run it locally, and I'll try and get to the root of why it's not loading online.
username_0: Thanks! Any idea why it's not running in Arch?
Status: Issue closed
username_1: Not sure about Arch, but likely some package dependency issue.
I just re-deployed the app and now it seems to be working on shinyapps.io again.
Thanks for catching the dependency issue! |
opengeospatial/teamengine | 271763277 | Title: NPE during execution of WFS 1.1 test suite
Question:
username_0: Stacktrace from teamengine.log:
```
SEVERE:
java.lang.NullPointerException
at com.occamlab.te.CtlEarlReporter.processResultAttributes(CtlEarlReporter.java:540)
at com.occamlab.te.CtlEarlReporter.processTestResults(CtlEarlReporter.java:415)
at com.occamlab.te.CtlEarlReporter.getSubtestResult(CtlEarlReporter.java:216)
at com.occamlab.te.CtlEarlReporter.generateEarlReport(CtlEarlReporter.java:132)
at com.occamlab.te.TECore.execute(TECore.java:354)
at com.occamlab.te.TECore.run(TECore.java:2406)
at java.lang.Thread.run(Thread.java:745)
```
The exception occurred during working on https://github.com/opengeospatial/ets-wfs11/issues/68<issue_closed>
Status: Issue closed |
kalexmills/github-vet-tests-dec2020 | 757870549 | Title: widnyana/nvltr: core/telegram.go; 25 LoC
Question:
username_0: [Click here to see the code in its original context.](https://github.com/widnyana/nvltr/blob/ae0b84e7c5079bb2f20c4a608a4b274d66a900ce/core/telegram.go#L39-L63)
<details>
<summary>Click here to show the 25 line(s) of Go which triggered the analyzer.</summary>
```go
for msg := range ch {
fmt.Printf("found msg: %s from: %d\n", msg.Text, msg.Chat.ID)
// react only to user sent messages
if msg.IsService() {
fmt.Print("msg is not service")
continue
}
// is message a bot command?
cmdname := msg.Command()
if cmdname == "" {
fmt.Print("msg is not command")
continue
}
// is the command even registered?
cmd := command.Lookup(cmdname)
if cmd == nil {
continue
}
// it is. cool, run it!
go cmd.Run(ctx, &bot, &msg)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: ae0b84e7c5079bb2f20c4a608a4b274d66a900ce |
explosion/spaCy | 245838884 | Title: Training word vectors and similarities
Question:
username_0: Just a question:
First an example:
I have a question, `What is the maximum of Email-Accounts I can create?`.
Then I have 2 questions in the FAQ:
`What is the number of E-Mail Accounts I can create?`
and
`What is the maximum of Calendar-Accounts I can create?`.
It seems that the second FAQ is more similar than the first one, but the first is the right one.
So I want to tell SpaCy that the first one is **more similar** than the second one.
In other words: I want to train the word vectors **with examples,** but how to do that?
Kind regards,
Timo
Answers:
username_1: We don't really have an example for this currently. There's an open enhancement issue now, which you can follow here: #2498 . Sorry for not replying to this sooner.
Status: Issue closed
|
freelawproject/free.law | 128989987 | Title: Figure out markdown extensions
Question:
username_0: Mentioned in [the documentation][1], but with very little context so far.
Settings from personal site are:
```python
MD_EXTENSIONS = ([
'codehilite(css_class=highlight)',
'extra',
'headerid',
'toc',
'footnotes',
])```
[1]: http://docs.getpelican.com/en/3.6.3/content.html<issue_closed>
Status: Issue closed |
Jhsmit/PyHDX | 712955275 | Title: Tensorflow release memory
Question:
username_0: This currently does not happen and probably TF fit should be in different process
See also https://github.com/tensorflow/tensorflow/issues/36465
Answers:
username_0: Maybe move to pytorch or other frameworks
username_0: moved to pytorch
Status: Issue closed
|
OpenMined/PyStatDP | 752868856 | Title: Fix benchmarkMinimal to run after refactor
Question:
username_0: ## Description
benchmarkMinimal does not run due to changes in file structur
## How to Reproduce
1. Got to the top level directory of the PyStatDP repo
2. python3 benchmarkMinimal.py
3. `Traceback (most recent call last):
File "benchmarkMinimal.py", line 33, in <module>
from pystatdp import detect_counterexample, ONE_DIFFER, ALL_DIFFER
ImportError: cannot import name 'detect_counterexample' from 'pystatdp' (/Users/patrickhunter/openmined/statdp/pystatdp/__init__.py)
`
Status: Issue closed
Answers:
username_1: This was refactored.
the correct way to import `detect_counterexample` is
```
#import pystatdp class from the module
from pystatdp import pystatdp
#class initiation
psd = pystatdp()
#detect_counterexample
psd.detect_counterexample(...)
``` |
ballerina-platform/ballerina-lang | 488018580 | Title: Ballerina staging central shows incorrect latest version for modules
Question:
username_0: **Description:**
The version shown as the latest version in modules and module templates is not the latest version (the oldest version is shown).
https://staging-central.ballerina.io/irushi/module0
**Related Issues:**
https://github.com/ballerina-platform/ballerina-lang/issues/18174
Answers:
username_1: Fixed with 1.0 release
Status: Issue closed
|
damonsk/onlinewardleymaps | 794531104 | Title: Support for De-evolving components?
Question:
username_0: I don't see it in the docs support for de-evolving components nor in the code at a glance. And using `evolve` with a position to the left is a workaround, but the arrowhead points to the right so is not ideal
Thanks! |
mysociety/fixmyblock-wordpress | 653315842 | Title: Styling for .wp-block-file
Question:
username_0: As seen here:
https://fixmyblock.staging.mysociety.org/citizen-journalism/

WordPress is including some default styles for `.wp-block-file`, but I think we could do better.
Feels like, if we link to files in the content of a page, the file should stand out as a _thing_ of its own. Perhaps with a nice distinct border, and a little icon suggesting the filetype?
<issue_closed>
Status: Issue closed |
gyng/save-in | 276482164 | Title: Save In displays the 'failed download' notification when the download panel / library is opened after moving a downlaoded file
Question:
username_0: On SaveIn 1.5.0, Nightly 59.0a1 (2017-11-22).
STR:
0) Enable SaveIn's notifications for both, successful and failed downloads.
1) Use SaveIn's context menu to save, say, an image to the download folder. Let the 'download complete notification' from SaveIn play out.
2) Move/delete/rename this downloaded file.
3) Open the download panel or the download window.
AR: SaveIn shows the 'download failed' notification.
ER: SaveIn shouldn't show this notification.
Answers:
username_1: Thanks for the report. 1.5.1 should fix this, please reopen if it doesn't seem to.
Status: Issue closed
username_1: I'll investigate it.
username_1: On SaveIn 1.5.0, Nightly 59.0a1 (2017-11-22).
STR:
0) Enable SaveIn's notifications for both successful and failed downloads.
1) Use SaveIn's context menu to save, say, an image to the download folder. Let the 'download complete notification' from SaveIn play out.
2) Move/delete/rename this downloaded file.
3) Open the download panel or the download window in the browser.
AR: SaveIn shows the 'download failed' notification.
ER: SaveIn shouldn't show this notification. |
ipython/ipython | 98368515 | Title: IOError: [Errno 2] No such file or directory: 'IPython/utils/submodule.py'
Question:
username_0: I downloaded ipython-3.0.0.zip from https://github.com/ipython/ipython/releases and ran the following in my MacBook's terminal:
MacBook-Pro-van-Tim-Eerd:~ timeerd$ python /users/timeerd/downloads/ipython-3.0.0/setup.py install
Response:
Traceback (most recent call last):
File "/users/timeerd/downloads/ipython-3.0.0/setup.py", line 56, in <module>
from setupbase import target_update
File "/Users/timeerd/Downloads/ipython-3.0.0/setupbase.py", line 530, in <module>
execfile(pjoin('IPython','utils','submodule.py'), globals())
IOError: [Errno 2] No such file or directory: 'IPython/utils/submodule.py'
I checked https://github.com/mattvonrocketstein/ipython/commit/ee35a6fe6e192630c433c8710c9db58513ee5fc1 but that doesn't seem to fix the problem. Any help?
Answers:
username_1: Installation won't work from Github tarballs, though what you're seeing is another problem. We recommend installing ipython using `pip install ipython`.
Status: Issue closed
|
par6n/ncov-19 | 575030951 | Title: Open Source Helps!
Question:
username_0: Thanks for your work to help the people in need! Your site has been added! I currently maintain the OpenSourceWuhan page, which collects all open source projects related to COVID-19, including maps, data, news, api, analysis, medical and supply information, etc. Please share to anyone who might need the information in the list, or will possibly contribute to some of those projects. You are also welcome to recommend more projects.
https://weileizeng.github.io/OpenSourceWuhan/world
Cheers!
Answers:
username_1: Please add this to readme file
Status: Issue closed
|
linuxdeepin/developer-center | 1177490120 | Title: sound effects problem
Question:
username_0: In the sound effects configuration box
the checkbox does not appear and when you drag it to the side it appears with a lot of difficulty due to the click due to the scroll bar
Status: Issue closed
Answers:
username_1: Sorry for the super late response since we were not looking into issues that were located outside the developer-center repo.
I believe this issue has been addressed in later updates. Currently, in `dde-control-center 5.5.11` the sound effect setting look like this:

Since it's resolved and we are processing a huge amount of issues that were transferred from other repos. If you still got similar issues, please consider creating a new issue with the necessary detailed information provided, so we could look into it. Thanks. |
void-linux/void-packages | 457603887 | Title: linux: missing bfq i/o scheduler
Question:
username_0: ### System
* xuname: `Void 5.1.11_1 x86_64-musl AuthenticAMD uptodate rFF`
* package: `linux4.14` and newer (only tested `5.1` but `bfq` was added in `4.12`)
### Expected behavior
`bfq` should be listed and selectable in `/sys/class/block/*/queue/scheduler`
### Actual behavior
it's not built into the kernel because the configuration option is unset
```
% cat /sys/class/block/sd[ab]/queue/scheduler
[mq-deadline] kyber none
[mq-deadline] kyber none
% zcat /proc/config.gz | grep BFQ
# CONFIG_IOSCHED_BFQ is not set
```
the default is `mq-deadline` but `bfq` is supposedly [better](https://algo.ing.unimo.it/people/paolo/disk_sched/results.php) for responsiveness under load
### Steps to reproduce the behavior
1. boot into `linux5.1`
2. look in `/sys/class/block/*/queue/scheduler`
Answers:
username_1: The scheduler is now enabled in the kernel configuration for all relevant kernel versions. I did not increase the revision of the package because the builders have been busy all day building kernels because of the SACK security vulnerability and I don't think we can afford another rebuild right now. So the change will only be effective in binary packages once the next version is released and the packages are rebuilt.
Status: Issue closed
|
CoinAlpha/hummingbot | 493251568 | Title: [BUG] Binance market fails with Discovery
Question:
username_0: **Describe the bug**
Running discovery strategy will sometimes start, and sometimes it won't. If successful, Binance market will eventually fail and does not restart.
**Steps To Reproduce**
Set up a bot and run with Discovery strategy on Binance. This is the `bn-hb_discovery` bot in `test-semiweekly.hummingbot.io`.
**Screenshots**

**Release version**
version: `dev-0.15.0`
commit: `<PASSWORD>`
date: `2019-09-13T00:38:51Z`
**Attachments**
[hummingbot_logs_2019-09-11-02-29-39.zip](https://github.com/CoinAlpha/hummingbot/files/3609579/hummingbot_logs_2019-09-11-02-29-39.zip)
[hummingbot_logs_2019-09-12-02-43-14.zip](https://github.com/CoinAlpha/hummingbot/files/3609583/hummingbot_logs_2019-09-12-02-43-14.zip)
[hummingbot_logs_2019-09-13-01-30-44.log](https://github.com/CoinAlpha/hummingbot/files/3609584/hummingbot_logs_2019-09-13-01-30-44.log)
[hummingbot_logs_2019-09-13-01-45-44.log](https://github.com/CoinAlpha/hummingbot/files/3609585/hummingbot_logs_2019-09-13-01-45-44.log)
Answers:
username_1: 418 on binance is temp ip ban. possibly from too many requests. is every bot a different ip / binance account?
username_0: @username_1 yes it's on a separate IP and Binance account. If you run the discovery strategy on Binance market you will get an error like this on many trading pairs.
```
2019-09-13 03:52:47,225 - hummingbot.market.binance.binance_api_order_book_data_source - ERROR - Error getting snapshot for WAVESBTC.
Traceback (most recent call last):
File "/hummingbot/market/binance/binance_api_order_book_data_source.py", line 143, in get_tracking_pairs
snapshot: Dict[str, Any] = await self.get_snapshot(client, trading_pair, 1000)
File "/hummingbot/market/binance/binance_api_order_book_data_source.py", line 124, in get_snapshot
raise IOError(f"Error fetching Binance market snapshot for {trading_pair}. "
OSError: Error fetching Binance market snapshot for WAVESBTC. HTTP status is 429.
```
The bot will keep trying and will eventually get the IP ban.
username_1: does this still happen on dev?
username_2: Unable to start binance-huobi on dev-0.15.0
config
```
#####################################
### Discovery strategy config ###
#####################################
# The following configuations are only required for the
# discovery strategy
# Exchange and token parameters
primary_market: binance
secondary_market: huobi
# Minimum profitability target required to place an order
# Expressed in decimals: 0.01 = 1% target profit
target_profitability: 0.0
target_amount: .inf
target_symbol_1: []
target_symbol_2: []
equivalent_tokens:
- - USDT
- USDC
- USDS
- DAI
- PAX
- TUSD
- USD
- - ETH
- WETH
- - BTC
- WBTC
# For more detailed information, see:
# https://docs.hummingbot.io/strategies/discovery/
```
Logs
```
2019-09-20 01:35:34,346 - hummingbot.notifier.telegram_notifier - INFO - Telegram is listening...
2019-09-20 01:35:36,052 - hummingbot.data_feed.coin_gecko_data_feed - INFO - Network status has changed to NetworkStatus.CONNECTED. Starting networking...
2019-09-20 01:35:36,278 - hummingbot.data_feed.coin_cap_data_feed - INFO - Network status has changed to NetworkStatus.CONNECTED. Starting networking...
2019-09-20 01:35:53,801 - hummingbot.client.hummingbot_application - ERROR - Error initializing strategy.
Traceback (most recent call last):
File "binance_market.pyx", line 194, in hummingbot.market.binance.binance_market.BinanceMarket.split_symbol
AttributeError: 'NoneType' object has no attribute 'group'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/hummingbot/strategy/discovery/start.py", line 98, in start
target_symbol_1 = process_symbol_list(market_1, target_symbol_1)
File "/hummingbot/strategy/discovery/start.py", line 91, in process_symbol_list
return filtered_trading_pair + filter_trading_pair_by_single_token(market_name, single_tokens)
File "/hummingbot/strategy/discovery/start.py", line 78, in filter_trading_pair_by_single_token
base_token, quote_token = MARKET_CLASSES[market_name].split_symbol(t)
File "binance_market.pyx", line 196, in hummingbot.market.binance.binance_market.BinanceMarket.split_symbol
ValueError: Error parsing symbol BNBBUSD: 'NoneType' object has no attribute 'group'
2019-09-20 01:37:04,071 - hummingbot.data_feed.coin_gecko_data_feed - WARNING - Coin Gecko API request failed. Unable to get prices.
```
Version
```
"branch": "development",
"commit": "<PASSWORD>",
"date": "2019-09-20T00:11:45Z",
```
username_0: @username_1 so far it's working and logged the output to a .csv file. But no arbitrage discovered yet. The error usually happens if you keep the bot running for a couple of hours. I will keep you posted.
<img width="947" alt="putty_ISdJrPeh20" src="https://user-images.githubusercontent.com/50150287/65293238-22123100-db8d-11e9-8fbe-3541ed494cc4.png">
username_0: @username_1 I'm getting the same errors with Binance.
<img width="947" alt="putty_idfQSPF3WR" src="https://user-images.githubusercontent.com/50150287/65295565-6523d280-db94-11e9-96ec-27260422882e.png">
username_3: I had the this error running the same strategy using the latest development branch as of this writing.

branch: development
Commit: 71c39c2
date: 2019-09-20T06:57:30Z
username_4: @username_3 I believe what you are seeing is a separate bug. Could you send me your log file?
username_3: @username_4 The error no longer exists after updating to
```
"branch": "development",
"commit": "<PASSWORD>",
"date": "2019-09-21T14:50:30Z",
```
username_0: This is confirmed working in version 0.15.0 release.
Status: Issue closed
|
shpasser/GaeSupport | 59185801 | Title: GAE php5.5
Question:
username_0: Hi,
I encounter a problem for the new Google App Engine runtime for the php 5.5 that supported Curl now.
Previously, in PHP 5.4, everything is running fine for the development and production as well.
After adjusted to PHP 5.5, it seems like the error logging is having some error.
First, it seems like it cannot detect our environment using the 'hostname' where i have to explicit set it as follow :
```
$env = $app->detectEnvironment(function(){
if (strpos($_SERVER['HTTP_HOST'],'.') === false) {
return 'local';
}
return 'production';
});
```
However, with this, it could detect it is in local mode now but it will have this error now.
**Error in exception handler: The stream or file "/GAE/server/app/storage/logs/laravel.log" could not be opened: failed to open stream: No such file or directory in /GAE/bookme/GAE/server/vendor/monolog/monolog/src/Monolog/Handler/StreamHandler.php:84**
If i revert it back to 5.4 and everything works fine again. I also make sure my storage folder got all the read and write access.
Any advise on this ?
Thanks.
Regards,
Nick
Answers:
username_1: I think that the issue is fully related to the environment detection.
The only supported log handler for GAE is 'syslog'. According to the posted error message the log handler used is using files, and was configured using `app/start/local.php`. When running on GAE the `production` environment is supposed to active and the log handler is configured using `app/start/production.php`, which configures log to use `syslog` handler.
I managed to run my test app on GAE without applying any changes.
I would suggest to revert the `bootstrap/start.php` to the original coding:
```php
$env = $app->detectEnvironment(array(
'local' => array('homestead'),
));
```
or something else, but using the same format i.e. `'env-name' => ['host1', 'host2', ...]`.
After that, please capture the error via GAE log and post it here for further discussion.
Best,
Ron.
username_0: Hi Ron,
Just to make sure, you did try it on the **runtime: php55** for the app.yaml right ?
This is my ```bootstrap/start.php``` which I converted back to the original files
```php
$env = $app->detectEnvironment(array(
'local' => array('Mac-mini.local'),
));
```
Some how. when i try to print out my monolog to get the current environment, it is still in production mode even though i set it to my local machine. I double check with my machine hostname to make sure it is not misspelled. Take note that it is working on php5.4 before the updates carried out by GAE.
Please advise.
Regards,
Nick
username_1: I used the following `app.yaml` file to push my app to GAE:
```yml
application: test1-app
version: 1
runtime: php55
api_version: 1
handlers:
- url: /favicon\.ico
static_files: public/favicon.ico
upload: public/favicon\.ico
- url: /packages
static_dir: public/packages
- url: /.*
script: public/index.php
skip_files:
- ^(.*/)?#.*#$
- ^(.*/)?.*~$
- ^(.*/)?.*\.py[co]$
- ^(.*/)?.*/RCS/.*$
- ^(.*/)?\.(?!env\.php).*$
- ^(.*/)?node_modules.*$
- ^(.*/)?_ide_helper\.php$
```
and I'm using `homestead` (via vagrant) for my local environment.
Anyway I would suggest you run PHP code like this one:
```php
<?php
var_dump(gethostname());
```
in order to obtain the correct hostname.
If you could send me a sample app code having the same issue I will be able to investigate.
In the mean time I would suggest to use `dd(var_name)`, while putting it in different places in the code to examine variables values. Another thing is to try creating a new Laravel 4 app, setup it for GAE and see if the issue reappears.
Best,
Ron.
username_0: Hi Ron,
I tried with the simplest project but somehow the issue is reappearing.
I have created a repositories in github as follow for the sample app code in order to debug:
https://github.com/username_0/gaelaravel55
The ```boostrap/start.php``` local files are using my PC hostname. I doublecheck again with the
PHP Code
```php
<?php
var_dump(gethostname());
```
and I am getting the same hostname.
Hence, I tried to switch between runtime 5.4 and 5.5 with the same settings and apparently, the 5.4 could detect my environment. For a further testing, I change the $env to wrong local name in the bootstrap/start.php in PHP 5.4 runtime and it could switch to production which means, the 5.4 is working as expected but not the PHP 5.5 runtime.
I have configure the route '/env' to printout the current environment.
Please help to test it out and let me know your input on this.
Thanks.
Nick
username_1: Hi Nick,
Just finished deploying your app on GAE, I had to change the app-id in `app.yaml` and GCS bucket name in `php.ini` in order to make it work, please take a look:
http://test-sql-l4.appspot.com
http://test-sql-l4.appspot.com/env
From my perspective the `production` environment works Ok in both PHP5.5. and PHP5.4.
Then I changed the `bootstrap/start.php` to include `homestead` instead of `ARMSs-Mac-mini.local` as a local environment and tested your app on `homestead` VM and also using `dev_appserver.py` and could not see any errors.
In my opinion there should be something in the workflow that you are using preventing the app from working properly or maybe something about your assumptions regarding the app. Please send me the exact set of commands, behaviours, assumptions and outputs you have in your workflow (including the deployment command, like `appcfg.py update ./`), so I could try to pinpoint the issue.
Best,
Ron.
username_0: Hi Ron,
I guess going into production is not getting me any problem as it will detect in production.
For me, the concern is at the local environment. As it can't do logging and selecting db based on environment unless I manually change DB settings.
Did you try on ur pc without using the homestead VM but use the 'hostname' instead?
My steps are quite easy. It would be great to send u the deployment command but I am just running the app folder locally using the GoogleAppEngineLauncher which I didnt use any of the deployment command at the terminal.
There is only one thing I change from PHP5.4 to PHP5.5,
```yaml
application: gaelaravelsql
version: 1
runtime: php55 --> changing this from php to php55
api_version: 1
handlers:
- url: /favicon\.ico
static_files: public/favicon.ico
upload: public/favicon\.ico
- url: /packages
static_dir: public/packages
- url: /.*
script: public/index.php
skip_files:
- ^(.*/)?#.*#$
- ^(.*/)?.*~$
- ^(.*/)?.*\.py[co]$
- ^(.*/)?.*/RCS/.*$
- ^(.*/)?\.(?!env\.php).*$
- ^(.*/)?node_modules.*$
- ^(.*/)?_ide_helper\.php$
```
Hence, I just relaunch the app using the GoogleAppEngineLauncher and use the '/env' route, and it just echo me back with production environment.
To conclude from PHP5.4 to PHP5.5, only one changes made which is the app.yaml from php to php55.
Nothing else is changed. Switching back to PHP5.4 by removing the '55' in the app.yaml and everything works fine for detecting environment. At PHP5.4, I could do logging in the ```storage/logs/laravel.log``` as well.
username_1: I downloaded the GoogleAppEngineLauncher and tried to run the app. It appears that `gethostname()` returns `localhost:8080` for php55, so that is the new hostname you have to fill in `bootstrap/start.php` in order to have the environment detection to work properly.
But still there is another big issue, on which I cannot help you, in php55 environment the local file system is read-only, meaning all the data previously written to `app/storage` directory including session data, compiled views and more is not allowed to be written there anymore.
Currently I don't know if this behaviour can be changed easily.
I suggest using Vagrant + VirtualBox + Homestead VM for local development.
Best Regards,
Ron.
username_0: Hi Ron,
Thanks for the feedback. Was busy over the weekend. I will try on using your suggestion for local development. However, for local development using Homestead VM, everything will be the same like in GAE environment ? With php54, everything is working fine in the local environment as if it is in GAE environment but sadly php55 changed the local file system to read only.
Thanks.
Regards,
Nick
username_1: Hi Nick,
I should clarify a little bit on the issue:
1. Both GAE production environments, `php45` and `php55`, have read-only file systems, meaning you can update the file system contents while deploying the app, but not while running on GAE.
2. The package detects when the app is running locally vs in production mode and sets the `storage` path appropriately. When running on GAE a default GCS bucket is used for as read-write storage.
3. It seems that in contrast with `local php45` dev server `local php55` dev server does not allow writing to the local file system. In particular it does not allow writing to `app/storage` directory, which is used by default by Laravel to store the compiled BLADE views etc.
4. You can use the `homestead` local environment to develop your app and then deploy and try it on GAE. The package is capable to apply local and production needed changes dynamically.
5. The only drawback is that not all the file operations are enabled on GCS buckets, so I suggest you consult the GAE documentation in order to understand what is really supported.
6. Some of PHP functions have to be enabled via `php.ini` to run on GAE.
7. Since PHP55 support for GAE is very new I suppose there will be more than one-two behavioural changes and bugs.
8. As for me, I do the development on my local `homestead` environment and I test on GAE. Of course this involves using different providers for mail, queue and database. The providers are switched automatically according to the local environment files and GaeSupport `knows` when not to alter them.
9. The support to run migrations on Cloud SQL when running in local environment is on the way. BTW it is already incorporated in GaeSupportL5 package.
Hope this helps,
Ron.
username_0: Hi Ron,
Thanks for the detail information. This has been useful.
Anyhow ,a quick question for u, do you have any example code to provide upload handling with laravel and gae? I notice that in GAE, if we want to upload picture or any files to the gae bucket, it is quite different.
Thanks.
Regards,
Nick
username_1: In general you could use something like proposed here:
https://cloud.google.com/appengine/docs/php/googlestorage/user_upload
But if you'd like my opinion on Laravel friendly way, I could get back to you in couple of days after some testing.
username_0: Hi Ron,
That could be great if you could come out with laravel friendly way .
In the mean while, I would try to use on the proposed method.
Thanks.
Regards,
Nick
username_1: I was very sick last week, it will take some time to respond.
username_0: Hi Ron,
You can close this one as I found a workaround for the file uploads.
Thanks for the help.
username_1: Hi Nick,
Thanks for your patience.
Actually I do have something in work about the issue, just couldn't get it right till now.
In order to use the same code in local and production environments I'll have to make
some changes to GaeSupport's Application class and create an alternative FormBuilder
with some changes. I'm going to include the feature in the next version, which will include
also some of the performance optimization features I added lately to the GaeSupportL5 package.
Sorry for the delay.
username_1: Some interesting explanation about the new GAE SDK:
https://gae-php-tips.appspot.com/2015/03/03/file-system-changes-in-app-engine-1-9-18/
I would recommend to take a look specifically at:
<pre>
To make the filesystem read/write, add the following to your php.ini file.
google_app_engine.disable_readonly_filesystem = 1
</pre>
Status: Issue closed
username_1: Since the relevance of `illuminate/html` is not that high, I decided to postpone the upload related activities, sorry, have to close this one. |
nannano-sys/furima-31163 | 745373382 | Title: ้็บ็ฐๅขใฎ็ฐๅขๅคๆฐใซbasic่ช่จผใ่จญๅฎใใ
Question:
username_0: ใณใผใใซ็ดๆฅ่ช่จผ็จใฎpassword็ญใ่จ่ฟฐใใใฎใงใฏใชใใ็ฐๅขๅคๆฐใซไปฃๅ
ฅใใฆใใณใผใไธใงใฏ็ฐๅขๅคๆฐใฎๅคใ่จ่ฟฐใใ
Answers:
username_0: ไปฅไธใฎใณใใณใใๅฎ่กใใ็ฐๅขๅคๆฐใซ่ช่จผๆ
ๅ ฑใไปฃๅ
ฅใใ
```
export BASIC_AUTH_USER='admin'
export BASIC_AUTH_PASSWORD='<PASSWORD>'
```
username_0: ่จญๅฎใใ็ฐๅขๅคๆฐใ่ชญใฟ่พผใใณใผใใซ่จ่ฟฐใๅคๆด
```
username == ENV["BASIC_AUTH_USER"] && password == ENV["BASIC_AUTH_PASSWORD"]
```
Status: Issue closed
|
stealjs/steal-bundler | 281202958 | Title: Dependencyโs assets not being bundled correctly
Question:
username_0: @rob4acre wrote up this forum post about Font Awesomeโs fonts not being bundled correctly with the steal-toolsโ `bundleAssets` option: https://forums.donejs.com/t/steal-tools-build-bundling-font-awesome-fonts/755 |
kjunelee/MetaOptNet | 568830859 | Title: About accuracy in CIFAR_FS 5-way 5-shot and how to implement MetaOptNet-SVM-trainval
Question:
username_0: I like you work very much and it give us many inspirement.
But when I use it at CIFAR_FS 5-way 5-shot and I get 83.89 ยฑ 0.51% accuracy. I think the reason maybe that I used MetaOptNet-SVM not MetaOptNet-SVM-trainval. So I want know how to implement MetaOptNet-SVM-trainval. Thank you very much.
Answers:
username_1: trainval set was generated by opening train set pickle file and val set pickle file in Python and concatenating them together.
Please email me at <EMAIL> if you wish to get python file for merging splits.
Status: Issue closed
username_1: This is the link to CIFAR_FS_train_val:
https://drive.google.com/open?id=1CI4kt7Plru6IX5Fp8dIocG1a5Yc5UcAk |
LeaVerou/awesomplete | 58848868 | Title: ๋ฒ์ผ์คํผ ์ข๋์คํผโ ใ๋ฐค ใใ์ใใ๋ท ์ปดใโ๊ธ์ ์คํผ
Question:
username_0: ๋ฒ์ผ์คํผ ์ข๋์คํผโ ใ๋ฐค ใใ์ใใ๋ท ์ปดใโ๊ธ์ ์คํผ๋ฒ์ผ์คํผ ์ข๋์คํผโ ใ๋ฐค ใใ์ใใ๋ท ์ปดใโ๊ธ์ ์คํผ๋ฒ์ผ์คํผ ์ข๋์คํผโ ใ๋ฐค ใใ์ใใ๋ท ์ปดใโ๊ธ์ ์คํผ๋ฒ์ผ์คํผ ์ข๋์คํผโ ใ๋ฐค ใใ์ใใ๋ท ์ปดใโ๊ธ์ ์คํผ๋ฒ์ผ์คํผ ์ข๋์คํผโ ใ๋ฐค ใใ์ใใ๋ท ์ปดใโ๊ธ์ ์คํผ๋ฒ์ผ์คํผ ์ข๋์คํผโ ใ๋ฐค ใใ์ใใ๋ท ์ปดใโ๊ธ์ ์คํผ๋ฒ์ผ์คํผ ์ข๋์คํผโ ใ๋ฐค ใใ์ใใ๋ท ์ปดใโ๊ธ์ ์คํผ๋ฒ์ผ์คํผ ์ข๋์คํผโ ใ๋ฐค ใใ์ใใ๋ท ์ปดใโ๊ธ์ ์คํผ๋ฒ์ผ์คํผ ์ข๋์คํผโ ใ๋ฐค ใใ์ใใ๋ท ์ปดใโ๊ธ์ ์คํผ๋ฒ์ผ์คํผ ์ข๋์คํผโ ใ๋ฐค ใใ์ใใ๋ท ์ปดใโ๊ธ์ ์คํผ |
mhils/gsoc-eval | 220185546 | Title: adding users to exports.auth
Question:
username_0: hi, once the app is deployed and there is live data on the server, and it's running, how does one add one user?
i.e. add (or delete) one username/password pair to the `exports.auth` entry in `data/config.js` .. but then how do i "push" that to the live server under docker and restart it to start accepting logins for those users?
(may be a good thing to add to the README)
Answers:
username_1: You adjust the file to your needs and then restart the app. :-)
Docker containers can be stopped and started like this:
```
docker-compose stop
docker-compose up -d
```
username_0: one quick/small suggestion is to make the links in the top right of proposals open with `target=_blank`.
fantastic, thanks!
Status: Issue closed
username_1: Fixed, thanks!
username_0: i wanted to make some style changes (e.g. usernames in comments be, say `#D70808` color) to make them more visible to the rest of the team, but app.css does not appear to be pushed with the stop + up -d. Is there are way to push that out?
Actually, that change to open in a separate tab does not appear to go through either after a pull + stop + up -d.
Sorry to bother you with these silly support questions.
username_1: Hey, no worries - a perfect opportunity to procrastinate.
I added dev instructions at https://github.com/username_1/gsoc-eval#development. Regarding the separate tab changes, see https://github.com/username_1/gsoc-eval#update. :)
username_0: thanks. that helps with local development. i have that working.
what happens when i want to deploy the changes? (assume for a moment that I have my own repo with my changes).
i tried to rebuild and did not work. ended up breaking things badly and i can't launch the server any more (all sort of errors with "device busy" and later as i tried to "down" and "stop" things with `ERROR: stat /var/lib/docker/overlay/93cbbfe0b44a06b09bbc337ccd161d9b5d16bfc1cad8a607a4c3a37e17c56e28: no such file or directory`).
assuming i get all this sorted out (new to docker itself), i want to be able to deploy changes **without** obliterating data. Or at the very least I want to be able to "extract" and back-up the comments and ratings, then later restore them before restarting the service. |
steadyfish/JustAnotherDataBlog | 31826896 | Title: R packaging all this?
Question:
username_0: Hi username_1! just bumped into your blog post and found my way here: R API for the India Open data portal looks very promising - if you would be interested to extend this into an R package, we would be happy to publicize it through the rOpenGov (see github/ropengov) which is a now emerging collection of R packages intended for open government data around the globe. We might also be able provide some support with the packaging if needed. Just drop a line if this sounds like something -
Answers:
username_1: Hey @username_0 ,
I have r-packaged this code here - https://github.com/username_1/ogdindiar.
Would appreciate your feedback.
Status: Issue closed
username_0: Great ! Let us continue this in the issue tracker of the package. |
stone-payments/ansible-rabbitmq | 532764208 | Title: Set erlang cookie regardless of cluster
Question:
username_0: In some cases, you want to set a specific Erlang cookie even if there's no cluster to be configured.
I suggest either moving or duplicating the task of setting the cookie file.
Answers:
username_1: @username_0, Hi Willian, could you bring some example of that situation, please?
username_1: @username_0! Version 3.0.0 has been changed to meet your request.
Status: Issue closed
|
fetchai/agents-aea | 613400057 | Title: Clean up usage of `os.chdir` in `tests/test_cli`
Question:
username_0: **Is your feature request related to a problem? Please describe.**
`os.chdir` is perhaps used improperly in `tests/test_cli`
**Describe the solution you'd like**
Cleanup the usage to ensure current working directory remains the same after each test as it was before.<issue_closed>
Status: Issue closed |
denoland/rusty_v8 | 687714236 | Title: Peculiar build error from func types
Question:
username_0: This code causes a strange build error:
```rs
let func: fn(&mut v8::HandleScope, v8::FunctionCallbackArguments, v8::ReturnValue) = someFunction;
v8::Function::new(scope, func).unwrap();
```
```
error[E0080]: could not evaluate constant
--> /home/snek/.cargo/registry/src/github.com-1ecc6299db9ec823/rusty_v8-0.9.0/src/support.rs:490:5
|
490 | [s][size_must_be_0]
| ^^^^^^^^^^^^^^^^^^^ index out of bounds: the len is 1 but the index is 8
error[E0080]: could not evaluate constant
--> /home/snek/.cargo/registry/src/github.com-1ecc6299db9ec823/rusty_v8-0.9.0/src/support.rs:484:22
|
484 | const SELF: Self = Self::new_checked();
| ^^^^^^^^^^^^^^^^^^^ referenced constant has errors
error: aborting due to 2 previous errors; 1 warning emitted
```
In practice the usage looks like this:
```rs
static BINDINGS: &[
(&str, fn(&mut v8::HandleScope, v8::FunctionCallbackArguments, v8::ReturnValue))
] = &[
("something", something),
("somethingElse", something_else),
];
for (name, f) in BINDINGS {
let f = v8::Function::new(scope, f).unwrap();
// ...
}
```
Status: Issue closed
Answers:
username_1: Function callbacks must have the 'function item' type and be zero-sized. That means you must provide either (1) the name of a function or (2) a closure that does not capture variables.
You're passing them as function pointers - that doesn't work.
There is also no way for us to make it work unfortunately, so I'll close this issue.
username_0: @username_1 what type annotation am I supposed to use? |
square/retrofit | 147659663 | Title: retrofit unable to create converter for byte[]?
Question:
username_0: - [ ] Question.
I use retrofit in my project.i wanna take a file from my server,that is a verification code.At first,everything is ok,but when i open the project again,it has a error.and i rebuild my project times but it doesn't fix.
Status: Issue closed
Answers:
username_1: Built-in support for a `byte[]` is done by specifying `ResponseBody` as your type and calling `body.source().readByteArray()`. If you want to return `byte[]` directly then you'll need to make your own converter which does that for you. |
FredHutch/Oncoscape | 112523462 | Title: Data package selective cloning feature for independent Data Package repository
Question:
username_0: Search for possible solution to achieve this goal
Answers:
username_1: problem: datapackages are large and not all users will want them all
# Options
## solution 1: sparse checkout
http://stackoverflow.com/questions/600079/is-there-any-way-to-clone-a-git-repositorys-sub-directory-only
## solution 2: submodules
http://git-scm.com/book/en/v2/Git-Tools-Submodules
username_0: ## solution 3: Git Large File Storage:
https://git-lfs.github.com
username_0: [experiment repository and README.md](https://github.com/username_0/submodules)
username_1: ## Separate out data packages into sparse submodule
http://stackoverflow.com/questions/359424/detach-subdirectory-into-separate-git-repository
https://git-scm.com/docs/git-submodule
-- create submodule on FredHutch/Oncoscape within /dataPackages that links to
new repo on FredHutch/Oncoscape-dataPackages
**Prepare old repo**
```
pushd <big-repo>
git subtree split -P <name-of-folder> -b <name-of-new-branch>
popd
```
**Create new repo**
```
mkdir <new-repo>
pushd <new-repo>
git init
git pull </path/to/big-repo> <name-of-new-branch>
```
**Link new repo to GitHub**
```
git remote add origin <<EMAIL>:my-user/new-repo.git>
git push origin -u master
```
**Cleanup**
```
popd # get out of <new-repo>
pushd <big-repo>
git rm -rf <name-of-folder>
```
**Add 'new-repo' as a git submodule to 'big-repo' **
```
git submodule add <EMAIL>:my-user/new-repo.git new-repo
```
**setup sparse checkout**
http://stackoverflow.com/questions/600079/is-there-any-way-to-clone-a-git-repositorys-sub-directory-only/13738951#13738951
```
cd dataPackages/TCGA
git config core.sparsecheckout true
echo DEMOdz > ../.git/modules/TCGA/info/sparse-checkout
echo TCGAbrain >> ../.git/modules/TCGA/info/sparse-checkout
...
```
### Default Oncoscape Clone
-- includes SttrDataPackage, PatientHistory, DGI, reference table
-- downloads DEMOdz
-- Has empty dataPackages/TCGA/ submodule
### Download TCGA data
-- edit sparse-checkout file with additional folders
username_1: @username_0 will you test the ability to use sparse-checkout of a submodule using the above code?
username_0: @username_1 Happy to do it! |
JonasBausch/APCSA | 543006258 | Title: Add default "notFound" image
Question:
username_0: When no thumbnail/image is present for a particular article, the height of the menubar on articleDetailView is unnecessarily big because the browser automatically uses a default not found image. We could either introduce a default image for such cases and store it permanently in the db or style it in a way that it doesn't increase the height of the menubar.<issue_closed>
Status: Issue closed |
dotnet/roslyn-analyzers | 324728635 | Title: Consider adding an analyzer to warn on await used in loops
Question:
username_0: `await` used in a loop may not be taking full advantage of parallelism if the operations can be run simultaneously; `Task.WhenAll` could be used instead. ESLint has such a rule here: https://eslint.org/docs/rules/no-await-in-loop
Status: Issue closed
Answers:
username_1: Async usage analyzers not implemented in this repo. |
LaunchMenu/LaunchMenu | 746834512 | Title: Increase SearchExecuter parallelism
Question:
username_0: Currently the search executer only handles 1 item at a time. This is problematic if a search requires some async behavior s;uch as getting data from a server. Code should be adapted to continue onto the next item already when this is the case.
I believe this should be doable without many changes.
Unit tests should be updated to properly test this parallelism.<issue_closed>
Status: Issue closed |
cerner/terra-framework | 522255823 | Title: [terra-tabs] Uplift to the controlled responsive element
Question:
username_0: # Feature Request
Investigate uplifting terra-tabs onto the controlled responsive element or alternative.
## Description
<!-- A clear and concise description of what the feature is. -->
The uncontrolled responsive element has been [flagged for deprecation](https://github.com/cerner/terra-core/pull/2729). The terra-tabs should either be uplifting onto the controlled responsive element implementation or alternatively the responsive element should be removed from the component.<issue_closed>
Status: Issue closed |
bmild/nerf | 723357601 | Title: Generating .json for Blender objects
Question:
username_0: Hello,
I was wondering how can I make the .json files and generate the poses of an Blender object, like the lego one. I have an object like *.blend that I have worked on and now I want to make a dataset from it.
Thank you and stay safe!
Answers:
username_1: You can use [this](https://github.com/bmild/nerf/files/4410324/360_view.py.zip) script as a starting point. You may need to make modifications for your specific blend file.
username_2: Alternatively, I discovered [BlenderProc](https://github.com/DLR-RM/BlenderProc) that allows you to generate images with set camera parameters as well. It requires more time to learn how to use it though.
username_3: Is this script runnable on Blender 3.x? I am using the MacOS system with an M1 chip, the script was just stuck:
 |
leanpub/translations | 105909993 | Title: de.yml not used?
Question:
username_0: The [current de.yml](https://github.com/leanpub/translations/blob/master/de.yml#L251) contains the correct text `... unterstรผtzt den Autor ...`.
Nevertheless, the generated book does not:

(`... unterstรผtzt de Autor ...`)
Is this still taken from [here](https://github.com/username_1/leanpub_localization/blob/master/de.yml#L189)?
Status: Issue closed
Answers:
username_1: We have corrected this, thanks. |
parse-community/parse-server | 303737753 | Title: Suggestion: Make A Blueprint of Architecture of Parse-Server
Question:
username_0: parse-server is a very large project now. It might help to have a blueprint of architecture of parse-server in CONTRIBUTE.md for the new contributor.
Suggestion Context
1. Concept of Architecture
2. Detail of Architecture. something like Adapter interface or module description.
3. Roadmap
Answers:
username_1: There's the [Development Guide](http://docs.parseplatform.org/parse-server/guide/#development-guide) in the Wiki. It could be overwhelming to new contributors.
We could have some kind of visual graph like in the [PHP Class Diagram](http://parseplatform.org/parse-php-sdk/graph_class.html) or [Developer Roadmap](https://github.com/kamranahmedse/developer-roadmap)
I usually follow this generic flow especially for queries. We can definitely expand on this for the detail of the architecture. We could leverage the test specs as they separated.
```
middleware -> router -> rest -> controller -> storageAdapter
```
As far as a Roadmap, sadly there isn't one. A previous discussion was opened [Roadmap](https://github.com/parse-community/parse-server/issues/3909)
I do have my own personal roadmap.
* Unique Indexes
* GeoPoint 2dsphere Indexes
* Add indexes to Parse Dashboard
* Ram Usage / MongoStats to Parse Dashboard
* LeftJoin on Pointer and Relation
* Sample (Random Object) Query
* MySql Adapter update
* Neo4J Adapter
* Adapter Test Spec
* New Adapter Guide
* Let's Encrypt
* Field Validation / Private Fields via Schema
* .NET Live Queries
username_2: @username_1, what a wonderful roadmap! Why don't you open the issues now around those feature? Maybe some +1's will let other people wanna help you or move things faster if we see a common urge.
username_1: @username_2 Most them already have issues opened. ๐
private fields via schema api https://github.com/parse-community/parse-server/issues/1547#issue-149363991
field validation via schema api https://github.com/parse-community/parse-server/issues/1547#issue-149363991
More database adapters https://github.com/parse-community/parse-server/issues/20
Live Query Subscribe to specific events https://github.com/parse-community/parse-server/issues/3841
LetsEncrypt https://github.com/parse-community/parse-server/issues/33
Randomly get objects from mongo https://github.com/parse-community/parse-server/issues/818
[.NET Live Queries](https://github.com/parse-community/Parse-SDK-dotNET/issues/199)
username_1: @username_2 If your interested I went through the opened and closed issues for all the repositories at one point in time and made a list of potential features. A lot of features were closed during the migration from parse.com because there wasn't time. |
libgdx/packr | 194648411 | Title: Exception running packr
Question:
username_0: Hi,
I'm getting the following exception when trying to use the latest version of packr (downloaded in binary form from the readme in this repo).
Nicks-MacBook-Pro:packr-test nick$ java -jar packr.jar --platform mac --jdk /Library/Java/JavaVirtualMachines/jdk1.8.0_111.jdk/Contents/Home --executable msv --classpath viewserver-mosaic.jar --mainclass com.mosaicsmartdata.msv.viewserver.sellside.demo.Main --vmargs Xmx1G --resources config.xml license.lic log4j2.xml viewserver-sellside-demo.h2.db --minimizejre soft --output out --icon GenericApp.icns --bundle com.mosaicsmartdata.msv.viewserver
Cleaning output directory '/Users/nick/Temp/packr-test/out' ...
Copying executable ...
Copying classpath(s) ...
Unpacking JRE ...
Copying resources ...
Minimizing JRE ...
Removing foreign platform libs ...
Exception in thread "main" org.zeroturnaround.zip.ZipException: java.io.IOException: File out/Contents/Resources/viewserver-mosaic.jar.tmp/license exists and is not a directory. Unable to create directory.
at org.zeroturnaround.zip.ZipUtil.rethrow(ZipUtil.java:2556)
at org.zeroturnaround.zip.ZipUtil.iterate(ZipUtil.java:379)
at org.zeroturnaround.zip.ZipUtil.unpack(ZipUtil.java:704)
at org.zeroturnaround.zip.ZipUtil.unpack(ZipUtil.java:689)
at com.badlogicgames.packr.PackrReduce.removePlatformLibs(PackrReduce.java:204)
at com.badlogicgames.packr.Packr.pack(Packr.java:65)
at com.badlogicgames.packr.Packr.main(Packr.java:308)
Caused by: java.io.IOException: File out/Contents/Resources/viewserver-mosaic.jar.tmp/license exists and is not a directory. Unable to create directory.
at org.apache.commons.io.FileUtils.forceMkdir(FileUtils.java:1482)
at org.zeroturnaround.zip.ZipUtil$Unpacker.process(ZipUtil.java:822)
at org.zeroturnaround.zip.ZipUtil.iterate(ZipUtil.java:368)
... 5 more
I'd be very grateful if you could help me get past this.
Thanks,
Nick
Answers:
username_1: Can you check if your .jar file contains both a "license" file and a "license" folder in the root directory? Just trying to understand what the ZipUtil class is complaining about.
username_2: Please try the latest version of packr and open a new issue if you run into any problems.
Status: Issue closed
|
Frege/try-frege | 48173116 | Title: Add newline after the REPL output
Question:
username_0: function g :: ((forall b.[b] -> [b]) -> Int) -> Int
frege> k1 (f :: forall a. a -> a) = 42
function k1 :: (forall a.a -> a) -> Int
frege> :t g k1
7: type error in expression k1
type is [t23262]
used as a
7: The inferred type must be at least as
polymorphic as the annotated one.
7: type `[t23262]` is not as polymorphic as suggested in the
annotation where just `a` is announced.
7: type error in expression k1
type is a
used as [t23262]
7: The inferred type must be at least as
polymorphic as the annotated one.
7: type `[t23262]` is not as polymorphic as suggested in the
annotation where just `a` is announced.
frege>
```<issue_closed>
Status: Issue closed |
MicrosoftDocs/azure-docs | 301586755 | Title: ARM Template section
Question:
username_0: Can this article be augmented to include non-GUI methods of enabling a backup during VM creation, such as through ARM templates
---
#### Document Details
โ *Do not edit this section. It is required for docs.microsoft.com โ GitHub issue linking.*
* ID: f7102948-dcb5-bc73-3a61-a4764ae6a77d
* Version Independent ID: 4a234f8c-1f05-ab83-a46c-add8b701346f
* [Content](https://docs.microsoft.com/en-us/azure/backup/backup-during-vm-creation)
* [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/backup/backup-during-vm-creation.md)
* Service: backup, virtual-machines
Answers:
username_1: @username_0 Thank you for the valuable feedback. I assigned case to the author.
username_2: #in-progress
username_2: @username_0, Thank you again for your excellent suggestion. I have added a work task for augmenting this article with content describing how to add backup protection to VMs deployed using an ARM template.
Status: Issue closed
|
c1rrus/gulp-svg-fill | 447950918 | Title: Specifying multiple colors does not generate multiple files
Question:
username_0: ```
gulp.task('svg', function() {
return gulp.src(['src/images/*.svg'])
.pipe(svg_fill({
colors: {
black: '#000000',
white: '#ffffff',
red: '#ff0000'
}
}))
.pipe(svg2png({
width: 72,
height: 72
}))
.pipe(gulp.dest('./extension/images'));
});
```
Answers:
username_1: Thanks for raising this. I'm not able to reproduce it though. What version of Gulp are you using?
Status: Issue closed
username_1: Closing this issue as I cannot reproduce it. |
infinum/eightshift-frontend-libs | 1097684383 | Title: Drawer A11y issues
Question:
username_0: ### Describe your bug
Currently, eightshift-frontend-libs provides a drawer component which uses a `transform: translate3d(100%, 0, 0);` rule (with `display: block;`) to hide itself and animate opening/closing. It also uses a page overlay component which renders a `.page-overlay` div to darken the rest of the page (and allow the drawer to be closed by clicking on the page itself)
However, this approach introduces some A11y issues:
- [ ] As the drawer has `display: block`, its elements are part of the tab order, even in cases where a drawer trigger isn't rendered (such as on desktop) or drawer is closed - this creates confusion for keyboard users
- [ ] Clickable elements must be focusable and should have interactive semantics - https://developer.mozilla.org/en-US/docs/Web/Accessibility/Understanding_WCAG/Keyboard
Generally speaking, keyboard control could be improved here (e.g., supporting `esc` to close the drawer)
### Steps to Reproduce
1. Set up a menu on the default eightshift-boilerplate.
### Expected behavior
Drawer elements aren't part of the tab order when the drawer is closed. Drawer is fully keyboard accessible.
### Screenshots, screen recording, code snippet
<img width="1037" alt="Screenshot 2022-01-10 at 10 14 42" src="https://user-images.githubusercontent.com/1742806/148748312-506f8f1e-155a-47ba-a336-3d2c2035ef35.png">
<img width="703" alt="Screenshot 2022-01-10 at 10 15 14" src="https://user-images.githubusercontent.com/1742806/148748321-2c0df3d7-2539-40eb-a697-85844c7a399a.png">
### Environment info
_No response_
### Please confirm that you have searched existing issues in this repo.
Yes
### Please confirm that you have searched in our documentation and couldn't find the answer.
Yes
### Please confirm that your bug occurs with all plugins uninstalled and with the default WordPress theme active.
No<issue_closed>
Status: Issue closed |
JCumin/Brachylog | 223966748 | Title: Suggestion for random labeling
Question:
username_0: There is an (undocumented) labeling option called `random_value(+Seed)`, and you can use it to produce more or less *random* solutions while still retaining completeness.
For example:
<pre>
<b>?- X in 0..5, labeling([random_value(0)], [X]).</b>
X = 3 ;
X = 0 ;
X = 2 ;
X = 1 ;
X = 4 ;
X = 5.
</pre>
Note that the use of `Seed` makes this completely reproducible, and act like a true relation.
I say "more or less random" because the solutions are not produced in such a way that each of them is equally likely. Rather, only the choice of value for each individual variable is affected.
---
One possible use case:
https://codegolf.stackexchange.com/a/117529/18434
Answers:
username_1: This is pretty slow when the domain is big, and even overflows the stack when it's too big (e.g. find a random 8-digits number). I suppose it's not easily fixable?
username_1: Implemented as of [this commit](https://github.com/username_1/Brachylog/commit/7c4ac494f00eddf9fe108ef12f0dd1cd67e0ea56).
The predicate symbol is `โโ`. This only works properly on finite domains, obviously.
```
?- run_from_atom('~lโโ โโ',2,Z).
Z = 23 ;
Z = 48 ;
Z = 60 ;
Z = 80 ;
โฆ
```
username_0: It seems sufficiently efficient now, for example:
<pre>
?- time((<b>U #= 10^40</b>,
X in 0..U,
labeling([random_value(0)], [X]))).
<b>% 169 inferences, 0.000 CPU in 0.000 seconds (99% CPU, 344898 Lips)</b>
U = 10000000000000000000000000000000000000000,
X = <b>2474137494053817230546329383082252009579</b> ;
<b>% 91 inferences, 0.000 CPU in 0.000 seconds (91% CPU, 1338235 Lips)</b>
U = 10000000000000000000000000000000000000000,
X = <b>2410725425306837646409597099865362473338</b> ;
<b>% 95 inferences, 0.000 CPU in 0.000 seconds (91% CPU, 1319444 Lips)</b>
U = 10000000000000000000000000000000000000000,
X = <b>1920885495182001178851451470251780102139</b> ;
<b>% 103 inferences, 0.000 CPU in 0.000 seconds (90% CPU, 1450704 Lips)</b>
U = 10000000000000000000000000000000000000000,
X = <b>2436945892123923177527233981283594389902</b> .
</pre>
It is such use cases that help me to improve the constraint solver, so please let me know if you find any case that needs to become more efficient. Thank you!
username_1: @username_0 Thanks, great improvement!
On that PPCG challenge, it now works well but seems to have pretty heavy variance on the execution time. I think this is caused by the "all digits are different" constraint though, because when I remove it it's pretty much instant:
```
?- time(run_from_atom('~lโโ โโ',8,Z)).
% 11,155,871 inferences, 0.967 CPU in 0.984 seconds (98% CPU, 11534119 Lips)
Z = 92318046 .
?- time(run_from_atom('~lโโ โโ',8,Z)).
% 454,873 inferences, 0.062 CPU in 0.070 seconds (89% CPU, 7289585 Lips)
Z = 80642917 .
?- time(run_from_atom('~lโโ โโ',8,Z)).
% 6,884,540 inferences, 0.577 CPU in 0.583 seconds (99% CPU, 11927401 Lips)
Z = 10284359 .
```
Without `โ `:
```
?- time(run_from_atom('~lโโโ',8,Z)).
% 115,501 inferences, 0.031 CPU in 0.040 seconds (78% CPU, 3701931 Lips)
Z = 45366017 ;
% 10,441 inferences, 0.016 CPU in 0.004 seconds (390% CPU, 669291 Lips)
Z = 16670246 ;
% 13,923 inferences, 0.000 CPU in 0.002 seconds (0% CPU, Infinite Lips)
Z = 18621401
```
username_0: I think one of the reasons here is that you are labeling the number as a whole instead of the list of digits. For example, I get much more uniform timings (but also less randomness on subsequent solutions) with the following patch:
<pre>
diff --git a/src/predicates.pl b/src/predicates.pl
index dd7e3d8..e423ed8 100644
--- a/src/predicates.pl
+++ b/src/predicates.pl
@@ -679,12 +679,13 @@ brachylog_different('default', [], []).
brachylog_different('default', [H|T], [H|T]) :-
maplist(prepend_integer, L, [H|T]),
all_distinct(L).
-brachylog_different('default', 'integer':I, 'integer':I) :-
+brachylog_different('default', 'integer':I, Out) :-
( integer_value('integer':_:[_], I) ->
- true
+ Out = integer:I
; H #\= 0,
integer_value('integer':_:[H,H2|T], I),
- all_distinct([H,H2|T])
+ all_distinct([H,H2|T]),
+ Out = [H,H2|T]
).
</pre>
Please note that I have not looked into this patch in detail, so please carefully review it before applying.
An alternative way out may be to strengthen or more efficiently implement the linking constraints between the whole number and its digits, so that the search space can be pruned more effectively.
Status: Issue closed
|
dotnet/runtime | 558377788 | Title: Make TypedReference useful
Question:
username_0: ### Rationale
`TypedReference` is currently a second class citizen because using it equates to using undocumented C# keywords. We could make it more useful by adding methods to it that let normal code work with it.
We would be adding APIs that replace (and enhance) undocumented C# keyword `__refvalue` and `__makref`.
### Proposed API
```csharp
namespace System
{
public struct TypedReference
{
public ref T AsRef<T>()
{
// Essentially "return ref __refvalue(this, T);" if C# allowed that.
}
}
}
```
Stretch goal:
```csharp
namespace System
{
public struct TypedReference
{
// This won't compile with C# because C# disallows returning a TypedReference
// We need a Roslyn feature to make TypedReference returnable like other byref-like types
public static TypedReference FromRef<T>(ref T value) { }
}
}
```
Answers:
username_1: Currently there is [active design work in 6.0 around fast invoke](https://github.com/dotnet/runtime/issues/45152), which may use `TypedReference` in the manner above (e.g. "AsRef", "FromRef") provided the current compiler restrictions around passing it can be removed (causing CS1601 or CS1955).
@username_0 are you aware of any reasons why the Roslyn compiler needs to continue to prevent `TypedReference` from being passed to other methods (either byref or byval). I see you [changed it to be a `ref struct` ](https://github.com/dotnet/runtime/pull/2216.
) a year ago which I hope addresses potential misuse, but I'm not sure about any other intrinsic dependencies or potential GC issues.
username_2: @username_1 `TypedReference` has the same problem as the internal `ByReference<T>` type. Both of them are structs with `ref` fields. The current C# lifetime rules for ref-structs are not able to model `ref` fields safely.
https://github.com/dotnet/csharplang/blob/master/proposals/low-level-struct-improvements.md talks about the problem and the proposed solution. |
mikec/kalamata | 210097195 | Title: Bug: Hooks only work when model name is the same as endpoint name
Question:
username_0: https://github.com/username_1/kalamata/commit/bdd45d6a6c92492afed5e3b3d69685fd05e65a20
Answers:
username_0: @username_1 I know this was a while back, but can you advise on the above? If there is a bug I would be happy to PR the fix, otherwise I think I will have to run off a forked version with lines 52 and 53 removed ๐
username_0: Here is an example where hooks are broken (they just don't run):
```
this.kalamata.expose(require('./some_bookshelf_model.js'), {
modelName: 'WorkingHour',
collectionName: 'WorkingHours',
endpointName: 'working_hours'
});
```
And one where it works properly:
```
this.kalamata.expose(require('./other_bookshelf_model.js'), {
modelName: 'User',
collectionName: 'Users',
endpointName: 'users'
});
```
In both situations the hooks are created (and are visible on the `kalamata` object), but they only run for ones like the second example.
To confuse things further, when I removed those lines as mentioned above at my fork https://github.com/username_0/kalamata, the first example started working with hooks but the second example stopped working. So confusing!
username_1: @username_0 thanks for finding this, that's a pretty nasty bug!
I think this will fix it https://github.com/username_1/kalamata/pull/7
Try that out for a bit, if it's working for you then I'll merge it in!
Wish that I had added better unit testing and more modularization in this code base. Having everything in index.js is a little ridiculous!
username_0: Hey @username_1, thanks for getting back to me! Really appreciate the maintenance. I'm trying your branch today and will let you know how it works ๐
username_0: Update: works well! Thanks for the help ๐ @username_1
username_1: @username_0 awesome! ๐ |
eslint/eslint | 146108217 | Title: Global vs Local version mismatch
Question:
username_0: This would be a good "workaround" for #3993
Answers:
username_1: May consider doing the same for plugins as well
username_2: How would you propose to do this, exactly?
username_0: This is how gulp handles it. https://github.com/gulpjs/gulp/blob/098e1825469ecb0b72c74e81db600678b625fea0/bin/gulp.js#L99
If this seems worthwhile, I don't mind looking into and creating a PR. Didn't want to work on it if it was out of scope.
username_3: I'm generally not a fan of dirtying the output of ESLint, especially when people rely on the output in CLI and other tools, as it's hard to know what people might be depending on.
Status: Issue closed
username_4: Thanks for your interest in improving eslint. Unfortunately, it looks like consensus couldn't be reached on this issue and so I'm closing it. While we wish we'd be able to accommodate everyone's requests, we do need to prioritize. We've found that issues failing to reach consensus after 21 days tend never to reach consensus, and as such, we close those issues. This doesn't mean the idea isn't interesting, just that it's not something the team can commit to. |
pythonindia/junction | 74245421 | Title: Ability to quote comments in review replies
Question:
username_0: Was trying to comment to a comment by the proposer when I noticed junction doesn't quote replies like email clients do.
This is forcing me to copy the sections and "quote" them manually one by one. Would be nice to have automatic quoting of comments I am replying to.
Didn't check for similar issues so feel free to close this as duplicate if already filed.
Answers:
username_1: This came up in initial conversation, we wanted to do it later.
@vigneshsarma can you take care of this ? |
Seagate/cortx-motr | 859019736 | Title: sudo ./scripts/install-build-deps completes with failures=1
Question:
username_0: When I try to build Motr on a single VM using CentOS 7.8.2003 when I run: `sudo ./scripts/install-build-deps` it completes but with failed=1 and then when I try to run `scripts/m0 rebuild` or `scripts/m0 make` this does not work which I presume is expected when the build deps command fails.
Answers:
username_1: @username_2 @madhavemuri, please address the query as raised?
username_2: Please attach detailed error messages here.
username_0: `Here is the error which causes the failure for the`
Here is the error when I run the command `sudo ./scripts/install-build-deps` that causes the failure:
```
fatal: [localhost]: FAILED! => {
"changed": true,
"cmd": [
"rpmbuild",
"-bb",
"--without",
"servers",
"--without",
"lustre-tests",
"--define",
"configure_args --disable-gss-keyring",
"~/rpmbuild/SPECS/lustre.spec"
],
"delta": "0:00:07.452584",
"end": "2021-04-19 09:11:53.886878",
"invocation": {
"module_args": {
"_raw_params": "rpmbuild -bb --without servers --without lustre-tests --define 'configure_args --disable-gss-keyring' ~/rpmbuild/SPECS/lustre.spec\n",
"_uses_shell": false,
"argv": null,
"chdir": null,
"creates": "/root/rpmbuild/RPMS/x86_64/lustre-client-devel-*.rpm",
"executable": null,
"removes": null,
"stdin": null,
"stdin_add_newline": true,
"strip_empty_ends": true,
"warn": true
}
},
"msg": "non-zero return code",
"rc": 1,
"start": "2021-04-19 09:11:46.434294",
"stderr": "+ umask 022\n+ cd /root/rpmbuild/BUILD\n+ cd /root/rpmbuild/BUILD\n+ rm -rf lustre-2.12.5\n+ /usr/bin/gzip -dc /root/rpmbuild/SOURCES/lustre-2.12.5.tar.gz\n+ /usr/bin/tar -xf -\n+ STATUS=0\n+ '[' 0 -ne 0 ']'\n+ cd lustre-2.12.5\n+ /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w .\n+ ln lustre/ChangeLog ChangeLog-lustre\n+ ln lnet/ChangeLog ChangeLog-lnet\n+ exit 0\n+ umask 022\n+ cd /root/rpmbuild/BUILD\n+ cd lustre-2.12.5\n+ cd /root/rpmbuild/BUILD/lustre-2.12.5\n+ export 'UTILS_CFLAGS=-O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector-strong --param=ssp-buffer-size=4 -grecord-gcc-switches '\n+ UTILS_CFLAGS='-O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector-strong --param=ssp-buffer-size=4 -grecord-gcc-switches '\n+ export 'UTILS_LDFLAGS=-Wl,-z,relro '\n+ UTILS_LDFLAGS='-Wl,-z,relro '\n+ CONFIGURE_ARGS=--disable-gss-keyring\n+ '[' -n --disable-gss-keyring ']'\n++ echo --disable-gss-keyring\n++ sed -e 's/\"\\?--with-linux=[^ ][^ ]* \\?//'\n+ CONFIGURE_ARGS=--disable-gss-keyring\n++ echo --disable-gss-keyring\n++ sed -e 's/\"\\?--with-linux-obj=[^ ][^ ]* \\?//'\n+ CONFIGURE_ARGS=--disable-gss-keyring\n++ echo --disable-gss-keyring\n++ sed -e 's/\"\\?--with-kmp-moddir=[^ ][^ ]* \\?//'\n+ CONFIGURE_ARGS=--disable-gss-keyring\n+ CFLAGS='-g -O2 -Werror'\n+ export CFLAGS\n+ CXXFLAGS='-g -O2 -Werror'\n+ export CXXFLAGS\n+ FFLAGS='-g -O2 -Werror -I/usr/lib64/gfortran/modules'\n+ export FFLAGS\n+ FCFLAGS='-g -O2 -Werror -I/usr/lib64/gfortran/modules'\n+ export FCFLAGS\n+ LDFLAGS='-Wl,-z,relro '\n+ export LDFLAGS\n+ '[' 1 == 1 ']'\n+ '[' x86_64 == ppc64le ']'\n++ find . -name config.guess -o -name config.sub\n+ for i in '$(find . -name config.guess -o -name config.sub)'\n++ basename ./config/config.guess\n+ '[' -f /usr/lib/rpm/redhat/config.guess ']'\n+ /usr/bin/rm -f ./config/config.guess\n++ basename ./config/config.guess\n+ /usr/bin/cp -fv /usr/lib/rpm/redhat/config.guess ./config/config.guess\n+ for i in '$(find . -name config.guess -o -name config.sub)'\n++ basename ./config/config.sub\n+ '[' -f /usr/lib/rpm/redhat/config.sub ']'\n+ /usr/bin/rm -f ./config/config.sub\n++ basename ./config/config.sub\n+ /usr/bin/cp -fv /usr/lib/rpm/redhat/config.sub ./config/config.sub\n+ eval ./configure --program-prefix= --disable-dependency-tracking --prefix=/usr --exec-prefix=/usr --bindir=/usr/bin --sbindir=/usr/sbin --sysconfdir=/etc --datadir=/usr/share --includedir=/usr/include --libdir=/usr/lib64 --libexecdir=/usr/libexec --localstatedir=/var --sharedstatedir=/var/lib --mandir=/usr/share/man --infodir=/usr/share/info --disable-gss-keyring --disable-tests --enable-utils --enable-modules --disable-ldiskfs --disable-server --without-zfs --disable-snmp --disable-gss --disable-gss-keyring --with-systemdsystemunitdir=/usr/lib/systemd/system --with-linux=/lib/modules/3.10.0-1127.el7.x86_64/source --with-linux-obj=/lib/modules/3.10.0-1127.el7.x86_64/build --with-kmp-moddir=extra/lustre-client\n++ ./configure --program-prefix= --disable-dependency-tracking --prefix=/usr --exec-prefix=/usr --bindir=/usr/bin --sbindir=/usr/sbin --sysconfdir=/etc --datadir=/usr/share --includedir=/usr/include --libdir=/usr/lib64 --libexecdir=/usr/libexec --localstatedir=/var --sharedstatedir=/var/lib --mandir=/usr/share/man --infodir=/usr/share/info --disable-gss-keyring --disable-tests --enable-utils --enable-modules --disable-ldiskfs --disable-server --without-zfs --disable-snmp --disable-gss --disable-gss-keyring --with-systemdsystemunitdir=/usr/lib/systemd/system --with-linux=/lib/modules/3.10.0-1127.el7.x86_64/source --with-linux-obj=/lib/modules/3.10.0-1127.el7.x86_64/build --with-kmp-moddir=extra/lustre-client\nwhich: no mpicc in (/sbin:/bin:/usr/sbin:/usr/bin)\ndirname: missing operand\nTry 'dirname --help' for more information.\nconfigure: error: Kernel source /usr/src/kernels/3.10.0-1127.el7.x86_64 could not be found.\nerror: Bad exit status from /var/tmp/rpm-tmp.swmRpA (%build)\n Bad exit status from /var/tmp/rpm-tmp.swmRpA (%build)",
"stderr_lines": [
"+ umask 022",
"+ cd /root/rpmbuild/BUILD",
"+ cd /root/rpmbuild/BUILD",
"+ rm -rf lustre-2.12.5",
"+ /usr/bin/gzip -dc /root/rpmbuild/SOURCES/lustre-2.12.5.tar.gz",
"+ /usr/bin/tar -xf -",
"+ STATUS=0",
"+ '[' 0 -ne 0 ']'",
"+ cd lustre-2.12.5",
"+ /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w .",
"+ ln lustre/ChangeLog ChangeLog-lustre",
"+ ln lnet/ChangeLog ChangeLog-lnet",
"+ exit 0",
"+ umask 022",
"+ cd /root/rpmbuild/BUILD",
"+ cd lustre-2.12.5",
"+ cd /root/rpmbuild/BUILD/lustre-2.12.5",
"+ export 'UTILS_CFLAGS=-O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector-strong --param=ssp-buffer-size=4 -grecord-gcc-switches '",
"+ UTILS_CFLAGS='-O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector-strong --param=ssp-buffer-size=4 -grecord-gcc-switches '",
[Truncated]
"checking size of unsigned long long... 8",
"checking for -Wno-format-truncation support... yes",
"checking for -Wno-stringop-truncation support... yes",
"checking for -Wno-stringop-overflow support... yes",
"checking whether to use OSD addon... no",
"checking whether to build Lustre docs... no",
"checking whether to build Lustre manpages... yes",
"checking whether to build Lustre utilities... yes",
"checking whether to build Lustre tests... no",
"checking whether to build Lustre client support... yes",
"checking whether mpitests can be built... no",
"checking whether to build Linux kernel modules... yes (linux-gnu)",
"checking for Linux sources... /usr/src/kernels/3.10.0-1127.el7.x86_64",
"checking for /usr/src/kernels/3.10.0-1127.el7.x86_64... no",
"",
"",
"RPM build errors:"
]
}
```
username_2: "/usr/src/kernels/3.10.0-1127.el7.x86_64 could not be found."
"configure: error: Kernel source /usr/src/kernels/3.10.0-1127.el7.x86_64 could not be found.",
Please install kernel source.
username_0: I tried this but they are already installed
username_3: @username_0 , can you please paste the output of 'ls -l /usr/src/kernels/3.10.0-1127.el7.x86\_64'?
username_0: ```
[root@localhost cortx-motr]# ls -l /usr/src/kernels/3.10.0-1127.el7.x86_64
ls: cannot access /usr/src/kernels/3.10.0-1127.el7.x86_64: No such file or directory
```
username_0: Ya so it is not there you are right, but when I tried this that I found on googleย
```
[root@localhost cortx-motr]# yum install kernel-devel
Loaded plugins: fastestmirror
Loading mirror speeds from cached hostfile
* base: mirrors.coreix.net
* epel: mirrors.coreix.net
* extras: mirrors.coreix.net
* updates: mirrors.coreix.net
Package kernel-devel-3.10.0-1160.24.1.el7.x86_64 already installed and latest version
Nothing to do
```
username_3: Looks like this thread might be what you need to figure this out:
[https://unix.stackexchange.com/questions/215694/empty-kernel-directory-but-kernel-headers-are-installed/215697](https://unix.stackexchange.com/questions/215694/empty-kernel-directory-but-kernel-headers-are-installed/215697)
Can you please look in there and try those suggestions and see if you can make more progress here? ย Maybe the install path got changed for some reason from what the script expects. ย Then we would need to modify the script probably to look in multiple locations. ย If you can find the missing resources, try changing the script to point to the new location and see if that fixes this problem. ย Then we can ask @username_2 and others what is the long term fix so that others don't run into this problem.
username_2: Please try to install required packages from
https://vault.centos.org/
username_2: @username_0 please confirm this issue is still requiring fixes or it is solved?
username_0: This issue can be solved by downloading the kernel version from this site [https://vault.centos.org/7.8.2003/os/x86_64/Packages/](https://vault.centos.org/7.8.2003/os/x86_64/Packages/) where 7.8.2003 can be changed to the CentOS version you are running. I have created a PR #628 with this information in the guide.
username_3: Fixed in #628
Status: Issue closed
|
flamencist/ldap4net | 560471992 | Title: Problem with รครถรผ
Question:
username_0: First of all. Thank you for your effort. Very easy to use library which works very well.
I have just a small problem when for example the username has an `รค` in his name. In this case i will get back "B๏ฟฝck" when the expected value is "Bรคck".
My Code:
```
private async Task<IEnumerable<LdapEntry>> QueryAsync(string filter)
{
using var ldapConnection = new LdapConnection();
ldapConnection.Connect(adOptions.LdapHost, ldapsPort, LdapForNet.Native.Native.LdapVersion.LDAP_VERSION3);
await ldapConnection.BindAsync(mechanism: LdapForNet.Native.Native.LdapAuthMechanism.SIMPLE, userDn: adOptions.LdapQueryUserDn, password: adOptions.LdapQueryUserPassword);
var entries = await ldapConnection.SearchAsync(adOptions.LdapBase, filter);
return entries;
}
```
I put a breakpoint on the return line and got the value "B๏ฟฝck" with the following query in the watcher window.
`entries.First().Attributes.Where(x => x.Key == "sn").First().Value.First()`
Expected Value is "Bรคck"
I have use Version 2.2.0 of your library which should according to your release notes contains support for unicode.
I can provide further details if needed.
Answers:
username_1: Could you please provide client's operating system, server's operating system ?
username_0: Currently i develop on a TermnalServer where Visual Studio is installed. The Terminal Server is a Windows Server 2012 R2 Standard.
The Active Directory Server is a Windows Server 2012R2 as well or maybe already 2016. But i cannot confirm this.
username_1: Hi! I've checked and could not to reproduce. I've created user wit sn=Bรคck in AD on Windows 2012 R2 via Users and Computers tool. Checked using LdapExample on netcoreapp3.1 and net48. On both platforms all good.
Could you please check your value in AD?
username_0: The value in AD is correct.
Also i have checked with the old (Windows Only) implementation (System.DirectoryServices). Where i get back the correct value "Bรคck".
And i have cloned your master branch here and i have tried everything again with the current master and i still get back the wrong value.
Both tests where made with .NET Core 3.1
username_0: I think i found "something". As soon as i set the `LdapAuthMechanism` so `SIMPLE` i got the wrong value. If i change to KERBEROS or GSSAPI it works as it should and i get back the right value.
username_1: Hm..i will check with simple authentication.. |
InseeFrLab/AdventOfCodeLeaderboard | 530973282 | Title: Allow timezone override
Question:
username_0: Currently, the timezone used to display solving time is the one defined by the system.
So docker users are stuck with default timezone (GMT probably).
I think we should add a timezone configuration option.<issue_closed>
Status: Issue closed |
OrifInformatique/gestion_questionnaires | 319812947 | Title: Ordrer les listes
Question:
username_0: rajouter des boutons sur les tables pour trier les รฉlรฉments dans question/index
Answers:
username_1: Faire รงa en JavaScript ou en PHP ?
En PHP, on peut directement utiliser une requรชte SQL "ORDER BY" mais on doit recharger la page.
username_0: l'erreur du dernier commit est surement liรฉ au #57, il y as du avoir des changements involontaires ou inadรฉquats avec le filtre "Module concernรฉ"
username_0: L'รฉrreurs php due au tri sans filtre module รฉtait due qu fait que le filtre module ajoutais des parenthรจses nรฉcessaire (pour utiliser les ORDER BY avec plusieurs filtres avant) ร la requรจte sql qui n'รฉtaient pas prรฉsentes avec les autres filtres
Status: Issue closed
|
home-assistant/core | 804576193 | Title: Sensative strips binary sensor isn't available in HA
Question:
username_0: **The problem**
My sensative stips (door/window sensors) look fine in zwavejs2mqtt. The status shows true/false as it should.
In the zwave js integration and in HA, the binary sensor is always unavailable. The other two - battery level and battery low - work ok. The binary sensor entity isn't visible in HA, even after wakeups and opening/closing windows.
Remove/add didn't help.
In the legacy zwave they worked fine.
**What is version of Home Assistant Core has the issue?**
2021.2.2
**What was the last working version of Home Assistant Core?**
2021.1.5
**What type of installation are you running?**
Home Assistant Supervised
**Integration causing the issue**
zwave js
**Link to integration documentation on our website**
_No response_
**Example YAML snippet**
```yaml
# Put your YAML below this line
```
**Anything in the logs that might be useful for us?**
```txt
# Put your logs below this line
```
[zwave_js.zip](https://github.com/home-assistant/core/files/5951779/zwave_js.zip)
Answers:
username_1: just letting you know they are working fine in my config, which is using both core zwavejs add-on and integration. Did you wake the strips manually? (using the magnet to slide 3 times towards the rounded edge)
<img width="980" alt="Schermafbeelding 2021-02-09 om 15 00 20" src="https://user-images.githubusercontent.com/33354141/107374298-b87b2580-6ae7-11eb-94d3-44fa050d4e51.png">
username_0: Yes, many times. They look fine in zwavejs2mqtt, but they seem to get lost on the way to HA.
username_1: I had one that was causing issues (but it is a long way from the controller, and between aluminum frames. I had to deleted it and re-add it again, and now is functioning properly too. Maybe give it a try and do that?
username_0: I tried. Same result
username_2: I had a similar issue. Look for one of the disabled binary_sensor entities "binary_sensor.[entityname].any" (or maybe any2), if you enable that, it actually works as on/off (open/close).
username_0: After leaving the system on it's own for a few days, the Strips-MaZw Any sensors started working!. I guess they just needed the time to finish their dialog with zwavejs. Now I've enabled them for all my sensative strips, and all is good.
Status: Issue closed
|
ntnn/vim-paperwork | 59392656 | Title: Image for readme
Question:
username_0: 
Status: Issue closed
Answers:
username_0: 
Status: Issue closed
|
nathanreyes/v-calendar | 1087289004 | Title: Events not Repopulating when Step=1
Question:
username_0: Super project. Thank you for sharing. I have seen one potential issue, however.
When I set the `:step=` property and navigate forwards or backwards, events seem to only be shown on the furthest month shown (going backward: earliest, going forward: furthest). This doesn't occur when the `step` property is not present. In case it's a math issue, my entire argument is:
```
<DatePicker v-model="date" is-dark color="indigo" :attributes="attrs" :columns="5" :rows="2" :step="1"/>
```
Answers:
username_1: Hi! I have the same issue.
Here you can reproduce this problem: https://codesandbox.io/s/intelligent-ptolemy-dvl3j
Here is short video with this problem: https://jmp.sh/WWIilVa |
quinnj/JSON3.jl | 845293962 | Title: Support Optional/Missing fields for StructType
Question:
username_0: ```
struct System
duration::Real
cwd::Union{Nothing, String}
environment::Union{Nothing, Dict}
System(duration, cwd=nothing, environment=nothing) = new(duration, cwd, environment)
end
StructTypes.StructType(::Type{System}) = StructTypes.Struct()
let system = """
{
"duration": 3600
}
"""
sys = JSON3.read(system, System)
@assert sys.duration == 3600
end
```
fails with:
```
ERROR: UndefRefError: access to undefined reference
Stacktrace:
[1] getindex at ./array.jl:809 [inlined]
[2] construct at /home/username_0/.julia/packages/StructTypes/MF8bN/src/StructTypes.jl:860 [inlined]
[3] #read#42 at /home/username_0/.julia/packages/JSON3/IvnMR/src/structs.jl:572 [inlined]
[4] read at /home/username_0/.julia/packages/JSON3/IvnMR/src/structs.jl:494 [inlined]
[5] read(::String, ::Type{System}; kw::Base.Iterators.Pairs{Union{},Union{},Tuple{},NamedTuple{(),Tuple{}}}) at /home/username_0/.julia/packages/JSON3/IvnMR/src/structs.jl:34
[6] read(::String, ::Type{System}) at /home/username_0/.julia/packages/JSON3/IvnMR/src/structs.jl:33
[7] top-level scope at REPL[6]:6
```
Additionally this fails similarly:
```
struct System
duration::Real
cwd::Union{Nothing, String}
environment::Union{Nothing, Dict}
System(duration, cwd=nothing, environment=nothing) = new(duration, cwd, environment)
end
StructTypes.StructType(::Type{System}) = StructTypes.Struct()
StructTypes.omitempties(::Type{System}) = (:cwd, :environment)
sys = JSON3.read(JSON3.write(System(3600)), System)
@assert sys.duration == 3600
```
Answers:
username_1: So the biggest question here is what would we do in the case where our json is like:
```json
{
"duration": 3600,
"environment": {"PATH": "/usr/lib"}
}
```
It makes me squirm a little to say we'd just end up calling `System(duration, environment)` because that could quickly cause weird issues for people if htey have a struct like:
```julia
struct T
id::Int
fname::String
lname::String
end
```
where `fname` is missing from the json. But I'm also not excited about the alternative of just ignoring any fields after the first `#undef`.
Hmmmm, what to do, what to do. We've talked at one point about having a `StructTypes.defaultvalues` overload where default values could be provided for missing fields; maybe that's the solution here.
username_2: I just ran into this exact problem.
username_1: Yeah, I guess this change is really a bug from the `StructTypes.Struct => StructTypes.UnorderedStruct` change, because the behavior before was that it would pass any arguments parsed in order ot the constructor. Hmmm, unfortunately that's quite at odds with the current code implementation where it's trying to put the fields in order. Let me stew on if we can get the old behavior here again.
username_1: Ok, here's a proposed alternative fix that wouldn't incur any performance hit: if a field wasn't found when parsing, we pass `nothing`. This is somewhat accurate, since `nothing` maps to the JSON `null` value, and it's often standard JSON to either _include_ the field with `null` value or omit the field entirely. So we'll just flip that coin when constructing: parse values, pass them in _field_ order to the constructor, if any field has a `null` field, it would be passed as `nothing`, if it's omitted from the parsing, we still pass `nothing`. How does that sound?
username_1: Ok, PR up if anyone cares to take a look: https://github.com/JuliaData/StructTypes.jl/pull/43
Status: Issue closed
|
couds/react-bulma-components | 656784254 | Title: Cannot select a tab in the storybook
Question:
username_0: It just stays on the same tab all the time.
Answers:
username_1: Hi @username_0 . Yes the storybook for tabs do not have any logic just show how the set the active tab (on the code) but now how to handle the state, that's our of the scope of the library
Status: Issue closed
username_0: Ok thanks. I was confused on how to react (excuse the pun) to a change in
state. Most other react libraries I've used accept some kind of handler
that is invoked when a tab changes.
I'm not an experienced react user.
What I did in the end was add an onclick handler to each of the tabs, but
that's a lot of duplication. |
hello-efficiency-inc/raven-reader | 460254072 | Title: Ability to rename subscriptions, or change their url
Question:
username_0: Would love to be able to rename a feed to something more easily identifiable, or change the url a bit instead of having to add a new subscription and remove the old one
Answers:
username_1: In upcoming new version I have included ability to edit name of the Feed.
username_1: Fixed in https://github.com/hello-efficiency-inc/raven-reader/releases/tag/v0.5.4
Status: Issue closed
|
rancher/rancher | 461227215 | Title: Backport: Server panic during startup when k8s v1.14 is not available
Question:
username_0: Backport https://github.com/rancher/rancher/issues/21129
Answers:
username_1: The bugfix is validated on Rancher:v2.2-head 5521557f
Steps:
- run Rancher:v2.2-head container
- add a cluster with k8s v1.14 using EC2, 1 etcd, 1 control plane, 1 worker
- stop the etcd node on AWS
- `docker stop` to stop the Rancher container
- wait to see the UI is gone and it shows no connection
- `docker start` to start the Rancher container
Results:
- Rancher container is restarted successfully
- in Rancher's UI, the cluster is still in the unavailable state
- restart the etcd node, then the cluster goes back to active and is functional
<img width="1453" alt="screenshot" src="https://user-images.githubusercontent.com/6218999/60300965-0fb58d80-98e5-11e9-855f-2b4ee369be6b.png">
Status: Issue closed
|
wix/react-native-navigation | 500012256 | Title: Push screen doesn't load UI
Question:
username_0: "react": "16.9.0",
"react-native": "0.61.1",
"react-native-navigation": "^3.2.0-snapshot.524"
### Issue Description
When I push screen with stack nav, it doesn't load UI. I see the log that the page loaded.
[ENTER DESCRIPTION HERE]
### Steps to Reproduce / Code Snippets / Screenshots
---
### Environment
* React Native Navigation version: 3.2.0-snapshot.524
* React Native version: 0.61.1
* Platform(s) (iOS, Android, or both?): ios (iphone 8)
* Device info (Simulator/Device? OS version? Debug/Release?): simulator ios 12 |
smooth-code/smooth.js | 431371078 | Title: Expose WP envs
Question:
username_0: It would be quite useful to be able to access to vars such as `WP_HOME` and others from Wordpress.
Status: Issue closed
Answers:
username_1: You can do it by extending the webpack configuration : https://github.com/smooth-code/smooth.js#customizing-webpack-config |
fuzhengwei/CodeGuide | 885327070 | Title: ๅคงๅญฆๆฏไธ่ฆๅๅคๅฐ่กไปฃ็ ๏ผๆ่ฝไธ็จ่ฑ้ฑๅน่ฎญๅฐฑๆพๅฐไธไปฝๅผๅๅทฅไฝ๏ผ - bugstack่ซๆดๆ
Question:
username_0: https://bugstack.cn/itstack-code-life/2021/05/09/%E5%A4%A7%E5%AD%A6%E6%AF%95%E4%B8%9A%E8%A6%81%E5%86%99%E5%A4%9A%E5%B0%91%E8%A1%8C%E4%BB%A3%E7%A0%81-%E6%89%8D%E8%83%BD%E4%B8%8D%E7%94%A8%E8%8A%B1%E9%92%B1%E5%9F%B9%E8%AE%AD%E5%B0%B1%E6%89%BE%E5%88%B0%E4%B8%80%E4%BB%BD%E5%BC%80%E5%8F%91%E5%B7%A5%E4%BD%9C.html
ๅไบซไธชไบบ็ปๅใๆป็ปๅญฆไน ่ฎกๅใๆดๆๅฐฑไธๅๆใๆๆฏ่ตๆๆฑๆป๏ผ่ฎฒ่ฎฒๆ13ๆฏไธๆถๆฏๆไนๆฟๅฐไธไธช7k็offer็๏ผไปฅๅๅๆฅๅๆฏๆไนไปๅนฒไบไธคๅนด็C#่ฝฌ่กๅปๅJavaไบ๏ผ |
TIY-Bracket/bracket_api | 117205553 | Title: Create Bracket Page
Question:
username_0: - [ ] Get fonts matching
- [ ] Try and get Email/Text as 1 input field
if using api view decorator, got to override how field is translated
- [ ] Longer Input Bracket Title Field
- [ ] Same length as all bottom fields, increase height as well;
- [ ] Could get it in as a table -
- [ ] Get rid of border box add in more samurai
- [ ] Get rid of instructions
- [ ] Fix alignment
- [ ] Change Button to Generate/Create - Make a little larger, Not same size as bracket title input field
- [ ] For border, if stays, make padding same all around
- [ ] darker color for background around border
- [ ] competitor - intermediate color
- [ ] submit has primary/secondary color
- [ ] Make sure colors are more distinct on projector
- [ ] Use samurai colors as a suggestion
- [ ] buttons - add color; |
ianstormtaylor/slate | 221183427 | Title: adding extra DOM handlers
Question:
username_0: I'm trying to add support for HTML5 DnD drop target event. As I understand it if I can set a drop target event handler on the editor DOM element, I'm fine. This would be passed on the editor. However the properties set on the Content component are explicitly listed. If instead it passed along all properties I would be able to add new event handlers to the editor that are not explicitly supported by Slate, that would be great I think.
```
function render(props, state, editor) {
return (
<Content
autoCorrect={props.autoCorrect}
autoFocus={props.autoFocus}
className={props.className}
children={props.children}
editor={editor}
onBeforeInput={editor.onBeforeInput}
onBlur={editor.onBlur}
onFocus={editor.onFocus}
onChange={editor.onChange}
onCopy={editor.onCopy}
onCut={editor.onCut}
onDrop={editor.onDrop}
onKeyDown={editor.onKeyDown}
onPaste={editor.onPaste}
onSelect={editor.onSelect}
readOnly={props.readOnly}
role={props.role}
schema={editor.getSchema()}
spellCheck={props.spellCheck}
state={state}
style={props.style}
tabIndex={props.tabIndex}
/>
)
}
```
should become
```
function render(props, state, editor) {
return (
<Content
{...editor}
autoCorrect={props.autoCorrect}
autoFocus={props.autoFocus}
className={props.className}
children={props.children}
editor={editor}
readOnly={props.readOnly}
role={props.role}
schema={editor.getSchema()}
spellCheck={props.spellCheck}
state={state}
style={props.style}
tabIndex={props.tabIndex}
/>
)
}
```
If that doesn't work perhaps all properties starting with on could be passed. Or a function that takes the name and value of a property and returns the value if it should be passed to the content.
Status: Issue closed
Answers:
username_0: I still need dragOver, but I can see it must be modified in the Content component |
ibm-cloud-architecture/refarch-kc | 580544924 | Title: Add an overview of services deployed in the demonstration script
Question:
username_0: In the demonstration script, it would be good if there was an overview at the top with a list of all the services you need to have deployed in order to run through the demo guide.
It could also link out to the quick installation set up from @djones6 as a recommendation for getting set up quickly with the required pre-requisites. |
ESGF/esgf-prepare | 386301116 | Title: stty errors
Question:
username_0: When stdin is a pipe, the `stty size` inside `MultilineFormatter` gives lots of copies of the error:
`stty: standard input: Inappropriate ioctl for device`
It should be made to just use the default values if stdin is not a terminal (`if not sys.stdin.isatty()`) instead of running `stty`
It might also be prudent to do the same regarding stdout also.
Thanks.<issue_closed>
Status: Issue closed |
valerauko/rusty | 296434920 | Title: Hangs mysteriously
Question:
username_0: For a m3u of 2000+ entries (so a total of about 4500 lines) it keeps hanging while copying around 1300 or so. It keeps the drive busy though so I'm not sure what's happening.
## To do
- [ ] add a debug message on an `else` branch in the copy status check `if` (then again, it doesn't seem to happen any earlier and it takes like 10 minutes to copy 1500 files so I'll need to use dummy m3us with empty text files for example...) |
grpc/grpc | 1112015071 | Title: Add base_namespace option to grpc_csharp_plugin
Question:
username_0: ### Is your feature request related to a problem? Please describe.
This is a follow-up to #13724. It was closed because a workaround was introduced in the `Grpc.Tools` NuGet package. However, I have a use case that isn't addressed by #17672.
I'm using [Buf](https://buf.build/) to generate sources (among other things), and it works for all languages except C# gRPC services (it _does_ work for C# protobuf messages). Buf generates sources directly using the `grpc_csharp_plugin` through `protoc`, and this ends up generating duplicate files (leaving me with incomplete generated sources) when two services are in files with the same name but in different directories. For example:
- `proto\v1\service.proto`
- `proto\v2\service.proto`
This isn't a problem with protobuf's `csharp_plugin` thanks to the `base_namespace` option. However, adding a similar option to `grpc_csharp_plugin` was rejected in #24426.
### Describe the solution you'd like
I'd love to see a native fix to this issue in `grpc_csharp_plugin` for scenarios where generating sources happens outside MSBuild/through `protoc`. A couple possible solutions:
- Add a `base_namespace` option that works like `csharp_plugin`'s (i.e. reconsider #24426).
- Add any other file name deconfliction algorithm (like random or hash-based file suffixes) to `grpc_csharp_plugin`, given how the C# compiler cares very little about the actual file names of the generated sources.
### Describe alternatives you've considered
I could assign unique file names to the different services. For example:
- `proto\v1\service_v1.proto`
- `proto\v2\service_v2.proto`
However, this would impact the output of other language plugins (e.g. Javascript) and introduce friction elsewhere.
Alternatively-- I could file this as an issue with the Buf project and request they investigate/implement their own workaround for this limitation. I thought about doing that first, but reconsidered after identifying that there are options in all other protobuf/gRPC plugins to work around duplicate file name issues like this. This just appears to be a limitation of the C# gRPC plugin. |
osalvador/ReplicaDB | 927624568 | Title: Replication not working for geometric type column
Question:
username_0: I have SDO_GEOMETRY column type that I am trying to replicate between oracle schemas and I see the following error.
ERROR ReplicaTask:72 ERROR in TaskId-3 inserting data to sink table: ORA-00932: inconsistent datatypes: expected MDSYS.SDO_GEOMETRY got CHAR
ERROR ReplicaDB:101 Got exception running ReplicaDB:
java.util.concurrent.ExecutionException: java.sql.BatchUpdateException: ORA-00932: inconsistent datatypes: expected MDSYS.SDO_GEOMETRY got CHAR
I checked data and all of them are SDO_GEOMETRY type. I even tried with 1 row and its the same. Have you seen this before?
Answers:
username_1: Hi @username_0 ,
Could you share the DDLs of the source and sink tables and the insert statement? I'm not familiar with the SDO_GEOMETRY data type, but with this I can reproduce the exception.
Thanks!
username_0: Here is the ddl. Its the same for source and sink tables just a different schema.
CREATE TABLE "TFPSYS_LOCATIONS"
( "ID" VARCHAR2(17 BYTE),
"TITLE" VARCHAR2(300 BYTE),
"FIRST_CHAR" VARCHAR2(1 BYTE),
"LOC_TYPE" NUMBER(6,0),
"GEN_CAT" NUMBER(6,0),
"SUB_CAT" NUMBER(6,0),
"GEO_ID" NUMBER(10,0),
"ADDRESS_LINE1" VARCHAR2(100 BYTE),
"ADDRESS_LINE2" VARCHAR2(100 BYTE),
"CITY" VARCHAR2(100 BYTE),
"STATE_CODE" VARCHAR2(2 BYTE),
"ZIP" VARCHAR2(15 BYTE),
"SORT_TITLE" VARCHAR2(300 BYTE),
"SOURCE" VARCHAR2(5 BYTE),
"PHONE" VARCHAR2(18 BYTE),
"FAX" VARCHAR2(15 BYTE),
"CITY_SORT" VARCHAR2(100 BYTE),
"INST_ID" NUMBER(9,0),
"LOC_ID" NUMBER(9,0),
"TIMESTAMP" DATE,
"ZIP_NUMBER" NUMBER(5,0),
"LATITUDE" NUMBER(10,6),
"LONGITUDE" NUMBER(10,6),
"GEOM" "SDO_GEOMETRY",
"REMARKS" VARCHAR2(2000 BYTE),
"STATUS" VARCHAR2(1 BYTE) DEFAULT 'A',
"INTERNAL_NOTES" VARCHAR2(4000 BYTE),
CONSTRAINT "PK_TFPSYS_LOCATIONS" PRIMARY KEY ("ID")
And here is the Insert statement
INSERT INTO /*+APPEND_VALUES*/ uat.tfpsys_locationsrepdb6310 (ID,TITLE,FIRST_CHAR,LOC_TYPE,GEN_CAT,SUB_CAT,GEO_ID,ADDRESS_LINE1,ADDRESS_LINE2,CITY,STATE_CODE,ZIP,SORT_TITLE,SOURCE,PHONE,FAX,CITY_SORT,INST_ID,LOC_ID,TIMESTAMP,ZIP_NUMBER,LATITUDE,LONGITUDE,GEOM,REMARKS,STATUS,INTERNAL_NOTES) VALUES ( ?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,?,? )
Hope this helps.
Thanks,
Udaya
username_1: Hi @username_0,
I've added the `sdo_geometry` data type for Oracle Database. Could you download the latest release and try it out?
Thanks!
username_0: I still see the same issue with version 10.2 and 10.5
ERROR ReplicaTask:72 ERROR in TaskId-4 inserting data to sink table: ORA-00932: inconsistent datatypes: expected MDSYS.SDO_GEOMETRY got CHAR
username_1: Hi @username_0,
`SDO_GEOMETRY` data type is tested on Oracle 11gR2 with this DDL file: https://github.com/username_1/ReplicaDB/blob/master/src/test/resources/oracle/oracle-sdo_geometry.sql
To reproduce the exception, could share with us?:
- Oracle database version for source and sink databases
- Replicadb config file
- DDL of source and sink tables
- Some insert statements with values for source table
username_1: Close because inactivity.
Status: Issue closed
|
thyamu/MGMNet | 234142504 | Title: metagenome bulk download script
Question:
username_0: Generate script to download metagenome EC numbers from https://img.jgi.doe.gov/
check
beautifulsoup python 2.7
https://stackoverflow.com/questions/2146383/https-connection-python
https://stackoverflow.com/questions/32754229/python-and-beautifulsoup-opening-pages |
silentmatt/expr-eval | 1123622530 | Title: Checking if all referenced variables exist?
Question:
username_0: If I have a case where the expression is always evaluated against an object containing a specific list of variables, what would be the intended way to sanitize the passed expression? I assumed I could use `variables()` in one way or another, however that will also return variables which are defined in the expression itself.
I have a hacky solution now where I recursively iterate over the tokens and account for every `IVARNAME` this way but surely there has to be a better way for this right? |
rust-windowing/winit | 550913883 | Title: [Windows] Not possible to create multiple windows (thread 'main' panicked at 'already mutably borrowed')
Question:
username_0: Hi ๐
Since https://github.com/rust-windowing/winit/commit/6a330a2894873d29fbbfdeebfc1a215577213996 it is not possible to create multiple windows on Windows (10). It can be reproduced with `multiwindow` example. The single `window` example works fine though.
```
PS C:\Users\tester\Documents\GitHub\winit> $ENV:RUST_BACKTRACE = "full"
PS C:\Users\tester\Documents\GitHub\winit> cargo run --example multiwindow
```
```
Finished dev [unoptimized + debuginfo] target(s) in 0.36s
Running `target\debug\examples\multiwindow.exe`
thread 'main' panicked at 'already mutably borrowed: BorrowError', src\libcore\result.rs:1165:5
stack backtrace:
0: 0x7ff6b2f461c9 - backtrace::backtrace::trace_unsynchronized
at C:\Users\VssAdministrator\.cargo\registry\src\github.com-1ecc6299db9ec823\backtrace-0.3.37\src\backtrace\mod.rs:66
1: 0x7ff6b2f461c9 - std::sys_common::backtrace::_print_fmt
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\sys_common\backtrace.rs:76 2: 0x7ff6b2f461c9 - std::sys_common::backtrace::_print::{{impl}}::fmt
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\sys_common\backtrace.rs:60 3: 0x7ff6b2f5adfb - core::fmt::write
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libcore\fmt\mod.rs:1030
4: 0x7ff6b2f433a4 - std::io::Write::write_fmt<std::sys::windows::stdio::Stderr>
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\io\mod.rs:1412
5: 0x7ff6b2f48960 - std::sys_common::backtrace::_print
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\sys_common\backtrace.rs:64 6: 0x7ff6b2f48960 - std::sys_common::backtrace::print
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\sys_common\backtrace.rs:49 7: 0x7ff6b2f48960 - std::panicking::default_hook::{{closure}}
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\panicking.rs:196
8: 0x7ff6b2f4858a - std::panicking::default_hook
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\panicking.rs:210
9: 0x7ff6b2f491b8 - std::panicking::rust_panic_with_hook
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\panicking.rs:473
10: 0x7ff6b2f48d24 - std::panicking::continue_panic_fmt
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\panicking.rs:380
11: 0x7ff6b2f48c09 - std::panicking::rust_begin_panic
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\panicking.rs:307
12: 0x7ff6b2f57469 - core::panicking::panic_fmt
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libcore\panicking.rs:85
13: 0x7ff6b2f5756f - core::result::unwrap_failed
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libcore\result.rs:1165
14: 0x7ff6b2f02aaf - core::result::Result<core::cell::Ref<core::option::Option<winit::platform_impl::platform::event_loop::runner::EventLoopRunner<()>>>, core::cell::BorrowError>::expect<core::cell::Ref<core::option::Option<winit::platform_impl::platform::event_loop::runner::
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\src\libcore\result.rs:960
15: 0x7ff6b2ef517c - core::cell::RefCell<core::option::Option<winit::platform_impl::platform::event_loop::runner::EventLoopRunner<()>>>::borrow<core::option::Option<winit::platform_impl::platform::event_loop::runner::EventLoopRunner<()>>>
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\src\libcore\cell.rs:788
16: 0x7ff6b2ef8508 - winit::platform_impl::platform::event_loop::runner::ELRShared<()>::send_event<()>
at C:\Users\tester\Documents\GitHub\winit\src\platform_impl\windows\event_loop\runner.rs:133
17: 0x7ff6b2e8196b - winit::platform_impl::platform::event_loop::SubclassInput<()>::send_event<()>
at C:\Users\tester\Documents\GitHub\winit\src\platform_impl\windows\event_loop.rs:104
18: 0x7ff6b2e858ec - winit::platform_impl::platform::event_loop::public_window_callback<()>
at C:\Users\tester\Documents\GitHub\winit\src\platform_impl\windows\event_loop.rs:1367
19: 0x7ffe8273b88a - DefSubclassProc
20: 0x7ffe8273b745 - DefSubclassProc
21: 0x7ffe97117446 - DispatchMessageW
22: 0x7ffe9711716c - DispatchMessageW
23: 0x7ffe97120a23 - EnumChildWindows
24: 0x7ffe99ea3494 - KiUserCallbackDispatcher
25: 0x7ffe96ed1184 - NtUserMessageCall
26: 0x7ffe97114b16 - GetWindowTextW
27: 0x7ffe97114488 - GetWindowTextW
28: 0x7ffe97117446 - DispatchMessageW
29: 0x7ffe9711716c - DispatchMessageW
30: 0x7ffe97120a23 - EnumChildWindows
[Truncated]
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\panicking.rs:292
68: 0x7ff6b2f4bcd2 - panic_unwind::__rust_maybe_catch_panic
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libpanic_unwind\lib.rs:80
69: 0x7ff6b2f494f2 - std::panicking::try
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\panicking.rs:271
70: 0x7ff6b2f494f2 - std::panic::catch_unwind
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\panic.rs:394
71: 0x7ff6b2f494f2 - std::rt::lang_start_internal
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\/src\libstd\rt.rs:48
72: 0x7ff6b2e7de1b - std::rt::lang_start<()>
at /rustc/4560ea788cb760f0a34127156c78e2552949f734\src\libstd\rt.rs:64
73: 0x7ff6b2e89280 - main
74: 0x7ff6b2f5ed24 - invoke_main
at d:\agent\_work\2\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:78
75: 0x7ff6b2f5ed24 - __scrt_common_main_seh
at d:\agent\_work\2\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:288
76: 0x7ffe98827974 - BaseThreadInitThunk
77: 0x7ffe99e6a271 - RtlUserThreadStart
error: process didn't exit successfully: `target\debug\examples\multiwindow.exe` (exit code: 0xc0000409, STATUS_STACK_BUFFER_OVERRUN)
```
Answers:
username_1: https://github.com/rust-windowing/winit/issues/1427#event-3016432338 has the beginning of an investigation into the issue.
username_2: This panic also appears to be occur for some windows users when resizing and toggling fullscreen.
As mentioned in #1427 it occurs in `send_event` where a `borrow` was added when previously only `try_borrow` was being used:
```rust
pub(crate) unsafe fn send_event(&self, event: Event<'_, T>) {
let handling_redraw = self
.runner
.borrow()
.as_ref()
.map(|r| RunnerState::HandlingRedraw == r.runner_state)
.unwrap_or(false);
let mut send = None;
if handling_redraw {
if let Event::RedrawRequested(_) = event {
send = Some(event);
} else {
self.buffer_event(event);
}
} else {
send = Some(event);
}
if let Some(event) = send {
if let Err(event) = self.send_event_unbuffered(event) {
// If the runner is already borrowed, we're in the middle of an event loop invocation. Add
// the event to a buffer to be processed later.
self.buffer_event(event);
}
}
}
```
The quickest solution would be to buffer the `RedrawRequested` event if the borrow fails. However, I don't know whether this behavior would be desired.
username_3: Just thought I'd add that we're also running into this on Windows in nannou.
We are able to create multiple windows before running the event loop without any issue, however we trigger this same `panic!` when attempting to `set_inner_size` after receiving a `RedrawRequested` event.
Here's the debug stacktrace:
```
Compiling nannou v0.12.0 (D:\rust\nannou)
Finished dev [unoptimized + debuginfo] target(s) in 1m 09s
Running `target\debug\examples\p_1_0_01.exe`
thread 'main' panicked at 'already mutably borrowed: BorrowError', src\libcore\result.rs:1084:5
stack backtrace:
0: backtrace::backtrace::trace_unsynchronized
at C:\Users\VssAdministrator\.cargo\registry\src\github.com-1ecc6299db9ec823\backtrace-0.3.34\src\backtrace\mod.rs:66
1: std::sys_common::backtrace::_print
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libstd\sys_common\backtrace.rs:47
2: std::sys_common::backtrace::print
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libstd\sys_common\backtrace.rs:36
3: std::panicking::default_hook::{{closure}}
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libstd\panicking.rs:200
4: std::panicking::default_hook
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libstd\panicking.rs:214
5: std::panicking::rust_panic_with_hook
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libstd\panicking.rs:477
6: std::panicking::continue_panic_fmt
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libstd\panicking.rs:384
7: std::panicking::rust_begin_panic
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libstd\panicking.rs:311
8: core::panicking::panic_fmt
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libcore\panicking.rs:85
9: core::result::unwrap_failed
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libcore\result.rs:1084
10: core::result::Result<core::cell::Ref<core::option::Option<winit::platform_impl::platform::event_loop::runner::EventLoopRunner<()>>>, core::cell::BorrowError>::expect<core::cell::Ref<core::option::Option<winit::platform_impl::platform::event_loop::runner::
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\src\libcore\result.rs:879
11: core::cell::RefCell<core::option::Option<winit::platform_impl::platform::event_loop::runner::EventLoopRunner<()>>>::borrow<core::option::Option<winit::platform_impl::platform::event_loop::runner::EventLoopRunner<()>>>
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\src\libcore\cell.rs:788
12: winit::platform_impl::platform::event_loop::runner::ELRShared<()>::send_event<()>
at C:\Users\josh\.cargo\registry\src\github.com-1ecc6299db9ec823\winit-0.21.0\src\platform_impl\windows\event_loop\runner.rs:133
13: winit::platform_impl::platform::event_loop::SubclassInput<()>::send_event<()>
at C:\Users\josh\.cargo\registry\src\github.com-1ecc6299db9ec823\winit-0.21.0\src\platform_impl\windows\event_loop.rs:104
14: winit::platform_impl::platform::event_loop::public_window_callback<()>
at C:\Users\josh\.cargo\registry\src\github.com-1ecc6299db9ec823\winit-0.21.0\src\platform_impl\windows\event_loop.rs:770
15: DefSubclassProc
16: DefSubclassProc
17: CallWindowProcW
18: CallWindowProcW
19: glPushClientAttrib
20: CallWindowProcW
21: DispatchMessageW
22: IsWindowVisible
23: KiUserCallbackDispatcher
24: NtUserMessageCall
25: SendMessageW
26: GetWindowTextW
27: MapWindowPoints
28: IsCompositionActive
29: IsCompositionActive
30: GetWindowTextW
31: CallWindowProcW
32: CallWindowProcW
[Truncated]
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libstd\panicking.rs:296
88: panic_unwind::__rust_maybe_catch_panic
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libpanic_unwind\lib.rs:80
89: std::panicking::try
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libstd\panicking.rs:275
90: std::panic::catch_unwind
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libstd\panic.rs:394
91: std::rt::lang_start_internal
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\/src\libstd\rt.rs:48
92: std::rt::lang_start<()>
at /rustc/625451e376bb2e5283fc4741caa0a3e8a2ca4d54\src\libstd\rt.rs:64
93: main
94: invoke_main
at d:\agent\_work\2\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:78
95: __scrt_common_main_seh
at d:\agent\_work\2\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:288
96: BaseThreadInitThunk
97: RtlUserThreadStart
error: process didn't exit successfully: `target\debug\examples\p_1_0_01.exe` (exit code: 0xc0000409, STATUS_STACK_BUFFER_OVERRUN)
```
username_1: @username_3 could you try the branch associated to this PR : https://github.com/rust-windowing/winit/pull/1461
But note that you should probably not change the window state when handling a `RedrawRequested`. You should only render/repaint the window content when handling `RedrawRequested`.
username_3: Yes you are right, this issue was the nudge we needed to fix this and we've since addressed it :)
Status: Issue closed
|
extras-evolution/multiTV | 351827134 | Title: MySQL 5.7 issue
Question:
username_0: Please replace
```
"elements": "@SELECT 'Unknown' AS username, 0 as id FROM [+PREFIX+]web_users LIMIT 1 UNION (SELECT username, id FROM [+PREFIX+]web_users ORDER BY username ASC)"
```
with
```
"elements": "@SELECT 'Unknown' AS username, 0 as id UNION (SELECT username, id FROM [+PREFIX+]web_users ORDER BY username ASC)"
```
in `assets/tvs/multitv/configs/dbtabledemo.config.json` and `assets/tvs/multitv/moduleconfigs/event_log.moduleconfig.json` |
mplusmuseum/mplusmuseum-tms-extract | 307346130 | Title: Check the contents of the XML/JSON files against the schema
Question:
username_0: When we have the XML processed based off the schema, or even before, we want to check to see if there's anything in the XML/JSON that _isn't_ in the scheme. So we can test to see if new fields have been entered or if we're missing them.
In short, I want to validate/check the XML against out expectations.<issue_closed>
Status: Issue closed |
nodejs/nodejs.dev | 944345587 | Title: StoryBook not working
Question:
username_0: ## Description
Storybook has stopped working
### Steps to reproduce
Go to [https://nodejs.dev/storybook/](url) ,we get a blank page
### Expected result
We should be able to see the storybook components page
### Actual result
We are getting a Blank page.
### Environment
- Browser: Chrome
Answers:
username_0: I have fixed this locally by upgrading the version of react-dom as well and it is working, Shall I make a pr?
username_1: A PR would be great, @username_0 !!
Status: Issue closed
|
aws/aws-cli | 233616478 | Title: SSL: CERTIFICATE_VERIFY_FAILED
Question:
username_0: Hi,
We recently implemented Palo Alto firewall in our network. The firewall does SSL decryption with the cert provided by the firewall itself. Since then, we can no longer use aws cli or aws web page to interact with aws services such as S3.
When working with CLI we get following error
[SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:645)
If I pass --no-verify-ssl, cli works.
I want to know if there is a way to specify our own ssl cert with aws cli or with web interface ?
Regards,
Saurabh
Answers:
username_1: Sounds a bit suspicious that it's decrypting your https traffic. Regardless, you can provide your own cert bundles using `ca_bundle` in addition to environment variable and command line options. See [the docs](https://github.com/aws/aws-cli#other-configurable-variables) for more details. Your web browsers probably also allow this sort of configuration, but I can't help you out there.
username_0: Thank you. I will try this and confirm once completed.
username_0: Thank you. This worked on scripts that makes use of aws cli.
I just need to figure out how to get s3 web pages to respond to this.
Regards,
Saurabh
username_1: Well it should just be configuration for your browsers. I'm pretty sure most of the big ones allow you to do this. Glad you got the CLI working though!
Status: Issue closed
|
fluxxus-nl/Test-Tool-Extension | 433107929 | Title: Test needed to test enabling/disabling within more complex filter
Question:
username_0: Currently the tests scenarios #0009 through #0016 only test the feature when filter is set on single test codeunit id.
Test are needed to also check when filter is complex. Not so much for multiple test codeununit ids, but more with filters on other columns.<issue_closed>
Status: Issue closed |
plotly/dash-table | 386705919 | Title: ImportError: cannot import name '_explicitize_args'
Question:
username_0: @chriddyp #error
I'm getting Import error when I run one of the examples of datatable. Could you please assist to get thru.
import dash
import dash_table
import pandas as pd
df = pd.read_csv('https://raw.githubusercontent.com/plotly/datasets/master/solar.csv')
app = dash.Dash(__name__)
app.layout = dash_table.DataTable(
id='table',
columns=[{"name": i, "id": i} for i in df.columns],
data=df.to_dict("rows"),
)
if __name__ == '__main__':
app.run_server(debug=True)

Answers:
username_1: @username_0 What version of Dash are you running?
username_1: Someone having a similar issue -- updating versions did the trick.
https://community.plot.ly/t/error-importing-dash-table-cannot-import-name-explicitize-args/16394
Status: Issue closed
username_2: **Hello, I just downloaded dash and I am still having the same issue.**
**When I am using the latest version of the packages and then I try to import ยดdash_tableยด I get this error :**
`import dash_table
Traceback (most recent call last):
File "<ipython-input-10-f48bca51a57b>", line 1, in <module>
import dash_table
File "C:\Users\tomk-\anaconda3\lib\site-packages\dash_table\__init__.py", line 18, in <module>
from ._imports_ import * # noqa: E402, F401, F403
File "C:\Users\tomk-\anaconda3\lib\site-packages\dash_table\_imports_.py", line 1, in <module>
from .DataTable import DataTable
File "C:\Users\tomk-\anaconda3\lib\site-packages\dash_table\DataTable.py", line 3, in <module>
from dash.development.base_component import Component, _explicitize_args
ImportError: cannot import name '_explicitize_args' from 'dash.development.base_component' (C:\Users\tomk-\anaconda3\lib\site-packages\dash\development\base_component.py)`
**The only way I manage to load the packages without problem is by installing those versions :**
` !pip install dash==0.30.0
!pip install dash-core-components==0.38.1
!pip install dash-html-components==0.13.2
!pip install dash-renderer==0.15.1
!pip install dash-table==3.1.7`
**Is it normal that the newest version does not contain the rigth file ?**
username_3: @username_2 those are extremely old package versions. Since `dash` v1.0, dash pins the versions of dcc, html, table, and renderer so the version should only be specified for `dash` itself. |
tuist/tuist | 772742087 | Title: Choose a path for the test xcresult when running tuist test
Question:
username_0: ## Need/problem
Some code analysis tools, for example SonarQube/Sonarcloud ask you to convert the files xcresult to their format before send them to the server for analysis.
## Motivation
xcresult files and stored under xcode DerivedData which on a CI server this directory changes all the time. If we could set a path where to export the results, could be easer for the developer to them extract them and do whatever we need with them, like when the command `xcodebuild` is run, we can setup an argument **derivedDataPath**
We could probably do something with the **find** command on the CI, but I think it's not ideal
## Detailed design
When running `tuist test` we should have an argument that accepts a path for the logs, I'm not sure if it could be worth for any other tuist command
# Drawbacks
Why should we *not* do this? Please consider:
Non comes to my head
## Alternatives
We could probably do something with the **find** command on the CI, but I think it's not ideal
# Adoption strategy
If we implement this proposal, how will existing Tuist developers adopt it? Is
this a breaking change?
Is not a breaking change, they will just need to update how they collect logs/coverage data
# How we teach this
Docs should be updated to reflect the new argument
# Unresolved questions
None
Answers:
username_1: Hi @username_0!
Sorry for the delayed response.
I think this can be solved in the same fashion as [tuist build](https://github.com/tuist/tuist/issues/2392).
We can do a follow-up PR for `tuist test` after `tuist build` is implemented (maybe @natanrolnik can chime in here if his PR can almost get us there to fix this)
Status: Issue closed
|
opengeospatial/ets-wfs11 | 157198437 | Title: tests wfs:wfs-1.1.0-Basic-DescribeFeatureType-tc3.1 failing
Question:
username_0: Hi,
I tried running the ets-wfs11 test suit against my WFS 1.1.0 webserver. All of the the test cases are passing except the 4 listed below:
1. wfs:wfs-1.1.0-Basic-DescribeFeatureType-tc3.1
2. wfs:wfs-1.1.0-Basic-DescribeFeatureType-tc3.2
3. wfs:wfs-1.1.0-Basic-DescribeFeatureType-tc3.3
4. wfs:wfs-1.1.0-Basic-DescribeFeatureType-tc3.4
These test cases expect a valid response even if the mandatory parameters 'SERVICE', 'VERSION' are missing in the POST request. But as per the OGC WFS 1.1.0 spec(http://portal.opengeospatial.org/files/?artifact_id=8339) section 14.4. These request parameters are mandatory for all operations(except 'VERSION' in GetCapabilities)
I verified the same with MapServer with request: http://demo.opengeo.org/geoserver/ows?service=wfs&version=1.1.0&request=DescribeFeatureType
It also returns <ExceptionReport> as my server is.
Please let me know if I am missing something or any update is required in the test suit!
Answers:
username_0: I even tried the reference server: http://cite.deegree.org/deegree-webservices-3.3.14/services/wfs110?service=WFS&request=DescribeFeatureType. It also returns Exception report
username_1: Thank you for reporting this issue. We will do further investigation.
username_2: For further investigations we need some more detailed information. Can you provide the URL of your service? Or at least the log files exported from the session?
We assume you executed the official WFS 1.1.0 Testuite [1] , selecting the profile 'Full'?
[1] http://cite.opengeospatial.org/teamengine
username_0: Hi
I realized the reason for this fix. Actually, there is difference in WFS 1.1.0 DescribeFeatureType POST and GET request. For DescribeFeatureType GET request both and 'service' and 'version' are mandatory. Whereas the DescribeFeatureType POST schema says 'service' and 'version' are optional and defaults to 'WFS' and '1.1.0'. So I have updated our service for POST and setting the respective default values in case they are missing.
If this difference is added explicitel, then this bug gets nullified.
I am using an onprem server. So not sure if the service can be hit publically.
Yes I am running the offcial WFS 1.1.0 TestSuite with'Full' profile.
username_1: Hi,
is your problem solved now or do you need any further help?
And would you propose any further actions regarding the ETS (e.g. updating the documentation)?
username_0: Hi,
My problem is solved but just wanted to confirm that why do we have this difference in GET and POST request. Is this difference documented in OGC spec for WFS 1.1.0.
username_1: Section 14 of the WFS 1.1.0 spec is about KVP.
Section 7.8 describes the common XML attributes including "version" and "service" (although, it is not clearly explained if they are mandatory or not). As the schema marks them as optional, they should be handled as optional parameters.
username_0: This issue was because of that missing clearification :) . I think this issue can be closed now!
Thanks for all your responses!
Status: Issue closed
|
lukeed/bee3d | 101570839 | Title: redirect to custom page on the last slide
Question:
username_0: i want to redirect to a custom page on the last slide by using the custom events found here : https://github.com/username_1/bee3d/wiki/Events
I cant figure out how to implement the script, can you please help me out ?
Best regards
Answers:
username_1: Hi prakashbholah, thanks for leaving your question here!
This is what you need to do:
1. cache the number of slides you have at the beginning of your script
2. on `activate` callback (every slide change), check the current active slides' index.
3. If on the last slide (# of slides minus 1), trigger your redirect
```javascript
var slider = new Bee3D(elem, {opts});
// (1)
var maxSlides = slider.el.slides.length;
// (2)
slider.el.on('activate', function(event) {
if (event.index == maxSlides - 1) {
// (3), we're on the last slide. go to new URL...
window.location.href = 'http://google.com';
}
});
```
username_0: hello username_1, i've tried this script and it does not work, maybe I am badly implementing the script.
Here's the full script used :
<script>
var element = document.querySelector('.slider_weeklies');
var slider = new Bee3D(element, {
effect: 'arc',
autoplay: {
enabled: true
},
listeners: {
keys: true,
touches: true
},
navigation: {
enabled: true,
},
/*
onInit: function(){
alert('Custom onInit callback function!')
},
*/
});
var slider = new Bee3D(elem, {opts});
// (1)
var maxSlides = slider.el.slides.length;
// (2)
slider.el.on('activate', function(event) {
if (event.index == maxSlides - 1) {
// (3), we're on the last slide. go to new URL...
window.location.href = 'http://google.com';
}
});
</script>
username_1: Delete this: `var slider = new Bee3D(elem, {opts});`
I just had that in there as an example.
username_0: cool man, now it works
thanks a lot :+1:
username_1: No problem!
Status: Issue closed
|
hmcmanus/yaml-validator-hook | 266848660 | Title: Support for multi-document YAML files
Question:
username_0: It may be as easy as changing [this line](https://github.com/username_1/yaml-validator-hook/blob/master/src/main/java/com/mcmanus/scm/stash/hook/YamlValidatorPreReceiveRepositoryHook.java#L133) to call `loadAll`. I can provide a PR if you see no problem.
Answers:
username_1: Hey Steve,
By all means please send a pull request. There is a shameful lack of tests in the plugin but lets do a bit of testing between us so that we have some confidence and I'll release a new version.
Thanks,
Hugh
username_2: Any update on this?
username_3: Hi, was wondering if this will be implemented? We support multi-document YAML files for our Spring Boot configuration and would be great if this plugin supported this.
username_1: Hey, I'm going to work on this separate to the latest upgrade. I attempted it with the snake yaml library I was using and had limited success even though it was in the documentation. Needs more testing but will come after the bitbucket server 6 upgrade.
username_4: Hi @username_1 ! I've submitted PR #19 that I think provides this feature. Please feel free to take a look and let me know what you think. :-)
username_1: Thanks @username_0, I'll take a look this week and cut a release so that people can start using it. Thanks for the contribution.
username_0: Thank @username_4!
Status: Issue closed
|
MDSplus/mdsplus | 400679388 | Title: Race condition in distributed client configuration
Question:
username_0: In the current alpha version a race condition arised creating several malfunctions and crashes in out current experiment configuration
After investigating the problem (thanks Cesare!!!) and discovered that at line 860 of treeshr/RemoteAccess.c the macro definition of IO_LOCK and IO_UNLOCK have been commented, therefore not using anymore the lock around the routines defined there. When we uncommented that macro definition, all the problems disappeared. Therefore the problem is very likely related to some race condition that has been uncovered removing those locks.
Answers:
username_1: I deactivated the lock as it did not seem to protect anything.
Yet to discover is what the lock actually protects and if it is the cure or if the real problem is somewhere else.
@username_0: Can you somewhat describe what you are doing when the issue is triggered. Can't be a simple read or write or dispatch as that would have been caught be the tests, helgrind at least.
username_0: The configuration that exposes the issue is a separate thread that writes on different nodes in distributed client configuration, concurrently with other data access programs. The effect is neat, so I strongly suspect that that lock is actually protecting something. In any case I will investigate further.
username_0: Looking at the routines that are protected via IO_LOCK macro, the number of routines recursively called is high and it is very difficult to state for sure that there are nowhere static structures that are not thread safe.
In any case, that lock protects several routines that call MdsIoRequest(). This routine sends a message over a mdsip connection and reads the response. If two MdsIoRequest() sharing the same connection id are interlaced (recall that two threads can share the same tree object) than you get unexpected responses. No surprise that in most cases the malfunctions occurred in in GetAnswerInfoTS().
username_1: But if this is the case the lock is simply in the wrong place.
Fist step would be to stick the lock into MdsIoRequest.
eventually, a individual lock for each connection would be better so multiple servers can be accessed in parallel.
username_0: I confirm that protecting MdsIoRequest alone resolves the issue. Let me know if you want we make the fix and issue a PR.
HOWEVER , UNLESS an exhaustive check of all the called routines has been done (I fear this is very difficult in practice, look at what happened now) AND a measurable improvement in performance is demonstrated, I feel a bit uncomfortable removing the other locks. As you have seen these are the worst errors to be detected and may happen after years of successful operation.
Timo, Tom, Josh, Andrea please advice!
Status: Issue closed
|
nrwl/nx | 772704036 | Title: @nrwl/nest:library dry-run is hanging extra long
Question:
username_0: <!-- Please do your best to fill out all of the sections below! -->
## Current Behavior
<!-- What is the behavior that currently you experience? -->
Takes about 1 min+ for dry run results to show
## Expected Behavior
<!-- What is the behavior that you expect to happen? -->
<!-- Is this a regression? .i.e Did this used to be the behavior at one point? -->
Can't recall which version has changed this behavior
## Steps to Reproduce
<!-- Help us help you by making it easy for us to reproduce your issue! -->
1. Create nx workspace
2. Install nrwl nest plugin
3. Run inside vscode `@nrwl/nest:library`
<!-- Can you reproduce this on https://github.com/nrwl/nx-examples? -->
<!-- If so, open a PR with your changes and link it below. -->
<!-- If not, please provide a minimal Github repo -->
<!-- At the very least, provide as much detail as possible to help us reproduce the issue -->
<!-- Remove this line -->
This issue may not be prioritized if details are not provided to help us reproduce the issue.
### Failure Logs
<!-- Please include any relevant log snippets or files here. -->
### Environment
<!-- It's important for us to know the context in which you experience this behavior! -->
<!-- Please paste the result of `nx report` below! -->
```
nx : 10.4.4
@nrwl/angular : Not Found
@nrwl/cli : 11.0.16
@nrwl/cypress : 11.0.16
@nrwl/devkit : 11.0.16
@nrwl/eslint-plugin-nx : 11.0.16
@nrwl/express : Not Found
@nrwl/jest : 11.0.16
@nrwl/linter : 11.0.16
@nrwl/nest : 11.0.16
@nrwl/next : 11.0.16
@nrwl/node : 11.0.16
@nrwl/react : 11.0.16
@nrwl/schematics : Not Found
@nrwl/tao : 11.0.16
@nrwl/web : 11.0.16
@nrwl/workspace : 11.0.16
typescript : 4.1.2
```
Status: Issue closed
Answers:
username_0: <!-- Please do your best to fill out all of the sections below! -->
## Current Behavior
<!-- What is the behavior that currently you experience? -->
Takes about 1 min+ for dry run results to show
## Expected Behavior
<!-- What is the behavior that you expect to happen? -->
<!-- Is this a regression? .i.e Did this used to be the behavior at one point? -->
Can't recall which version has changed this behavior
## Steps to Reproduce
<!-- Help us help you by making it easy for us to reproduce your issue! -->
1. Create nx workspace
2. Install nrwl nest plugin
3. Run inside vscode `@nrwl/nest:library`
<!-- Can you reproduce this on https://github.com/nrwl/nx-examples? -->
<!-- If so, open a PR with your changes and link it below. -->
<!-- If not, please provide a minimal Github repo -->
<!-- At the very least, provide as much detail as possible to help us reproduce the issue -->
<!-- Remove this line -->
This issue may not be prioritized if details are not provided to help us reproduce the issue.
### Failure Logs
<!-- Please include any relevant log snippets or files here. -->
### Environment
<!-- It's important for us to know the context in which you experience this behavior! -->
<!-- Please paste the result of `nx report` below! -->
```
nx : 10.4.4
@nrwl/angular : Not Found
@nrwl/cli : 11.0.16
@nrwl/cypress : 11.0.16
@nrwl/devkit : 11.0.16
@nrwl/eslint-plugin-nx : 11.0.16
@nrwl/express : Not Found
@nrwl/jest : 11.0.16
@nrwl/linter : 11.0.16
@nrwl/nest : 11.0.16
@nrwl/next : 11.0.16
@nrwl/node : 11.0.16
@nrwl/react : 11.0.16
@nrwl/schematics : Not Found
@nrwl/tao : 11.0.16
@nrwl/web : 11.0.16
@nrwl/workspace : 11.0.16
typescript : 4.1.2
```
username_1: Seeing the same. It's not only on dry run but also on the real run.
I think this deleteFile method is the issue. It goes over all the files in the repo, including all files in node_modules, .git, dist and tmp.
And it does it twice as it deletes 2 files.
Ideally this deleteFile method would take .gitignore into account.
https://github.com/nrwl/nx/blob/ddf765bbb7211f2e024af38ba6489ce64c4c2b01/packages/nest/src/schematics/library/library.ts#L53
username_2: That definitely looks like it's the culprit. We could probably just make a regular function that takes the host as an argument and call ```host.delete(`${options.projectRoot}/src/lib/${options.fileName}.spec.ts`)```.
Status: Issue closed
|
yowko/picsbed | 976616923 | Title: csharp-socket-proxy
Question:
username_0: 









 |
kubernetes-monitoring/kubernetes-mixin | 478714774 | Title: Change Alert Rule Duration ex. KubeNodeNotReady from 1h to 15m
Question:
username_0: Is it possible to change the duration of the alerts. We would like to lower the duration of alerts such as the KubeNodeNotReady from 1h to 15m where 1h is too long to wait to get notified of a bad node.
Currently we are just duplicating the rule so it alerts earlier but looking to see if it can be modified directly using jsonnet config.
Answers:
username_1: I think just lowering the default to 15m sounds fine :)
username_0: I just submitted PR #237
username_2: This can be closed as #237 is merged.
username_0: Closing since PR was merged.
Status: Issue closed
username_3: I think 15m is still too long for a few scenarios, such as network partition or node loss. The node controller already is responsible for assessing a delay (which is also technically tunable and in the future should probably be node specific) - once that condition is met on the node object or on pods there should be a firing alert. That is *not* the same as NotReady, but node.kubernetes.io/unreachable.
I'd suggest we add an alert that fires as soon as that taint is added - anyone who is experiencing continuous network disruption should already be tuning their grace period up so that it doesn't fire, or their workloads will get evicted. |
sul-dlss/libsys-webforms | 192360132 | Title: Incorrectly marking some barcodes as duplicates
Question:
username_0: Barcodes should be considered duplicates only if the barcode AND date match another barcode-date combination in the uni_updates table. I think perhaps now just the barcode is being considered a dup?
(I submitted a batch of 23,000 items as the second batch submitted today. A few thousand were rejected, but none should have been, there were no dups with the only other batch submitted the same day.)
Answers:
username_1: Key does not have to be up to minute, just today's date (YYYYMMDD).
Status: Issue closed
|
getsentry/sentry-fullstory | 729915720 | Title: FS.getCurrentSessionURL not ready
Question:
username_0: I'm trying to integrate our Sentry and Fullstory accounts using this package, but after following the setup instructions it appears that any Sentry exceptions that are captured on initial page load happen before Fullstory is fully bootstrapped. When I throw after the app has loaded, things appear to work correctly.
For more context: this is a React app, using Sentry and Fullstory JavaScript SDKs (via NPM), and manually calling `Sentry.captureException` from inside a component. When I load the app homepage and navigate to the route that mounts this component, I see correct links between Sentry/Fullstory. But when this component is mounted on initial page load (visit URL directly), I see the following:
- `FS.getCurrentSessionURL not ready` shown in JS console
- Sentry error page shows `current session URL API not ready` under fullStoryUrl
How can I make sure Fullstory is fully bootstrapped so I can capture errors that may happen on page load?
Answers:
username_1: Hey @username_0 thanks for raising this issue and apologies for the delayed reply. You've got it right, there are errors being thrown before FullStory is fully bootstrapped, and our `getCurrentSessionURL` function isn't defined until bootstrapping is complete - thus the "not ready" messages that are showing up in Sentry. There is a server call that needs to happen before `getCurrentSessionURL` is ready (see [this article ](https://help.fullstory.com/hc/en-us/articles/360032975773-How-does-FullStory-recording-work-to-recreate-my-users-experience-) for more details). However, calls to `FS.event` can be made prior to bootstrapping and will be recorded once FullStory is ready. We record the Sentry URL for all errors captured by Sentry, so you can search for them and find them in FullStory. Does this help you?
username_0: Hey @username_1 yes sir, very helpful, thank you! |
argoproj/argo-cd | 597453128 | Title: Ignore CRD status diff by default
Question:
username_0: # Summary
For some reason some Helm charts include CRD definitions with an empty status. Examples:
* https://github.com/helm/charts/blob/master/incubator/jenkins-operator/CRD/jenkins_v1alpha2_jenkinsinstance.yaml#L158
* https://github.com/helm/charts/blob/master/incubator/sparkoperator/templates/crds.yaml#L3971
* https://github.com/banzaicloud/logging-operator/blob/master/charts/logging-operator/crds/logging.banzaicloud.io_flows.yaml#L770
I suspect this is required to CRD delete conditions on every `kubectl apply`. The status is immediately modified by Kubernetes and Argo CD detecting it as a deviation ( https://github.com/argoproj/argo-cd/issues/3391, https://github.com/argoproj/argo-cd/issues/2239).
# Motivation
The workaround is to configure Argo CD to ignore CRD status. Because an issue is happening for multiple users lets include that configuration by default.
Status: Issue closed
Answers:
username_0: Agree. This can be closed. Thank you @darshanime ! |
tensorflow/tensorflow | 307873657 | Title: Fail to convert TransposeConv into tflite
Question:
username_0: I try to convert my model into tflite, shows the below error:
Check failed: weights_shape.dims(0) == 1 && weights_shape.dims(3) == 1 TransposeConv weights dimensions must begin and end with 1. Input weights "Variable_39/read_transposed" had shape [ 128, 3, 3, 64 ].
Does the TransposeConv weights only support [1, X, X, 1] ?
Answers:
username_1: Thank you for your post. We noticed you have not filled out the following field in the issue template. Could you update them if they are relevant in your case, or leave them as N/A? Thanks.
Have I written custom code
OS Platform and Distribution
TensorFlow installed from
TensorFlow version
Bazel version
CUDA/cuDNN version
GPU model and memory
Exact command to reproduce
username_0: System information
Have I written custom code (as opposed to using a stock example script provided in TensorFlow):
OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 16.04
TensorFlow installed from (source or binary):source
TensorFlow version (use command below): latest
Python version: 2.7
Bazel version (if compiling from source): Build label: 0.7.0
GCC/Compiler version (if compiling from source):N/A
CUDA/cuDNN version:N/A
GPU model and memory:N/A
Exact command to reproduceN/A:
username_2: This question is better asked on [StackOverflow](http://stackoverflow.com/questions/tagged/tensorflow) since it is not a bug or feature request. There is also a larger community that reads questions there. Thanks!
Status: Issue closed
username_1: It has been 14 days with no activity and the `awaiting response` label was assigned. Is this still an issue?
username_3: @aselle I have the same problem with a frozen and optimized semantic segmentation network that has transposed convolution by design. Any advices?
`F tensorflow/contrib/lite/toco/graph_transformations/propagate_fixed_sizes.cc:252] Check failed: weights_shape.dims(0) == 1 && weights_shape.dims(3) == 1 TransposeConv weights dimensions must begin and end with 1. Input weights "score2_full_weight/read/_108__cf__108_transposed" had shape [ 5, 16, 16, 5 ].`
Status: Issue closed
username_4: @username_0 have this issue resolved or not . I have the similar issue on Tensorflow R1.8 GPU version on Ubuntu16.04:
#2018-06-04 18:11:22.606761: I tensorflow/contrib/lite/toco/import_tensorflow.cc:1265] Converting unsupported operation: SplitV
#2018-06-04 18:11:22.606840: I tensorflow/contrib/lite/toco/import_tensorflow.cc:1265] Converting unsupported operation: Prod
#2018-06-04 18:11:22.606859: I tensorflow/contrib/lite/toco/import_tensorflow.cc:1265] Converting unsupported operation: Prod
#2018-06-04 18:11:22.617876: I tensorflow/contrib/lite/toco/graph_transformations/graph_transformations.cc:39] Before Removing unused ops: 328 operators, 590 arrays (0 quantized)
#2018-06-04 18:11:22.627740: I tensorflow/contrib/lite/toco/graph_transformations/graph_transformations.cc:39] After Removing unused ops pass 1: 326 operators, 586 arrays (0 quantized)
#2018-06-04 18:11:22.650427: I tensorflow/contrib/lite/toco/graph_transformations/graph_transformations.cc:39] Before general graph transformations: 326 operators, 586 arrays (0 quantized)
#2018-06-04 18:11:22.666719: F **tensorflow/contrib/lite/toco/graph_transformations/propagate_fixed_sizes.cc:252] Check failed: weights_shape.dims(0) == 1 && weights_shape.dims(3) == 1 TransposeConv weights dimensions must begin and end with 1. Input weights "import/transpose_23_transposed" had shape [ 64, 4, 4, 32 ].
#Aborted (core dumped)** |
galatea-associates/fuse-test-data-gen | 507285751 | Title: Extend tests of all objects to check number of attributes
Question:
username_0: Issue Description
---
If ever the number of attributes defining a domain object is increased, then this attribute will go untested and tests will still pass.
Design
---
To restrict this, add a new test for each domain object where the number of attributes (keys) in the dictionary is compared to an expected amount.
Some domain objects will have a configuration-provided-amount of "dummy fields" in future. This is not the immediate issue, and another can be raised at the relevant time, or this one amended.
Documentation Changes
---
n/a
Test Evidence
---
Evidence tests for each domain object passing with this measure included.
Validation in Develop
---<issue_closed>
Status: Issue closed |
MassBank/RMassBank | 724729064 | Title: QA/QC: Masses in peak list should be checked if in scan range
Question:
username_0: As soon the scan range is known (#216), we could make use of it to enhance data quality. Peaks below or above scan range limits should no be end up in the final peak list.
Answers:
username_1: This is _usually_ but not always valid, we have seen occasions where adducts were formed in the MS/MS and there are also cases of 2+ charged precursors (z=2) losing part of the charge and resulting in peaks above the precursor mass (m/z with z=1). Thus, care should be taken here, otherwise one might remove valuable information.
username_0: Yes, I did not say that we remove them automatically, but give the user the chance to decide similar to the failed peak list. We may develop some filtering rules to check the plausibility of peaks and it could be a starting point for a more in deep search on possible adducts as a side product. |
keybase/client | 836250067 | Title: kbfuse.fs was blocked from use because it is not from an identified developer on macos big sur
Question:
username_0: Keybase GUI Version: 5.6.2-20210202191343+d72cc00cd3
Macbook Air M1
macOS Big Sur 11.2.3 (20D91)
kbfuse.fs was blocked from use because it is not from an identified developer.
When I click on Open Anyway nothing happens and I can't get rid of it or install other apps that require me to allow them from system preferences. Only solution is to uninstall keybase |
canab/fd-ant-plugin | 160531019 | Title: Start debugger in HaxeDevelop
Question:
username_0: On of the other ant plugins had a checkbox that, when checked, would start the debugger for flash. It would be nice to have that
Answers:
username_1: You'd be better off requesting on SlavaRa/fdplugin-antpanel, this one's pretty much dead. |
OfficeDev/office-js-docs-pr | 389215049 | Title: getFileAsync returns empty PDF
Question:
username_0: <!---
Welcome to the Office Add-ins documentation repository.
To report an issue with the Office-Add-ins documentation, please provide the article URL and describe the issue below. Alternatively, if you want to submit a pull request with your recommended documentation changes, we will review your contributions and update our documentation accordingly.
If your issue is not related to the Office Add-ins documentation, please post it to one of the following channels instead:
- To ask a question about using the Office.js API, post your question to Stack Overflow and tag it with the "office-js" tag (http://stackoverflow.com/questions/tagged/office-js).
- To report an issue with the Office.js API or platform, create the issue in the OfficeDev/office-js repository (https://github.com/OfficeDev/office-js), which members of the product team monitor for customer-reported issues.
- To submit a feature request for the Office.js API or platform, post your idea to our User Voice page (https://officespdev.uservoice.com/), or if the feature request already exists there, add your vote for it.
-->
<!--- Provide a general summary of the documentation issue in the Title above -->
## Article URL
<!-- Provide the URL of the article that this documentation issue relates to -->
https://docs.microsoft.com/en-us/office/dev/add-ins/word/get-the-whole-document-from-an-add-in-for-word
## Issue
<!-- Provide a thorough description of the documentation issue -->
I'm trying to generate the file from PowerPoint presentation or Word document using the Add-in as described [here](https://docs.microsoft.com/en-us/office/dev/add-ins/word/get-the-whole-document-from-an-add-in-for-word), using exactly the example code from [getFileAsync()](https://docs.microsoft.com/en-us/javascript/api/office/office.document?view=office-js#getfileasync-filetype--options--callback-). I receive the PDF with correct document dimensions and number of pages, however all the pages are blank white. No text and background styling.
It is worth to mention that `Office.FileType.Text` returns correct text content. The file that I'm trying to save is just 1 slice. I'm using `msSaveOrOpenBlob(new Blob([fileContent], { type: 'application/pdf' }), 'test.pdf');` to open the document.
Office version: 16.0.11001.20091 64-bit, Office 365
Status: Issue closed
Answers:
username_0: <!---
Welcome to the Office Add-ins documentation repository.
To report an issue with the Office-Add-ins documentation, please provide the article URL and describe the issue below. Alternatively, if you want to submit a pull request with your recommended documentation changes, we will review your contributions and update our documentation accordingly.
If your issue is not related to the Office Add-ins documentation, please post it to one of the following channels instead:
- To ask a question about using the Office.js API, post your question to Stack Overflow and tag it with the "office-js" tag (http://stackoverflow.com/questions/tagged/office-js).
- To report an issue with the Office.js API or platform, create the issue in the OfficeDev/office-js repository (https://github.com/OfficeDev/office-js), which members of the product team monitor for customer-reported issues.
- To submit a feature request for the Office.js API or platform, post your idea to our User Voice page (https://officespdev.uservoice.com/), or if the feature request already exists there, add your vote for it.
-->
<!--- Provide a general summary of the documentation issue in the Title above -->
## Article URL
<!-- Provide the URL of the article that this documentation issue relates to -->
https://docs.microsoft.com/en-us/office/dev/add-ins/word/get-the-whole-document-from-an-add-in-for-word
## Issue
<!-- Provide a thorough description of the documentation issue -->
I'm trying to generate the file from PowerPoint presentation or Word document using the Add-in as described [here](https://docs.microsoft.com/en-us/office/dev/add-ins/word/get-the-whole-document-from-an-add-in-for-word), using exactly the example code from [getFileAsync()](https://docs.microsoft.com/en-us/javascript/api/office/office.document?view=office-js#getfileasync-filetype--options--callback-). I receive the PDF with correct document dimensions and number of pages, however all the pages are blank white. No text and background styling.
It is worth to mention that `Office.FileType.Text` returns correct text content. The file that I'm trying to save is just 1 slice. I'm using `msSaveOrOpenBlob(new Blob([fileContent], { type: 'application/pdf' }), 'test.pdf');` to open the document.
Office version: 16.0.11001.20091 64-bit, Office 365
username_1: @username_0 We use this repo to track problems with the documentation. For technical problems like this, we recommend that you raise the issue on Stack Overflow. That way, more experts will see it and the whole community will benefit from the answers that you get. Be sure to tag the question `office-js`.
Status: Issue closed
username_0: I consider it as a docs issue as well. This question is on stack as well. But I get your point. Thanks. |
type-challenges/type-challenges | 1101674512 | Title: 8 - MyReadonly
Question:
username_0: <!--
Notes:
๐ Congrats on solving the challenge and we are happy to see you'd like to share your solutions!
However, due to the increasing number of users, the issue pool would be filled by answers very quickly.
Before you submit your solutions, please kindly search for similar solutions that may already be posted. You can "thumb up" on them or leave your comments on that issue. If you think you have a different solution, do not hesitate to create the issue and share it with others. Sharing some ideas or thoughts about how to solve this problem is greatly welcome!
Thanks!
-->
```ts
type MyReadonly<T> = { readonly [K in keyof T]: T[K] }
``` |
justinmk/vim-gtfo | 515365553 | Title: [FR] allow opening tmux pane vertically
Question:
username_0: Currenlty `go{t,T}` always opens vertically split tmux pane; could we have an option to open horizontal split?
For bonus points there could also be an option that calculates the most reasonable split on the go automatically. |
jotegui/statReports | 124250997 | Title: Monthly VertNet data use report for 2015-9, resource ansp_orn
Question:
username_0: Your monthly VertNet data use report is ready!
You can see the HTML rendered version of the reports with this link:
http://localhost:8080/reports/8658262c-f762-11e1-a439-00145eb45e9a/201509/
Raw text and JSON-formatted versions of the report are also available for download from this link.
You can find more information on the reporting system, along with an explanation of each metric, here: http://www.vertnet.org/resources/usagereportingguide.html
Please post any comments or questions to http://www.vertnet.org/feedback/contact.html
Thank you for being a part of VertNet. |
chrislim2888/IP2Location-PHP-Module | 357169876 | Title: Not able to get CountryName, RegionName, CityName due to Warning: unpack() & Warning: fread()
Question:
username_0: I am geting below warning message. Due to these, I am not able to get CountryName, RegionName, CityName & Zip code. Even i am using DB9 BIN database.
Warning: unpack(): Type C: not enough input, need 1, have 0 in D:\Xampp\htdocs\IP2Location.php on line 1118
Warning: fread(): Length parameter must be greater than 0 in D:\Xampp\htdocs\iplocation\IP2Location.php on line 1049
```
/*
Default file I/O lookup
*/
$db = new \IP2Location\Database('./databases/IPV6-COUNTRY-REGION-CITY-LATITUDE-LONGITUDE-ZIPCODE.BIN', \IP2Location\Database::FILE_IO);
$records = $db->lookup('172.16.58.3', \IP2Location\Database::ALL);
```
I have placed the error screenshot in the attachment.

Answers:
username_1: Can you try to reupload your .BIN file and make sure it's under binary mode? It seems like your .BIN file is corrupted.
username_0: Yes, .BIN file is corrupted, I have downloaded a fresh DB9 BIN database. and uploaded again .BIN file on a server. After reupload I checked the script and it was started working fine.
Status: Issue closed
|
ballerina-platform/ballerina-lang | 727148418 | Title: Typo in `function` keyword produces a Bad Sad instead of a compilation error
Question:
username_0: **Description:**
When there is a typo in `function` keyword instead of a compilation failure, a Bad Sad is produced.
**Steps to reproduce:**
Create a ballerina file with the following code and run the build command.
```ballerina
public functio main() {
}
```
Answers:
username_1: Fixed in #26949
Status: Issue closed
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.