repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
Alluxio/alluxio
423755769
Title: S3 '_$folder$' files and Presto/Hive on read from Alluxio Question: username_0: **Is your feature request related to a problem? Please describe.** Assume we have a directory with Parquet files in S3 called `mydir`. Alluxio setup in EMR cluster with this directory as the UFS with read-only permission. The data in this directory is generated by S3 Hadoop related components that create _$folder$ files in the directory. These _$folder$ files should not be deleted. Presto and Hive in the EMR cluster query a table with `LOCATION 'alluxio://master_hostname:port/mydir'` When trying to query the data with Presto or Hive, I'm getting this error: `Query 20190321_132537_00026_4enx4 failed: Error opening Hive split alluxio://master_hostname:port/year=2019/month=01_$folder$ (offset=0, length=0): alluxio://master_hostname:port/year=2019/month=01_$folder$ is not a valid Parquet File` Hive doesn't have an option to ignore specific files based on regex, neither Alluxio. These files shouldn't be deleted. **Describe the solution you'd like** Add configuration to ignore files based on regex, or Add configuration to ignore _$folder files. **Describe alternatives you've considered** Can't find any solution. Answers: username_1: @username_0 Alluxio has a similar concept of using empty placeholder objects to represent directories. Instead of `_$folder$` we use `/` by default. The suffix is controlled by the `alluxio.underfs.s3a.directory.suffix` property. Can you try setting the property to `_$folder$`? Then Alluxio will understand that `month=01_$folder$` is a folder, not a file. To update the property, update `alluxio-site.properties` on all servers `alluxio.underfs.s3a.directory.suffix=_$folder$` then restart the cluster username_2: this feature request can be already achieved by existing alluxio properties. We will close this Issue in a few days if you don't have further request . username_0: It worked. Thanks! Status: Issue closed
fatih/vim-go
437955041
Title: gopls is very slow Question: username_0: ### What did you do? (required. The issue will be **closed** when not provided.) After the go modules update in Kubernetes, I've been noticing that :GoDef has been really slow. I assumed it was because guru/godef was searching the module cache instead of vendor. I tried setting `GOFLAGS=-mod=vendor` but that didn't help all that much. After trying to improve guru/godef I figured I'd try gopls with the new 1.20 release (before that I was on v1.16). After upgrading, `gopls` was even slower and consumed a lot of CPU on my machine. I saw the `gopls` proces running on my machine but vim was not jumping to the definition. If I wait about 3 - 5 mins it will sometimes go to the right place. I'm assuming that my environment is misconfigured but figured I'd open an issue to make sure I'm not missing anything. ### What did you expect to happen? :GoDef to be faster with the gopls update. ### What happened instead? :GoDef was much slower after using gopls. ### Configuration (**MUST** fill this out): * vim-go version: v1.20 * `vimrc` you used to reproduce (use a *minimal* vimrc with other plugins disabled; do not link to a 2,000 line vimrc): ``` let g:go_highlight_functions = 1 let g:go_highlight_methods = 1 let g:go_highlight_structs = 1 let g:go_highlight_operators = 1 let g:go_highlight_build_constraints = 1 let g:go_def_mode='gopls' let g:go_info_mode='gopls' let g:go_disable_autoinstall = 0 let g:auto_save = 0 let g:auto_save_in_insert_mode = 0 ``` * Vim version (first three lines from `:version`): ``` VIM - Vi IMproved 8.1 (2018 May 18, compiled Feb 19 2019 12:07:46) macOS version Included patches: 1-950 ``` I use macvim, not sure if that changes anything. * Go version (`go version`): ``` $ go version go version go1.12.1 darwin/amd64 ``` * Go environment (`go env`): ``` $ go env GOARCH="amd64" GOBIN="/Users/akim/go/bin" GOCACHE="/Users/akim/Library/Caches/go-build" GOEXE="" GOFLAGS="-mod=vendor" GOHOSTARCH="amd64" [Truncated] GOOS="darwin" GOPATH="/Users/akim/go" GOPROXY="" GORACE="" GOROOT="/usr/local/Cellar/go/1.12.1/libexec" GOTMPDIR="" GOTOOLDIR="/usr/local/Cellar/go/1.12.1/libexec/pkg/tool/darwin_amd64" GCCGO="gccgo" CC="clang" CXX="clang++" CGO_ENABLED="1" GOMOD="" CGO_CFLAGS="-g -O2" CGO_CPPFLAGS="" CGO_CXXFLAGS="-g -O2" CGO_FFLAGS="-g -O2" CGO_LDFLAGS="-g -O2" PKG_CONFIG="pkg-config" GOGCCFLAGS="-fPIC -m64 -pthread -fno-caret-diagnostics -Qunused-arguments -fmessage-length=0 -fdebug-prefix-map=/<KEY>build004257707=/tmp/go-build -gno-record-gcc-switches -fno-common" ``` Answers: username_1: I've been able to use gopls successfully in kubernetes without such a long delay, but it does take gopls a while to initialize, so the first operation can be slow. After that it should be much faster. But perhaps you're working in a different kubernetes package than I've tested in. Can you give more detail about what definition you're trying to jump to and from where? Also, gopls is moving pretty fast. Have you updated it recently (`:GoUpdateBinaries gopls`)? username_0: What's the usual expected time for this? I updated to the latest gopls binary and `:GoDef` still does not work. The `gopls` process has been running on my machine for the last 3 hours though, is that how long it usually takes to initialize? A simple example that doesn't work for me is running `:GoDef` on [app.NewControllerManagerCommand](https://github.com/kubernetes/kubernetes/blob/ba81a5409affb6b861d5994c1e2f7c74a05826e5/cmd/kube-controller-manager/controller-manager.go#L38) in the kube-controller-manager. But generally I can't get `:GoDef` in the repository to work. Does gopls still work for you after you pulled in the latest go modules changes in the Kubernetes repo or are you using `:GoDef` from Kubernetes vendored in another repository? username_1: Hey, Andrew! I tested this after my first response using latest kubernetes. It took a couple of minutes for gopls to initialize, but after that it was snappy. A couple of things to be aware of: gopls expects your working directory to be within the module you're working in while in module mode. You can get some useful logs with `:let g:go_debug=['gopls']` and then using `:echo g:go_lsp_log`. It could be quite extensive, but may help us understand what's happening on your system. Have you tried gopls with any non-k8s repos? username_0: Yes, it works with other repos. I can GoDef to kubernetes from another repo as a vendored dep. I'll try enabling logs like you mentioned and get back to you. Thanks! username_0: Did you mean `:let g:go_debug=['lsp']`? `gopls` wasn't working so wanted to double check :) username_1: Yes. Sorry about that. username_1: I think I'm able to duplicate what you're seeing in module mode. But GOPATH mode it works fine: initialization takes about 30-40 seconds. I was able to get module to work by running `go list` before starting up Vim. I'll clear my cache and see if I can recreate the problem in module mode, though, in case there's something that needs to change in vim-go or a bug needs to be created for gopls. username_1: After clearing my cache, I can't duplicate the failures I thought I'd seen. I'll wait for your output from `:echo g:go_lsp_log` before continuing. username_0: Here's the output from `:echo g:go_lsp_log` https://gist.github.com/username_0/452d371f760b2af5b49c2532c7593c8f username_1: Thanks, Andrew! It looks like you tried jumping to the definition really quickly after opening the first Go file. I'd expect that first jump to take about 30-40 seconds (because it has to wait on gopls to initialize). How long did wait for it to complete? Also, for us to get an idea of what's a reasonable time for `gopls` to initialize, can you try running `time gopls query definition /Users/akim/go/src/k8s.io/kubernetes/cmd/kube-controller-manager /controller-manager.go:#1138`? (you might need to adjust `#1138` - it's the byte offset to evaluate in the source file. You can use `g ctrl+g` in vim to get the byte offset of the cursor). username_0: I ran it for about 30m and it never returned so I killed the process. I ran it on byte offset 1193 which should GoDef to `NewControllerManagerCommand`. ``` $ time gopls query definition /Users/akim/go/src/k8s.io/kubernetes/cmd/kube-controller-manager/controller-manager.go:#1193 ^C real 28m4.396s user 28m3.322s sys 0m5.765s ``` I ran it with `-logfile` as well but there were no logs either (in fact, the log file wasn't even created either). username_1: The good news is that the problem doesn't seem to be vim-go, then. The bad news is that I'm not quite sure what's wrong. Have you tried removing your `$GOPATH/pkg` directory and also cleaning your module cache (`go clean -modcache`)? If you do either of those, it's probably a good idea to run `GO111MODULE=on go list` in `/Users/akim/go/src/k8s.io/kubernetes` before trying the `gopls query definition` again. I'd recommend updating to the latest `gopls` before doing any of that (`:GoUpdateBinaries gopls` in vim-go should make that pretty easy). If none of that works, it's probably time to create an issue for `gopls` in github.com/golang/go. username_0: Thanks Billie, let me try that and get back to you! username_0: This did it for me. Thanks Billie! Status: Issue closed username_2: hey @username_1, I met the same problem. if I work with module mode, it lags for a quite long time. Once I `export GO111MODULE=off`, it runs much faster. But i will met a error: `could not import k8s.io/kubernetes/cmd/kube-apiserver/app (invalid package name: "")`. Is there anyway to detect the vendor folder firstly and faster when in module mode? username_2: ![Screen Shot 2021-06-21 at 11 55 41 AM](https://user-images.githubusercontent.com/18364341/122705291-fd7a4a00-d287-11eb-9a24-ba9d334cb2b6.png) code looks like.
Bouni/luxtronik
646314210
Title: Wrong device_class for temperature sensors Question: username_0: I just noticed that all my luxtronik temperature sensors have the state attribute `device_class: pressure`. Before using your integration, I had an integration written by myself, which had the same name and same entity IDs. So there is a chance that it might be a leftover from back then (many months ago). Can you confirm the device_class in your setup? I currently don't have any customizations in Home Assistant. Answers: username_1: That's inded a bug, I'll fix it asap! Status: Issue closed
reiserm/Xana
667934621
Title: comparing notebook cells with nbval Question: username_0: We have an issue with the jupyter notebook we use for testing. Cell comparison fails. I think it is still related to the absolute paths and progress bars that are printed by the notebook. I tried to add another regex to the sanitize file to exclude the home folder on macs `/Users/` but apparently there is another issue. There is another issue with the newest pytest version 6.0 but as far as I understand, Thomas is dealing with that already in PR #152 of nbval. I set the requirement in `setup.py` to <= 5.4 to make the test work, but the cell comparison issue persisits. Answers: username_0: closed with PR #9. Status: Issue closed
jlippold/tweakCompatible
421987016
Title: `FloatyDock` working on iOS 12.1.1 Question: username_0: ``` { "packageId": "com.synnyg.floatydock", "action": "working", "userInfo": { "arch32": false, "packageId": "com.synnyg.floatydock", "deviceId": "iPhone8,1", "url": "http://cydia.saurik.com/package/com.synnyg.floatydock/", "iOSVersion": "12.1.1", "packageVersionIndexed": true, "packageName": "FloatyDock", "category": "Tweaks", "repository": "BigBoss", "name": "FloatyDock", "installed": "1.4.1", "packageIndexed": true, "packageStatusExplaination": "This package version has been marked as Likely working based on feedback from users in the community. The current positive rating is 66% with 4 working reports.", "id": "com.synnyg.floatydock", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.1.4", "shortDescription": "iPad dock power, for all !", "latest": "1.4.1", "author": "SynnyG", "packageStatus": "Likely working" }, "base64": "<KEY> "chosenStatus": "working", "notes": "" } ```<issue_closed> Status: Issue closed
smartcontractkit/chainlink
321485897
Title: if statement logic in BeginRunAtBlock function Question: username_0: Should this if statement: ``` if input.Amount != nil && store.Config.MinimumContractPayment.Cmp(input.Amount) > 0 { msg := fmt.Sprintf( "Rejecting job %s with payment %s below minimum threshold (%s)", job.ID, input.Amount, store.Config.MinimumContractPayment.Text(10)) run = run.ApplyResult(input.WithError(errors.New(msg))) } ``` Be this instead? ``` if input.Amount == nil || store.Config.MinimumContractPayment.Cmp(input.Amount) > 0 { msg := fmt.Sprintf( "Rejecting job %s with payment %s below minimum threshold (%s)", job.ID, input.Amount, store.Config.MinimumContractPayment.Text(10)) run = run.ApplyResult(input.WithError(errors.New(msg))) } ``` Currently runs with nil input.Amount values will always be started. Answers: username_1: We are planning on allowing payment for other types of initiators soon, so I think it makes sense to leave it there for now. Status: Issue closed
stal111/Forbidden-Arcanus
653394730
Title: [1.16.1] Crash while rendering Forbidden And Arcanus items in JEI. Question: username_0: Title^ **MC Version:** 1.16.1 **Mod Version:** 1.16.1-1.0-beta-1 **Forge:** 32.0.59 Crash-Report: https://hastebin.com/udogipoliz.pl Answers: username_1: Does this still crash in the newest beta? username_0: Nope looks like its fixed! Status: Issue closed
codalab/codalab-competitions
1139184980
Title: How to host a private competition on public codelab? Question: username_0: Is it possible to host a private competition on the public codalab instance? I would like to host a competition for my course (i.e., only students in the course can compete) but I do not need/want to host my own server. Thanks. Status: Issue closed Answers: username_0: Nevermind - I see that competitions are private by default. Thanks!
oppia/oppia
1182817385
Title: Cyclic Imports Question: username_0: The linter found cyclic imports when run locally: ``` ---------------------------------------- Please fix the errors below: ---------------------------------------- ************* Module core.domain.topic_fetchers_test core/domain/topic_fetchers_test.py:1:0: R0401: Cyclic import (core.domain.exp_services -> core.domain.opportunity_services -> core.domain.suggestion_services -> core.domain.suggestion_registry) ``` In this case: * `core.domain.exp_services` imports `core.domain.opportunity_services`: https://github.com/oppia/oppia/blob/d010f1abfa257eac4be09b34ba4f5986cfb6f4ae/core/domain/exp_services.py#L52 * `core.domain.opportunity_services` imports `core.domain.suggestion_services`: https://github.com/oppia/oppia/blob/d010f1abfa257eac4be09b34ba4f5986cfb6f4ae/core/domain/opportunity_services.py#L29 * `core.domain.suggestion_services` imports `core.domain.suggestion_registry`: https://github.com/oppia/oppia/blob/d010f1abfa257eac4be09b34ba4f5986cfb6f4ae/core/domain/suggestion_services.py#L33 * `core.domain.suggestion_registry https://github.com/oppia/oppia/blob/d010f1abfa257eac4be09b34ba4f5986cfb6f4ae/core/domain/suggestion_registry.py#L29 This creates a cycle of imports. How has this gone undetected until now? Our linter should have caught this before the changes were merged, and this cycle of imports should have caused problems before now. Answers: username_1: Just a note -- @username_2 has kindly offered to take a quick look at this. Thanks @username_2! username_2: @username_0 Getting this error in other files as well: https://github.com/oppia/oppia/runs/5715305440?check_suite_focus=true And in two other files while running linter locally. ************* Module core.platform.bulk_email.mailchimp_bulk_email_services_test core/platform/bulk_email/mailchimp_bulk_email_services_test.py:1:0: R0401: Cyclic import (core.platform.auth.firebase_auth_services -> core.platform.models) ************* Module core.domain.object_registry core/domain/object_registry.py:1:0: R0401: Cyclic import (core.domain.event_services -> core.domain.feedback_services) username_2: @username_0 @username_1 This is what I discovered: To check for cyclic import all the files causing cyclic import must be loaded in the linter, only then linter will identify the cyclic error. Right now, I think we run linter in shards, so there's a possibility that all the files causing cyclic import error might not be in that shard's filegroup and therefore our linter does not detect any error. username_0: @username_2 thanks, that makes sense. I started digging into this because of cyclic import errors I'm seeing on #15219, which just changes the codeowners file. Do you have any thoughts as to why changes to the codeowners file would cause these errors to surface? username_0: @username_2 Actually, I don't think the shards can explain this. I also see cyclic imports intermittently when running the linter on all files. Further, all of `core/domain` is in one shard, so sharding shouldn't have affected the cycle at the top of this issue. I think this error might be arising despite our code running correctly because pylint considers imports to be cyclic even if some imports are not at the top level of the file. For example, we sometimes put imports in functions specifically to avoid cyclic imports, but pylint still considers these to be cyclic imports. To see this, create a folder `tmp` and inside, create `a.py` and `b.py`. Fill `a.py` with this: ```python CONSTANT_A = 'constant a' def func_a(): from b import func_b func_b() print('a') if __name__ == '__main__': func_a() ``` Fill `b.py` with this: ```python from a import CONSTANT_A def func_b(): print('b') print(CONSTANT_A) ``` Then if you `cd` into `tmp`, you should be able to execute `a.py` just fine: ```console $ python a.py b constant a a ``` However, pylint will find errors: ``` $ pylint * ************* Module a a.py:1:0: C0114: Missing module docstring (missing-module-docstring) a.py:3:0: C0116: Missing function or method docstring (missing-function-docstring) a.py:4:4: C0415: Import outside toplevel (b.func_b) (import-outside-toplevel) ************* Module b b.py:1:0: C0114: Missing module docstring (missing-module-docstring) b.py:3:0: C0116: Missing function or method docstring (missing-function-docstring) b.py:1:0: R0401: Cyclic import (a -> b) (cyclic-import) ----------------------------------- Your code has been rated at 4.55/10 ``` (To get pylint to successfully import your modules, you may need to set your `PYTHONPATH` to the path to `tmp`.) Notice that even though Python doesn't raise a cyclic import error and `a.py` runs just fine, pylint finds a cyclic import because it doesn't account for the fact that the `from b import func_b` import is inside a function. Status: Issue closed
cmangos/issues
468741167
Title: 🐛 [Bug Report]Bot as priest stop buffing my main char shaman with Power Word: Fortitude Question: username_0: ## 🐛 Bugreport <!-- Describe your issue in detail. Include screenshots if needed. Give us as much information as possible. --> Bot as priest stop buffing my main char shaman with Power Word: Fortitude But other bots he buffs with Power Word: Fortitude ### Expected behavior <!-- How should it work + proof --> ### Version & Environment <!-- Client Version - is required Valid values are: - "3.3.5a" (WOTLK) --> Client Version: <!-- Commit Hash - is required Valid values are: - [WOTLK](https://github.com/cmangos/mangos-wotlk/tree/13992) To find XXXX use "git log -1 --format=%H" in your local CMaNGOS repo --> CMaNGOS Repo & Commit Hash: <!-- Database Version - is required Valid values are: World DB: WoTLKDB v1.2+ 'Ymiron' for CMaNGOS-WOTLK 13992 - [WOTLK](https://github.com/cmangos/wotlk-db/tree/XXXX) To find XXXX use "git log -1 --format=%H" in your local Database repo --> Database Repo & Commit Hash: <!-- Operating System - optional Valid values are: - Win 10 --> Operating System: ### Steps to reproduce 1. 2. 3. 4. ... 5. Profit ### Crashlog <!-- If this is a crash report, include the crashlog from a debug build with https://gist.github.com/) --> - None Answers: username_1: @username_0 Please provide **all** of the information requested in the bug form. username_1: Closing this out as we have heard no response. We will consider reopening it if the requested information can be provided. Status: Issue closed
anuraghazra/github-readme-stats
736034073
Title: Github Status card data not updated Question: username_0: ![image](https://user-images.githubusercontent.com/32858162/98105115-78d7e880-1ebd-11eb-9cb3-8b4f48937820.png) - I did some new commits and went to this url https://github-readme-stats.vercel.app/api?username=username_0&show_icons=true&theme=tokyonight - It shows the updated git stat card ![image](https://user-images.githubusercontent.com/32858162/98105365-e421ba80-1ebd-11eb-8656-85fcd8fde912.png) The data's are updated here but if i refreshed the profile page https://github.com/username_0 the new data's are not updated i also waiting more than 30 mins to check if updates or not. But still not updated. Answers: username_0: I tried in online markdown viewer and i shows me the updated data ![image](https://user-images.githubusercontent.com/32858162/98106912-c7868200-1ebf-11eb-98a0-35f2fe3bff39.png) username_0: sorry guys i forgot to add private repo count my bad Thanks Naveen Status: Issue closed
KonstantinEger/Bau-Abrechnungsprogramm
787694488
Title: Cannot close header edit inputs when no change Question: username_0: **Describe the bug** When double clicking on a property in the header of the project view, it turns into an input field. When the `onchange` event doesn't get fired, it can never turn back. **To Reproduce** Steps to reproduce the behavior: - double click on a property in the header of the project view - **don't** change the data, just press enter when the input is focused - nothing changes **Expected behavior** When pressing enter, even when the input doesn't change, it should turn back into text.<issue_closed> Status: Issue closed
type-challenges/type-challenges
927859478
Title: 10 - Tuple to Union Question: username_0: <!-- Notes: 🎉 Congrats on solving the challenge and we are happy to see you'd like to share your solutions! However, due to the increasing number of users, the issue pool would be filled by answers very quickly. Before you submit your solutions, please kindly search for similar solutions that may already be posted. You can "thumb up" on them or leave your comments on that issue. If you think you have a different solution, do not hesitate to create the issue and share it with others. Sharing some ideas or thoughts about how to solve this problem is greatly welcome! Thanks! --> ```ts // your answers type TupleToUnion<T extends any[]> = T[number] ```
roccopung/export-radio
781415240
Title: Chat doesn't work online Question: username_0: In the meantime that we are waiting for Max reply, today I have tried to put the website online on a free hosting platform, to simulate how it would be when we go online in a week. For some reason, the chat still doesn't work, and I can't log in to social media to join the conversation. Do you eventually have any idea of why this happens? In case that can help, you can find the website here [](http://roccomaster.altervista.org/) You can access the Panel here [](https://en.altervista.org/) with User and Pss and these are the credentials to go in via FTP: Host: ftp.roccomaster.altervista.org User: roccomaster Pss: Quomorasato1? Port: 21 Answers: username_1: Don't post credentials here, this is a public website. I removed them from your comment for you. The chat works fine for me (including logging in via Twitter) once I disable my Adblocker and tracking blocker in Firefox. I suspect this is an issue on their side, I don't know how well-maintained that project is. Is there a reason you're not embedding the Twitch chat? That one works fine in my experience. username_0: Thanks for removing it, I know it is public but just didn't consider it an issue. Very bad idea anyway! The reason was that you would need a Twitch account to log in to the chat and participate, while here was easier to join in... However, it looks like it creates some problems, and even if I disable Adblocker and tracking in Chrome, it doesn't work. I know you used it on another website, did it create any problem there? username_1: I made a [test case here](https://maxkoehler.com/test/) which seems to work in Chrome, I suspect because my site uses https. You could set that up for the Export Radio site and probably improve your chances, but honestly the component just seems a little dead to me. Might make your life easier to go back to Twitch. The fact that it forces you to log in isn't that bad in my opinion - it's actually a safety feature. If people have to log in through Twitch, they can be banned by Twitch for spam, abuse, illegal content and so on. This is why we used it on [this site](https://www.loveactually.works/), and it worked just fine. username_0: Ooh so that wasn't tlk.io that you used! Ok then, what you say makes total sense to me and I will forward the information. I stop bothering you for today. Thank you as always and have a nice evening! username_0: Changed to Twitch. Just much better. I close this issue. Status: Issue closed
hyrise/hyrise
355074740
Title: Deleting values inserted in the same transaction fails Question: username_0: <pre> (debug)!> <strong>load src/test/tables/int.tbl t</strong> Loading src/test/tables/int.tbl into table "t" ... (debug)> SELECT * FROM t; === Columns | a| | int| === Chunk 0 === | 123| | 1234| | 12345| === 3 rows total Execution info: [PARSE: 4 µs, TRANSLATE: 184 µs, OPTIMIZE: 95 µs, COMPILE: 39 µs, EXECUTE: 104 µs (wall time) | QUERY PLAN CACHE HITS: 0/1 statement(s)] (debug)> <strong>begin</strong> New transaction (2) started. (debug)> <strong>INSERT INTO t VALUES (100);</strong> === 0 rows total Execution info: [PARSE: 3 µs, TRANSLATE: 57 µs, OPTIMIZE: 65 µs, COMPILE: 79 µs, EXECUTE: 164 µs (wall time) | QUERY PLAN CACHE HITS: 0/1 statement(s)] (debug)> <strong>DELETE FROM t WHERE a = 100;</strong> <strong>libc++abi.dylib: terminating with uncaught exception of type std::logic_error: src/lib/sql/sql_pipeline.cpp:187 No result tables</span> Abort trap: 6 </pre>
gkopff/logback-raygun
103496394
Title: Make stack for log without an exception configurable Question: username_0: We are using `logback-raygun` in an [Akka](akka.io) project. Akka provides a logging interface that wraps SLF4J, meaning that calling `log` in an akka project results in stack traces containing only the SLF4J call from inside Akka. For us this looks like: ``` akka.event.slf4j.Slf4jLogger$$anonfun$receive$1$$anonfun$applyOrElse$2.apply$mcV$sp in Slf4jLogger.scala:71 ``` It would be good if there was some means to configure `RaygunAppender.locateCallSite`, either to return the full stack trace or to alter the filters. This is non-urgent for us just now so I've not started working on anything. Do you think this would be useful? How would you like to see it implemented? Answers: username_1: Would it be sufficient, do you think, to add to the configuration a set of classname prefixes that simply also get examined in the `for (StackTraceElement ...` loop within `locateCallSite()`? Class names (or prefixes) for all of the Akka logging glue code would have to be included in your configuration file - when we hit the first stack trace element that wasn't flagged, we'd declare that the call site ... Effectively, instead of the static `if` statements for `FBI` and `LOGBACK`, we iterate over the set of `(FBI, LOGBACK, custom)`. username_0: I'm going to close this to clear it from my context. Clearly there's not a demand for it! Status: Issue closed
vpc-ccg/pamir
644486013
Title: Pamir hanging at mrsfast_anchor_wg_map Question: username_0: Hi, I am using your latest pamir commit 148e2cffbdc9d241129830950466d234ab727735 and the snakemake pipeline broke down with an **Error in rule mrsfast_anchor_wg_map** `Error: Cannot Open the file <path/to/reference>/chr21_ins.fa.index` In the folder where I provided the reference (chr21) is the chromosome's FASTA and it seems that pamir created a FAI (FASTA index) but there is no ".index" file. Can you see what is going on here? Also, since the fix of issue https://github.com/vpc-ccg/pamir/issues/47 I noticed tons of warnings about inconsistent paths in the snake rules like `path /<workdir>//analysis/<project-name>/001-pamir-remove-concordants/samplename17/samplename17.stat contains double '/'. This is likely unintended. It can also lead to inconsistent results of the file-matching approach used by Snakemake.` Not sure if this is related to the downstream breakdown of mrsfast_anchor_wg_map. My config.yaml looks like: ``` path: /<path/to/workdir>/ raw-data: raw-data-links reference: /<path/to/reference>/chr21_ins.fa population: project-name input: "samplename1": - S0001.bam [...] "samplename50": - S0050.bam ``` Answers: username_1: Hi, early in the development we decided not to index the provided index in the snakemake pipeline, because it was likely to override existing index. Can you try indexing the reference by running ``` mrsfast --index genome.fa ``` I thought I resolved the double "/" errors. I will look in to it. username_0: Okay, I sent a pull request https://github.com/vpc-ccg/pamir/pull/53 that successfully handled the missing `reference.fa.index`. --- The other matter was the relative paths: if `path` in the config.yaml does not end with a "/" then the Snake-pipeline merges the relative paths of e.g. `path+reference` without the "/" delimiter and messes up the output repositories. The quick solution is to hard-code a "/" to the end of `path` in config.yaml. Status: Issue closed
rex706/SAM
511174617
Title: Suggestions: Ban Check, Drop-Down Timeout Question: username_0: Hey, had a couple ideas to enhance the Account Manager. 1) Add an input field for a [Steam API Key](https://steamcommunity.com/dev/registerkey) which would then check all accounts for a Game Ban and mark the accounts. A single API request can check up to 100 accounts at a time and requires only the ID64 though a full URL works, too. 2) A dropdown menu for the most common cooldowns you can have in CS:GO. * 30 minutes * 2 hours * 24 hours * 7 days In addition to those a 21 hours option for when you got 2 wins in matchmaking and can't play for a day. Looking forward to your thoughts on this. Answers: username_1: Having an input field for the user to input their on API key is a great idea as I'm sure the one that has been hard coded from a long time ago is most likely over used. By 'Game Ban', do you mean VAC? Or does the API have awareness of other kinds of bans as well? I'll take a look into the documentation for this to see what options there are. Good to know the request can take up to 100 accounts at a time as well. You can already set a 'timeout' on any given account through the context menu, but I guess I could add some quick options like those to make it easier than fiddling with the spin boxes. username_0: Those spin boxes are a pain in the ass, which makes me use them literally never. It's annoying to have to set the times individually. A drop down for the most common would be awesome to have (I could finally get rid of my Google Spreadsheets lmao) username_2: This would be super awesome. username_0: Will there be an ETA for a new update? username_1: I'm almost finished with #65. I think the last bit I'm missing is deleting selected accounts while in this view mode, which I'll work on today. I'll likely push out an update after that. I forgot to include the issue in the commit so it would display in this thread, but I included the quick timeout options (including 21 hours) as well. Status: Issue closed
getdnsapi/getdns
216996183
Title: Address Selection API? Question: username_0: So I have this code that calls getdns_address() and obtains a response. Parsing the response, I can get a list of addresses. What I want to do next, of course, is set up a connection to the selected server. Which API am I supposed to use to choose among the multiple addresses returned from the DNS? Am I supposed to write an implementation of RFC 6724 in the application? Status: Issue closed Answers: username_1: Yes, I agree, we've started work on this too. We have a requirements document here: https://docs.google.com/document/d/1gCUk-7vd5Vo10-VFD5jHjpI9QUi1j-Tm7JtWEK1V4_o/edit?usp=sharing . I've pushed some initial work here: https://github.com/getdnsapi/connectbyname A first implementation of https://tools.ietf.org/html/draft-ietf-v6ops-rfc6555bis-00 would be a good start I think. Want to join?
react-native-picker/picker
945230949
Title: Picker on Android leads to recursive loop Question: username_0: <Picker.Item label="Male" value="male" /> <Picker.Item label="Female" value="female" /> <Picker.Item label="Other" value="other" /> </Picker> ``` And my change handler looks like this: ``` handlePickerChange = (t, _) => { console.log("handlePickerChange", t); this.setState({gender: t}); } ``` I've set an initial value in the state via the constructor ``` constructor(props) { super(props); this.state = { isLoading: false, patient_id: null, formData: { mobile_number: "", full_name: "", age: "" }, gender: "other" }; } ``` And once data is fetched and the state is updated, the picker goes into an infinite loop ``` LOG handlePickerChange other LOG handlePickerChange female LOG handlePickerChange other LOG handlePickerChange female LOG handlePickerChange other LOG handlePickerChange female LOG handlePickerChange other LOG handlePickerChange female LOG handlePickerChange other LOG handlePickerChange female LOG handlePickerChange other LOG handlePickerChange female LOG handlePickerChange other LOG handlePickerChange female LOG handlePickerChange other ``` Answers: username_1: We are having this same issue. Also with a "gender" picker by the way, but I'm guessing that's coincidence :) username_1: Looking at the commit history, this one might be related: https://github.com/react-native-picker/picker/commit/363baa88614f44422eebf41333a3a2d68914e0b5. Not sure though. That's your commit @username_2, what do you think? username_2: Hi, that commit only replaced legacy ref usage in android picker, so I doubt it has broken sth, as similar logic is included in ios picker. I have tested that commit before, in example app, and it was working as expected. username_0: @username_1 I worked around it by rendering the form only after the state value fetched and set ``` {!this.state.isLoading ? ( <PickerControl label="Gender" selectedValue={this.state.gender} onValueChange={this.handlePickerChange} /> ) : ( <></> )} ``` username_1: @username_0 Thanks for the additional information. I'm not sure what the essence of your work-around is and how it would translate to the environment I'm using. @username_2 I just tried with 1.16.2 (instead of 1.16.3) and there I don't seem to have the issue... I guess that narrows it down to two commits ;)
MathijsBlok/ngx-alerts
604605782
Title: Import warning with Angular 9 Question: username_0: I'm getting this import warning with Angular 9: `Warning: Entry point 'ngx-alerts' contains deep imports into '/Users/*****/*****/*****/node_modules/rxjs/internal/operators'. This is probably not a problem, but may cause the compilation of entry points to be out of order.` The correct import path should be `import { debounceTime, map, startWith, tap } from 'rxjs/operators';` Answers: username_1: Thanks, fixed it in version 9.0.2 Status: Issue closed
mmistakes/minimal-mistakes
172945080
Title: add sidebar nav instructions to documentation Question: username_0: <!-- Before submitting please search open and closed issues at https://github.com/username_1/minimal-mistakes/issues to avoid duplication. Feel free to use the following as a template and remove or add fields as you see fit. You can convert `[ ]` into `[x]` to check boxes. --> - [ x ] This is a question about using the theme. - [ ] I believe this to be a bug with the theme --- not Jekyll, GitHub Pages or one of the bundled plugins. - [ ] This is a feature request. - [ ] I have updated all gems with `bundle update`. - [ ] I have tested locally with `bundle exec jekyll build`. --- ## Environment informations <!-- Please include theme version, `github-pages --version`, and the operating system you are on or tested with. --> - **Minimal Mistakes version: 3.4.3** - **`github-pages`** --- ## Expected behavior <!-- I cannot seem to find how to enable the sidebar navigation the way it is done in your docs collection. Could this be added to the documentation? --> Answers: username_1: Currently undocumented because I built it as a "one off" for the theme's documentation section. Wasn't really meant to be something as part of the theme as it's not the most straight forward thing to do. It involves manually listing out all of the links in a `_data` file and then referencing that set of navigation links with `sidebar: navName` in the YAML Front Matter of whatever post/page/collection you want it to appear on. For example, the menu for the [theme's documentation](https://username_1.github.io/minimal-mistakes/docs/quick-start-guide/) was built by: **`_data/navigation.yml`** ```yaml # documentation links docs: - title: Getting Started children: - title: "Quick-Start Guide" url: /docs/quick-start-guide/ - title: "Structure" url: /docs/structure/ - title: "Installation" url: /docs/installation/ - title: "Upgrading" url: /docs/upgrading/ - title: Customization children: - title: "Configuration" url: /docs/configuration/ - title: "Navigation" url: /docs/navigation/ - title: "UI Text" url: /docs/ui-text/ - title: "Authors" url: /docs/authors/ - title: "Layouts" url: /docs/layouts/ - title: Content children: - title: "Working with Posts" url: /docs/posts/ - title: "Working with Pages" url: /docs/pages/ - title: "Working with Collections" url: /docs/collections/ - title: "Helpers" url: /docs/helpers/ - title: "Utility Classes" url: /docs/utility-classes/ - title: Extras children: - title: "Stylesheets" url: /docs/stylesheets/ - title: "JavaScript" url: /docs/javascript/ - title: Meta children: - title: "History" url: /docs/history/ - title: "Contributing" [Truncated] And then in each `_doc` collection page the following was added to the YAML Front Matter to pull in the navigation in the sidebar. ```yaml sidebar: nav: "docs" ``` The better solution for adding the same YAML to multiple pages is with Front Matter Defaults in `_config.yml` like so: ```yaml defaults: # _docs - scope: path: "" type: docs values: sidebar: nav: "docs" ``` username_0: Thanks a lot. I had figured out the part about `_config.yml` but for some reason I didn't think to look at `navigation.yml`, which of course is a logical place. I am making a minisite to accompany my recurrent neural network package which means adding a lot of documentation so this is perfect. http://qua.st/rnn/docs/quick-start-guide/ Status: Issue closed username_1: Documentation has been added https://username_1.github.io/minimal-mistakes/docs/layouts/#custom-sidebar-navigation-menu username_0: very cool, thanks a lot! username_0: I just went through the documentation, I believe the step about making a collection is missing. username_1: @username_0 Didn't include that as collection creation is a basic of using Jekyll that can be learned about from the [official docs](http://jekyllrb.com/docs/collections/). username_0: Ok I agree. I guess I meant to say that people might not realise that the files need to a collection for the sidebar nav to work. Or perhaps this is obvious to people who use jekyll more frequently than I do. username_1: Sidebar nav should work with posts and pages, not just collections.
inpsyde/Wonolog
627083464
Title: [2.x] Q&A improvments Question: username_0: We need to do following for version 2.x of Wonolog: - Add Github Actions - Add PSALM - Add Inpsyde Coding Standards 1.x All tests should pass, which also contains renaming methods and adding return types. Answers: username_0: Starting to work with 9f5aa71 and e8ae174 on it. username_1: Continued working on it #53 username_2: This is all done in `2.x` branch: - Github action for static QA (PHPCS & Psalm) + Unit tests in the same workflow (2 separate jobs): https://github.com/inpsyde/Wonolog/blob/2.x/.github/workflows/qa.yml - Github action for integration tests (running in Docker with full WP installation): https://github.com/inpsyde/Wonolog/blob/2.x/.github/workflows/integration-tests.yml - Psalm config: https://github.com/inpsyde/Wonolog/blob/2.x/psalm.xml - PHPCS config with Inpsyde CS: https://github.com/inpsyde/Wonolog/blob/2.x/phpcs.xml.dist - Min PHP ver is 7.2 https://github.com/inpsyde/Wonolog/blob/2.x/composer.json#L31 and we're running tests on versions from 7.2 to 8.0 https://github.com/inpsyde/Wonolog/blob/2.x/.github/workflows/qa.yml#L43 - Tests are green and we've a good coverage: https://github.com/inpsyde/Wonolog/actions/runs/1466660872 (coverage report downloadable as HTML from action artifact) - Pretty much all methods have type declarations, the `phpcs:disable` is reduced to the very minimum So I guess we can close this. I would like to release a "beta" version of v2 soon. Status: Issue closed
jlippold/tweakCompatible
302371879
Title: `TranslucentMessages` working on iOS 10.3.3 Question: username_0: ``` { "packageId": "applebetas.ios.tweak.translucentmessages", "action": "working", "userInfo": { "arch32": false, "packageId": "applebetas.ios.tweak.translucentmessages", "deviceId": "iPhone7,2", "url": "http://cydia.saurik.com/package/applebetas.ios.tweak.translucentmessages/", "iOSVersion": "10.3.3", "packageVersionIndexed": false, "packageName": "TranslucentMessages", "category": "Tweaks", "repository": "BigBoss", "name": "TranslucentMessages", "packageIndexed": true, "packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.", "id": "applebetas.ios.tweak.translucentmessages", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.0.6", "shortDescription": "beautiful blurred effect for Messages app", "latest": "1.0.2", "author": "AppleBetas", "packageStatus": "Unknown" }, "base64": "<KEY> "chosenStatus": "working", "notes": "" } ```<issue_closed> Status: Issue closed
kiwiupover/ember-cli-spinjs
262381677
Title: Cannot set zIndex Question: username_0: Hi guys, There is a PR #12 which allows to pass in `zIndex` but this was removed in #13 most likely by mistake. If so - can it be added back? Answers: username_1: Sounds like a good plan. Status: Issue closed username_1: @username_0 thanks for the heads up fixed by #136 username_0: Awesome, thanks @username_1 !
WFCD/warframe-items
925471462
Title: :bug: Primary Weapon Kuva Bamma missing wikia thumbnail field Question: username_0: **Describe the bug** The Primary Weapon **Kuva Bamma** is missing the wikia thumbnail field "_wikiaThumbnail_" **To Reproduce** Fetching jsondata for primary weapons or all weapon data **Expected behavior** wikiaThumbnail should be set to "https://static.wikia.nocookie.net/warframe/images/a/a7/Kuva_Bramma.png/revision/latest?cb=20210609234150" **Additional context** That should be the proper url, but any url with the right image will do of course. Answers: username_1: Still not quite sure on this one username_1: eyyyy, found it. the imageUrls map that we're using for getting images from wikia (it's a map from them) mapped the image title for it as `Kuva Bramma.png`, whereas the image name in the weapons lua is `Kuva_Bramma.png`... gg wikia :D Status: Issue closed username_1: **Describe the bug** The Primary Weapon **Kuva Bamma** is missing the wikia thumbnail field "_wikiaThumbnail_" **To Reproduce** Fetching jsondata for primary weapons or all weapon data **Expected behavior** wikiaThumbnail should be set to "https://static.wikia.nocookie.net/warframe/images/a/a7/Kuva_Bramma.png/revision/latest?cb=20210609234150" **Additional context** That should be the proper url, but any url with the right image will do of course. Status: Issue closed
i18next/react-i18next
748823753
Title: Same 'previous' and 'new' value of language property Question: username_0: <!-- Before you submit an issue we recommend you visit [docs](https://www.i18next.com/) or [docs](https://react.i18next.com/) or [StackOverflow](https://stackoverflow.com/) or similar and ask any questions you have or mention any problems you've had getting started with i18next. **Please read this entire template before posting any issue. If you ignore these instructions and post an issue here that does not follow the instructions, your issue might be closed, locked.** --> ## 🐛 Bug Report Hi! Faced an issue when using withTranslation HOC. Previous value of language property of i18n object is exact same as new one when i'm trying to access it within ComponentDidUpdate method. Is it a bug or i'm doing something wrong? I've found a similar old issue: https://codesandbox.io/s/great-moore-04ob5 But is seems like that solution is not actual for new versions. ## To Reproduce https://codesandbox.io/s/great-moore-04ob5 ## Expected behavior Previous value and new value not same. ## Your Environment i18next 19.7.0 react 16.12.0 react-i18next 11.7.3 Answers: username_1: `i18n` is the prop -> that i18n instance did not change -> so accessing something like i18n.language will always show the current value username_0: @username_1 so i need to implement some logic to track that at my side. Ok! I have no problem with that :) Just wanted to check if there any other more natural solution. Thank you for reply! Status: Issue closed
ryck/MMM-Humanize-Duration
424544806
Title: Syntax error re-occuring Question: username_0: I would love to get this module installed, however no matter how I re-edit the config text (or even leave the default values for testing), my MM won't load due to syntax error. Here is my config text: ``` { module: 'MMM-Humanize-Duration', position: 'bottom_left', header: 'David', config: { date: "2018-11-01 07:51", options: { units: ["y", "mo", "w", "d", "h", "m", "s"], round: true, largest: 5, language: config.language }, updateInterval: 1 * 1 * 1000, animationSpeed: 250, initialLoadDelay: 0, debug: false } }, Not sure if I did this in the right format, proper noob here. ```
eBay/skin
1085284645
Title: primitives: update spacing tokens to use proportional labeling Question: username_0: The design system team is starting to integrate tokens into the system, starting at the source Figma file. The current token system in Skin uses a "t-shirt sizing" labeling style. Unfortunately, this style isn't scalable, especially when adding values between others. Instead, we can move to a proportional labeling style anchored to our base 8px grid, where 8px is 100% and is represented as `spacing-100` in the system. All of the other spacing values can be derived from there. Example ![image](https://user-images.githubusercontent.com/35155833/146845203-953763a9-c285-49af-a43d-4bb8c4510d4d.png) Answers: username_1: A subset of issue https://github.com/eBay/skin/issues/1646 which has latest specs/links. Status: Issue closed username_1: Done in 13.6.0 branch
daquexian/onnx-simplifier
853079596
Title: Simplify gather on constant indices of input shape even if shape has a dynamic index Question: username_0: **Describe the bug** Thank you for this tool, it's great for optimizing performance! I noticed that there's the potential to possibly optimize a bit further. If you consider the following situation: ![image](https://user-images.githubusercontent.com/10335022/113972902-4fb2df80-97f0-11eb-84af-059f90b411e3.png) Currently since the first dimension of the input to shape is variable, the entire operator is considered dynamic so no simplification is done. However, because indices 1-3 are still constant, we can pre-compute the gather. For the mask-rcnn model attached below, this pattern occurs quite a bit so simplifying all of these would lead to a sizable gain I think. **Model** https://github.com/username_0/onnxruntime-riscv/releases/download/v0.01/mask_rcnn_unquantized.onnx
perfsonar/pscheduler
216108507
Title: Improve no-tools-in-common reporting Question: username_0: Improve the error report when there is no common set of tools on the participants that can run a task. Maybe something like this: ``` Participants had no tools in common that can run this task. foo.bar.org: tool1 - Parameter 'blortz' is not supported. tool2 - OK tool3 - Not on your life. baz.quux.net: tool1 - No can do. tool2 - No can do for another reason. tool4 - OK ``` This will require a change to the REST API. Answers: username_0: Can-run methods now return JSON: ``` { "can-run": false, "reasons": [ "Cannot do 50 hops", "Source routing is not supported" ] } ``` (See the developer's guide.) username_0: Added basic diagnostics in d8629a7c33308fa0fb69d4e80e829958a3f538b1 (cited wrong ticket #326). Leaving open and in the icebox for the more-enhanced version.
hcmiya/opuscomment
418685963
Title: 1.4.7: stable? Question: username_0: Flac support was redundant. Metaflac does everything at a high level. Answers: username_1: 1.4.8 was released. Have fun Japanese: 1.4系列はOpus以外のコーデックをサポートすることを目標に開発をしていました。しかし、FLACはメタデータの取り扱いが独特で、対応がめんどくさくて放置していました。今後、1.4系列はFLACサポートを削除し、その対応は1.5系列で続けたいと考えています。 Status: Issue closed username_0: Flac support was redundant. Metaflac does everything at a high level. username_1: However, metaflac has difficulty handling line breaks... This is one of the motivation to add FLAC support to opuscomment. username_0: May be. Waiting for 1.5.0 username_0: METADATA_BLOCK_PICTURE and flac. Why not ignore like SoX?
Team-Tomato/Learn
687304310
Title: Ruby Db table associations Question: username_0: Learn the different types of rails associations between tables listed out here and implement the same in a sample application. https://guides.rubyonrails.org/association_basics.html belongs_to has_one has_many has_many :through has_one :through has_and_belongs_to_many The above link is an official one so it may be difficult to understand, so refer any youtube links or websites incase needed... Upon Completion, Show us a quick demo on your implementation (you should be able to create and display the record) and when its done create a Pull Request to Learn repo. Answers: username_1: <NAME> - Rails association task username_2: Deepalakshmi.B - Rails Association Task Status: Issue closed username_0: Deferred
kubernetes/dns
314012857
Title: DNS resolution for externalName services broken in v1.14.9 Question: username_0: After updating to v1.14.9, externalName services are not resolving anymore. I tested with v1.14.8 and it works. ``` / # dig @172.16.17.32 mysql-external.default.svc.cluster.local ; <<>> DiG 9.11.2-P1 <<>> @172.16.17.32 mysql-external.default.svc.cluster.local ; (1 server found) ;; global options: +cmd ;; Got answer: ;; WARNING: .local is reserved for Multicast DNS ;; You are currently testing what happens when an mDNS query is leaked to DNS ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 45659 ;; flags: qr aa rd ra; QUERY: 1, ANSWER: 0, AUTHORITY: 1, ADDITIONAL: 0 ;; QUESTION SECTION: ;mysql-external.default.svc.cluster.local. IN A ;; AUTHORITY SECTION: cluster.local. 60 IN SOA ns.dns.cluster.local. hostmaster.cluster.local. 1523602800 28800 7200 604800 60 ;; Query time: 1 msec ;; SERVER: 172.16.17.32#53(172.16.17.32) ;; WHEN: Fri Apr 13 07:58:15 UTC 2018 ;; MSG SIZE rcvd: 110 ``` kube-dns pod logs (I've replaced the correct domain with example.com but it exists and resolves normally in v1.14.8): ``` [kube-dns-aps-78b75f775c-b2j8p kubedns] I0413 08:12:22.197655 1 dns.go:612] Query for "mysql-external.default.svc.cluster.local.", exact: false [kube-dns-aps-78b75f775c-b2j8p kubedns] I0413 08:12:22.197682 1 dns.go:841] Not a federation query: len(["mysql-external" "default" "svc" "cluster" "local"]) != 4+len(["local" "cluster"]) [kube-dns-aps-78b75f775c-b2j8p kubedns] I0413 08:12:22.197700 1 dns.go:732] Found 1 records for [local cluster svc default mysql-external] in the cache [kube-dns-aps-78b75f775c-b2j8p kubedns] I0413 08:12:22.197715 1 dns.go:739] getRecordsForPath retval=[{Host:db.aurora.external.x.example.com Port:0 Priority:10 Weight:10 Text: Mail:false Ttl:30 TargetStrip:0 Group: Key:/skydns/local/cluster/svc/default/mysql-external}], path=[local cluster svc default mysql-external] [kube-dns-aps-78b75f775c-b2j8p kubedns] I0413 08:12:22.197739 1 dns.go:641] Records for mysql-external.default.svc.cluster.local.: [{db.aurora.external.x.example.com 0 10 10 false 30 0 /skydns/local/cluster/svc/default/mysql-external}] [kube-dns-aps-78b75f775c-b2j8p kubedns] I0413 08:12:22.197754 1 dns.go:612] Query for "db.aurora.external.x.example.com.", exact: false [kube-dns-aps-78b75f775c-b2j8p kubedns] I0413 08:12:22.197789 1 dns.go:860] Not a federation query: "x" != "svc" (serviceSubdomain) [kube-dns-aps-78b75f775c-b2j8p kubedns] I0413 08:12:22.197797 1 dns.go:732] Found 0 records for [com example x external aurora db] in the cache [kube-dns-aps-78b75f775c-b2j8p kubedns] I0413 08:12:22.197807 1 dns.go:739] getRecordsForPath retval=[], path=[com example x external aurora db] [kube-dns-aps-78b75f775c-b2j8p kubedns] I0413 08:12:22.197816 1 dns.go:645] No record found for db.aurora.external.x.example.com. [kube-dns-aps-78b75f775c-b2j8p kubedns] I0413 08:12:22.197830 1 logs.go:41] skydns: incomplete CNAME chain from "db.aurora.external.x.example.com.": no nameservers configured can not lookup name ``` I initially thought it's regression from #210 but I'm not sure. Please let me know how I can help with any more debug info. Answers: username_1: I can confirm DNS resolution for externalName service is broken in 1.14.9 but working in 1.14.8. Though with 1.14.9, I can also see the externalName record is generated in kubedns logs: ``` I0413 18:43:54.632820 1 dns.go:593] newExternalNameService: storing key test1 with value &{www.google.com 0 10 10 false 30 0 } as test1.default.svc.cluster.local. under [local cluster svc default] ``` Will spend more time. username_1: Built an image with d522d10e2e5e0797637e89ddb0a68a318cc8be60 (one commit before https://github.com/kubernetes/dns/pull/220) and the externalName service DNS resolution is working. @grayluck username_1: Got some pointers, seems like the nameserver for skydns somehow gets reset to empty, instead of using the ones listed in /etc/resolv.conf. ``` I0413 23:13:58.916300 1 dns.go:612] Query for "test1.default.svc.cluster.local.", exact: false I0413 23:13:58.916350 1 dns.go:841] Not a federation query: len(["test1" "default" "svc" "cluster" "local"]) != 4+len(["local" "cluster"]) I0413 23:13:58.916400 1 dns.go:732] Found 1 records for [local cluster svc default test1] in the cache I0413 23:13:58.916413 1 dns.go:739] getRecordsForPath retval=[{Host:www.google.com Port:0 Priority:10 Weight:10 Text: Mail:false Ttl:30 TargetStrip:0 Group: Key:/skydns/local/cluster/svc/default/test1}], p ath=[local cluster svc default test1] I0413 23:13:58.916462 1 dns.go:641] Records for test1.default.svc.cluster.local.: [{www.google.com 0 10 10 false 30 0 /skydns/local/cluster/svc/default/test1}] I0413 23:13:58.916485 1 dns.go:612] Query for "www.google.com.", exact: false I0413 23:13:58.916493 1 dns.go:841] Not a federation query: len(["www" "google" "com"]) != 4+len(["local" "cluster"]) I0413 23:13:58.916505 1 dns.go:732] Found 0 records for [com google www] in the cache I0413 23:13:58.916516 1 dns.go:739] getRecordsForPath retval=[], path=[com google www] I0413 23:13:58.916523 1 dns.go:645] No record found for www.google.com. I0413 23:13:58.916555 1 logs.go:41] skydns: incomplete CNAME chain from "www.google.com.": no nameservers configured can not lookup name ``` username_1: Also found out why our e2e test didn't catch this. Turned out in the externalName test we deliberately dig for just CNAME record. So the CNAME -> A (or AAAA) record path in upstream server is not examined. username_1: @grayluck is working on a fix. username_1: For the record, query for PTR records that only exist in upstream server seems to be broken as well. Status: Issue closed username_1: kube-dns 1.14.10 is released with the fix. username_0: Confirming that it works with v1.14.10.
ReactiveX/RxJava
170630023
Title: filter + map in a single operator Question: username_0: Hello there! I've seen several times (specially when working with Buses or generic events) this kind of method chain: `observable.filter(clazz::isInstance).map(clazz::cast)` Would make sense to have another operator that does this already? Something like a `filterClass` or similar? If so I could gladly create a PR with all the needed changes. Cheers! Answers: username_1: There is already operator for your case: ofType(). username_0: Ouch, I must be blind, thank you Artem! Status: Issue closed username_2: This has come up before and i always forget it exists. For discoverability purposes I'd support the addition of an alias method (preferably beginning with filter).
cadon/ARKStatsExtractor
323983570
Title: Mutation not picking up in extractor Question: username_0: Please forgive my ignorance, as I am still trying to learn this complex process. It is my understanding that it is possible to get a color mutation only, but this will *NOT* count in the mutation count x/20, right? Because the game is showing me there is a mutation now present (my very first one) but when this creatures stats were input into the tool, it's showing me ZERO mutations present?? Answers: username_1: In addition to that, the color mutation can be invisible as well if it mutates a color region that is not shown on that species of dino. username_0: I am aware of those possibilities based on the creature being bred. This is a Megatherium which does only use two color regions, but according to the Wiki does not have any stat's locked (like a flier does) The body (region 0 color) definitely does NOT show in it's natural color pallete, so it must be a mutation, hense my confusion as to where the mutation is, and why cant I see it in the stats. Cadon, are you referring to using the capture tool vs manually entering stats? if so, I cannot use it due to I am not using a supported resolution on my monitor, so I never have tried it that way. I do make sure I enter the parents into the tool before extracting. username_2: The stat-mutation should be visible in a higher torpor stat, which increases even if a stat without increase was mutated (e.g. speed). Compare the torpor of the offspring with the torpor of both parents to see if there is an increase. I was refering to the export feature in ARK (look at creature, hold E, choose options - export). The exported file can be imported in Smart Breeding. These file may lack the mutations-info if you didn't open the pedigree of that creature ingame. username_1: While it is true that the Megatherium doesn't have a stat that is "locked" (ie you can still increase the speed post tame) that only applies after the dino is raised and you start leveling it. Every Megatherium that is bred or tamed will have 100% as their base movement speed. If you get a mutation in speed it will still be 100% movement speed. This is true for all dinos that I am aware of. username_0: Thank you! So, if I understand this correctly, if It was a movement speed mutation, if I level it, I *SHOULD* see greater than the default 1% change for the Megatherium. That's my clue there is where the mutation occurred. They sure don't make this easy do they. LOL. What we would ever do without this tool i'll never know. Good thing you like coffee. ;-) username_2: Unfortunately, the mutation in speed will not change anything about the speed-stat, not the current value and not the amount you get when leveling it. The only change that will be noticeable is an increase of the torpor-level and the total level, each by two. username_3: @username_0 Not sure if this was covered, but if you have the parents of the offspring in ASB and the offspring has them listed as the parents, you can used the pedigree tab to see what muted. In the example below this baby pulled a speed mutation from one of its parents. ![image](https://user-images.githubusercontent.com/28239481/40269481-e2b5e3f4-5b44-11e8-820a-910f41573119.png) Pedigree is super helpful to figure out what mutated. Sometimes they will pull the stat from the lower parent and mutate that, which if they don't get a pretty color with it, puts em on the chopping block for me! username_0: username_3, THATS IT!! I did not clearly understand WHERE that mutation circle icon showed up, which was what I was looking for! Like I said initially, forgive my ignorance. BTW, It was in fact speed that my first ever mutation occurred (of course candon was spot on!). My most humble thanks to all! Status: Issue closed
soflyy/oxygen-bugs-and-features
821278477
Title: oxygen_lock_post_edit_mode error when saving Custom Fields Group Question: username_0: **Describe the bug** When I save a custom field group I get this error. ``` Notice: Undefined index: oxygen_lock_post_edit_mode in /Users/raphael/Documents/hamak/elexpertise/app/bedrock/web/app/plugins/oxygen/component-framework/admin/admin.php on line 107 ``` **Have you tried all the steps at https://oxygenbuilder.com/documentation/troubleshooting/troubleshooting-guide/?** The only relevant one is the log file and I haven't found anything interesting in them. **To Reproduce** 1. Install Oxygen 2. Install ACF (I have the pro version) 3. Create a new field group and save. **Expected behavior** Reload the page saying it saved. **What actually happens** Shows an error page even though it actually saved **Versions** - Oxygen Version 3.7 - ACF Pro Version 5.9.5 - PHP Version 7.4.1 I also see the "Edit with Oxygen" section on my custom field group, which doesn't make sense at all. Answers: username_0: Disabling the oxygen editor on acf custom fields solves the problem. Status: Issue closed username_1: still all index should be checked before accessing.
AIRLegend/aitrack
714072714
Title: Application crash Question: username_0: **Describe the bug** Application does not respond / crashed, in to 2 cases : - At launch - When starting the tracking **To Reproduce** Aitrack is crashing one time over 2 When successfull started, it is crashing everytime when launching the tracking In any cases, application is never working as expected. **Screenshots** 2 logs available : Crash at start : [2020-10-03 12:41:07.032] [aitrack] [info] ---------- AITRACK LOG ---------- [2020-10-03 12:41:07.265] [aitrack] [info] Created/Found prefs.ini [2020-10-03 12:41:07.265] [aitrack] [info] Searching for cameras... Crash at tracking start : [2020-10-03 12:45:21.465] [aitrack] [info] ---------- AITRACK LOG ---------- [2020-10-03 12:45:21.714] [aitrack] [info] Created/Found prefs.ini [2020-10-03 12:45:21.714] [aitrack] [info] Searching for cameras... [2020-10-03 12:45:23.378] [aitrack] [info] Number of recognized cameras: 2 [2020-10-03 12:45:23.379] [aitrack] [info] UDP sender reinitialized. IP: 127.0.0.1 PORT: 4242 [2020-10-03 12:45:23.379] [aitrack] [info] Building Tracker with selected camera: 0 [2020-10-03 12:45:23.420] [aitrack] [info] Tracker initialized. [2020-10-03 12:45:23.420] [aitrack] [info] Updated stabilizer. [2020-10-03 12:45:23.420] [aitrack] [info] Updating camera parameters... [2020-10-03 12:45:23.420] [aitrack] [info] Updated camera parameters. 1280x720@30 [2020-10-03 12:45:23.420] [aitrack] [info] Checking for updates [2020-10-03 12:45:23.457] [aitrack] [info] App initialized [2020-10-03 12:45:29.839] [aitrack] [info] Starting camera 0 capture **Environment (please complete the following information):** - Windows 10 Professionnal, Version 19041.508 (latest) - Webcam model: Logitech C920 Pro - AITrack version: 0.6.2 alpha **Additional context** Resolution method tried : - Stopping AV - Stopping Windows Defender - Checked privacy settings for Camera and AiTrack - Disabling/unistalling any third party software using the webcam (including logitech's one) - Changing all AiTrack options (camera supported resolution and FPS, model, etc....) - Granting administrator rights to AiTrack - Deleting the webcam from windows device manager - Changing USB port for Webcam Answers: username_1: Hello! Thanks for the info! Seems to be the same bug some people are experiencing (some other issues are already posted here). I don't think it's a hardware issue as there are quite a few of people who are using that camera. However, until now, I've been unable to get to the root of the problem. My guess for the moment is that windows could be interfering in some way with OpenCV while accessing the camera device. That would explain both types of crash you described. Quick questions: - Does the crashing behaviour change at random or did you do something in order to get to the 2nd step? - Do you have only one camera available under Options>Camera? If not, try with others. - In the second case, does it crash if you set other resolution? (1080p, for example). username_0: Hello ! Thanks for quick reply ! - The crash seems random, usually after 1 or 2 (max) crashes at start, the application is starting as usual. I confirm I do not do anything in between my attempts. - I don't have a second camera. I had previously a software (Snap Camera) installed emulating a 2nd device that was available in the AiTrack camera menu. I removed it to troubleshoot. - I tried different resolutions supported by the camera (1920x1280, or even 640x480, all with 30 FPS), with no different behaviour that I could notice.... Hope this is helping. username_1: Do you get the same error if you use Snap Camera (or any other virtual camera)? Thanks! username_0: Indeed working with snap Camera selected. username_1: Strange then... Do you have your camera connected via USB hub? username_0: No, directly to pc, I've tried different ports, make no difference. Status: Issue closed
react-figma/react-figma
1040540243
Title: Fix typo Question: username_0: Can we write Github as GitHub in line no 134 of [README.md](https://github.com/react-figma/react-figma/blob/master/README.md) file? For your reference, I have attached a screenshot of the same. ![Screenshot (382)](https://user-images.githubusercontent.com/60184229/139595743-b40cb453-4a77-4075-97ed-f0ee3763da87.png)<issue_closed> Status: Issue closed
swoole/swoole-src
434707816
Title: 协程里面mysql和lock 为何死锁了 Question: username_0: ``` $lock = new swoole_lock(SWOOLE_MUTEX); function get( $lock ) { $lock->lock(); echo "lock\n"; $Mysql = new Swoole\Coroutine\MySQL(); $flag = $Mysql->connect( [ 'host' => '127.0.0.1', 'port' => 3306, 'user' => 'root', 'password' => '123', 'database' => 'ty_blast', 'charset' => 'utf8', ] ); echo "unlock\n"; $lock->unlock(); } go(function() use ( $lock ){ get($lock); }); go(function() use ( $lock ){ get($lock); }); ``` swoole 4.3.2 Answers: username_1: ```php <?php class lock { private $chan; public function __construct() { $this->chan = new Chan; } public function lock() { $this->chan->push(true); } public function unlock() { $this->chan->pop(); } } function get(lock $lock) { echo "lock\n"; $lock->lock(); Co::sleep(0.1); echo "unlock\n"; $lock->unlock(); } $lock = new lock(); go(function () use ($lock) { get($lock); }); go(function () use ($lock) { get($lock); }); ``` username_0: @username_1 谢谢帮助, 但还是不是很理解. 为何把我上面的代码, 那段mysql的代码, 换成sleep(2); 就得到了预想的结果呢? ``` $lock = new swoole_lock(SWOOLE_MUTEX); function get( $lock ) { $lock->lock(); echo "lock\n"; sleep(2); echo "unlock\n"; $lock->unlock(); } go(function() use ( $lock ){ get($lock); }); go(function() use ( $lock ){ get($lock); }); ``` Status: Issue closed
m-farokhtabar/CitywareAndroidClient
427911754
Title: مشکل نمایش امتیازات در عکس های ارسالی Question: username_0: چرا امتیازات این دو تا فرم با هم نمی خونه مگه مال یک نفر نیست ؟؟ ![0c7de823-e7e2-40e6-b00d-62fef4f798cd](https://user-images.githubusercontent.com/46286579/55359501-3c752700-54e7-11e9-90e7-acbdadb19ffe.jpg) ![bc859a0f-f0e2-4a41-b7d8-4c8d30e07d4b](https://user-images.githubusercontent.com/46286579/55359533-4f87f700-54e7-11e9-8c9e-14cef049b3a1.jpg)<issue_closed> Status: Issue closed
rancher/rancher
195607264
Title: java.lang.NullPointerException when service.update is executed on upgraded setup from 1.1.4 -> v1.2.1-rc2 Question: username_0: Rancher-server version - upgraded setup from 1.1.4 -> v1.2.1-rc2 Fresh install of 1.1.4. Had 1 cattle env with few services and 1 k8s env with few services/pods/rc. Upgraded to v1.2.1-rc2 Lot of java.lang.NullPointerException seen in logs when service.update is done. ``` 2016-12-14 18:15:37,358 ERROR [342340ce-82c4-4474-9a9c-9dceddce0390:83428] [instance:4396] [instance.start->(InstanceStart)] [] [cutorService-16] [i.c.p.process.instance.InstanceStart] Failed to Scheduling for instance [4396] 2016-12-14 18:15:37,931 ERROR [03ce4a1a-33b1-445c-91ef-1a55b6d817c3:83441] [instance:4399] [instance.start->(InstanceStart)] [] [ecutorService-4] [i.c.p.process.instance.InstanceStart] Failed to Scheduling for instance [4399] 2016-12-14 18:15:37,957 ERROR [e2515a61-1243-4693-a12e-7e38e5c21039:36671] [service:30] [service.update] [] [ecutorService-3] [c.p.e.p.i.DefaultProcessInstanceImpl] Unknown exception java.lang.NullPointerException: null at io.cattle.platform.servicediscovery.deployment.impl.unit.DefaultDeploymentUnitInstance.scheduleCreate(DefaultDeploymentUnitInstance.java:377) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.unit.DeploymentUnit.create(DeploymentUnit.java:293) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.ServiceDeploymentPlanner.deploy(ServiceDeploymentPlanner.java:153) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.DeploymentManagerImpl.startUnits(DeploymentManagerImpl.java:375) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.DeploymentManagerImpl$2.run(DeploymentManagerImpl.java:340) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.activity.ActivityService.run(ActivityService.java:43) ~[cattle-activity-log-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.DeploymentManagerImpl.activateDeploymentUnits(DeploymentManagerImpl.java:337) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.DeploymentManagerImpl.deploy(DeploymentManagerImpl.java:268) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.DeploymentManagerImpl.reduceScaleAndDeploy(DeploymentManagerImpl.java:217) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.DeploymentManagerImpl.incremenetScaleAndDeploy(DeploymentManagerImpl.java:180) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.DeploymentManagerImpl.deployWithScaleAdjustement(DeploymentManagerImpl.java:169) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.DeploymentManagerImpl.reconcileDeployment(DeploymentManagerImpl.java:136) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.DeploymentManagerImpl$1.doWithLock(DeploymentManagerImpl.java:123) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.DeploymentManagerImpl$1.doWithLock(DeploymentManagerImpl.java:117) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.AbstractLockManagerImpl$3.doWithLock(AbstractLockManagerImpl.java:40) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:13) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:37) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.DeploymentManagerImpl.activate(DeploymentManagerImpl.java:117) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.deployment.impl.DeploymentManagerImpl.activate(DeploymentManagerImpl.java:103) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.process.ServiceUpdateActivate$1.run(ServiceUpdateActivate.java:73) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.activity.ActivityService.run(ActivityService.java:43) ~[cattle-activity-log-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.servicediscovery.process.ServiceUpdateActivate.handle(ServiceUpdateActivate.java:70) ~[cattle-iaas-service-discovery-server-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runHandler(DefaultProcessInstanceImpl.java:448) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$4.execute(DefaultProcessInstanceImpl.java:399) ~[cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$4.execute(DefaultProcessInstanceImpl.java:393) ~[cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.idempotent.Idempotent.execute(Idempotent.java:42) ~[cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runHandlers(DefaultProcessInstanceImpl.java:393) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runLogic(DefaultProcessInstanceImpl.java:495) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runWithProcessLock(DefaultProcessInstanceImpl.java:326) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$2.doWithLockNoResult(DefaultProcessInstanceImpl.java:243) ~[cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.LockCallbackNoReturn.doWithLock(LockCallbackNoReturn.java:7) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.LockCallbackNoReturn.doWithLock(LockCallbackNoReturn.java:3) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.AbstractLockManagerImpl$3.doWithLock(AbstractLockManagerImpl.java:40) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:13) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:37) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.acquireLockAndRun(DefaultProcessInstanceImpl.java:240) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runDelegateLoop(DefaultProcessInstanceImpl.java:182) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.executeWithProcessInstanceLock(DefaultProcessInstanceImpl.java:155) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$1.doWithLock(DefaultProcessInstanceImpl.java:114) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$1.doWithLock(DefaultProcessInstanceImpl.java:111) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.AbstractLockManagerImpl$3.doWithLock(AbstractLockManagerImpl.java:40) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:13) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:37) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.execute(DefaultProcessInstanceImpl.java:111) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.server.impl.ProcessInstanceDispatcherImpl.processExecuteWithLock(ProcessInstanceDispatcherImpl.java:98) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] [Truncated] at io.cattle.platform.engine.server.impl.ProcessInstanceDispatcherImpl.processExecuteWithLock(ProcessInstanceDispatcherImpl.java:98) ~[cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.server.impl.ProcessInstanceDispatcherImpl$1$1.doWithLockNoResult(ProcessInstanceDispatcherImpl.java:71) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.LockCallbackNoReturn.doWithLock(LockCallbackNoReturn.java:7) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.LockCallbackNoReturn.doWithLock(LockCallbackNoReturn.java:3) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.AbstractLockManagerImpl$4.doWithLock(AbstractLockManagerImpl.java:50) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.AbstractLockManagerImpl.tryLock(AbstractLockManagerImpl.java:25) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.lock.impl.AbstractLockManagerImpl.tryLock(AbstractLockManagerImpl.java:47) [cattle-framework-lock-0.5.0-SNAPSHOT.jar:na] at io.cattle.platform.engine.server.impl.ProcessInstanceDispatcherImpl$1.doRun(ProcessInstanceDispatcherImpl.java:68) [cattle-framework-engine-0.5.0-SNAPSHOT.jar:na] at org.apache.cloudstack.managed.context.NoExceptionRunnable.runInContext(NoExceptionRunnable.java:15) [cattle-framework-managed-context-0.5.0-SNAPSHOT.jar:na] at org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49) [cattle-framework-managed-context-0.5.0-SNAPSHOT.jar:na] at org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:55) [cattle-framework-managed-context-0.5.0-SNAPSHOT.jar:na] at org.apache.cloudstack.managed.context.impl.DefaultManagedContext.callWithContext(DefaultManagedContext.java:108) [cattle-framework-managed-context-0.5.0-SNAPSHOT.jar:na] at org.apache.cloudstack.managed.context.impl.DefaultManagedContext.runWithContext(DefaultManagedContext.java:52) [cattle-framework-managed-context-0.5.0-SNAPSHOT.jar:na] at org.apache.cloudstack.managed.context.ManagedContextRunnable.run(ManagedContextRunnable.java:46) [cattle-framework-managed-context-0.5.0-SNAPSHOT.jar:na] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_72] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_72] at java.lang.Thread.run(Thread.java:745) [na:1.8.0_72] ``` Answers: username_1: Some race condition in the situation when 0 hosts are available for scheduling in the environment, shouldn't impact things. Moving the target release username_1: Please reopen if you see it again on the -> 1.6 upgrade. Status: Issue closed
jlippold/tweakCompatible
418636129
Title: `ColorBanners 2 (Packix)` working on iOS 12.1.1 Question: username_0: ``` { "packageId": "com.golddavid.colorbanners2-new", "action": "working", "userInfo": { "arch32": false, "packageId": "com.golddavid.colorbanners2-new", "deviceId": "iPad5,3", "url": "http://cydia.saurik.com/package/com.golddavid.colorbanners2-new/", "iOSVersion": "12.1.1", "packageVersionIndexed": false, "packageName": "ColorBanners 2 (Packix)", "category": "Tweaks", "repository": "Packix", "name": "ColorBanners 2 (Packix)", "installed": "1.2.2", "packageIndexed": true, "packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.", "id": "com.golddavid.colorbanners2-new", "commercial": true, "packageInstalled": true, "tweakCompatVersion": "0.1.4", "shortDescription": "Color your notifications", "latest": "1.2.2", "author": "<NAME>", "packageStatus": "Unknown" }, "base64": "<KEY>", "chosenStatus": "working", "notes": "" } ```<issue_closed> Status: Issue closed
SkaceKamen/vscode-sqflint
343977052
Title: invalid semicolon error Question: username_0: inside RscDefines.hpp file: As an example: ![image](https://user-images.githubusercontent.com/13535616/43132540-7fed5616-8f7f-11e8-864b-ff95a0298c22.png) Error thrown on first line at the semi colon: [sqflint] Expected ":", "{", comment, end of line, or whitespace but ";" found. Answers: username_1: Oh, you're right, I forgot about empty classes. I'll try to add it in future.
Yelp/paasta
172805795
Title: NoSlavesAvailable message should be more helpful with incorrect deploy_whitelist/blacklist Question: username_0: When you've got a deploy_whitelist with values that don't exist on your cluster (say, you typo an attribute value), `paasta status` will show something like: ``` $ paasta status -s mycoolservice -c norcal-prod Pipeline: https://jenkins.yelpcorp.com/view/services-mycoolservice cluster: norcal-prod instance: coolinstance Git sha: 916bba50 (desired) ERROR:__main__:Exception raised while looking at service mycoolservice instance coolinstance: ERROR:__main__:Traceback (most recent call last): File "/usr/bin/paasta_serviceinit", line 122, in main delta=args.delta, File "/usr/share/python/paasta-tools/local/lib/python2.7/site-packages/paasta_tools/marathon_serviceinit.py", line 351, in perform_command app_id = job_config.format_marathon_app_dict()['id'] File "/usr/share/python/paasta-tools/local/lib/python2.7/site-packages/paasta_tools/marathon_tools.py", line 396, in format_marathon_app_dict 'constraints': self.get_calculated_constraints(service_namespace_config), File "/usr/share/python/paasta-tools/local/lib/python2.7/site-packages/paasta_tools/marathon_tools.py", line 322, in get_calculated_constraints constraints.extend(self.get_routing_constraints(service_namespace_config)) File "/usr/share/python/paasta-tools/local/lib/python2.7/site-packages/paasta_tools/marathon_tools.py", line 331, in get_routing_constraints whitelist=self.get_deploy_whitelist()) File "/usr/share/python/paasta-tools/local/lib/python2.7/site-packages/paasta_tools/mesos_tools.py", line 523, in get_mesos_slaves_grouped_by_attribute raise NoSlavesAvailable("No mesos slaves were available to query. Try again later") NoSlavesAvailable: No mesos slaves were available to query. Try again later ssh -A -o LogLevel=QUIET -t 10.1.2.3 sudo paasta_serviceinit -s mycoolservice -i coolinstance status exited with non-zero return code. ``` The exception message should indicate that this can be caused by too-restrictive `deploy_whitelist` / `deploy_blacklist`.
allenai/allennlp
267380622
Title: Embedding POS/NER/arbitrary features Question: username_0: First of all, I want to apologize if this question is not suitable for an issue tracker (but I could not find any other place). I am trying to incorporate POS-tag and NER-type features (one-hot encoded vectors mainly) to the word&character embeddings and I wanted to ask what is the best way to do this using this library? Thank you in advance! Answers: username_1: Using POS tags is currently easy, but we don't have NER tags currently available for this. It was easy to add, so I just submitted #430 that adds it. That PR also modifies one of the test configuration files to show how to use POS tag and NER embeddings as additional components of the word representations. Just look at the modifications to `tests/fixtures/encoder_decoder/simple_seq2seq/experiment.json` in that PR, and make similar modifications to your configuration file. Let me know if you have any questions. username_0: Thank you very much, I will take a look at your PR. I am still curious how it is possible to incorporate arbitrary features (e.g. binary flag if token is capitalized, token frequence etc). As I understood TextFieldEmbedders operate on a list of indices, but somehow it would be nice to map Token -> feature vector directly? username_1: For this, there are two options. The quickest option is to do something like what the SRL model does to incorporate the verb indicator. The `DatasetReader` adds a `SequenceLabelField` with binary values indicating whether each token is the verb to find arguments for: https://github.com/allenai/allennlp/blob/ed37843d528b0902d73e4d5dde931812fd65835d/allennlp/data/dataset_readers/semantic_role_labeling.py#L277 The model then takes that field, embeds it, and concatenates it manually to the word embedding: https://github.com/allenai/allennlp/blob/ed37843d528b0902d73e4d5dde931812fd65835d/allennlp/models/semantic_role_labeler.py#L119-L122 You could do this for each binary feature you want, but just skip the feature embedding. If you have a lot of binary features you want to add, though, this could get annoying. So the second option is to write your own `TokenIndexer` that converts word tokens into feature vectors using whatever feature extractors you want to use, and a `TokenEmbedder` that just passes through the given feature vector. If you want to go this route, I'm happy to look at a work-in-progress PR to let you know if you're doing it the right way. username_0: Great, thank you for the very detailed answer. I will try to make a PR for a second option. username_0: I managed to create a combination of a token indexer that does arbitrary feature extraction and bypass embedder - I will do PR a bit later after I polish the code a bit. One more thing I was interested to do is to extract features for a particular token given also its context - let's say the whole sequence (in the simplest case, for instance, a part-of-speech tag of a next token). I am not sure if it is possible given the current design of the data processing pipeline, but maybe there is a way? username_1: I'd be a bit of a hack, but you could just save a reference to the whole sentence inside each `Token` object, then use that info in your `TokenIndexer`. It wouldn't be that bad to add a `context` field to `Token`, and let callers populate it however they want. We'll have to think about whether it's worth adding that to the official API, but the nice thing about Python is you can do it yourself with duck typing, even if we don't add it. If this isn't clear enough, I can give an example of what I mean. username_0: @username_1 Thx, for the advice - adding reference worked very well :) username_2: I'm closing this as I don't see any clear follow on. Feel free to re-open and specify what follow-on you are looking for, if needed! Status: Issue closed username_1: We addressed the POS/NER feature embeddings, and there's an open PR to address the arbitrary features. It needs a little bit of work, but I'm planning on fixing it up soon, because we'll probably need it for re-implementing the WikiTables parser. We're pretty close to having everything else done, so it should be pretty soon that we finally finish that PR. No strong opinion on keeping this issue open or closed.
folke/ultra-runner
708483331
Title: Unable to run more than 10 concurrent scripts Question: username_0: Hi, I'm trying to run this scripts: ``` ultra -r --concurrency 20 --filter "@my-workspace/*" start ``` Unfortunately, even if I specified a `concurrency` of `20` I'm not able to run it on every package. It seems the concurrency is always limited to `10` (the default). Do you have any advice? Answers: username_1: Just tested this, and it seems to work just fine on my end. If you have packages depending on other packages in your repo, then those will be built before the dependents. This is by design, so increasing concurrency makes no difference here. This is needed, since otherwise you'd be building packages that depend on other packages that might not have been built yet. Maybe that's the reason? Status: Issue closed
electron/electron
1010577238
Title: [Bug]: webrtc screen/window capturer cannot capture other app without creating mainWindow first Question: username_0: ### Preflight Checklist - [X] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/main/CONTRIBUTING.md) for this project. - [X] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/main/CODE_OF_CONDUCT.md) that this project adheres to. - [X] I have searched the [issue tracker](https://www.github.com/electron/electron/issues) for a feature request that matches the one I want to file, without success. ### Electron Version 15.0.0 ### What operating system are you using? macOS ### Operating System Version macOS Catalina ### What arch are you using? x64 ### Last Known Working Electron version _No response_ ### Expected Behavior I'm using electron's desktopCapturer to do app window capture & then stream it to mediasoup webrtc server. However, i find that desktopCapturer can only be used to capture other app's window AFTER mainWindow is created, which means I have to click once on electron app to activate(?) I try to use desktopCapturer directly without creating any renderer window, that is to say, try to use it in headless CLI mode, but failed. ``` ReferenceError: navigator is not defined ``` Can i do app window screen sharing without click on electron? ### Actual Behavior see description above. ### Testcase Gist URL _No response_ ### Additional Information _No response_ Answers: username_0: my js code using screenCapturer API and mediasoup client lib: ``` //https://www.electronjs.org/docs/api/desktop-capturer const { desktopCapturer } = require('electron') const mediasoup = require('mediasoup-client'); const socketClient = require('socket.io-client'); const socketPromise = function(socket) { return function request(type, data = {}) { return new Promise((resolve) => { socket.emit(type, data, resolve); }); } }; //signaling is based on https://github.com/mkhahani/mediasoup-sample-app: const webrtcServerUrl = 'http://127.0.0.1:3000'; let socket; let device; let producer; desktopCapturer.getSources({ types: ['window', 'screen'] }).then(async sources => { for (const source of sources) { console.log(source); if (source.name === 'myapp') { try { const stream = await navigator.mediaDevices.getUserMedia({ audio: false, video: { mandatory: { chromeMediaSource: 'desktop', chromeMediaSourceId: source.id, minWidth: 1280, maxWidth: 1280, minHeight: 720, maxHeight: 720 } } }) await handleStream(stream) } catch (e) { handleError(e) } return } } }) async function loadDevice(routerRtpCapabilities) { try { device = new mediasoup.Device(); } catch (error) { if (error.name === 'UnsupportedError') { console.error('browser not supported'); } } await device.load({ routerRtpCapabilities }); return device } [Truncated] const track = stream.getVideoTracks()[0]; const params = { track }; //if ($chkSimulcast.checked) { params.encodings = [ { maxBitrate: 100000 }, { maxBitrate: 300000 }, { maxBitrate: 900000 }, ]; params.codecOptions = { videoGoogleStartBitrate : 1000 }; } producer = await transport.produce(params); } function handleError (e) { console.log(e) } ``` username_1: Thanks for reporting this and helping to make Electron better! Because of time constraints, triaging code with third-party dependencies is usually not feasible for a small team like Electron's. Would it be possible for you to make a standalone testcase with only the code necessary to reproduce the issue? For example, [Electron Fiddle](https://www.electronjs.org/fiddle) is a great tool for making small test cases and makes it easy to publish your test case to a [gist](https://gist.github.com) that Electron maintainers can use. Stand-alone test cases make fixing issues go more smoothly: it ensure everyone's looking at the same issue, it removes all unnecessary variables from the equation, and it can also provide the basis for automated regression tests. I'm adding the `blocked/need-repro` label for this reason. After you make a test case, please link to it in a followup comment. This issue will be closed in 10 days if the above is not addressed.
t3knoid/MagyarTV
412239040
Title: Use a separate thread outside of UI when grabbing video stream Question: username_0: When the user clicks on a new channel to watch, do not shut off current channel. Instead, run a separate thread to get the new stream. Only change the stream once its verified that the new stream is accessible. Answers: username_0: When the user clicks on a new channel to watch, do not shut off current channel. Instead, run a separate thread to get the new stream. Only change the stream once its verified that the new stream is accessible. username_0: Use backgroundworker to call Python code. Status: Issue closed
awslabs/service-workbench-on-aws
960887943
Title: [Misc] is golang a requirement Question: username_0: **What's on your mind?** scripts/environment-deploy.sh works for me but I noticed errors in the log regarding go. ![image](https://user-images.githubusercontent.com/148409/128239317-28269a4e-bb89-436a-8405-53eb0e3dd3a4.png) Can these be ignored or should I install go and redeploy? **Versions (please complete the following information):** - Release Version installed - mainline Answers: username_1: Hi @username_0, golang is a requirement since it is used for the multipart S3 downloader tool EC2 Windows based research environments. Although it is recommended to have it installed, you can safely ignore the error if you don't expect to use EC2-Windows workspaces. username_1: Also, the fix for this bug is in [progress](https://github.com/awslabs/service-workbench-on-aws/pull/641). Feel free to install from the latest release after this has been merged. username_2: The [fix](https://github.com/awslabs/service-workbench-on-aws/pull/641) for this issue has been merged. Closing this issue now. Status: Issue closed
spatie/laravel-searchable
394493549
Title: Passing searchable attributes as array does not work Question: username_0: When I pass the searchable attributes as an array (like in the examples) the following error is thrown: `Argument 1 passed to Spatie\\Searchable\\SearchableAttribute::__construct() must be of the type string, array given` Code looks like this: ```php $searchResults = (new Search()) ->registerModel(Ticket::class, ['title', 'body', 'from_address', 'from_name']) ->perform(request('q')); ``` This works: ```php $searchResults = (new Search()) ->registerModel(Ticket::class, 'title', 'body', 'from_address', 'from_name') ->perform(request('q')); ``` Status: Issue closed Answers: username_1: Fixed in 1.0.1! Thanks
zhilizheng/qsubshcom
755740103
Title: qsubshcom in Westlake university Question: username_0: Dear Zhili, I'm not sure whether this bug was caused by the version of 'SGE' or some specific parameter settings in the Westlake university cluster. Line 109: echo "#$ -pe **onehost** $2" >> $scriptname echo "#$ -pe **thread** $2" >> $scriptname (WLU) Best wishes, Liyang Answers: username_1: Thanks. This is a specific setting for SGE. You can change it to work, since SGE is too old. username_0: I see, thanks for your reply. Status: Issue closed
Aminerman/vue-mouse-parallax
540402418
Title: Missing CSS classes to make parallax-element overlap Question: username_0: Could you provide the CSS to make parallax-element overlap? It would help if you released the code you used to make the demo. Answers: username_1: `parallax-element` are actually just normal DOM elements. You can add a class to them and give them a `position: absolute` and you can therefore make them overlap. Does it help solving your issue? username_0: As soon as I use `position: absolute` everything stop moving. No more parallax effect. I'm missing some CSS magic somewhere. username_1: My guess is that, if all your `parallax-element` are in `position: absolute`, therefore your `parallax-container` has a **width** and **height** equal to **0**. Therefore there can't be any movement as the `parallax-container` is the one that detects the mouse movement. So depending on your case and the result you're trying to achieve you need to find a way to give your `parallax-container` a **width** and **height**. Status: Issue closed
commercetools/commercetools-dotnet-core-sdk-v2
1045821134
Title: Deserialize ImportOperation: Failed to find the required 'Type' discriminator property error Question: username_0: **Describe the Bug** While ImportOperation Deserialization which has error of type "DuplicateAttributeValue", it's failing to deserialize the attribute because it doesn't have type property in the json. **To Reproduce** Try to Deserialize the below json to IImportOperation: { "version": 1, "importContainerKey": "productsImport", "resourceKey": "MG-Iphone11", "id": "d2c20ea3-657b-4314-b312-cd527asss", "state": "validationFailed", "errors": [ { "code": "DuplicateAttributeValue", "message": "Attribute can't have the same value in a different variant.", "attribute": { "name": "phonecolor", "value": "Black" } } ], "createdAt": "2021-11-05T09:38:28.878Z", "lastModifiedAt": "2021-11-05T09:38:29.204Z", "expiresAt": "2021-11-07T09:38:28.878Z" }<issue_closed> Status: Issue closed
lz4/lz4-java
40782914
Title: LZ4BlockInputStream cannot read two consecutive write-close operations from two different LZ4BlockOutputStream Question: username_0: **How to reproduce:** Run the test case testWriteCloseWriteCloseRead(). In pseudo code: 1. Write some data to a file with a LZ4BlockOutputStream and close the stream 2. Write some more data to the same file with a new LZ4BlockOutputStream and close the stream. 3. Read the sum of the data with one single instance of LZ4BlockInputStream ``` java /** * Write and close two stream instances to the same file. Read the entire data with one * LZ4BlockInputStream. */ @Test public void testWriteCloseWriteCloseRead() throws IOException { final byte[] testBytes = "Testing!".getBytes(Charset.forName("UTF-8")); //Write the first time ByteArrayOutputStream bytes = new ByteArrayOutputStream(); LZ4BlockOutputStream out = new LZ4BlockOutputStream(bytes); out.write(testBytes); out.close(); //Write the second time out = new LZ4BlockOutputStream(bytes); out.write(testBytes); out.close(); ByteArrayInputStream in = new ByteArrayInputStream(bytes.toByteArray()); LZ4BlockInputStream lz4In = new LZ4BlockInputStream(in); DataInputStream dataIn = new DataInputStream(lz4In); byte[] buffer = new byte[testBytes.length]; dataIn.readFully(buffer); assertArrayEquals(testBytes, buffer); // in.skip(LZ4BlockOutputStream.HEADER_LENGTH); //This test case can only be passed if 21 bytes (the footer) is skipped buffer = new byte[testBytes.length]; dataIn.readFully(buffer); assertArrayEquals(testBytes, buffer); } ``` **Actual:** An java.io.EOFException is thrown **Expected:** The sum of the data should be read and returned. **Analysis:** The LZ4BlockOutputStream will write a header, data and a footer. The footer is very similar to the header. Two LZ4BlockOutputStreams will create this: Header | Compressed Data | Footer | Header |Compressed Data | Footer One instance of LZ4BlockInputStream will read the header and the compressed data. If the user tries to read more data it will try to read a header again. But since it has not skipped the previous footer it will read the footer instead. The footer, although similar to the header contains a 0 length and will therefore return -1 from the read() method and the DataInputStream will thus throw a EOFException. If the user manually skips 21 bytes (the length of the header/footer) the LZ4BlockInputStream will happily continue to read another “frame” (se the out-commeted row in the test case). **Workaround:** The user can manually call in.skip(21). **Suggested fix:** I think it would be appropriate if a LZ4BlockInputStream consumes all bytes related the one frame: that is the footer should be consumed when the end of the frame has been reached I’m guessing the solution might be a bit trickier because the footer is related to the frame and the header to the block? (I’m probably using the term block and frame wrong) Another approach would be to just say that this should not be possible. But this “feature” works with a normal GZIPOutputStream/GZIPInputStream so it would be good if it also works with LZ4. Answers: username_1: Fixed by #105. Thanks much for your suggestion! Status: Issue closed
MozillaFoundation/plan
59518633
Title: Reduce run costs on AWS and Heroku Question: username_0: We suspect our monthly AWS and Heroku costs are higher than required. Let's do an inventory of spending across each of our properties and propose (when required) or implement optimizations/reductions. Solution vectors: * Instance pre-allocation * Transitioning to smaller instances * Sharing of instances across applications * Reduced-cost storage on Heroku instances (SSD -> normal) ### RACI Phase: Build / Ship Owner: @username_1 Decision: @username_0 Lead design: N/A Dev: @username_3 Quality: @username_0 Answers: username_1: I've trimmed off a bunch of unused volumes, and have killed four RDS instances that we don't really need or don't know what they're doing (phabricator). We of course took final snapshots, so if we hear any screams, we'll be able to restore. The plan moving forward: 1) Combine nearly all staging EC2 nodes into a big honkin' cluster(maybe 2), with separate ELBs pointing to distinct node ports per app. 2) Put some selected apps (based on priority, current load averages, and frequency of updates recently) into smaller clusters. IE, Badgekit, Badgekit-api, Openbadges into a cluster, maybe Goggles, Thimble, Butter into a cluster. Meanwhile, we'll keep central/often changed apps separate, such as Webmakerorg, Makeapi, Login. 3) Reserve a swath of instances for these new clusters with the no-up-front cost option recently announced by AWS. 4) Sip champagne on a beach. username_0: @username_1 could you post an update? How was the champagne? username_1: The manifests should be set to reduce a bunch of our apps into clusters, but this week will see me launching first staging then production versions of that. Meanwhile, we've scaled a ton of instances down last week. I'm working on the costing server but so far it seems hosed still. The app sounds like it should be working fine 2015-03-16 16:30:26,643 [cost_daily_elasticache] INFO basic.BasicDataManager - cost_daily_elasticache start polling... 2015-03-16 16:30:26,643 [cost_monthly_elasticache] INFO basic.BasicDataManager - cost_monthly_elasticache start polling... 2015-03-16 16:30:32,974 [usage_monthly_glacier] INFO basic.BasicDataManager - usage_monthly_glacier start polling... 2015-03-16 16:30:32,975 [usage_hourly_glacier] INFO basic.BasicDataManager - usage_hourly_glacier start polling... 2015-03-16 16:30:32,975 [usage_daily_glacier] INFO basic.BasicDataManager - usage_daily_glacier start polling... 2015-03-16 16:30:32,975 [cost_monthly_glacier] INFO basic.BasicDataManager - cost_monthly_glacier start polling... 2015-03-16 16:30:32,975 [usage_weekly_glacier] INFO basic.BasicDataManager - usage_weekly_glacier start polling... I've got a new run going today which should take a few hours, and I'll know more. username_1: Here's so far (without cluster benefits or reserved instances yet) (~14% cost reduction) ![screen shot 2015-03-16 at 12 02 21 pm](https://cloud.githubusercontent.com/assets/1373628/6671614/784b183e-cbd4-11e4-8369-8b7fc55052c6.JPG) ![screen shot 2015-03-16 at 12 02 38 pm](https://cloud.githubusercontent.com/assets/1373628/6671618/7d967842-cbd4-11e4-9214-b7436d93b69d.JPG) username_1: Today's update: wmappcluster and wmapp2cluster are live, combining: goggles, popcorn, thimble, webmaker-profile2(wmappcluster), wmscreenshot, events, events-api, and wmpublisher(wmapp2cluster). All the deploy scripts have been repointed. ![screen shot 2015-03-20 at 9 42 23 am](https://cloud.githubusercontent.com/assets/1373628/6753649/9c6a8fc4-cee5-11e4-996c-4a712fd2489f.JPG) username_1: We're at about 25% savings so far from where we started. Next week, I'll knock it down a bunch more (or over the weekend) by combining 5 apps into 2 clusters (badgekit, badgekit-api, openbadges, badgekit-mozilla, badgekit-api-mozilla). Then, I can likely cluster another number apps together in the same cluster architecture, and we'll be able to do at least some reserved instances for some really big cost reductions. username_0: Awesome, thanks for pushing on this, @username_1! username_1: This morning I migrated all the badging infra to the new clusters, and have scaled down the old clusters. I'll get a cost update for this change tomorrow when our costing bucket gets its files, and I'll have my eye on this infra today. Over the weekend, I got alerted that we were filling up logs. wmappcluster was storing a couple gigs of logs for thimble and goggles, and put together, it was bringing the drive to 97%. Since we send syslog and app logs to loggins, I trimmed down all app rsyslog configs to be "delete each day". username_1: ![screen shot 2015-03-24 at 9 46 50 am](https://cloud.githubusercontent.com/assets/1373628/6804257/bbc6e43a-d20a-11e4-9603-b2781c0b215f.JPG) username_1: So far, ~26% reduction in cost prior to reserving instances. username_2: Taking this ticket out of the March 13 milestone, since it's passed. username_2: @username_1 @username_0 - is this ticket still useful? should we assign to a 2016 milestone? Status: Issue closed
MetaMask/eth-phishing-detect
1183292436
Title: [Legitimate Site Blocked] officialnft.xyz Question: username_0: https://officialnft.xyz/ We don't collect any user signing keys and we don't Impersonate other known and established sites. The [tool](https://metamask.github.io/eth-phishing-detect/) provided to check the reason for blocking says there is no problem with our website. Please check screenshot below. We request to stop showing the phishing error on our website ![Screenshot from 2022-03-28 16-50-58](https://user-images.githubusercontent.com/29797105/160387525-ab38a2e3-1306-462a-ac83-c99a2643facc.png) Answers: username_1: Please update us on the whitelisting of our site. We don't collect any user signing keys and we don't Impersonate other known and established sites. username_1: @devenblake Any chance you can assist us with this? username_1: or... @AlexHerman1 Any chance you can assist us with this? username_1: I feel bad, @ing everybody but one last one... @Gudahtt , any chance you an assist us with this? username_2: https://officialnft.xyz/ We don't collect any user signing keys and we don't Impersonate other known and established sites. The [tool](https://metamask.github.io/eth-phishing-detect/) provided to check the reason for blocking says there is no problem with our website. Please check screenshot below. We request to stop showing the phishing error on our website ![160387525-ab38a2e3-1306-462a-ac83-c99a2643facc](https://user-images.githubusercontent.com/89087401/160627624-608ef86d-a855-4551-868c-eea8bc2f9f73.png) username_3: Hi, the domain does not resolve. Can you fix this so I can take a look. username_1: @username_3 our users are still getting the phishing detect warning. username_2: @username_3 @mcmire Any chance you can assist us with this? username_1: @username_3 @mcmire Hope we're not seeming inpatient. I know you guys are stacked with work. This phishing detect is completely killing the vibe for our site. New user adoption has dropped to zero and existing users are now afraid to login. We don't collect any user signing keys and we don't Impersonate other known and established sites. We're legit and we'd appreciate any effort to expedite us onto the allowlist/whitelist. If there's anything we need to do, please let us know but I'm hoping you can take 5 mins to look at our case and determine that we can get this phishing-detect removed. username_1: @devenblake Deven, any chance you an assist on this? We're don't seem to be able to get a response from the person assigned to this. Grateful for whatever you can do.
adafruit/circuitpython
769211065
Title: storage.remount("/", True) resulting in RuntimeError on 6.1.0-beta.2 Question: username_0: When working through [this CircuitPython storage example](https://learn.adafruit.com/circuitpython-essentials/circuitpython-storage) on QT Py M0, running 6.1.0-beta.2, it fails to work properly - the temperature logging file is never created and the filesystem never switches to read-only. Checking `boot_out.txt`, I find the following traceback: ``` Traceback (most recent call last): File "boot.py", line 18, in <module> RuntimeError: Cannot remount '/' when USB is active. ``` I tried attaching it to a USB battery pack, thinking it meant the USB data connection - this did not resolve the issue, the results are the same. Tried the example on a Trinket running 6.1.0-beta.2, and the results are the same. I tried the QT Py and Trinket running 6.0.0, and the example works properly. It appears that something changed between 6.0.0 and 6.1.0, and I am unsure whether it is a bug or a feature. Answers: username_1: dup of https://github.com/adafruit/circuitpython/issues/3709 ? username_0: Yes. Moved info there. Closing. Status: Issue closed
google/exposure-notifications-android
626650639
Title: Why the team preferred a native android app as reference instead of a cross-platform Flutter app? Question: username_0: It should be possible to implement a single app (android and iOS) and only change the EN calls to invoke the native APIs, right? Answers: username_1: This repo is to show how to use the API, not how to build an app using Flutter. Keep it simple. username_0: If anyone is looking for a cross-platform solution to simplify the implementation, you can check [CovidShield](https://github.com/CovidShield). They have a repo (https://github.com/CovidShield/mobile) with an implementation in ReactNative.
Chia-Network/chia-blockchain
941005313
Title: Where are all my 10000 plots? (1000Tb / 1Pb) [BUG] Question: username_0: Hello everybody! We have been making plots for 2 months. Since we have installed the new version 1.2 when we open the program there is not a single plot. We have added a pool like space.pool to see if this was the problem but no. The 10000 plots that we have been doing during these 2 months do not work? They said there would be no re-plots... someone can help me? Thanks! #bramcohen @bramcohen It cannot be that they have made us plot for months with chia blockchain, that they tell us that it would not be necessary to return to plott and now it seems that we have to start over? Explain a little more. are we going to have to re-plot every time they release a new version? NOTE: very ecological it is not with the light that is used to plot Answers: username_1: My plots aren't recognized via the CLI in version 1.2.0 (running Ubuntu) but they are recognized via the GUI. username_2: i currently facing the same issue, i did complete removal of chia and install 1.2 and while i was adding my "plot directories" in windows gui, it loaded some plots like 130 then adding more "plot directories" made my farm completely empty also the plot count is 0 i cant find any of my plots in the gui i checked the confgue file and my "plot directories" are listed there even in windows gui i can see some of my "plot directories" but no plots are showing at all! note: while running "chia plots check" or "chia plots show" they are listed!, someone please help me! username_3: Same issue here. Reverted to beta 1.1.795 then works again.... username_4: Re-update via command line and launch GUI via command line. Update instructions on GitHub that haven’t changed. I had an issue as well but somehow was on 1.2.0.dev version. Check your version if interested and then reinstall everywhere. You don’t need to delete you .chia folder which keeps your config intact. About 5 minutes per harvester to update. Should work smoothly. Good luck username_0: I have found the solution. You have to uninstall version 1.2 and delete the following folders: C: \ Users \ username \ .chia and C: \ Users \ username \ AppData \ Local \ Chia-Blockchain \ restart the pc and reinstall version 1.2. folders with plots and recognizes them. It is a bug of version 1.2 that when you install with the previous version it is as if it blocked your plots folders. username_0: however I certify that the plots of version 1.7 do not work with the pools. They clearly said that the plots from version 1.7 would work in version 2.0 but it is not totally true because they do not work for pools.... username_5: `2021-07-11T00:52:00.605 chia.plotting.plot_tools : ERROR Failed to open file E:\@FINALPLOTS\plot-k32-2021-07-07-21-11-54179b37bbbd16674b4d7e3c2d23bd60852822623ed1e56a5126b0a6aa12f637.plot. Invalid file E:\@FINALPLOTS\plot-k32-2021-07-07-21-11-54179b37bbbd16674b4d7e3c2d23bd60852822623ed1e56a5126b0a6aa12f637.plot Traceback (most recent call last): File "chia\plotting\plot_tools.py", line 189, in process_file ValueError: Invalid file E:\@FINALPLOTS\plot-k32-2021-07-07-21-11-54179b37bbbd16674b4d7e3c2d23bd60852822623ed1e56a5126b0a6aa12f637.plot ` what does this mean? username_1: I reverted to 1.1.6 and my plots are recognized via CLI. username_6: After I upgraded to 1.2.1 I don't see my plots, when I any command it shows some output (most likely incomplete) but they return ¨Segmentation fault (core dump)" message at the end, like this one: ![image](https://user-images.githubusercontent.com/8117289/125202589-3df33500-e242-11eb-870f-533fe306d2c7.png) username_0: parcels appear until the chia blockchain syncs. once the program is synchronized it is blocked. The program has an error and does not allow deleting even the locations (folders) of the plots. They also do not allow you to add the old plots to your own pool ... they said they would not work for the pools but they also do not work for solo plotting. BRAVO username_6: That looks bad username_6: I got the latest from github and compile, but I keep getting the Segmentation fault message when trying the chia command, I assume it is related to my issue of not displaying all my plots. username_6: I found that I was using Python 3.6 and not 3.7, that seems to be the whole issue all the time :) my bad. Now I have chia 1.3.0 and Python 3.7 (as indicated as the minimum version) and no more errors, back in business :) username_7: Per earlier comments, this issue appears solved - marking closed Status: Issue closed
Downquark7/ftc_app
190363615
Title: Need multiple simultaneous log files. Question: username_0: This would help with things because it would allow a file for debugging everything and other files for graphing large amounts of numerical data. Also, it would be helpful if the main debug file was an XML separated into different states then into separate parts of states. However, it might be too much effort for the XML file so the other part is more important.<issue_closed> Status: Issue closed
tomopy/tomopy
451557409
Title: tiff stack reconstruction issue Question: username_0: Hey, I was trying to read projection files as tiff files from a folder and somehow it only reads 2 of them even though there are 141 images inside. I used dxchange.reader.read_tiff_stack, and then tried to transpose matrix to make as in examples but it didn't work out well. Maybe I am missing something but I have checked again and everything seems ok. Any suggestion? from __future__ import print_function import tomopy import dxchange import numpy as np import math import matplotlib.pyplot as plt if __name__ == '__main__': # Set path to the micro-CT data to reconstruct. fname = '/Users/xx/Desktop/xx/xx_0000.tif' # Select the sinogram range to reconstruct. start = 0 end = 140 proj=dxchange.reader.read_tiff_stack(fname, ind=(start,end)) plt.imshow(proj[:, 0, :], cmap='Greys_r') plt.show() print("proj: ", len(proj)) proj.shape --> 2 theta = np.linspace(0, 140, num=141) theta = np.deg2rad(theta) # Flat-field correction of raw data. proj = tomopy.minus_log(proj) # Find rotation center. rot_center = tomopy.find_center(proj, theta, init=0, ind=0, tol=0.5) print("Center of rotation: ", rot_center) # Reconstruct object using Gridrec algorithm. rec = tomopy.recon(proj, theta, center=rot_center, algorithm='gridrec') # Mask each reconstructed slice with a circle. rec = tomopy.circ_mask(rec, axis=0, ratio=0.95) # Write data as stack of TIFs. #dxchange.write_tiff_stack(rec, fname='recon_dir/recon') Answers: username_1: Hi @username_0: From the docs of [`dxchange.read_tiff_stack`](https://dxchange.readthedocs.io/en/latest/source/api/dxchange.reader.html#dxchange.reader.read_tiff_stack)`: ``` ind : list of int Indices of the files to read. ``` This means you should provide _all_ of the indices of the files. You provided two indices, so you got two images. P.S. @username_2, please transfer this issue to the dxchange repository. Status: Issue closed username_2: what is the size [x, y] of each tiff file? francesco > username_0: @username_1 Thanks for the prompt reply and solution. And sorry for wasting your time with simple indexing mistake. @username_2 The image size is 1300x1600. username_1: @username_0, no worries. Brain farts happen! :smile:
nohaibogdan1/meteor-tuts-tutorial
359339064
Title: 4. Post upgrade Question: username_0: New to posts: a. A post ca be view by all app users(registered and unregistered) or only by registered. b. Logged in users can add reactions to posts. Similar to facebook: like, love, happy, wow, sad, angry
ethereum/tests
1038735012
Title: [documentation] Broken link in test/README.md documentation Question: username_0: In the README.md in the paragraph https://github.com/ethereum/tests/blob/develop/README.md#contribute-to-the-test-suite shall be update the link "[section](https://ethereum-tests.readthedocs.io/en/latest/generating-tests.html)" should be this https://ethereum-tests.readthedocs.io/en/latest/how2contribute.html Answers: username_1: Thank you, I'm about to submit a PR to fix it. Status: Issue closed
ioBroker/ioBroker.mihome
430781771
Title: Xiaomi Video Klingel Question: username_0: Hey Leute, Da die Video Tür Klingel von Xiaomi auch über die Mi Home App eingerichtet wird, würde ich es cool finden wenn ich zum beispiel den Bewegungssensor der Klingel auch im ioBroker abgreifen könnte um damit etwas triggern zu können, oder natürlich auch das Bild um es zum bespiel in VIS oder Telegram weiter verwenden zu können :) Hier ein paar nähere Details und ein Video von mir über die Klingel... https://dealheros.de/produkt/xiaomi-smarthome-video-tuerklingel-720p-ir-nachtsicht-%E2%9C%AA/ Liebe Grüße, Dennis Answers: username_1: Das wäre eine feine Sache, wäre schön wenn es umgesetzt werden könnte username_2: und ob man das Braucht würde mich super freuen über ne coole API :) username_3: We need it !!! username_4: Jaaa ich kann @username_0 nur zustimmen! Bitte umsetzen username_5: Bin auch dafür :-) username_6: I need urgently. Bitte umsetzten. username_7: Ich kann mich nur anschließen: Es wäre super, wenn eine Einbindung möglich wäre. username_0: am besten Teilt Ihr alle nochmal den Beitrag damit er mehr Beachtung findet und diejenigen unter euch die hier schon kommentiert haben das sie auch für solch eine Einbindung sind, sollten meinen Hauptbeitrag mit einem Daumen versehen, damit dieser eben entsprechend Beachtung findet :) username_6: Hat schon wer Fortschritte erzielt oder geht es sogar schon? username_8: wer hat denn eine im Einsatz und kann ein Logfile liefern? Abgesehen davon, dass ein "einrichten in der App" nicht bedeutet, dass es mit dem MiHome Adapter ueberhaupt auszuwerten ist. Kommt auf das Protokoll an, wenn die Klingel miio 2.0 nutzt, wird der mio-adapter benoetigt und hier waere das falsch. username_4: @username_8 Ich habe eine im Einsatz und könnte sogar eine zur Verfügung stellen falls das noch erforderlich ist? Hat schon jemand das Logfile geliefert? LG username_6: Ich habe noch keine und kann leider das logfile nicht liefern. username_4: @username_8 Wenn du mir verrätst wie ich an das Logfile komme, lasse ich es dir gerne zukommen :) username_8: in iobroker auf Instanzen, Expertenansicht, Logstufe auf Debug stellen. Dann an der Klingel mal rumspielen, dann das Logfile von /opt/iobroker/log/iobroker.datum.log mit "less iobroker.2019-08-07.log | grep mihome" filtern und hier hochladen. Dann wissen wir, ob die Klingel ueberhaupt was ueber den mihome adapter schickt, ich denke eher nicht, die verwendet wahrscheinlich das miio2.0 protokoll, dafuer gibts n anderen adapter. Aber abwarten. username_4: @username_8 aktuell läuft bei mir leider der Mi Home Adapter nicht, weil ich den Port öffnen muss - dies werde ich aber nun machen. Sobald der Adapter läuft werde ich das mal probieren. username_6: Schön dass Ihr an der Sache dran seit. Bin gespannt wie’s ausgeht. username_0: @lovegym66 Welchen anderen Adapter gibt's denn wo wir eventuell Signale von der Klingel bekommen? username_8: Hi, es gibt mindestens 2 Varianten, einmal den MiHome Adapter hier, und einmal den mioo adapter, der viele andere Geraete unterstuetzt, die ueber das miioo Protokoll laufen. Es kann der mioo Adapter problemlos dazu installiert werden. Adapter auf Debug stellen, ganz wichtig, dann ein paarmal auf die klingel druecken, wenn wir Glueck haben, ist was im Protokoll zu sehen, ansonsten hilft nur ein Auszug mit Wireshark. Ich hatte mir vor ein paar Monaten so ne billige China Videoklingel geholt, dachte ich koennte da wenigstens den Videostream oder n Signal abgreifen, aber die sendet alles verschluesselt ueber UDP mit rolling Ports... keine Chance.. schon doof, wenn die Sicherheit so gut ist.. :-) username_6: @lovegym66 den mioo adapter kann ich nicht finden. der mihome adapter wird nicht grün da ich auch keinen key habe. auf debug kann ich ihn stellen. Was kann ich machen (habe keine Xiaomi Gateway) Würde dir gerne das Protokoll schicken. username_8: wenn die Klingel nicht am Gateway eingebunden wird, kann hier der Adapter nix machen, da ein Gateway zwingend da sein muss. Dann beim mioo Adapter ein Issue erstellen. https://github.com/smarthomefans/ioBroker.miio username_4: @username_8 Soo ich habe es gerade endlich mal geschafft im Protokoll nach "less iobroker.2019-08-07.log | grep mihome" zu suchen - wird aber nichts gefunden... Habe mir nun den miio Adapter installiert. Was genau muss ich nun eingeben? Wie finde ich die IP Adresse der Klingel raus? username_6: Das ist gut sonst hätte ich demnächst geschaut. Die ip findest du im Router. Name müsste sein DLing-DoorBell. username_8: Das issue hier ist im falschen Adapter, bitte schliessen. username_0: Kann man es dann nicht irgendwie verschieben, wir wünschen uns irgendwie einen Adapter dazu, wäre so cool wenn da jemand einen Plan hätte wie man das hinbekommen könnte username_8: was willst du denn verschieben, hier steht absolut nix wissenswertes ... ausser, dass zwei Leute das Teil haben.. Zum debuggen braucht man zumindest mehr Informationen, Ports, Protokoll, Datenverkehr.. also wireshark anwerfen und mitschneiden und auswerten. Aber hier der Adapter kann nur devices, die über das Gateway gehen. username_6: https://github.com/smarthomefans/ioBroker.miio/issues/21 Hab mir erlaubt in dem anderen Adapter zu posten. Bitte folgt alle um die Anbindung voran zu treiben.
neuml/codequestion
678387730
Title: Vector model file not found Question: username_0: Hello, Thank you very much for the project. But I have one small issue, right now it seems that when you run `python -m codequestion.download` it downloads a configuration file that will be used by codequestion to load the model. The path to the model seems hardcoded to `/home/dmezzett/.codequestion/vectors/stackexchange-300d.magnitude` How can we specify to codequestion to use our home or modify the config file? Best Answers: username_1: Thank you for reporting this, I'm working on an updated version of codequestion backed by txtai. I will make sure to resolve this issue with that release. In terms of a workaround, the config file is just a dict object that was pickled to a file. If you open a python shell in ~/.codequestion/models and edit/run the following code, if should workaround the issue: ``` import pickle with open("config", "rb") as handle: config = pickle.load(handle) config["path"] = "/<your home directory>/.codequestion/vectors/stackexchange-300d.magnitude" with open("config", "wb") as handle: pickle.dump(config, handle, protocol=pickle.HIGHEST_PROTOCOL) ``` username_0: Thanks for your quick response ! Status: Issue closed username_2: I just want to point out to newcomers that the script proposed by username_1 should be run in the "stackexchange folder". Btw, thanks for this amazing package!
magma-hackers/finden
394449881
Title: Create table from visorias - B Question: username_0: ### *Aceptance Criteria* * La tabla debe de contener los siguientes campos: fecha lugar horario club (club logeado) * visores (combo de visores que tiene el club) descripción para el jugador categorías (combo 2009 - 2018) ****** campos nuevos - opción de que seleccionen la categoría x horario que va a estar - tipo de visoria (abierta o privada) - division (categoria sub17, premier, liga interna, etc) ****** * El club que este logeado será el que publique la visoria. * El campo descripción para el jugador debe ser un campo texto para que el club pueda poner información referente a la visoria por ejemplo: " traer playera blanca y short de color negro, asiste con copia de tu credencial de electro, busca al visor <NAME> para registrar tu asistencia, etc". ### Points * 5 ### Type of task _Normal_<issue_closed> Status: Issue closed
conjure-up/conjure-up
236358598
Title: Exception: Unable to determine controller: ERROR controller conjure-up-controller not found Question: username_0: https://sentry.io/canonical-pj/conjure-up/issues/295577769/ ``` Exception: Unable to determine controller: ERROR controller conjure-up-controller not found (8 additional frame(s) were not displayed) ... File "conjureup/controllers/lxdsetup/common.py", line 45, in setup self.next_screen() File "conjureup/controllers/lxdsetup/common.py", line 32, in next_screen return controllers.use('controllerpicker').render() File "conjureup/controllers/controllerpicker/tui.py", line 9, in render self.finish(app.argv.controller) File "conjureup/controllers/controllerpicker/common.py", line 33, in finish c_info = juju.get_controller_info(app.current_controller) File "conjureup/juju.py", line 679, in get_controller_info sh.stderr.decode('utf8'))) Exception: Unable to determine controller: ERROR controller conjure-up-controller not found ``` A little more context is that we should be able to provide `conjure-up spell cloud controller model` and if controller and model do not exist they are created during bootstrap. This has recently changed in the last 15 or so days<issue_closed> Status: Issue closed
telesoho/vscode-markdown-paste-image
988206578
Title: command 'telesoho.MarkdownPaste' not found Question: username_0: ![image](https://user-images.githubusercontent.com/51902242/132082195-4dcceff6-06b3-4d81-b7a0-b93fb2cf445b.png) Previously, it still working fine but not today. In particular, the cmd + alt + v keybinding is not working. while it is registered at `default keybindings.json`, see below ![image](https://user-images.githubusercontent.com/51902242/132082267-86a61f07-6def-4f7a-a605-3707b88973ef.png) Version: 1.57.1 Commit: <PASSWORD> Date: 2021-06-17T13:28:32.912Z Electron: 12.0.7 Chrome: 89.0.4389.128 Node.js: 14.16.0 V8: 8.9.255.25-electron.0 OS: Darwin x64 20.6.0 Answers: username_1: @username_0 You can toggle vscode develop tools on and try to use username_1.MarkdownPaste again to see what's going on with the extension like this: ![image](https://user-images.githubusercontent.com/10979091/132098572-8faa5fd5-5fc7-4cad-87ce-9c065adbada8.png) username_0: ![image](https://user-images.githubusercontent.com/51902242/132104784-ee6560a0-7dc1-450e-b172-a1d0ad577a7d.png) username_1: @username_0 You installed extension is ver 0.13.6, please update it to lastest ver 0.13.8, then try again. username_0: Work perfectly. Thanks. Status: Issue closed
JuliaLang/julia
218286656
Title: `convert(Tuple{Int,Int,Int}, (1,2))` doesn't throw on julia-0.5 Question: username_0: On Julia 0.5 we seem to have the following: ```julia $ julia _ _ _ _(_)_ | A fresh approach to technical computing (_) | (_) (_) | Documentation: http://docs.julialang.org _ _ _| |_ __ _ | Type "?help" for help. | | | | | | |/ _` | | | | |_| | | | (_| | | Version 0.5.1 (2017-03-05 13:25 UTC) _/ |\__'_|_|_|\__'_| | |__/ | x86_64-linux-gnu julia> convert(Tuple{Int,Int,Int}, (1,2)) (1,2) ``` A more complex issue (that showed up only when inlining was off) was picked up over at https://github.com/JuliaImages/ImageFiltering.jl/pull/28. Here's a simple reproducer. On Julia 0.5: ```julia $ julia --inline=no _ _ _ _(_)_ | A fresh approach to technical computing (_) | (_) (_) | Documentation: http://docs.julialang.org _ _ _| |_ __ _ | Type "?help" for help. | | | | | | |/ _` | | | | |_| | | | (_| | | Version 0.5.1 (2017-03-05 13:25 UTC) _/ |\__'_|_|_|\__'_| | |__/ | x86_64-linux-gnu julia> abstract AbstractBorder julia> immutable Pad{N} <: AbstractBorder style::Symbol lo::Dims{N} # number to extend by on the lower edge for each dimension hi::Dims{N} # number to extend by on the upper edge for each dimension end julia> (::Type{Pad{N}}){N}(style, lo::AbstractVector, hi::AbstractVector) = Pad{N}(style, (lo...), (hi...)) julia> Pad{3}(:replicate, [1,2], [5,3]) Pad{3}(:replicate,(1,2,140604813857440),(5,3,1)) ``` On master it behaves properly: ```julia $ julia-0.6 --inline=no _ _ _ _(_)_ | A fresh approach to technical computing (_) | (_) (_) | Documentation: http://docs.julialang.org _ _ _| |_ __ _ | Type "?help" for help. | | | | | | |/ _` | | | | |_| | | | (_| | | Version 0.6.0-pre.alpha.306 (2017-03-29 09:24 UTC) _/ |\__'_|_|_|\__'_| | Commit 1eb<PASSWORD> (1 day old master) |__/ | x86_64-linux-gnu julia> abstract AbstractBorder WARNING: deprecated syntax "abstract AbstractBorder" at REPL[1]:1. [Truncated] julia> immutable Pad{N} <: AbstractBorder style::Symbol lo::Dims{N} # number to extend by on the lower edge for each dimension hi::Dims{N} # number to extend by on the upper edge for each dimension end julia> (::Type{Pad{N}}){N}(style, lo::AbstractVector, hi::AbstractVector) = Pad{N}(style, (lo...), (hi...)) julia> Pad{3}(:replicate, [1,2], [5,3]) ERROR: MethodError: Cannot `convert` an object of type Tuple{} to an object of type Tuple{Int64} This may have arisen from a call to the constructor Tuple{Int64}(...), since type constructors fall back to convert methods. Stacktrace: [1] convert(::Type{Tuple{Int64,Int64}}, ::Tuple{Int64}) at ./essentials.jl:136 (repeats 2 times) [2] Pad{3}(::Symbol, ::Tuple{Int64,Int64}, ::Tuple{Int64,Int64}) at ./REPL[2]:2 [3] Pad{3}(::Symbol, ::Array{Int64,1}, ::Array{Int64,1}) at ./REPL[3]:1 [4] eval(::Module, ::Any) at ./boot.jl:235 ``` Since it works on master, I recognize this may not have the highest priority, but if there's an easy fix... Answers: username_1: My guess is that this probably won't happen... username_2: worth reverse-bisecting to find what fixed it? and is it tested on master if we want it to stay this way? username_1: It's due to the corrected type behavior of `Type{Tuple{Vararg{T}}} where T`. No need to bisect, but worth adding tests. Status: Issue closed username_0: #21275
i-net-software/jlessc
615348437
Title: :extend() not working with elements Question: username_0: Hi, I'm currently switching over from https://github.com/marceloverdijk/lesscss-java to your less compiler and had pretty good results so far. However, I stumbled upon an issue which the old compiler solved without a problem: The [`:extend()`](http://lesscss.org/features/#extend-feature) feature does not seem to work with elements as a parameter value: ```less h2 { &:extend(h1); } ``` This throws an error message: ``` Undefined mixin: &:extend on line 1320, column 14 on line 1318, column 24 ``` The examples in the Less docs state, that extending an element selector is supported. It works fine though with a class selector: `&:extend(.heading1);` Any chance to get this fixed? Thanks :-) Status: Issue closed Answers: username_1: I have create a snapshot version that you can test it. username_0: Awesome! I built `1.9-SNAPSHOT` locally and it works perfect! Thanks for the quick fix! Do you have an ETA for the next stable release?
jsdom/whatwg-url
928720723
Title: "ReferenceError: Can't find variable: whatwgURL" on Safari 14.1 Question: username_0: 1. Navigate to https://jsdom.github.io/whatwg-url/#url=aHR0cHM6Ly9leGFtcGxlLnRlc3Qv&base=YWJvdXQ6Ymxhbms= 2. See "ReferenceError: Can't find variable: whatwgURL" in the "jsdom/whatwg-url's components" section 3. See "[Error] ReferenceError: Can't find variable: SharedArrayBuffer" in the console It would be very useful if this worked in Safari. Is the dependency on SharedArrayBuffer required? Answers: username_1: We've tried to work around this a few times; I guess it keeps breaking? https://github.com/jsdom/whatwg-url/commit/2083a47cd0d9d981bb7bb228606e5a9f077d7a3b was the latest attempt... username_0: What is the dependency on SharedArrrayBuffer used for? If this issue keeps coming up we could consider an alternative primitive. username_1: It's part of a support library for our Web IDL bindings. It's not really used directly, but messing with the dependency (which is Node.js-focused) is going to be tricky. Status: Issue closed
conikeec/HelloShiftLeft
382907208
Title: Ocular - detected Path Travesal Vulnerability Question: username_0: # Path Traversal vulnerability discovered ## Flow sequence ``` PARAMETER : LINE_NUMBER : METHOD_NAME : FILE_NAME categoryName : 30 : getRssForCategoryUnsafe : io/shiftleft/controller/RSSController.java categoryName : 32 : getRssForCategoryUnsafe : io/shiftleft/controller/RSSController.java param0 : SYSTEM : append : java/lang/StringBuffer.java this : SYSTEM : append : java/lang/StringBuffer.java filePath : 32 : getRssForCategoryUnsafe : io/shiftleft/controller/RSSController.java filePath : 33 : getRssForCategoryUnsafe : io/shiftleft/controller/RSSController.java this : SYSTEM : append : java/lang/StringBuffer.java this : SYSTEM : append : java/lang/StringBuffer.java filePath : 33 : getRssForCategoryUnsafe : io/shiftleft/controller/RSSController.java filePath : 36 : getRssForCategoryUnsafe : io/shiftleft/controller/RSSController.java this : SYSTEM : toString : java/lang/StringBuffer.java $ret : SYSTEM : toString : java/lang/StringBuffer.java filePath.toString() : 36 : getRssForCategoryUnsafe : io/shiftleft/controller/RSSController.java param1 : SYSTEM : <operator>.assignment : N/A param0 : SYSTEM : <operator>.assignment : N/A $r1 : 36 : getRssForCategoryUnsafe : io/shiftleft/controller/RSSController.java $r1 : 36 : getRssForCategoryUnsafe : io/shiftleft/controller/RSSController.java param0 : SYSTEM : <init> : java/io/FileInputStream.java ``` ### Remediation proposal: Input Validation Assume all input is malicious. Use an “accept known good” input validation strategy, i.e., use a whitelist of acceptable inputs that strictly conform to specifications. Reject any input that does not strictly conform to specifications, or transform it into something that does. Do not rely exclusively on looking for malicious or malformed inputs (i.e., do not rely on a blacklist). However, blacklists can be useful for detecting potential attacks or determining which inputs are so malformed that they should be rejected outright. ### Remediation proposal: Reference Map When the set of filenames is limited or known, create a mapping from a set of fixed input values (such as numeric IDs) to the actual filenames, and reject all other inputs. For example, ID 1 could map to “technology.xml” and ID 2 could map to ” economy.xml ” etc.
MartinKamenov/GoSport
359755996
Title: App crashes when map for creating event opens on new device. Question: username_0: When application has been installed and map on create event is opened the application crashes. ### Steps to reproduce 1. Fresh install on new device 2. Open new event activity 3. Press button for selecting place.<issue_closed> Status: Issue closed
shoppingflux/shoppingfluxexport
279005529
Title: Incomplete cleaning of tag's name - should not start with digit and match should be global Question: username_0: ## Expected behaviour XML tag's name cannot starts with a digit and should not contains other characters than aA-zZ and digits. ## Actual behaviour The current regex in charge of cleaning the tag name is not removing the starting digits. It's creating issues with product's attributes and features that may start with a number and will therefore create an invalid XML tag. From the regex in the code, there is also other ## Steps to reproduce the behaviour 1 - Create a product's attribute or feature starting with a digit. 2 - Start XML feed generation Status: Issue closed Answers: username_0: **Reopened** In some specific cases we have tag's names containing only numbers, in which case 60e0b80 is not efficient. We need to review the tag's cleaning logic. The idea would be to restablish the initial regex before 60e0b80 and to add the other regex that remove starting digits by replacing them with a string such as _rep-2_ username_0: ## Expected behaviour XML tag's name cannot starts with a digit and should not contains other characters than aA-zZ and digits. ## Actual behaviour The current regex in charge of cleaning the tag name is not removing the starting digits. It's creating issues with product's attributes and features that may start with a number and will therefore create an invalid XML tag. ## Steps to reproduce the behaviour 1 - Create a product's attribute or feature starting with a digit. 2 - Start XML feed generation Status: Issue closed
oklas/react-app-rewire-alias
630900374
Title: typo in eslint filter apply Question: username_0: @username_0 This almost works, but there is one bug: ```diff - if(rule.use && 0 < rule.use.filter(isRuleOfEslint)) return true + if(rule.use && 0 < rule.use.filter(isRuleOfEslint).length) return true ``` Otherwise, this function always returns false. You can easily test it like this: ```js module.exports = function override(config) { const configCopy = JSON.parse(JSON.stringify(config)); aliasDangerous({ ...configPaths('tsconfig.paths.json') })(config); console.log("These should be equivalent:") console.log(configCopy.module.rules[1].include); console.log(config.module.rules[1].include); return config; } ``` Also note that you might have cache pollution and need to delete the `.cache` folder in `node_modules`. I opened a new PR with the fix for the fix and some documentation. _Originally posted by @JollyGoodHolly in https://github.com/username_0/react-app-rewire-alias/issues/3#issuecomment-638808510_<issue_closed> Status: Issue closed
ikedaosushi/tech-news
649016066
Title: 銀行での不正検知の精度向上へ 暗号のままデータ連携機械学習ができるわけ | DG Lab Haus Question: username_0: &#37504;&#34892;&#12391;&#12398;&#19981;&#27491;&#26908;&#30693;&#12398;&#31934;&#24230;&#21521;&#19978;&#12408; &#26263;&#21495;&#12398;&#12414;&#12414;&#12487;&#12540;&#12479;&#36899;&#25658;&#12539;&#27231;&#26800;&#23398;&#32722;&#12364;&#12391;&#12365;&#12427;&#12431;&#12369; | DG Lab Haus<br> <br> https://ift.tt/2CNoxW4
readium/r2-streamer-kotlin
297801405
Title: handle bad values in dc:date Question: username_0: In https://github.com/edrlab/readium-desktop/issues/54 it appears that bad values in dc:date of the form <dc:date>--T00:00:00Z</dc:date> only raise a warning in epub-check. The kotlin parser should handle such an error with a fallback. note: the same to be applied to the iOS codebase. Answers: username_0: The dc:date is optional in EPUB. It is the date of publication of the EPUB (not the book). Therefore if the date is badly formatted, we must do as it was not present in the EPUB. question: What is the in-memory value of the date if it is not present in the EPUB ? username_1: Fixed in PR #89. If present, dc:date is tried to be parsed and null in case of failing. Status: Issue closed
mimblewimble/grin
428714390
Title: Owner_api configuration in confiuration file Question: username_0: Owner_api is listen in 127.0.0.1 . how can i access through ip . All the rest of the server listen addresses are in the configuration except that one. It is for compatibility with docker/cloud. Because need to access the owner_api form different container. Answers: username_1: You should be able to expose the port of the container running the owner api and you get access to the wallet at the container network level. Then, you need an additional route that you should be able to setup via the loadbalancing method offered by the platform you use. username_2: See https://github.com/mimblewimble/grin/issues/2663 (and some others) It's recurrent question, perhaps we need to address it somehow. However the wallet has been moved to https://github.com/mimblewimble/grin-wallet, I'm closing this issue, feel free to open a discussion there. Status: Issue closed
jiangchenguang/blog
396267991
Title: vue组件原理 Question: username_0: - - - - vue组件 - - - - # 概要 vue允许使用组件的方式 拆分和复用模块,本文将解释vue组件实现的原理。 我们已经知道vue挂载是**从根节点到叶子节点**、**深度优先**这样的一个过程,将组件标签理解成一个特殊的html标签,需要特殊的创建方法。 我们知道vue更新dom节点分为两步: 1. 生成新的虚拟dom树 2. 对比新旧虚拟dom树,以“打补丁”将旧的dom树修改为新的dom树 ## 生成render函数 这个阶段就是将**字符串模版翻译成render函数**,此阶段父组件不关心字符串模版中的标签是html标签还是组件名称。唯一需要注意的是,如果使用了`is`属性,那么需要将`is`对应的值设置成标签。 ## 生成虚拟dom 这个阶段父组件执行`render`函数生成最新的虚拟dom树,我们看下创建一个虚拟dom对应的函数,即:`_createElement`函数: ``` javascript // src/core/vnode/create-element.ts function _createElement( context: Vue, tag?: string, data?: VNodeData, children?: any, normalizeType?: number ): VNode { if ((Ctor = resolveAsset( context.$options, “components”, tag)) ) { vnode = createComponent(Ctor, data, context, tag); } else { vnode = new VNode(tag, data, children, undefined, context); } return vnode; } ``` 简化的逻辑非常简单:`resolveAsset`函数判断当前`tag`是否是组册的组件名,返回该组件的值,并调用`createComponent`创建组件对应的虚拟dom,否则创建html标签对应的虚拟dom。 下面看下创建组件对应的虚拟dom的函数,`createComponent`: ``` javascript // src/core/vnode/create-component.ts export function createComponent( Ctor: typeof Vue | Object, data: VNodeData, context: Vue, tag: string ): VNode { const base = context.$options._base; if (isPlainObject(Ctor)) { Ctor = base.extend(Ctor); } if (typeof Ctor !== 'function') { return; } resolveConstructorOptions(Ctor); data = data || {}; [Truncated] parent: Vue, parentElm: HTMLElement, refElm: HTMLElement ): Vue { const componentOption = vnode.componentOptions; const options = { _isComponent: true, propsData: componentOption.propsData, parent, _parentVnode: vnode, _parentElm: parentElm, _refElm: refElm } return new componentOption.Ctor(options); } ``` 调用了两个函数,其实就是实例化一个vue子类并挂载。 # 更新dom节点 # 删除dom节点 # 附录<issue_closed> Status: Issue closed
NeurodataWithoutBorders/pynwb
320623084
Title: `ImageSeries.dimensions` should be `int` Question: username_0: from the tutorial: ```python image_series = TwoPhotonSeries(name='test_iS', source='Ca2+ imaging example', dimension=[2], external_file=['images.tiff'], imaging_plane=imaging_plane, starting_frame=[0], format='tiff', timestamps=list()) ``` I think `dimensions` should be an `int`. It is `int` in [nwb-schema](https://github.com/NeurodataWithoutBorders/nwb-schema/blob/76fdcc0051568ed3ace184da751dc20d2d202407/core/nwb.image.yaml): ```yaml - rank doc: Number of pixels on x, y, (and z) axes. dtype: int32 name: dimension quantity: '?' shape: - null ```
vuetifyjs/vuetify
480956515
Title: [Bug Report] The max-height prop for v-card does not prevent contents from overflowing Question: username_0: ### Environment **Vuetify Version:** 2.0.7 **Vue Version:** 2.6.10 **Browsers:** Chrome 76.0.3809.100 **OS:** Windows 10 ### Steps to reproduce Add `max-height` to a `v-card` with several elements inside. Make sure the `max-height` is less than what is required to display all of the elements. ### Expected Behavior If elements would overflow the `max-height`, then it should scroll. ### Actual Behavior By specifying the `max-height` prop on a `v-card`, the height of the card is modified, but any elements inside the card do not appear affected and remain where they were as though the card's height did not change. ### Reproduction Link <a href="https://codepen.io/username_0/pen/BaBKQQp" target="_blank">https://codepen.io/username_0/pen/BaBKQQp</a> <!-- generated by vuetify-issue-helper. DO NOT REMOVE --> Answers: username_1: Odd, this has gone on since it was introduced in like 1.2 (at least can reproduce it in 1.5). Work around really is just to set overflow-y explicitly. Technically in your example, you would be better off setting that though on `v-card-text` as if you set overflow on the `v-card` you scroll **all** contents, including your fab and title. There's also a FR for adding overflow props to `v-card` so kinda in conflict as best approach to fix this. @vuetifyjs/core-team whats your thoughts on this? username_0: Yeah, the workaround I came up with was to conditionally apply `overflow-y: scroll` to the `v-card` and then perform some logic to change the positioning of the fab from `absolute` to `fixed` depending on whether or not the number of items in the list would overflow the viewport. Works in a pinch for what I want it to do, but isn't ideal. username_2: Any updates on this?
autozimu/LanguageClient-neovim
452986764
Title: Project root not updated when editing file from different project Question: username_0: ### Environment - Plugin version: 3e31d01 - Plugin binary version: 0.1.146 - Neovim version: 0.3.5 - Language: C - Language server: `ccls` latest release (0.20190314.1) ### Describe the bug After launching Neovim and editing a file belonging to a project in my workspace, LanguageClient successfully detects the project root, using my custom root marker. The project root is printed and logged correctly, and all the language client+server functionalities work. When I start editing another file belonging to a different project, the project root is not updated, in fact is remains the same as before thus breaking all the functionalities. The workspace tree is something like the following: ``` workspace │ ├── project_a │   ├── build │   │   └── compile_commands.json │   └── source │   └── main.c │ └── project_b ├── build │ └── compile_commands.json └── source └── main.c ``` My root marker is `build/compile_commands.json`. LanguageClient is set to automatically start. To recap, I start Neovim inside `workspace`, then `:edit project_a/source/main.c` and LanguageClient outputs ``` [LC] Project root: workspace/project_a ``` and everything works. Then I `:edit project_b/source/main.c` but the project root is not updated, thus also the `compile_commands.json` database is not loaded, and nothing works as expected. Am I missing something or is this a bug/non-feature of LanguageClient? Answers: username_1: This is how this plugin works as of now. In order to associate files of same type + project to the same language server instance, filetype is used as key mapping to language server instances. This might be updated to use project root as key in the future. I cannot guarantee when it will come or whether it will. As a workaround, when switching project, you can run `:LanguageClientStop`, `:LanguageClientStart` to run the language server for project b. Alternatively running different neovim session for different project.
lowlighter/metrics
860772426
Title: Merge commits are not condensed Question: username_0: When merging commits form upstream, Metrics does not condense merge commits, instead spamming the recent activity section. It would be better if Metrics removes all other commits if one of them matches "Merge branch Y into Z". ![Screenshot_20210419-094345](https://user-images.githubusercontent.com/42429413/115161851-03bb4100-a0f4-11eb-82c8-0597f9569ac5.png) Answers: username_0: More of a feature request than a bug, I just didn't remove the label manually when creating (there's no "generic" issue template) username_0: I assume this could be this be done by just adding something like: ```js if (/Merged branch '.+?' into .+?/.test(commits[commits.length-1].message)) commits = commits[commits.length-1]; ``` to these lines: https://github.com/username_1/metrics/blob/ec03286c90b7c1f6b4a1174a2106c362a9db7921/source/plugins/activity/index.mjs#L108-L111 username_1: Yes that would be nice 👍🏼 I'm actually wondering if we should add a `.reverse()` too on the commits list, so pushed commits would appear from most recent to older instead Status: Issue closed username_1: Thanks a lot for your contributions in #242 and #244 !
jashkenas/coffeescript
52959891
Title: Enhancement: functions that by default have no return result, aka 'undefined' Question: username_0: [email protected] As of now, you will always have to include an undefined on the last line of a function's block, i.e. ``` a = -> do_something undefined ``` otherwise, coffee-script will try to compile a return result from either the last invocation or assignment, or, if that fails, it will compile an array, injecting the required code and thus making things overly complex. How about adding the following syntax extension to functions that will instruct the compiler to just return undefined, aka void 0? ``` a = ->> do_something ``` or, for bound functions ``` a = =>> do_something ``` I think that this extra syntactic sugar will make things more easier for all of us. Please note that I settled on doubling the ```>``` instead of introducing yet another character. Having tried out multiple alternatives, for example ```->:``` or ```->u```, simply doubling the ```>``` seems to be more readable and also easier to type. Answers: username_1: If this discussion keeps being brought up, maybe that's a sign that something in the language should be changed.
karawin/Ka-Radio32
1082456700
Title: Changing to another LCD Question: username_0: I'm sorry, but I'm not familiar with programming and compilation. I read in various files that I can change LCD after changing the sys.lcd("x") parameter in the HardwareConfig.md file. In another place I read that a .csv or .bin file has to be generated. I even zaisntalled Eclipse, esp-idf-v4.0.4 and MSYS64bit. However, I do not understand much of this. Can I count on a hint? Answers: username_1: sys.lcd("x") is the command to change the lcd type. It must be done one time only and registered in the configuration. This command can be sent with telnet or the serial I/O username_0: Big thanks for the tip. I works great. Now I can work on the nice cover. I'll going to send to you final effect. Great work! Status: Issue closed
telstra/open-kilda
391036633
Title: Omitting initial exception context on handling exceptions in some Storm bolts Question: username_0: On guard exception handlers like this: ``` } catch (Exception e) { throw new MessageException(message.getCorrelationId(), System.currentTimeMillis(), ErrorType.UPDATE_FAILURE, errorType, e.getMessage()); } ``` initial exception context not passed as cause of new exception nor logged. That makes unexpected behavior investigation much complex.
jlippold/tweakCompatible
497581640
Title: `SB EW21` working on iOS 12.4 Question: username_0: ``` { "packageId": "com.evynw.sbew21", "action": "working", "userInfo": { "arch32": false, "packageId": "com.evynw.sbew21", "deviceId": "iPhone10,5", "url": "http://cydia.saurik.com/package/com.evynw.sbew21/", "iOSVersion": "12.4", "packageVersionIndexed": true, "packageName": "SB EW21", "category": "Homescreen Widgets", "repository": "Evelyn's Collection", "name": "SB EW21", "installed": "1.0", "packageIndexed": true, "packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.", "id": "com.evynw.sbew21", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.1.5", "shortDescription": "A widget. Like what else do you think this is?", "latest": "1.0", "author": "Evelyn (ev_ynw)", "packageStatus": "Unknown" }, "base64": "<KEY> "chosenStatus": "working", "notes": "" } ```<issue_closed> Status: Issue closed
a2develop/bugTracker
562494156
Title: Корректировать автонумерацию по последним цифрам, даже если есть дробь (/) Question: username_0: Кейс: префикс для счетов не настраивается? Например, счета имеют формат ДР-20/ххх. Те ДР-20/ в пределах года не меняется, а меняется ххх. Буккипер меняет первое число, которое нашел те второй счёт он делает просто ДР-21. А должен менять ххх. Я ставлю ДР-20/001. Следующий счёт создается ДР-21.
jupyterlab/jupyterlab
720487852
Title: Stylesheet failed to load in case of a path segment that includes a : Question: username_0: Hi. We are integrating JupyterLab into a cloud environment that has some special path handling. It uses urls like this: https://host-name/path1/abc:xyz/lab JupyterLab loads fine but at the place it generates absolute urls for CSS files it seems to make an error that leads to Error Loading Theme Stylesheet failed to load: abc:/path1/abc:xyz/lab/api/themes/@jupyterlab/theme-light-extension/index.css It might be related to the way that urls are generated here: https://github.com/blink1073/jupyterlab/blob/e633ff77d39b00284359b6ae8eea5ec7f7bdd813/packages/coreutils/src/url.ts#L55 Answers: username_1: I'll take a look at this username_1: @username_0 I'm trying to recreate this and wanted to confirm you only see this when there's a `:` in the url path and not in other cases? username_1: Opened #9169 with a fix and tagged it for backport username_0: @username_1 Thanks for looking into this. I only checked with : as that it what appears in our infra. Status: Issue closed
ModellingWebLab/cellmlmanip
578550787
Title: Parse RDF in external files Question: username_0: And coordinate with PMR so that the format is similar? Answers: username_0: What format would we want, @username_1 ? Everything in a centralised triple store? A separate file per model? username_1: A file per model, primarily. If they're packaged up in a COMBINE Archive (or a Git repo with equivalent content) you'd also want the separate file listed in the manifest for that model. username_0: Would that give us something like ``` <rdf:RDF xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"> <rdf:Description rdf:about="./model-file.cellml#some_variables_id"> ``` ? username_1: Yep, that's the kind of thing. (Though the "./" isn't strictly necessary, and shouldn't be assumed!) username_0: Cross-posting from https://github.com/ModellingWebLab/project_issues/issues/75 (thanks @skeating ) https://drive.google.com/file/d/1AngoIjQovM7DpCwR0W6TP0admMloGa9k/view The draft has `is` and `isVersionOf` and while a full implementation might have some complications creating something that reads the bits we like and ignores the rest should be fairly staightforward, I think username_0: E.g. 1. *Load model*: - If it's a cellml model then parse it as before - If it's anything else then assume a zip file or a directory, look inside for a manifest file and treat it at a combine archive, then look for a meta data file and continue pretty much as before username_0: Would that make sense @username_1 @skeating ? And who wants to have a first go at implementing something? username_0: We'll have to discuss with @nickerso if that would make us compatible with PMR in the short run :D username_1: That seems reasonable to me. So the load model function can take any of: 1. CellML file - current implementation 2. Directory or Zip file - treat as (maybe unpacked) Combine Archive, look for the manifest.xml file, error if not found 3. manifest.xml file directly? When reading the manifest.xml, it looks for the entry with `master='true'` and checks it's a `format="http://identifiers.org/combine.specifications/cellml"` (or sub-version 1.0 thereof). What format would we allow for metadata files? `http://identifiers.org/combine.specifications/omex-metadata` at least I'm guessing, maybe a generic RDF mimetype, maybe also .ttl or anything rdflib can read?
dcmk1mr2/ebarc-org
309232013
Title: Q - can the photo thumbs be exposed more or in a better way? Question: username_0: I changed it to be taller. I know it breaks the nice spacing but people complained (go figure - Berkeley). Answers: username_1: Can you elaborate on what you mean by "exposed more"? Also, I'm not sure what you did, exactly, as I can't find it in a specific commit.
ddnionio/news
271032789
Title: @realDonaldTrump: The rigged Dem Primary one of the biggest political stories in years got ZERO coverage on Fake News Network TV last night. Disgraceful! Question: username_0: @realDonaldTrump: The rigged Dem Primary, one of the biggest political stories in years, got ZERO coverage on Fake News Network TV last night. Disgraceful!<br> via Twitter http://twitter.com/realDonaldTrump/status/926481563214376961<br> November 04, 2017 at 12:09AM
cloudinary/cloudinary_php
811537836
Title: Media::fromParams doesn't default to secure, nor respect configuration Question: username_0: ## Bug report for Cloudinary PHP SDK ## Describe the bug in a sentence or two. I've upgraded to version 2 of the Cloudinary PHP SDK and used the migration guide to convert from `cloudinary_url` to `Media::fromParams`. However, this method doesn't default to secure URLs (odd behaviour), and it doesn't respect a global secure configuration option either. ```php Cloudinary::instance([ 'cloud' => [ 'cloud_name' => config('services.cloudinary.cloud'), 'api_key' => config('services.cloudinary.key'), 'api_secret' => config('services.cloudinary.secret'), ], 'url' => [ 'secure' => true, 'analytics' => false, ], ]); ``` From reading the source it looks like that is where the secure option is meant to go, but it doesn't appear to affect `Media::fromParams`. It's not documented, but it probably should be. ## Issue Type (Can be multiple) - [ ] Build - Can’t install or import the SDK - [ ] Performance - Performance issues - [x] Behaviour - Functions aren’t working as expected (such as generate URL) - [x] Documentation - Inconsistency between the docs and behaviour - [ ] Other (Specify) ## Steps to reproduce Create a URL using `Media::fromParams` and it'll return an HTTP URL. ## Error screenshots or Stack Trace (if applicable) … ## Operating System - [ ] Linux - [ ] Windows - [ ] macOS - [x] All ## Environment and Frameworks (fill in the version numbers) - PHP Cloudinary SDK version - 2.0.2 - PHP Version - 8.0.2 - Framework (Laravel, Symphony, etc) - Laravel 8.28.1 Answers: username_1: @username_0 , thank you for reporting the issue. The reason for this behavior is that now the default value for secure is `true`, and in v1 it was set to `false`. The primary goal of Media::fromParams() is to provide backwards compatibility and produce the same URL. Right now it is impossible to know whether `'secure' => true` comes from the default value or it was set explicitly. One of the solutions would be to track the keys that were set and respect their values. We'll release a fix for this issue soon. username_0: Alright - that makes sense. Re-reviewing the migration docs now I see that I had missed that `fromParams` is more a of a migration assistant rather than the way of doing things. Still, I was getting secure images before and after the change I wasn't. Status: Issue closed username_1: @username_0 The issue should be fixes in [2.0.3](https://github.com/cloudinary/cloudinary_php/releases/tag/2.0.3)
dKvale/aqi-watch
464577751
Title: 1-hr AQI at 132 for PM25 Question: username_0: **AQI Watch** </br>5 monitors are reporting a 1-hr AQI above 90&#46; A value of **132** for PM2&#46;5 was reported at **Apple Valley** (MPCA)&#46; For more details visit the <a href=http://mpca-air&#46;github&#46;io/aqi-watch> AQI Watch</a>&#46; </br>_Jul 05, 2019 at 05:35 CDT_ </br> </br>Attention: &#64;monikav21 &#64;lcharpentier &#64;Mr-Frank &#64;krspalmer &#64;K-ander &#64;waqu-cat
hedefalk/atom-vue
146566415
Title: Incorrect syntax highlighting for ES6 script? Question: username_0: <img width="950" alt="2016-04-07 17 50 50" src="https://cloud.githubusercontent.com/assets/1210282/14347378/6d6cd4ba-fce9-11e5-884f-65e545b65835.png"> Answers: username_1: @username_0 ![selection_003](https://cloud.githubusercontent.com/assets/3037661/14349023/c43c737c-fc85-11e5-97b4-32421d590cd0.png) ![selection_002](https://cloud.githubusercontent.com/assets/3037661/14349024/c5bf9d14-fc85-11e5-8ba1-33a3c6de2b1d.png) username_2: They say an a picture is worth a thousand words, but I don't get it :) username_0: I think you use 'JavaScript' to show ES6 syntax highlighting. But we prefer 'JavaScript with JSX' or something like that which support ES6 syntax. ![image](https://cloud.githubusercontent.com/assets/1210282/14372881/42e34308-fd77-11e5-9faf-5f09e8d46631.png) ![image](https://cloud.githubusercontent.com/assets/1210282/14372884/48bdd702-fd77-11e5-92ae-d90c711ae75c.png) username_1: @username_2 Haha, you got me. That's actually what I was trying to say: "I don't understand." Status: Issue closed username_1: @username_0 Now I have understood what's the problem. I'm sorry that we only support the core package `language-javascript` other than `language-javascript-jsx`. To get you want, you have to hack this package loaclly: 1. open the grammar file of `language-vue` package (~/.atom/packages/language-vue/grammars/vue.cson) 2. replace `include: "source.js"` with `include: "source.js.jsx.react"` 3. reload Atom by `ctrl + alt + R` 4. done Note: package update will reset this temporal hacking. username_0: I think hacking this package loaclly is not the right way... Maybe it's better to tell users to add `lang=babel` in `script` and make this package support `(?:^\\s+)?(<)((?i:script))\\b(?=[^>]*lang=[\"']babel[\"'])`, just like how we support CoffeeScript by adding 'lang=coffee'. username_1: It seems better but against a philosophy when I write packages: do only the minimum. In my opinion, 1. CoffeeScript is not in the same situation as this one, since it's supported by the core package `language-coffee-script`. Here `language-javascript-jsx` is a community one and has a scope named `source.js.jsx.react` which smells bad and brings the second issue. 2. There are several community packages supporting ES6 syntax highlighting. Which name should we use? `babel` is obviously a bad choice, as there exists a package named `language-babel` with another scope `source.js.jsx`. username_0: OK. I understand the philosophy. By the way, `vue-loader` uses `babel-loader` by default. So If we use `lang=babel`, it will not affect anything. username_1: Thanks for pointing out that. I forgot the `lang` attribute affects loaders too. What I said is it's a bit awkward that scripts annotated by `lang=babel` are highlighted by `language-javascript-jsx` rather than `language-babel`. Isn't it? username_0: You are right. My bad.😝
mpls-landlord-db/landlord-lookup
619361519
Title: Make website usable for non english speakers Question: username_0: The web UI text will need to be translated into the common languages spoken in Minneapolis. **Steps** - [ ] Find people willing and able to do translations - [ ] Compile list of languages to which site content should be translated
helexy22/helexy22.github.com
368669130
Title: Will done Question: username_0: - [Atom飞行手册(中文版)](https://wizardforcel.gitbooks.io/atom-flight-manual-zh-cn/content/1.3-Atom-Basics.html) - [Atom编辑器入门到精通(一) 安装及使用基础](https://blog.csdn.net/u010494080/article/details/50372857) - [Atom 使用教程](https://doc.yonyoucloud.com/doc/wiki/project/atom/index.html) - [Atom 使用手册](http://notes.11ten.net/atom.html) Answers: username_0: 已切换到 VScode Status: Issue closed
CadQuery/cadquery
388016484
Title: add Color property for solid objects Question: username_0: <a href="https://github.com/easyw"><img src="https://avatars1.githubusercontent.com/u/3032347?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [easyw](https://github.com/easyw)** _Saturday Aug 08, 2015 at 09:04 GMT_ _Originally opened as https://github.com/username_0/cadquery/issues/105_ ---- It would be useful to add Color property for solid objects and keep color for all functions (e.g. union cut etc.) thanks Answers: username_0: <a href="https://github.com/hyOzd"><img src="https://avatars2.githubusercontent.com/u/1479166?v=4" align="left" width="48" height="48" hspace="10"></img></a> **Comment by [hyOzd](https://github.com/hyOzd)** _Sunday Aug 09, 2015 at 02:39 GMT_ ---- I was going to create an issue for "exporting STEP with colors". Since I know we are both here for the same thing I thought I would just add this here : ) Yeah we want to export colorfuls STEPs from cadquery. username_0: <a href="https://github.com/easyw"><img src="https://avatars1.githubusercontent.com/u/3032347?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [easyw](https://github.com/easyw)** _Saturday Aug 08, 2015 at 09:04 GMT_ _Originally opened as https://github.com/username_0/cadquery/issues/105_ ---- It would be useful to add Color property for solid objects and keep color for all functions (e.g. union cut etc.) thanks username_0: <a href="https://github.com/hyOzd"><img src="https://avatars2.githubusercontent.com/u/1479166?v=4" align="left" width="48" height="48" hspace="10"></img></a> **Comment by [hyOzd](https://github.com/hyOzd)** _Sunday Aug 09, 2015 at 04:14 GMT_ ---- Let me correct; I myself only care for colors when exporting or displaying. But it would be nice to be able to fuse parts, and conserve the colors of different faces. Like freecad does; ![freecad_fuse](https://cloud.githubusercontent.com/assets/1479166/9153511/bdbbb39a-3e63-11e5-9f87-98893fbf9145.png) I know this has some limitations. For example refine shape operation of FreeCAD can merge different colored faces; ![freecad_fuse2](https://cloud.githubusercontent.com/assets/1479166/9153533/86d4ccd4-3e65-11e5-9532-9d2b1305b2b1.png) But I think such glitches should be expected. User can avoid them by carefully designing the part, cutting before fusing for example; ![after_cutfuse](https://cloud.githubusercontent.com/assets/1479166/9153539/fc753d34-3e65-11e5-9d30-71f07c54cfad.png) username_0: <a href="https://github.com/username_4"><img src="https://avatars0.githubusercontent.com/u/1015439?v=4" align="left" width="48" height="48" hspace="10"></img></a> **Comment by [username_4](https://github.com/username_4)** _Sunday Aug 09, 2015 at 03:23 GMT_ ---- @easyw and @hyOzd - Thanks for the ideas. The CadQuery module for FreeCAD allows you to specify the color and transparency for an object when it's rendered. However, it sounds like you are both saying that you want the color to be embedded as an intrensic part of a solid. My thinking is that the color property would only matter when exporting the solid (STLs/STEPs with color), or displaying it (in FreeCAD). Are these the only use cases you both see for it? On the surface this doesn't sound too bad, but operations that result in multiple solids (split), or operations on compound solids might get a little interesting. I'd need to research it a bit. username_0: <a href="https://github.com/easyw"><img src="https://avatars1.githubusercontent.com/u/3032347?v=4" align="left" width="48" height="48" hspace="10"></img></a> **Comment by [easyw](https://github.com/easyw)** _Sunday Aug 09, 2015 at 06:46 GMT_ ---- @username_4 and @hyOzd we are using cadquery to generate a parametric lib of ICs models for kicad... may be you could also consider to merge hyOzd fork that seems to have some enhancement... https://github.com/hyOzd/cadquery thank you Maurice username_0: <a href="https://github.com/hyOzd"><img src="https://avatars2.githubusercontent.com/u/1479166?v=4" align="left" width="48" height="48" hspace="10"></img></a> **Comment by [hyOzd](https://github.com/hyOzd)** _Sunday Aug 09, 2015 at 07:47 GMT_ ---- @easyw Sorry for confusion. I guess you have seen some of my earlier notes somewhere : ) My repository is already merged to upstream. You should be able to run my scripts with the latest version of cadquery. username_1: How would one apply the color to the solid, and how would that tie into show_object https://github.com/CadQuery/CQ-editor/issues/53? I'd really like this, as I use cadquery for making models for kicad :) username_2: Previously, I had the some problem. Now, I'm using cqparts to create separate parts with different colors for pins, housing, screws, etc. and assemble them into one component which can be exported. ![PTDA-1,5_1x12_P3 50mm_45-Degree](https://user-images.githubusercontent.com/1370732/60836459-01f5e700-a1c6-11e9-87f4-1914012501a5.png) username_3: Shouldn't this be closed now that Assembly exists and supports colors? username_4: The blending of color is still not supported, but AFAIK that's not an in-demand feature. I think a strong case could be made for closing this. To get a simple model's color to export to STEP will it be a requirement to wrap that model in an assembly and then set it's color? username_3: Yes, one needs to use the `cq.Assembly` class to be able to specify colors (and export colored STEP). username_4: Ok, I think it's fine to close this. Status: Issue closed
AbdulRahmanAlHamali/flutter_typeahead
822323079
Title: AutocompleteCore in Flutter 2 Question: username_0: Flutter 2 now has an official autocomplete, https://github.com/flutter/flutter/pull/62927 -- just wondering what's the plan for this plugin in the future? Answers: username_1: Good headsup -- I think when this feature has been fully made in GA -- then in most liklihood this package will be deprecated. That's my guess anyway username_2: I must be missing something but I don't see async functionality with Flutters Autocomplete for on the fly api calls. Am I overlooking something ? username_3: Its a correct observation. They kinda missed the entire point. username_4: `optionsBuilder` is probably what you are looking for? ```dart Autocomplete<T>( optionsBuilder: (TextEditingValue input) async {} ... ) // typedefed as AutocompleteOptionsBuilder<T extends Object> = FutureOr<Iterable<T>> Function( TextEditingValue textEditingValue ) ```