repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
adjust/rmq
113685646
Title: Can you expose the queue.RemoveConsumer function through Queue interface? Question: username_0: Hi, Thanks a lot for this package. Could you expose the queue.RemoveConsumer function through the Queue interface? Thanks! Answers: username_1: @username_0: Thanks for your interest! I'm not sure that function is doing what you think it does. Could you give me a bit more context about your use case? Why and when do you want to remove a consumer from a queue, and what exactly do you expect to happen? Once we're on the same page we can see if `RemoveConsumer` fits the bill or what we want to do instead. username_0: @username_1 : Thanks for taking the time to respond. We're using the rmq package as the core of one of our microservices. This microservice is responsible for creating the queues given the name and registering the consumers to specific queues. We have default/predefined queues and consumers into the microservices and the rmq package fits well this context. The other context is that we expose some HTTP endpoints for specific apps responsible for sending and receiving messages from the microservice. When an app needs to receive/process specific messages, it sends an HTTP request to the microservice and this one creates or reuses the necessary consumers to approach this task. We would like to have access to the RemoveConsumer function to remove those consumers that are not being used or won't be used anymore once we send back the messages to apps clients. Thanks! username_1: Ok, so it would be used like this: - open queue `q1` - start consuming on `q1` - add consumers `c1`, `c2`, `c3` to consume on `q1` - all three consumers receive deliveries - remove consumer `c2` from consuming on `q1` - consumers `c1` and `c3` should still continue to consume from `q1` - consumer `c2` should not receive any deliveries anymore Can you confirm this is what you need? Because `RemoveConsumer` currently does not actually stop consuming, it just removes the consumer from some bookkeeping data structure. This would need to be properly implemented before exposing it. username_0: @username_1 : yep exactly what I need. username_0: Hi @username_1, we implemented this feature through a ClosableConsumer interface, so the consumer provides a `closer channel` that we use it to stop the goroutine. You can check the implementation in [our fork](https://github.com/adjust/rmq/compare/master...Spatially:master) Here's an example of our consumer: ```go type Consumer struct { .... closer chan bool } func (c *Consumer) Close() error { c.closer <- true return nil } func (c *Consumer) Closer() chan bool { return c.closer } func (c *Consumer) ConsumeOneTime() bool { return true // true: if you want to consume only one delivery } ``` username_1: Interesting! Thanks for your work, some quick comments: Why do you use a type switch in `consumerClosableConsume` instead of passing a `ClosableConsumer` into `AddClosableConsumer`? Why would one ever want to call that function on a normal `Consumer`? You would need to include `Consumer` in `ClosableConsumer` to make this work (or add `Consume` to `ClosableConsumer`). Also you might want to switch the case statements to allow closing the channel even if the `deliveryChan` never gets empty. I believe currently it only closes if no delivery can be consumed. Why is `Close` part of the interface, it's not used there. username_0: I'm not sure if I understand this part. It closes when you call `consumer.Close()`. username_0: It's just a little restriction by the fact that we're operating over Closable consumer. I believe it provides a proper place to introduce any logic related with closing the consumer. username_1: Lets continue in #7. username_1: I believe this is no longer needed since #33. Status: Issue closed
PopGoesTheWza/swgoh-tb-sheets
374575006
Title: Offer alternative algorithms for assigning platoons Question: username_0: ✔ Enhancements → Be specific. Suggest implementation when possible. The current algorithm assigns a platoon to the person with the weakest unit for that platoon. This system can be unfair to those members with lowest gp for two reasons: 1. They will tend to be asked to fill in more platoons 2. They are more likely to need some of the units they are asked to platoon, necessitating the use of a separate exclusion spreadsheet. I propose the following algorithm: 1. For each member, standardize the power values for their units (subtract the mean, divide by the standard deviation). Then very negative scores indicate units that are weak relative to the units on this member's roster (and hence less likely to be used by that member). 2. Assign platoons based on these standardized scores. This will tend to ask members to platoon units that are low within their rosters. Answers: username_1: @username_2 so on the Breakdown tab, you could include the Light/Darkside character count, then also the Average there so can be referenced during the Platoon assignments instead of calculating it during the Assignment function. username_1: @username_2 maybe another hidden sheet for the Z scores. I believe I have the arrayformula working, but need to store the values somewhere else. I'll keep you updated username_2: @username_1 i’m Working on a OO design to better address platoon recommendation.the current code is an terrifying function which loops over platoonsat least three times and is cluttered with specific tests an rules. username_1: I need to investigate a little more... but the Algorithm assigned 3 of the bottom 4 of a character in a test. it chose number 5 instead of number 4. But it might be because he filled 10 other spots where he had a weaker character.... username_1: Implemented a zScore option in the latest version of TB Sheet Version: 2.200409 username_0: Oooh nice! I'll check it out.
Alluxio/alluxio
427162018
Title: Embedded journal test hardening Question: username_0: **Is your feature request related to a problem? Please describe.** Now we enabled embedded journal by default, but most of the tests are still using UFS journal type. We need embedded journal test coverage **Describe the solution you'd like** Add a maven property that can switch between tests using EMBEDDED JOURNAL or UFS JOURNAL. Embedded journal by default. **Urgency** It will be good that this feature can go in before 2.0 release<issue_closed> Status: Issue closed
Azure/azure-functions-vs-build-sdk
535186202
Title: Build V3 functions throws NullReferenceException Question: username_0: Building V3 function app throws during build: ``` Microsoft.NET.Sdk.Functions.Build.targets(41, 5): System.NullReferenceException: Object reference not set to an instance of an object. at MakeFunctionJson.TypeUtility.ToReflection(CustomAttribute customAttribute) at MakeFunctionJson.TypeUtility.GetResolvedAttribute(ParameterDefinition parameter, CustomAttribute customAttribute) at MakeFunctionJson.ParameterInfoExtensions.<>c__DisplayClass1_0.<ToFunctionJsonBindings>b__1(CustomAttribute a) at System.Linq.Utilities.<>c__DisplayClass2_0`3.<CombineSelectors>b__0(TSource x) at System.Linq.Utilities.<>c__DisplayClass2_0`3.<CombineSelectors>b__0(TSource x) at System.Linq.Enumerable.WhereSelectEnumerableIterator`2.ToList() at System.Linq.Enumerable.ToList[TSource](IEnumerable`1 source) at MakeFunctionJson.ParameterInfoExtensions.ToFunctionJsonBindings(ParameterDefinition parameterInfo) at MakeFunctionJson.MethodInfoExtensions.<>c.<ToFunctionJson>b__6_1(ParameterDefinition p) at System.Linq.Enumerable.WhereSelectEnumerableIterator`2.MoveNext() at System.Linq.Enumerable.SelectManySingleSelectorIterator`2.ToArray() at System.Linq.Enumerable.ToArray[TSource](IEnumerable`1 source) at MakeFunctionJson.MethodInfoExtensions.ToFunctionJson(MethodDefinition method, String assemblyPath) at MakeFunctionJson.FunctionJsonConverter.GenerateFunctions(IEnumerable`1 types)+MoveNext() at System.Collections.Generic.List`1..ctor(IEnumerable`1 collection) at System.Linq.Enumerable.ToList[TSource](IEnumerable`1 source) at MakeFunctionJson.FunctionJsonConverter.TryGenerateFunctionJsons() at MakeFunctionJson.FunctionJsonConverter.TryRun() Error generating functions metadata ``` Answers: username_0: Project is https://github.com/username_0/NuGetTypeSearch, after applying the attached patch. [Update_to_functions_v3.patch.txt](https://github.com/Azure/azure-functions-vs-build-sdk/files/3941438/Update_to_functions_v3.patch.txt) username_0: Seems to be caused by having a custom trigger binding attribute username_1: I just experienced the same issue. username_2: I was able to get a repro of this. If you create a function app that targets `netcoreapp3.1` and add a custom trigger, then use it, you'll hit this. If you target `netcoreapp3.0` it should work. When we try to load that type via reflection (to generate function.json), the version of System.Runtime that it tries to load is different than the one being used by the FunctionsGenerator, which targets `netcoreapp3.0`. I'm working on a fix now. username_2: @username_1 committed a fix and I've just pushed 3.0.2 to nuget. This should take care of the issue, please let me know if it doesn't! Status: Issue closed username_0: Works, thanks folks!
jupyterlab/jupyterlab
842750376
Title: JuypterLab re-opens wrong file when opening file via command Question: username_0: <!-- Welcome! Before creating a new issue: * Search for relevant issues * Follow the issue reporting guidelines: https://jupyterlab.readthedocs.io/en/latest/getting_started/issue.html --> I'm using JupyterLab with Chromium under Linux Mint. In `jupyter_notebook_config.py`, I have these settings: c.ServerApp.browser = '/usr/bin/chromium --app=%s' c.LabApp.browser = '/usr/bin/chromium --app=%s' When I open an ipynb file directly with JupyterLab -- be it by double-clicking it or via command line `$ jupyter lab myfile.ipynb` -- close the browser and open another file `myotherfile.ipynb`, that resides in the same directory, only the first file `myfile.ipynb` is opened by JupyterLab. I can, of course, then open the wanted file from within JupyterLab, but it is not exactly what I expected. This behaviour doesn't occur when I open a file from a different directory. I think that this first happened with the new "ServerApp" setting in the config file and already led to some confusion, as I don't always look which file name is open and assume that the presented file is the one I selected earlier. Is there some setting I have to change or is this a bug? Answers: username_0: In my current setup (JupyterLab 3.0.14, Python 3.9.5, Chrome 92.0.4515.107), this bug is still present. It also does not appear to depend on the browser or the mode in which it is started (same happened with disabled app-mode and in Firefox). I came a little closer to the problem in that I found that in the ".jupyter" config directory, in the sub-dir lab/workspaces, a "default-37a8.jupyterlab-workspace" is generated automatically, in which some json-like contents is stored. The "widgets" and "current" keys contain the recent notebooks, which remain there even after closing the browser (why shouldn't they). However, on re-starting jupyterlab, these are not overwritten but instead override the parameter with the notebook I actually want to be loaded. I now found an - admittedly ugly - workaround by setting the "c.LabApp.workspaces_api_url" parameter in "jupyter_lab_config.py" to a nonsense value. I suspect that this will lead to further problems at some point, thus I really hope to find a better solution soon.
Enelar/EJS
39700791
Title: Async rendering Question: username_0: Right now all render work in synonymous/blocking mode. Even if resource still on server. We need smooth rendering. Answers: username_0: ```Synchronous XMLHttpRequest on the main thread is deprecated because of its detrimental effects to the end user's experience.``` Status: Issue closed username_0: Fixed https://github.com/phoxy/ENJS/releases/tag/v2.1.8
ant-design/ant-design
556732876
Title: [4.0.0-rc.3] Icon inside tooltip: Function components cannot be given refs. Attempts to access this ref will fail. Did you mean to use React.forwardRef Question: username_0: - [x] I have searched the [issues](https://github.com/ant-design/ant-design/issues) of this repository and believe that this is not a duplicate. ### Reproduction link [![Edit on CodeSandbox](https://codesandbox.io/static/img/play-codesandbox.svg)](https://codesandbox.io/s/antd-reproduction-template-zpogx) ### Steps to reproduce I **cannot** reproduce it on Codesandbox, most likely it's related to the middleware I use (webpack 4.41.5). I use the middleware from a popular boilerplate https://github.com/flexdinesh/react-redux-boilerplate I followed the example from https://next.ant.design/components/form/#components-form-demo-register to add a tooltip in a form item label. After a few hours of investigating I have cleaned up everything that could possibly cause this and left with the following code: ```js const MOUNT_NODE = document.getElementById('app') const render = () => { ReactDOM.render( <Form> <Form.Item label={( <span> Nickname&nbsp; <Tooltip title="What do you want others to call you?"> <QuestionCircleOutlined /> </Tooltip> </span> )}> <Input type="text" /> </Form.Item> </Form>, MOUNT_NODE, ) } render() ``` And it produces the following warning: ``` react-dom.development.js?61bb:530 Warning: Function components cannot be given refs. Attempts to access this ref will fail. Did you mean to use React.forwardRef()? Check the render method of `Trigger`. in QuestionCircleOutlined (created by Trigger) in Trigger (created by ForwardRef(Tooltip)) in ForwardRef(Tooltip) (created by Context.Consumer) in Tooltip in span in label (created by Context.Consumer) in div (created by Context.Consumer) in Col (created by Context.Consumer) in FormItemLabel (created by FormItem) in div (created by Context.Consumer) in Row (created by FormItem) in FormItem in form (created by ForwardRef(Form)) in ForwardRef(Form) (created by ForwardRef(InternalForm)) in SizeContextProvider (created by ForwardRef(InternalForm)) in ForwardRef(InternalForm) warningWithoutStack @ react-dom.development.js?61bb:530 warning @ react-dom.development.js?61bb:1018 validateFunctionComponentInDev @ react-dom.development.js?61bb:18885 [Truncated] ``` If I add any additional text inside `Tooltip` there will be no warning. ### What is expected? No warning ### What is actually happening? A warning is showing up | Environment | Info | |---|---| | antd | 4.0.0-rc.3 | | React | 16.12.0 | | System | macOS | | Browser | Chrome 79.0.3945.130 (x64) | <!-- generated by ant-design-issue-helper. DO NOT REMOVE --> Answers: username_1: @username_0 try ``` <Tooltip title="What do you want others to call you?"> <> <QuestionCircleOutlined /> </> </Tooltip> ```
XePeleato/HwComposer_Research
225371906
Title: Hi6210sft HwComposer? Question: username_0: I found one from HiKey, I am not sure, but I think it might help: https://github.com/xin3liang/hwcomposer Hope this can help you. Answers: username_0: @username_3 @<NAME> @username_1 username_1: I think is compatible only with lollipop roms and also we should enable drm on kernel to get it work.. username_0: @username_1 So it is impossible or very hard for 6.0 and above to have a properly working HwComposer? username_2: Who summoned me ? :smile: username_0: Sorry, I want to @Huawei-Dev … Status: Issue closed username_3: @username_0 Actually, if you've read the README from my HwC you'd have seen that I've used some code from there.
Daltron/NotificationBanner
570531355
Title: Use light status bar when banner is shown? Question: username_0: I feel like the banner (other than the floating style) looks best when the status bar is light (white). The text in the banner is already white. Is it possible to force status bar color when the banner is visible?
moment/moment-timezone
304143783
Title: Can't format a literal Z Question: username_0: i'm selecting : date : 2018-03-11T15:00Z // this format come with i'm selecting date and which is not changes because framework give me ISO 8601 Datetime Format: YYYY-MM-DDTHH:mmZ after convert moment(2018-03-11T15:00Z ).format("YYYY-MM-DDTHH:mm"); output // 2018-03-11T21:00Z my time zone is IST Answers: username_1: Hi @username_0, I just tried ``` moment('2018-03-11T15:00Z' ).format("YYYY-MM-DDTHH:mm"); ``` and it returns `2018-03-11T17:00` Status: Issue closed username_2: If you want to leave it in UTC, just: ```js moment.utc('2018-03-11T15:00Z' ).format("YYYY-MM-DDTHH:mm"); ```
networkedsystems/PercEvite
562751317
Title: [IF mode] question Question: username_0: https://github.com/networkedsystems/PercEvite/blob/0dbe2496b3c9ce5f409a838d17fc047f1bd4163b/ESP32_firmware/Only_BS/Only_BS.ino#L87 esp suggests to use a different mode: ![image](https://user-images.githubusercontent.com/22093651/74178980-ce3bdd80-4c3c-11ea-8a39-00808d26c467.png) Is this okay?
SUI-Components/sui
337822330
Title: [sui-deploy] use safe brach name Question: username_0: Error: Invalid name: "@sui-deploy/milanuncios-feature/ci" at ensureValidName (/snapshot/repo/node_modules/normalize-package-data/lib/fixer.js:335:15) at Object.fixNameField (/snapshot/repo/node_modules/normalize-package-data/lib/fixer.js:215:5) at /snapshot/repo/node_modules/normalize-package-data/lib/normalize.js:32:38 at Array.forEach (<anonymous>) at normalize (/snapshot/repo/node_modules/normalize-package-data/lib/normalize.js:31:15) at module.exports.pathType.dir.then.then.x (/snapshot/repo/node_modules/read-pkg/index.js:24:38) at <anonymous> Error: Invalid name: "@sui-deploy/milanuncios-feature/ci" at ensureValidName (/snapshot/repo/node_modules/normalize-package-data/lib/fixer.js:335:15) at Object.fixNameField (/snapshot/repo/node_modules/normalize-package-data/lib/fixer.js:215:5) at /snapshot/repo/node_modules/normalize-package-data/lib/normalize.js:32:38 at Array.forEach (<anonymous>) at normalize (/snapshot/repo/node_modules/normalize-package-data/lib/normalize.js:31:15) at module.exports.pathType.dir.then.then.x (/snapshot/repo/node_modules/read-pkg/index.js:24:38) at <anonymous> ```<issue_closed> Status: Issue closed
m-lab/etl
564260593
Title: some scamper tests filtered by code change, cause tarball not reporcessed Question: username_0: SELECT * FROM `mlab-oti.ndt.traceroute` WHERE Parseinfo.TaskFileName = "gs://archive-measurement-lab/ndt/traceroute/2019/11/02/20191102T020000.909108Z-traceroute-mlab4-lhr05-ndt.tgz" LIMIT 1000 Row | partition_date | uuid | TestTime | Parseinfo.TaskFileName | Parseinfo.ParseTime | Parseinfo.ParserVersion | Parseinfo.Filename | start_time | stop_time | scamper_version | Source.IP | Source.Port | Source.IATA | Source.Geo.continent_code | Source.Geo.country_code | Source.Geo.country_code3 | Source.Geo.country_name | Source.Geo.region | Source.Geo.metro_code | Source.Geo.city | Source.Geo.area_code | Source.Geo.postal_code | Source.Geo.latitude | Source.Geo.longitude | Source.Geo.radius | Source.Network.IPPrefix | Source.Network.Systems.ASNs | Destination.IP | Destination.Port | Destination.Geo.continent_code | Destination.Geo.country_code | Destination.Geo.country_code3 | Destination.Geo.country_name | Destination.Geo.region | Destination.Geo.metro_code | Destination.Geo.city | Destination.Geo.area_code | Destination.Geo.postal_code | Destination.Geo.latitude | Destination.Geo.longitude | Destination.Geo.radius | Destination.Network.IPPrefix | Destination.Network.Systems.ASNs | ProbeSize | ProbeC | Hop.Source.IP | Hop.Source.City | Hop.Source.CountryCode | Hop.Source.Hostname | Hop.Source.ASN | Hop.Linkc | Hop.Links.HopDstIP | Hop.Links.TTL | Hop.Links.Probes.Flowid | Hop.Links.Probes.Rtt | exp_version | cached_result |   -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- 1 | 2019-11-02 | ndt-qzlhp_1571972644_00000000000148CB | 2019-11-02 01:45:00 UTC | gs://archive-measurement-lab/ndt/traceroute/2019/11/02/20191102T020000.909108Z-traceroute-mlab4-lhr05-ndt.tgz | 2019-11-05 00:33:18.164853 UTC | https://github.com/m-lab/etl/tree/prod-v2.0.1 | null | 1572659100 | 1572659100 | 0.1 | fc00:e968:6179::de52:7100:212.113.31.50 | 0 |   | EU | GB |   | United Kingdom | ENG | 0 | York | 0 | YO10 | 53.9573 | -1.0837 | 1 |   | 3356 | fc00:e968:6179::de52:7100:35.225.75.192 | 0 | NA | US |   | United States | VA | 556 |   | 0 |   | 38.6583 | -77.2481 | 0 |   | 15169 | 60 | 0 |   |   |   |   |   |   |   |   |   |   | null | null |   2 | 2019-11-02 | ndt-qzlhp_1571972644_00000000000145CD | 2019-11-02 00:10:01 UTC | gs://archive-measurement-lab/ndt/traceroute/2019/11/02/20191102T020000.909108Z-traceroute-mlab4-lhr05-ndt.tgz | 2019-11-03 18:28:10.211337 UTC | https://github.com/m-lab/etl/tree/prod-v2.0.1 | null | 1572653281 | 1572653281 | 0.1 | fc00:e968:6179::de52:7100:172.16.31.10 | 0 |   | EU | GB |   | United Kingdom | ENG | 0 | York | 0 | YO10 | 53.9573 | -1.0837 | 1 |   | 3356 | fc00:e968:6179::de52:7100:172.16.31.10 | 0 | NA | US |   | United States | VA | 556 |   | 0 |   | 38.6583 | -77.2481 | 0 |   | 15169 | 60 | 0 |   |   |   |   |   |   |   |   |   |   | null | null |   The reason is that ::ffff handling code change, make the majority of tests in that tarball filtered, thus failed Gardener 99% sanity check. Answers: username_0: PR: #830 && m-lab/etl-gardener#232 And https://github.com/m-lab/etl/releases/tag/prod-v2.2.3 Fix this problem. SELECT Parseinfo.ParseTime, Parseinfo.TaskFileName FROM `mlab-staging.base_tables.traceroute` WHERE _PARTITIONTIME > "2019-01-24" AND Parseinfo.ParseTime < "2019-12-12" And there is 0 result. Yay :) Status: Issue closed username_0: Reopen issue: gs://archive-mlab-oti/host/traceroute/2019/11/15/20191115T034951.000655Z-traceroute-mlab1-tpe01-host.tgz was not reprocessed since 2019/11/15 and trigger the same error again. username_0: for archive-mlab-staging bucket, the host/traceroute dir has all mlab4 data, which is different from mlab-oti (with NO overlapping) I will use mlab-testing to reproduce the reprocessing process for above tarball. username_0: SELECT * FROM `mlab-oti.ndt.traceroute` WHERE Parseinfo.TaskFileName = "gs://archive-measurement-lab/ndt/traceroute/2019/11/02/20191102T020000.909108Z-traceroute-mlab4-lhr05-ndt.tgz" LIMIT 1000 Row | partition_date | uuid | TestTime | Parseinfo.TaskFileName | Parseinfo.ParseTime | Parseinfo.ParserVersion | Parseinfo.Filename | start_time | stop_time | scamper_version | Source.IP | Source.Port | Source.IATA | Source.Geo.continent_code | Source.Geo.country_code | Source.Geo.country_code3 | Source.Geo.country_name | Source.Geo.region | Source.Geo.metro_code | Source.Geo.city | Source.Geo.area_code | Source.Geo.postal_code | Source.Geo.latitude | Source.Geo.longitude | Source.Geo.radius | Source.Network.IPPrefix | Source.Network.Systems.ASNs | Destination.IP | Destination.Port | Destination.Geo.continent_code | Destination.Geo.country_code | Destination.Geo.country_code3 | Destination.Geo.country_name | Destination.Geo.region | Destination.Geo.metro_code | Destination.Geo.city | Destination.Geo.area_code | Destination.Geo.postal_code | Destination.Geo.latitude | Destination.Geo.longitude | Destination.Geo.radius | Destination.Network.IPPrefix | Destination.Network.Systems.ASNs | ProbeSize | ProbeC | Hop.Source.IP | Hop.Source.City | Hop.Source.CountryCode | Hop.Source.Hostname | Hop.Source.ASN | Hop.Linkc | Hop.Links.HopDstIP | Hop.Links.TTL | Hop.Links.Probes.Flowid | Hop.Links.Probes.Rtt | exp_version | cached_result |   -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- | -- 1 | 2019-11-02 | ndt-qzlhp_1571972644_00000000000148CB | 2019-11-02 01:45:00 UTC | gs://archive-measurement-lab/ndt/traceroute/2019/11/02/20191102T020000.909108Z-traceroute-mlab4-lhr05-ndt.tgz | 2019-11-05 00:33:18.164853 UTC | https://github.com/m-lab/etl/tree/prod-v2.0.1 | null | 1572659100 | 1572659100 | 0.1 | fc00:e968:6179::de52:7100:172.16.31.10 | 0 |   | EU | GB |   | United Kingdom | ENG | 0 | York | 0 | YO10 | 53.9573 | -1.0837 | 1 |   | 3356 | fc00:e968:6179::de52:7100:172.16.31.10 | 0 | NA | US |   | United States | VA | 556 |   | 0 |   | 38.6583 | -77.2481 | 0 |   | 15169 | 60 | 0 |   |   |   |   |   |   |   |   |   |   | null | null |   2 | 2019-11-02 | ndt-qzlhp_1571972644_00000000000145CD | 2019-11-02 00:10:01 UTC | gs://archive-measurement-lab/ndt/traceroute/2019/11/02/20191102T020000.909108Z-traceroute-mlab4-lhr05-ndt.tgz | 2019-11-03 18:28:10.211337 UTC | https://github.com/m-lab/etl/tree/prod-v2.0.1 | null | 1572653281 | 1572653281 | 0.1 | fc00:e968:6179::de52:7100:172.16.31.10 | 0 |   | EU | GB |   | United Kingdom | ENG | 0 | York | 0 | YO10 | 53.9573 | -1.0837 | 1 |   | 3356 | fc00:e968:6179::de52:7100:172.16.31.10 | 0 | NA | US |   | United States | VA | 556 |   | 0 |   | 38.6583 | -77.2481 | 0 |   | 15169 | 60 | 0 |   |   |   |   |   |   |   |   |   |   | null | null |   The reason is that ::ffff handling code change, make the majority of tests in that tarball filtered, thus failed Gardener 99% sanity check. username_0: The reason is: back to 2019/11/155, we played with all kinds of alternatives, such as new version of paris traceroute, and for very short time, there were ill-formatted paris traceroute tests generated in gs://archive-measurement-lab/paris-traceroute/2019/11/15/ which block the reprocessing of all 2019/11/15 data. After we manually delete the current BQ data with PartitionTime = 20191115, this problem will be fixed automatically.
department-of-veterans-affairs/va.gov-team
1178595854
Title: Research - Authentication issues Question: username_0: ## Issue Description _What details are necessary for understanding the specific work or request tracked by this issue?_ --- ## Tasks - [ ] _What work is necessary for this story to be completed?_ ## Acceptance Criteria - [ ] _What will be created or happen as a result of this story?_ --- ## How to configure this issue - [ ] **Attached to a Milestone** (when will this be completed?) - [ ] **Attached to an Epic** (what body of work is this a part of?) - [ ] **Labeled with Team** (`product support`, `analytics-insights`, `operations`, `service-design`, `Console-Services`, `tools-fe`) - [ ] **Labeled with Practice Area** (`backend`, `frontend`, `devops`, `design`, `research`, `product`, `ia`, `qa`, `analytics`, `contact center`, `research`, `accessibility`, `content`) - [ ] **Labeled with Type** (`bug`, `request`, `discovery`, `documentation`, etc.) Answers: username_0: Lighthouse team stated that Oauth Proxy and SAML Proxy will still be necessary in the future configuration, but they have no plan to move them from the dsvagovcloud AWS account. This may need a plan for who will manage and/or own these applications moving forward. We may also need to decide how the routing to these applications will work. It sounded like the Lighthouse plan was to route to these via the *api.va.gov URL, through Apigee, out to the dvp-gov-internal account, and then to the dsvagovcloud account. We should verify this and see if we can optimize this more. username_1: The Lighthouse team specifically mentioned plans to remove the incidental dependency on vets-api for OAuth token validation and migrate that functionality to be under their own control. @patrickvinograd, please let us know any insights you might have into authentication flows, or any risks you can identify in that area.
Rfam/rfam-website
203096285
Title: Inaccurate PDB mapping Question: username_0: It looks like some PDB chains do not match the right family. For example, chains 1A and 2A from 5DOX should match the Bacterial LSU but they match the 23S pseudoknot (RF01118). Answers: username_0: This was resolved in release 12.3. Status: Issue closed
typedefs/tdlc
222392951
Title: Support for fantasy/static-land Question: username_0: Is this something we can support in JS extraction? Possibly in combination with TS/Flow support... - https://github.com/rpominov/static-land - https://github.com/fantasyland/fantasy-land (posting this as a reminder to myself)
GoogleCloudPlatform/google-cloud-visualstudio
332600560
Title: Publish GCE OpenWebsite does not open to correct port. Question: username_0: On the Publish Google Compute Engine, the Open Website checkbox does not cause the browser to open to a specific port. This can be a problem if publishing to a website that is using a non default port, especially if multiple websites are on the specific VM.
dropbox/dropbox-sdk-python
210812231
Title: How to read a file inside the dropbox folder using python dropbox api V2? Question: username_0: i need to know how files inside a dropbox folder can be read. For ex: if there are text and csv files inside a dropbox folder, then i need to read the contents of those file using python dropbox api V2? Answers: username_1: Hi @username_0. Thanks for the question. The issues for this repository should be bug reports or feature requests for Dropbox's Python SDK. If you'd like to learn how to use our APIs and SDKs, please take a look at the following resources available via the [developers section of Dropbox's website](https://www.dropbox.com/developers): - [Dropbox Python SDK with examples](https://www.dropbox.com/developers/documentation/python) and [accompanying docs](http://dropbox-sdk-python.readthedocs.io/en/master/) - [Dropbox general API endpoint documentation](https://www.dropbox.com/developers/documentation/http/documentation) - [`/2/files/download` endpoint documentation](https://www.dropbox.com/developers/documentation/http/documentation#files-download) - [`/2/files/list_folder` endpoint documentation](https://www.dropbox.com/developers/documentation/http/documentation#files-list_folder) _et seq._ See also: - [Dropbox Python SDK tutorial](https://github.com/username_1/dropbox-api-primer) and [accompanying docs](https://dropbox-api-primer.readthedocs.io/en/master/tutorial.html) Status: Issue closed username_2: Also, for reference, the examples use the Python SDK `file_download` methods here: - https://github.com/dropbox/dropbox-sdk-python/blob/master/example/updown.py#L158 - https://github.com/dropbox/dropbox-sdk-python/blob/master/example/back-up-and-restore/backup-and-restore-example.py#L55 username_1: @username_0, the [developer support area](https://www.dropbox.com/developers/support) may be useful to you as well.
HackIllinois/api
347767153
Title: Allow config to specify a domain for Staff & a single Admin user Question: username_0: Right now adding the first `Admin` user to the API requires manually modifying the database. We should have a config variable that allows an `Admin` or `Superuser` role to be granted to a single user. It would also be useful to automatically grant the `Staff` role to all users with an email at a specified domain. This needs to avoid the potential security issue of an OAuth provider allowing an email without verification allowing anyone to obtain admin rights. We need to verify that the user actually owns the email we are getting from the OAuth provider. One solution to this is to send a verification email to the user with a link that triggers a granting of the role. - [ ] The admin role should be granted the user user with the specified email - [ ] The staff role should be granted to all users with an email at the specified staff domain in the Answers: username_0: GitHub and Google provide a `verified` field in the response which provides us with the user's email. And LinkedIn only allows verified emails to be the primary emails which is the one returned by the API. If the user has their email verified and has the correct domain, it is safe to grant the staff role. Status: Issue closed
frenzymadness/specfile_generator
964646996
Title: Test the tool Question: username_0: For MVP find or prepare good testing examples containing: - pure Python package - alpha/beta releases, rc's - package with non-Python dependencies - ... Setup CI (or reuse the existing one) to run the tests on commit.<issue_closed> Status: Issue closed
OpenNebula/one
742579434
Title: Isolated Front-end Deployment Model Question: username_0: **Description** Design and implement a new model to distribute and run OpenNebula front-end which - uses a run-time environment isolated from the host OS - simplifies configuration and maintenance - provides easy rollback on upgrade failure <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches<issue_closed> Status: Issue closed
jdanylko/Tuxboard
1029255562
Title: Suggestion Rename Tuxboard to something more catchy Question: username_0: Hello, we should consider renaming Tuxboard to something more catch. For e.g. something like... ### _...Core Dynamic_ Dashboards Answers: username_1: The naming was a one-day idea that just popped into my head when I was looking for a name with some type of branding. "Core Dynamic Dashboards" doesn't quite roll off the tongue. :-) username_0: You have a point, we need something that sticks & conveys its a dashboard, any ideas.. Status: Issue closed
Drake53/War3Api
1040347157
Title: War3Api.Objct UnitType and ItemType enums are not generated for campaign objects Question: username_0: For instance, there is no UnitType generated for Slave Master and no ItemType generated for Scroll of Resurrection. This means that I'm unable to read my map object data which features modifications for both of these. Answers: username_0: I've described the issue incorrectly; Illidan has a UnitType enum so the campaign data must be being generated. I'm just not sure why a few things like Slave Master and Scroll of Resurrection don't appear. Status: Issue closed username_0: Closed because this issue was fixed in https://github.com/Drake53/War3Net/commit/d5caa7136ac8fe85c864deddc30907465344adc0.
microsoft/STL
565626682
Title: CMake build failure in presence of gcc's c++.exe in the PATH Question: username_0: **Describe the bug** I've synced (on Windows) to latest (as of today): git clone https://github.com/microsoft/STL --recursive **Command-line test case** Per documentation, that was my command: cmake -G Ninja -DCMAKE_TOOLCHAIN_FILE=d:\p\microsoft\STL\vcpkg\scripts\buildsystems\vcpkg.cmake -S . -B out123 I've made sure "out123" is not present. The problem now was that CMake build tool found gcc's c++.exe (installed from chocolatey), although I was in the x64 VS2019 Command line prompt, and "cl.exe" was first in the PATH (though it does not matter much in this case, as they have different names). Since I'm not much familiar with CMAKE, googled, and found that this should improve it: ```cmake -G Ninja -DCMAKE_TOOLCHAIN_FILE=d:\p\microsoft\STL\vcpkg\scripts\buildsystems\vcpkg.cmake -DCMAKE_C_COMPILER="cl.exe" -DCMAKE_CXX_COMPILER="cl.exe" -DMSVC_TOOLSET_VERSION=142 -S . -B out123 ``` e.g. explicitly specifying the compiler driver ("cl.exe"), and the toolset (142 for VS2019) With this I've got ninja generated (though then I realized I need the preview 19.25 compiler, and the highest I had was 19.14). Answers: username_1: @BillyONeal @barcharcraz @username_2 Should we update the README here, or can/should we modify the CMake build system to specify that we always want MSVC and it has to be the 2019 release series (if multiple toolsets are installed)? username_2: @username_0 did you have the environment variable CC or CXX set to something? CMake respects those. Otherwise the preferred compiler does seem to be c++ as per [this](https://gitlab.kitware.com/cmake/cmake/blob/master/Modules/Platform/Windows-Determine-CXX.cmake). It looks like we could set `CMAKE_CXX_COMPILER_NAMES` before we call project to force cl to be selected. It isn't clear to me how to force a vs2019 compiler to be selected if there is a non vs2019 compiler in the path but we could fast fail pretty easily by checking the compiler version. username_1: https://github.com/microsoft/STL/blob/2b2746dd78d06493cb204e8616beec7991d72840/stl/inc/yvals_core.h#L454-L456 will cause an attempted build with an outdated MSVC to fail quickly, but checking in CMake might be better. username_0: Thanks! That seems better approach (less typing, easier to remember). I'm just not that familiar with CMake that well, and that was the first thing I've found in stack overflow (to actually work). username_2: @username_1 having `yvals_core.h` fail the build sounds fine. Maybe we could do more niceness and fail a little earlier in the CMake but I don't feel like it's worth any additional complexity unless contributors seem to start having trouble often. username_1: We talked about this and we're uncertain whether this is a vcpkg bug. If it's not, we have a weak consensus that the README should mention this category of failures and how to avoid it. username_2: It's unclear to me how this issue could be related to `vcpkg`. Are we somehow hooking into the `vcpkg` compiler detection logic? username_1: See the command line in the original report: `-DCMAKE_TOOLCHAIN_FILE=d:\p\microsoft\STL\vcpkg\scripts\buildsystems\vcpkg.cmake` (I am a total novice here, this is just what I recall from our weekly meeting.) username_2: Ah, of course, blanked on that. username_2: Just took a look through the toolchain file (I may have been working on it not even 3 months ago but I couldn't remember the details). I don't see anything in there which should affect the selected compiler. I don't think this has anything to do with vcpkg. username_3: It works for me now ![изображение](https://user-images.githubusercontent.com/4289847/149288770-bac1d645-76ad-49f9-988d-1d1266783c4d.png) username_3: Actually it matters. If i have gcc first then I got the error. ![image](https://user-images.githubusercontent.com/4289847/149294754-384c8cc3-9415-4eb9-aae8-32349eed0a9c.png) username_3: I found the PR: https://gitlab.kitware.com/cmake/cmake/-/merge_requests/4625 So it should not be an issue because "Developer Command Prompt for VS 2022 Preview" add cl to first place in the PATH. So I think the issue could be closed... Status: Issue closed username_1: Thanks @username_3 - we talked about this at the weekly maintainer meeting and believe that there's no further action to take here - we've removed our vcpkg dependency, so the only remaining question is how CMake/Ninja behave with multiple compilers. It seems reasonable for them to default to the first compiler on the path when none is explicitly specified, and if the VS 2022 Dev Command Prompt is placing `cl.exe` first on the path, then it would require a very unusual setup to have `gcc.exe` prepended to the path at a later time. So, there seems to be no need to explicitly document how to override CMake/Ninja's choice of compiler. Please let us know if there are any remaining issues we need to consider. Thanks!
dotnet/roslyn-analyzers
656614104
Title: CA1810: false positive when field is used in static ctor Question: username_0: #### Analyzer package [Microsoft.CodeAnalysis.FxCopAnalyzers](https://www.nuget.org/packages/Microsoft.CodeAnalysis.FxCopAnalyzers) #### Package Version v3.3.0-beta1.final (Latest pre-release) #### Diagnostic ID Example: [CA1810](https://docs.microsoft.com/visualstudio/code-quality/ca1810) #### Repro steps ```csharp [Fact] public async Task CA1810_FieldMethodCall() { await new VerifyCS.Test { ReferenceAssemblies = AdditionalMetadataReferences.DefaultWithWinForms, TestCode = @" using System; using System.Windows.Forms; public class SKWindow { public void CreateControl() {} } public class SendKeys { private static readonly SKWindow messageWindow; static SendKeys() { Application.ThreadExit += new EventHandler(OnThreadExit); messageWindow = new SKWindow(); messageWindow.CreateControl(); } private static void OnThreadExit(object sender, EventArgs e) {} } ", }.RunAsync(); } ``` #### Expected behavior No CA1810 diagnostic. #### Actual behavior CA1810 diagnostic. See https://github.com/dotnet/winforms/pull/3561/files Answers: username_0: Looking at the issue, I think that there are a couple of cases/questions raised here. ## Scenario 1 ```csharp public class C { private static readonly SKWindow messageWindow; static C() { messageWindow = new SKWindow(); messageWindow.CreateControl(); } } ``` In this scenario, I think that an issue shall be reported because nothing prevents you from creating a static method which returns the object initialized. For example a solution could be: ```csharp public class C { private static readonly SKWindow messageWindow = GetMessageWindow(); private static SKWindow GetMessageWindow() { var window = new SKWindow(); window.CreateControl(); return window; } } ``` ## Scenario 2 This scenario is IMO the actual problem here. As long as you have a subscription (`+=` or `-=`) for an event (what about actions?) in the ctor, don't report no matter what. Anyways you need to have the static ctor. ## Scenario 3 Initialization order matters, let's say you have a couple of fields and you need to chain some calls on these fields in a particular order. In this case, I don't think there is anything that can be one automatically so users will have to suppress the diagnostic. ```csharp public class C { private static readonly A fieldA; private static readonly B fieldB; static C() { fieldA = new A(); fieldA.DoSomething(); fieldB = new B(); fieldB.DoSomething(); fieldA.DoSomethingElseWhichDependsUpon(fieldB); } } ``` Status: Issue closed
iwown/IVUIKit
444716050
Title: **screenShotWithScrollView** method return nil object. Question: username_0: I got following issues when use this method: `UIImage *image2 = [UIImage screenShotWithScrollView:strongSelf->_scroll];` The image2 is a nil object return. ``` ... 2019-05-16 09:39:04.479054+0800 ZeronerHealth[1333:242721] [Unknown process name] CGContextDrawPath: invalid context 0x0. If you want to see the backtrace, please set CG_CONTEXT_SHOW_BACKTRACE environmental variable. 2019-05-16 09:39:04.479096+0800 ZeronerHealth[1333:242721] [Unknown process name] CGContextRestoreGState: invalid context 0x0. If you want to see the backtrace, please set CG_CONTEXT_SHOW_BACKTRACE environmental variable. 2019-05-16 09:39:04.479141+0800 ZeronerHealth[1333:242721] [Unknown process name] CGContextRestoreGState: invalid context 0x0. If you want to see the backtrace, please set CG_CONTEXT_SHOW_BACKTRACE environmental variable. ... ``` Answers: username_0: The input **strongSelf->_scroll** had contentSize: {0, 0}; init _scroll with setting to fix this issue。 `_scroll.contentSize = CGSizeMake(SCREEN_WIDTH, SCREEN_HEIGHT - NavigationBarHeight);` Status: Issue closed
karriereat/json-decoder
599754883
Title: Generate Transformers based on annotations Question: username_0: If a class already has type hints inside the property doc-block, this information can be used to automatically generate the transformer for the given class. There should also be a `generateMultiple` method that does the same for multiple classes. Answers: username_1: I've tested this too but i can't make it work My AudienceListTransformer is ```php <?php /** * Project: szopen\mailchimp * * Author: <NAME> <<EMAIL>> * Created: 10/04/2020 10:05 */ namespace Szopen\Mailchimp\Helper\Transformer\Audience; use Karriere\JsonDecoder\Bindings\AliasBinding; use Karriere\JsonDecoder\Bindings\ArrayBinding; use Karriere\JsonDecoder\Bindings\CallbackBinding; use Karriere\JsonDecoder\Bindings\DateTimeBinding; use Karriere\JsonDecoder\Bindings\FieldBinding; use Karriere\JsonDecoder\ClassBindings; use Karriere\JsonDecoder\Transformer; use Szopen\Mailchimp\Audience\AudienceList; use Szopen\Mailchimp\Audience\CampaignDefaults; use Szopen\Mailchimp\Audience\Contact; use Szopen\Mailchimp\Audience\Link; use Szopen\Mailchimp\Audience\Stats; class AudienceListTransformer implements Transformer { /** * @inheritDoc */ public function register(ClassBindings $classBindings) { /* $classBindings->register(new AliasBinding('id', 'id')); $classBindings->register(new AliasBinding('webId', 'web_id')); $classBindings->register(new AliasBinding('permissionReminder', 'permission_reminder')); $classBindings->register(new AliasBinding('useArchiveBar', 'use_archive_bar')); $classBindings->register(new AliasBinding('notifyOnSubscribe', 'notify_on_subscribe')); $classBindings->register(new AliasBinding('notifyOnUnsubscribe', 'notify_on_unsubscribe')); $classBindings->register(new AliasBinding('emailTypeOption', 'email_type_option')); $classBindings->register(new AliasBinding('doubleOptin', 'double_optin')); $classBindings->register(new AliasBinding('marketingPermission', 'marketing_permission')); $classBindings->register(new AliasBinding('listRating', 'list_rating')); $classBindings->register(new AliasBinding('subscribeUrlShort', 'subscribe_url_short')); $classBindings->register(new AliasBinding('subscribeUrlLong', 'subscribe_url_long')); $classBindings->register(new AliasBinding('beamerAddress', 'beamer_address')); $classBindings->register(new AliasBinding('hasWelcome', 'has_welcome'));*/ /*$classBindings->register(new FieldBinding("contact", "contact", Contact::class)); $classBindings->register(new FieldBinding("campaignDefaults", "campaign_defaults", CampaignDefaults::class)); $classBindings->register(new FieldBinding("stats", "stats", Stats::class));*/ [Truncated] int(0) ["last_sub_date"]=> string(25) "2019-07-15T05:42:05+00:00" ["last_unsub_date"]=> string(25) "2015-12-23T23:51:49+00:00" } ``` In AudienceList i've declared the variable stats (it's in the same namespace, but i've tried to add also the whole path) ```php /** * Stats for the list. Many of these are cached for at least five minutes. * * @var Stats */ private $stats; ``` Where am i wrong? username_0: Hey, sorry for the late response. You are registering a custom Transformer for `AudienceList` after calling `scanAndRegister`. This will overwrite the generated Transformer. Status: Issue closed
ndharasz/onthefly-ios
201862523
Title: Report Screen Question: username_0: ## Story/Task Details - [ ] Create screen for where report will be shown - [ ] Add create/send/save buttons ## Acceptance Scenarios - Given: a basic user of the application - When: they transition to the "New Flight" screen - Then: they will see the above UI components, no functionality ## Done Done Criteria When the user can see the basic UI components mentioned above, without any actual functionality yet.<issue_closed> Status: Issue closed
OpenNMT/OpenNMT-py
373142599
Title: Error in Gigawords summarization preprocess Question: username_0: The current preprocess code does not allow special tokens, like <unk>, shows at the beginning of sentence(in OpenNMT-py/onmt/inputters/dataset_base.py(73)extract_text_features(): `assert all([special != split_token[0] for special in specials])` ). This causes problems, since gigawords has 46811 titles begin with <unk>. Is can be solved by remove unk from special tokens, but is it the proper way to solve this ? Best, Guoyin
MD-Studio/cerise
387274204
Title: Failed .cerise/ environment setup on remote HPC resource Question: username_0: I'm experiencing incomplete setup of remote .cerise/ environments on GT and LISA HPC resources. For GT the .cerise/ directory and all of the files in it where copied to the remote resource but the setup procedure failed at the miniconda stage. Miniconda was downloaded and installed but the cerise virtual env was not created. After running the respective install script manually the environment was created successfully and MD jobs could be launched by lie_md. For LISA only the .cerise/api and .cerise/jobs directories where created without any files in them. The system hanged in this state indefinite. I tried a few times always with the same result. Answers: username_0: This is the cerise_backend log from the last run using the cerise-mdstudio-lisa specialization: ``` [2018-12-04 12:43:30.014] [INFO] Starting up [root] [2018-12-04 12:43:30.022] [DEBUG] protocol: sftp, location: lisa.surfsara.nl, credential: <cerulean.credential.PasswordCredential object at 0x7f4f2c2434a8> [cerise.config] [2018-12-04 12:43:30.369] [INFO] Connecting to lisa.surfsara.nl on port 22 [cerulean.ssh_terminal] [2018-12-04 12:43:30.369] [DEBUG] Authenticating using a password [cerulean.ssh_terminal] [2018-12-04 12:43:30.369] [DEBUG] starting thread (client mode): 0x2c2431d0 [paramiko.transport] [2018-12-04 12:43:30.370] [DEBUG] Local version/idstring: SSH-2.0-paramiko_2.4.2 [paramiko.transport] [2018-12-04 12:43:30.375] [DEBUG] Remote version/idstring: SSH-2.0-OpenSSH_7.4p1 Debian-10+deb9u4 [paramiko.transport] [2018-12-04 12:43:30.376] [INFO] Connected (version 2.0, client OpenSSH_7.4p1) [paramiko.transport] [2018-12-04 12:43:30.377] [DEBUG] kex algos:['curve25519-sha256', '[email protected]', 'ecdh-sha2-nistp256', 'ecdh-sha2-nistp384', 'ecdh-sha2-nistp521', 'diffie-hellman-group-exchange-sha256', 'diffie-hellman-group16-sha512', 'diffie-hellman-group18-sha512', 'diffie-hellman-group14-sha256', 'diffie-hellman-group14-sha1'] server key:['ssh-rsa', 'rsa-sha2-512', 'rsa-sha2-256', 'ssh-ed25519'] client encrypt:['<EMAIL>', 'aes128-ctr', 'aes192-ctr', 'aes256-ctr', '<EMAIL>', '<EMAIL>'] server encrypt:['<EMAIL>', 'aes128-ctr', 'aes192-ctr', 'aes256-ctr', '<EMAIL>', '<EMAIL>'] client mac:['<EMAIL>', '<EMAIL>', 'hmac-sha2-256-etm@open<EMAIL>', 'hmac-sha2-512-etm<EMAIL>', '<EMAIL>-<EMAIL>', '<EMAIL>', '<EMAIL>', 'hmac-sha2-256', 'hmac-sha2-512', 'hmac-sha1'] server mac:['<EMAIL>', '<EMAIL>', '<EMAIL>-sha2-256-et<EMAIL>', 'hmac-sha2-512-etm@<EMAIL>', 'hmac-sha1-etm@open<EMAIL>', '<EMAIL>', '<EMAIL>', 'hmac-sha2-256', 'hmac-sha2-512', 'hmac-sha1'] client compress:['none', '<EMAIL>'] server compress:['none', '<EMAIL>'] client lang:[''] server lang:[''] kex follows?False [paramiko.transport] [2018-12-04 12:43:30.377] [DEBUG] Kex agreed: ecdh-sha2-nistp256 [paramiko.transport] [2018-12-04 12:43:30.377] [DEBUG] HostKey agreed: ssh-ed25519 [paramiko.transport] [2018-12-04 12:43:30.378] [DEBUG] Cipher agreed: aes128-ctr [paramiko.transport] [2018-12-04 12:43:30.378] [DEBUG] MAC agreed: hmac-sha2-256 [paramiko.transport] [2018-12-04 12:43:30.378] [DEBUG] Compression agreed: none [paramiko.transport] [2018-12-04 12:43:30.425] [DEBUG] kex engine KexNistp256 specified hash_algo <built-in function openssl_sha256> [paramiko.transport] [2018-12-04 12:43:30.425] [DEBUG] Switch to new keys ... [paramiko.transport] [2018-12-04 12:43:30.425] [DEBUG] Attempting password auth... [paramiko.transport] [2018-12-04 12:43:30.429] [DEBUG] userauth is OK [paramiko.transport] [2018-12-04 12:43:30.562] [INFO] Auth banner: b' SURFsara\n \n Welcome to SURFsara\n\n This is a private computer facility. Access for any reason must be\n specifically authorized by the owner. Unless you are so authorized,\n your continued access and any other use may expose you to criminal\n and/or civil proceedings.\n\n Information: http://www.surfsara.nl\n\n' [paramiko.transport] [2018-12-04 12:43:30.562] [INFO] Authentication (password) successful! [paramiko.transport] [2018-12-04 12:43:30.562] [INFO] Connection (re)established [cerulean.ssh_terminal] [2018-12-04 12:43:30.562] [INFO] Connecting to SFTP server [cerulean.sftp_file_system] [2018-12-04 12:43:30.562] [DEBUG] [chan 0] Max packet in: 32768 bytes [paramiko.transport] [2018-12-04 12:43:30.813] [DEBUG] Received global request "<EMAIL>" [paramiko.transport] [2018-12-04 12:43:30.813] [DEBUG] Rejecting "<EMAIL>" global request from server. [paramiko.transport] [2018-12-04 12:43:30.815] [DEBUG] [chan 0] Max packet out: 32768 bytes [paramiko.transport] [2018-12-04 12:43:30.815] [DEBUG] Secsh channel 0 opened. [paramiko.transport] [2018-12-04 12:43:30.819] [DEBUG] [chan 0] Sesch channel 0 request ok [paramiko.transport] [2018-12-04 12:43:31.051] [INFO] [chan 0] Opened sftp connection (server version 3) [paramiko.transport.sftp] [2018-12-04 12:43:31.051] [INFO] Connected to SFTP server [cerulean.sftp_file_system] [2018-12-04 12:43:31.056] [INFO] Connecting to lisa.surfsara.nl on port 22 [cerulean.ssh_terminal] [2018-12-04 12:43:31.056] [DEBUG] Authenticating using a password [cerulean.ssh_terminal] [2018-12-04 12:43:31.056] [DEBUG] starting thread (client mode): 0x309ffb00 [paramiko.transport] [2018-12-04 12:43:31.056] [DEBUG] Local version/idstring: SSH-2.0-paramiko_2.4.2 [paramiko.transport] [2018-12-04 12:43:31.062] [DEBUG] Remote version/idstring: SSH-2.0-OpenSSH_7.4p1 Debian-10+deb9u4 [paramiko.transport] [2018-12-04 12:43:31.062] [INFO] Connected (version 2.0, client OpenSSH_7.4p1) [paramiko.transport] [2018-12-04 12:43:31.064] [DEBUG] kex algos:['curve25519-sha256', '<EMAIL>25519-<EMAIL>', 'ecdh-sha2-nistp256', 'ecdh-sha2-nistp384', 'ecdh-sha2-nistp521', 'diffie-hellman-group-exchange-sha256', 'diffie-hellman-group16-sha512', 'diffie-hellman-group18-sha512', 'diffie-hellman-group14-sha256', 'diffie-hellman-group14-sha1'] server key:['ssh-rsa', 'rsa-sha2-512', 'rsa-sha2-256', 'ssh-ed25519'] client encrypt:['<EMAIL>', 'aes128-ctr', 'aes192-ctr', 'aes256-ctr', '<EMAIL>', '<EMAIL>'] server encrypt:['<EMAIL>', 'aes128-ctr', 'aes192-ctr', 'aes256-ctr', '<EMAIL>', '<EMAIL>'] client mac:['<EMAIL>', '<EMAIL>', 'hmac-sha2-256-etm<EMAIL>', '<EMAIL>-sha2-512-etm<EMAIL>', 'hmac-sha1-etm<EMAIL>', '<EMAIL>', '<EMAIL>', 'hmac-sha2-256', 'hmac-sha2-512', 'hmac-sha1'] server mac:['<EMAIL>-64-etm<EMAIL>', '<EMAIL>-128-et<EMAIL>', '[email protected]', '<EMAIL>-sha2-512-etm@<EMAIL>', '<EMAIL>-sha1-et<EMAIL>', '<EMAIL>', '<EMAIL>', 'hmac-sha2-256', 'hmac-sha2-512', 'hmac-sha1'] client compress:['none', '<EMAIL>'] server compress:['none', '<EMAIL>'] client lang:[''] server lang:[''] kex follows?False [paramiko.transport] [2018-12-04 12:43:31.064] [DEBUG] Kex agreed: ecdh-sha2-nistp256 [paramiko.transport] [2018-12-04 12:43:31.064] [DEBUG] HostKey agreed: ssh-ed25519 [paramiko.transport] [2018-12-04 12:43:31.065] [DEBUG] Cipher agreed: aes128-ctr [paramiko.transport] [2018-12-04 12:43:31.065] [DEBUG] MAC agreed: hmac-sha2-256 [paramiko.transport] [2018-12-04 12:43:31.065] [DEBUG] Compression agreed: none [paramiko.transport] [2018-12-04 12:43:31.071] [DEBUG] kex engine KexNistp256 specified hash_algo <built-in function openssl_sha256> [paramiko.transport] [2018-12-04 12:43:31.072] [DEBUG] Switch to new keys ... [paramiko.transport] [2018-12-04 12:43:31.072] [DEBUG] Attempting password auth... [paramiko.transport] [2018-12-04 12:43:31.075] [DEBUG] userauth is OK [paramiko.transport] [2018-12-04 12:43:31.209] [INFO] Auth banner: b' SURFsara\n \n Welcome to SURFsara\n\n This is a private computer facility. Access for any reason must be\n specifically authorized by the owner. Unless you are so authorized,\n your continued access and any other use may expose you to criminal\n and/or civil proceedings.\n\n Information: http://www.surfsara.nl\n\n' [paramiko.transport] [2018-12-04 12:43:31.209] [INFO] Authentication (password) successful! [paramiko.transport] [2018-12-04 12:43:31.209] [INFO] Connection (re)established [cerulean.ssh_terminal] [2018-12-04 12:43:31.210] [DEBUG] [chan 0] stat(b'/') [paramiko.transport.sftp] [2018-12-04 12:43:31.213] [DEBUG] [chan 0] stat(b'/home') [paramiko.transport.sftp] [2018-12-04 12:43:31.217] [DEBUG] [chan 0] stat(b'/home/mvandijk') [paramiko.transport.sftp] [2018-12-04 12:43:31.221] [DEBUG] [chan 0] stat(b'/home/mvandijk/.cerise') [paramiko.transport.sftp] [2018-12-04 12:43:31.224] [DEBUG] [chan 0] mkdir(b'/home/mvandijk/.cerise', 511) [paramiko.transport.sftp] [2018-12-04 12:43:31.227] [DEBUG] [chan 0] chmod(b'/home/mvandijk/.cerise', 511) [paramiko.transport.sftp] [2018-12-04 12:43:31.231] [DEBUG] [chan 0] stat(b'/home/mvandijk/.cerise/api') [paramiko.transport.sftp] [2018-12-04 12:43:31.234] [DEBUG] [chan 0] mkdir(b'/home/mvandijk/.cerise/api', 511) [paramiko.transport.sftp] [Truncated] [2018-12-04 12:43:31.860] [DEBUG] Command executed successfully [cerulean.ssh_terminal] [2018-12-04 12:43:31.860] [DEBUG] sbatch --version exit code: 0 [cerulean.slurm_scheduler] [2018-12-04 12:43:31.860] [DEBUG] sbatch --version output: slurm-wlm 18.08.3 [cerulean.slurm_scheduler] [2018-12-04 12:43:31.860] [DEBUG] sbatch --version error: [cerulean.slurm_scheduler] [2018-12-04 12:43:31.861] [DEBUG] Slots per node set to 4 [cerise.back_end.job_runner] [2018-12-04 12:43:31.861] [DEBUG] [chan 0] stat(b'/home/mvandijk/.cerise/api/cerise/version') [paramiko.transport.sftp] [2018-12-04 12:43:31.866] [CRITICAL] Traceback (most recent call last): File "cerise/run_back_end.py", line 42, in <module> manager = ExecutionManager(config, apidir) File "cerise/../cerise/back_end/execution_manager.py", line 78, in __init__ self._update_available = self._remote_api.update_available() File "cerise/../cerise/back_end/remote_api.py", line 57, in update_available return self._updatable_projects() != [] File "cerise/../cerise/back_end/remote_api.py", line 140, in _updatable_projects project_name)) RuntimeError: Project "steps" in local API definition is missing a "version" file. [root] [2018-12-04 12:43:31.866] [INFO] Shutting down [root] ``` username_1: Uh oh, this is not good. You're using the latest develop Cerise, and the specialisation hasn't been updated to that yet. The new version has a `version` file to facilitate the CWL API update mechanism. It looks like the problem is in the cerise-mdstudio-lisa Dockerfile, it says `FROM mdstudio/cerise:develop` at the top, which should be `FROM mdstudio/cerise:0.1.0`. The GT and DAS5 versions are correct. username_0: I see, makes sense, I will update the version username_0: Updated the docker file and rerun a workflow with the newly build docker image. Got a bit further but running into the following error in the cerise_backend log: ``` [2018-12-04 14:05:18.362] [DEBUG] Staging API install script to /home/mvandijk/.cerise/api/install.sh from /home/cerise/cerise/../api/install.sh [cerise.back_end.xenon_remote_files] [2018-12-04 14:05:21.858] [CRITICAL] Traceback (most recent call last): File "cerise/run_back_end.py", line 48, in <module> manager = ExecutionManager(config, apidir, xenon_) File "cerise/../cerise/back_end/execution_manager.py", line 64, in __init__ api_files_path, api_install_script_path) File "cerise/../cerise/back_end/xenon_job_runner.py", line 30, in __init__ self._sched = config.get_scheduler() File "cerise/../cerise/config.py", line 224, in get_scheduler scheme, location, credential, properties) jpype._jexception.nl.esciencecenter.xenon.XenonExceptionPyRaisable: nl.esciencecenter.xenon.XenonException: slurm adaptor: Got invalid key/value pair in output: Cgroup Support Configuration: [root] [2018-12-04 14:05:21.859] [INFO] Shutting down [root] ``` The api is indeed not build, no conda installation. username_1: That looks like it could be that Lisa has the very latest version of Slurm, and that Xenon 1 doesn't support it. So I guess it'll have to wait for the new Cerise with Cerulean... username_1: The current Cerulean works fine with 18.08, and this will be backed up by tests in the next version.
kubernetes/kubernetes
916745461
Title: k8s.io/kubectl/pkg/cmd/version TestNewCmdVersionWithoutConfigFile always fails locally Question: username_0: /sig cli run `make test` (or just `make test WHAT=k8s.io/kubectl/pkg/cmd/version`. observe failure: ``` --- FAIL: TestNewCmdVersionWithoutConfigFile (0.06s) version_test.go:43: Cannot execute version command: could not parse pre-release/metadata (-master+$Format:%H$) in version "v0.0.0-master+$Format:%H$" FAIL FAIL k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/version 0.307s ``` xref: https://github.com/kubernetes/kubernetes/issues/102607#issuecomment-854985308 Answers: username_0: https://github.com/kubernetes/kubernetes/issues/102607#issuecomment-858300307 following up here in the original issue for now. Status: Issue closed username_0: /sig cli run `make test` (or just `make test WHAT=k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/versionn`). observe failure: ``` --- FAIL: TestNewCmdVersionWithoutConfigFile (0.06s) version_test.go:43: Cannot execute version command: could not parse pre-release/metadata (-master+$Format:%H$) in version "v0.0.0-master+$Format:%H$" FAIL FAIL k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/version 0.307s ``` xref: https://github.com/kubernetes/kubernetes/issues/102607#issuecomment-854985308 username_1: @username_0 Hi, I have filed #102824 to address this issue. username_0: Thank you!
jsdevtom/take-until-destroy
284718981
Title: Component method ngOnDestroy overrided multiple times Question: username_0: # I'm submitting a... Bug report # Current behavior Calling takeUntilDestroy overrides ngOnDestroy method on the target prototype. When there are multiple component instances of the same class, it causes overriding the method multiple times, each time calling the previous one. Destroying the last component instance, will complete all previously created subjects. # Expected behavior ngOnDestroy should be changed on the prototype only one time. One instance of the component should not affect other instances. # Minimal reproduction of the problem with instructions 1. Create component and use takeUntilDestroy 2. Create another component that uses the first one in ngFor<issue_closed> Status: Issue closed
dshawul/Scorpio
1067069735
Title: Ponder support in UCI mode? Question: username_0: Hi Daniel, can you add support for ponder in UCI mode? I see that the engine can ponder in xboard mode already, but there's no option for it in UCI. Is there a reason for this? Is it difficult to add? Answers: username_1: It was supposed to work in UCI mode but I admit I never tested it well. Will get back to you once I fix this. Thanks! username_0: Well maybe it works, but it doesn't even present the "ponder" option (when you type "uci") so I can enable it, and so the GUI doesn't know Scorpio can actually ponder. username_1: I tested ponderng in Arena with scorpio as UCI engine and it actually seems to work there just fine. For some reason though, winboard with -fUCI option for scorpio doesn't seem to work with pondering.
JuliaDocs/Documenter.jl
171240560
Title: Use HTML renderer for docs Question: username_0: Now that #171 is merged it would be good to start using it for our own docs. @username_1 would you like to do the honours? Answers: username_1: I'm looking into this by the way, just trying stuff out to make sure that nothing breaks with the HTML build. However, in principle, setting ``` target = "build", deps = nothing, make = () -> nothing, ``` in `deploydocs` should actually sufficient, right? username_0: `cleandir` is run on the `target` prior to `make` (https://github.com/JuliaDocs/Documenter.jl/blob/e20b1f8187cc140f568c424d07d4bf04117c0d28/src/Documenter.jl#L308), so I don't think that would be sufficient. I'm also trying this out on DocStringExtensions (https://github.com/JuliaDocs/DocStringExtensions.jl/compare/mh/htmldocs?expand=1) to see what would work. username_0: If we just remove the `cleandir` call things should probably be fine, since a fresh travis clone shouldn't have anything in that directory I suppose. That would mean we can use your approach from above. username_0: And we could probably add a `make === nothing` check like for `deps` so that it would just be ``` target = "build", deps = nothing, make = nothing, ``` username_0: #197 appears to do the right thing I think. Tested on https://michaelhatherly.github.io/DocumenterTests.jl/latest/index.html. username_1: Yup. I tested it with http://username_1.eu/Hafta.jl/latest/ and it also seems to work. Status: Issue closed
PyDataPune/Talks
473863707
Title: Forecasting time series data effectively Question: username_0: In this session i will introduce audience to the concept of forecasting and live code demo for forecasting and showcase time series example that have strong seasonal effects and several seasons of historical data. We will be using Facebook's Prophet library for forecasting purpose. I want to target this session to audience having intermediate python developers. I expect my session to be atleast 45 min long. I will bring my own laptop for session. Thanks Answers: username_1: Thanks, Shoaib for the proposal!!!. I am quietly excited about time series analysis. Do let me know the talk line up once your talk is ready Will try to arrange an entire event focussed on time series data and associated analysis. Thanks <NAME> username_1: Hey Shoaib, Please let me know when we can have your talk on board. How about this month. Thanks <NAME>
SharePoint/sp-provisioning-service
718289720
Title: Need a Yes/No editor prompt that returns True/False to template parameter Question: username_0: **Is your feature request related to a problem? Please describe.** The Custom Learning PnP template includes a parameter to enable or disable telemetry. We would like to make this parameter visible when the template is installed. **Describe the solution you'd like** When provisioning the template, the user is presented with Yes or No radio buttons, which they can select in response to a text prompt. **Describe alternatives you've considered** We tried using the Boolean editor, in settings.json as follows: ``` { "name":"Telemetry", "caption":"Improve learning pathways by sending usage telemetry to Microsoft?", "description":"Microsoft uses this data to help improve future Microsoft 365 learning pathways solutions. To learn more about Microsoft privacy policies see https://go.microsoft.com/fwlink/?LinkId=521839.", "editor": "Boolean" } ``` But it returned this error: There was an unexpected issue. The model item passed into the dictionary is of type 'SharePointPnP.ProvisioningApp.WebApp.Models.TemplateParameterModel', but this dictionary requires a model item of type 'System.Boolean'. **Additional context** Bert suggested I tag @PaoloPia in my inquiry. Answers: username_0: @VesaJuvonen, is the right place to ask for feature requests, or would you like me to submit this elsewhere? @MHollinshead
MatterHackers/MatterControl
226679521
Title: the second extrudeur clone the first Question: username_0: we see on the print the PVa is a copy of the PLA ![img_5535](https://cloud.githubusercontent.com/assets/26421644/25762610/b3531ca6-31df-11e7-8f91-d67dd980f4b9.JPG) Answers: username_0: ok it s not an issu, it just the temperatur set too hight PVA must be around 180 ( the wihte ) PLA must be at 200 or less ( the black) thanks for your super help guys Status: Issue closed username_0: we see on the print the PVa is a copy of the PLA ![img_5535](https://cloud.githubusercontent.com/assets/26421644/25762610/b3531ca6-31df-11e7-8f91-d67dd980f4b9.JPG) the original is like that ![img_5460](https://cloud.githubusercontent.com/assets/26421644/25762680/05d1bbb8-31e0-11e7-8a76-d6172f0a9222.JPG) see the .stl ( change extention) [Bras roue droite.txt](https://github.com/MatterHackers/MatterControl/files/980212/Bras.roue.droite.txt) now it clone in inverse ....... ![img_6059](https://cloud.githubusercontent.com/assets/26421644/25763652/68e41328-31e4-11e7-8a35-926137c2514c.JPG) username_0: but it s still an insu ...because this PVA come out at 165 degres and matter control can't do that
einaregilsson/Redirector
396288981
Title: Change repository description Question: username_0: Please change "Firefox extension to redirect urls ..." To something like "multi browser" or "Firefox, Chrome, Opera, Edge" Answers: username_1: Done :) Does it work in Edge though? Status: Issue closed username_0: I didn't try, but it is possible: https://www.google.com/search?q=edge+chrome+extension :-) username_2: MS Edge is now Chromium based browser, it’s now accept installing extensions from Chrome Web Store. username_1: Done. I tested it in Edge and it does work, although I'm not sure how to get to the logs and stuff, but the redirecting does work.
pingcap/tiup
784962199
Title: Command will trigger to stop cluster services(tiup cluster clean --log) Question: username_0: ``` # tiup --version v1.3.1 tiup Go Version: go1.13 Git Branch: release-1.3 GitHash: d51bd0c # tiup cluster --version Starting component `cluster`: /root/.tiup/components/cluster/v1.3.1/tiup-cluster --version tiup version v1.3.1 tiup Go Version: go1.13 Git Branch: release-1.3 GitHash: d51bd0c ``` Answers: username_1: Nodes will be ignored: [] Roles will be ignored: [] Files to be deleted are: xxx.xx.x.51: /data2/dylan/deploy/prometheus-19090/log/*.log /data2/dylan/deploy/tiflash-9000/log/*.log /data2/dylan/deploy/pd-12379/log/*.log /data2/dylan/deploy/alertmanager-19093/log/*.log /data2/dylan/deploy/grafana-13000/log/*.log xxx.xx.x.52: /data2/dylan/deploy/tidb-4008/log/*.log /data2/dylan/deploy/pd-12379/log/*.log xxx.xx.x.53: /data2/dylan/deploy/tidb-4008/log/*.log /data2/dylan/deploy/tikv-11160/log/*.log /data2/dylan/deploy/pd-12379/log/*.log xxx.xx.x.54: /data2/dylan/deploy/tikv-11160/log/*.log xxx.xx.x.55: /data2/dylan/deploy/tikv-11160/log/*.log Do you want to continue? [y/N]: y IMO, we can highlight the stop message `This operation will stop tidb v4.0.9 cluster tidb-dylan and clean its' log`. In the current implementation, we delete all log files and the process does not write to a new file without restarting, so the stop is necessary. Maybe we need another command to delete the old logs without cluster stopping @lucklove username_2: We'll need to wait tidb to support rotating current log file without restarting the process.
aodn/content
355442297
Title: Update link to collection on Portal in metadata record Question: username_0: In each record in catalogue-imos corresponding to a collection on the Portal, in the distribution section there is a link to the collection on the portal - labelled: View and download data though the AODN Portal. The link is the former URL of the portal, for example: **https://imos.aodn.org.au/imos123/home**?uuid=e952bcee-79e7-4995-91ad-a7e6408d29ce It does resolve, but would be neater to update them all: **https://portal.aodn.org.au/search**?uuid=e952bcee-79e7-4995-91ad-a7e6408d29ce Answers: username_1: There are two variations to the link so we could also automate a check on the label associated to any of these uuid links. username_0: The appropriate online resource in each record is easily identified by the URL. If there is variation in the text in description, this won't affect the identification of the online resource. At the same time that the link is updated, the description text can be standardised to "View and download data though the AODN Portal".
smalruby/smalruby3-gui
347540908
Title: ScratchやCatの問題が残っている。 Question: username_0: #21 のマージでタイトルにCatが入ってしまっています。 また、ローカライズの関係かもしれませんが、トップモーダルで『Scratch 3.0にようこそ』と出ています。 Answers: username_0: https://github.com/LLK/scratch-l10n こちらのリポジトリが影響しているみたいです。 node_modulesの中にあるファイルが影響しているみたいです。 解決するには、Smalruby-10n ファイルを作成しそちらを`npm install`するパターン。もしくは必要なところだけ独自のtranslationファイルを作成するの2通りが考えられます。 ローカライズは、Scratch本家がreact-intlを使っています。https://github.com/yahoo/react-intl ローカライズの中ではトップ2に入るほど有名なパッケージですね。 username_0: Smalruby独自のローカライズid一覧 `gui.smalruby3.unsupportedBrowser.description` `gui.smalruby3.previewInfo.welcome` `gui.smalruby3.previewInfo.invitation` `gui.smalruby3.webglModal.description` `gui.smalruby3.previewInfo.label` username_1: @username_0 #26 でこのissueで想定していた作業は完了ですかね?もしそうならば、このissueはcloseします。 username_0: そうですね。issue的には終わりなのでこちらをClosedして新しいissueを作る流れでしょうか? メッセージカタログの追加 みたいな username_1: をお願いします。 Status: Issue closed
Chip-L/ColmarAcademy
281861677
Title: long selector chains Question: username_0: You are using some long selectors chains here which makes your css very rigid. Any minor change to design in this section and all this code is going to break. The more chained selectors the less adaptable the styling, and websites are going to change. One solution to this is to create new selectors so you don't have to be so specific in this location. https://github.com/Chip-L/ColmarAcademy/blob/master/resources/css/style.css#L373-L384
joachimruhs/myleaflet
554027893
Title: htmlspecialchars() creates problems with the rendering Question: username_0: Hi, in the rendering of the discription the html code is showing in the frontend. when the htmlspecialchars() is remove everything works fine. is this code important? https://github.com/username_1/myleaflet/blob/94746fdde2445a5262e3ba388fc854a0b60c2516/Classes/Controller/AjaxController.php#L354 Answers: username_1: Hi, may be you can run into XSS issues without htmlspecialchars. Otherwise you can use $locations[$i]['infoWindowDescription'] ... username_1: Ok, fixed. Changed the sequence of desciption and infoWindowDescription in AjaxController. username_0: thanks!! It works know Status: Issue closed
blockchain/My-Wallet-V3-Android
133166280
Title: App does not close and return transaction hash to the initial bip21 intent. Question: username_0: After sending a bip21 bitcoin URI to the blockchain app for payment, the app does not return transaction hash in the response back to the first app that sends the intent. It would be great to follow what Mycelium and Bitcoin android app does as follows: `// Include the transaction hash in the response Intent result = new Intent(); result.putExtra(Constants.TRANSACTION_HASH_INTENT_KEY, _transaction.getHash().toString()); setResult(RESULT_OK, result);`
Shwetaphatale/vita-assignment
679583828
Title: *Hive Basic Lab2- Analyzing Big Data with Hive Question: username_0: ![image](https://user-images.githubusercontent.com/63608018/90313885-94ba9680-df2d-11ea-8902-a98d732c8ea6.png) ![image](https://user-images.githubusercontent.com/63608018/90313890-98e6b400-df2d-11ea-9a09-717a5e6f5b6f.png) ![image](https://user-images.githubusercontent.com/63608018/90313923-c59acb80-df2d-11ea-93b3-6cf97600b6e6.png) ![image](https://user-images.githubusercontent.com/63608018/90313927-cb90ac80-df2d-11ea-8749-f8dc4731a84c.png) ![image](https://user-images.githubusercontent.com/63608018/90313954-ecf19880-df2d-11ea-99c8-f0434f394f5e.png) ![image](https://user-images.githubusercontent.com/63608018/90313964-f24ee300-df2d-11ea-93a8-d4d37c3c9890.png) ![image](https://user-images.githubusercontent.com/63608018/90314004-3510bb00-df2e-11ea-85e6-522bd116284e.png) ![image](https://user-images.githubusercontent.com/63608018/90314009-3e9a2300-df2e-11ea-8e64-4aff62cf45f1.png) ![image](https://user-images.githubusercontent.com/63608018/90314030-64272c80-df2e-11ea-81f8-8b1aeb3ffb96.png) ![image](https://user-images.githubusercontent.com/63608018/90314040-6d17fe00-df2e-11ea-9d95-81347a0997e8.png) ![image](https://user-images.githubusercontent.com/63608018/90314072-989ae880-df2e-11ea-8ced-e57cf380fe0e.png) ![image](https://user-images.githubusercontent.com/63608018/90314078-a2bce700-df2e-11ea-973c-02c1dc24a40a.png) ![image](https://user-images.githubusercontent.com/63608018/90314369-dd278380-df30-11ea-98a6-b5b62ccea866.png) ![image](https://user-images.githubusercontent.com/63608018/90314374-e7e21880-df30-11ea-95c4-1445aedc343f.png) ![image](https://user-images.githubusercontent.com/63608018/90314398-22e44c00-df31-11ea-910c-d38eb67d32cd.png) ![image](https://user-images.githubusercontent.com/63608018/90314405-2972c380-df31-11ea-810d-88fa6c7f218e.png) ![image](https://user-images.githubusercontent.com/63608018/90314427-545d1780-df31-11ea-9f3e-428830a9e2c8.png) ![image](https://user-images.githubusercontent.com/63608018/90314431-5d4de900-df31-11ea-8453-5ee4843ed46c.png) ![image](https://user-images.githubusercontent.com/63608018/90314468-9ab27680-df31-11ea-9244-552827954379.png) ![image](https://user-images.githubusercontent.com/63608018/90314471-a4d47500-df31-11ea-8da0-d1f115283d70.png)<issue_closed> Status: Issue closed
ulab-committee/ulab-website
427468412
Title: ActionView::Template::Error in spina/conferences/conferences#show Question: username_0: ## Error in ulab-website **ActionView::Template::Error** in **spina/conferences/conferences#show** undefined method `year' for #<Spina::Conferences::Conference:0x000055a14719c240> [View on Bugsnag](https://app.bugsnag.com/ulab/ulab-website/errors/5ca15ec275057d001954078c?event_id=5ca15ec2003796ed8cfb0000&i=gh&m=ci) ## Stacktrace app/views/spina/conferences/conferences/show.ics.haml:9 - block in _app_views_spina_conferences_conferences_show_ics_haml__1087225668375819926_47075585515860 app/views/spina/conferences/conferences/show.ics.haml:3 - _app_views_spina_conferences_conferences_show_ics_haml__1087225668375819926_47075585515860 [View full stacktrace](https://app.bugsnag.com/ulab/ulab-website/errors/5ca15ec275057d001954078c?event_id=5ca15ec2003796ed8cfb0000&i=gh&m=ci) *Created automatically via Bugsnag*<issue_closed> Status: Issue closed
KnowledgeLinks/Plains2PeaksPilot
289443641
Title: Title duplicated in subject field Question: username_0: Title is consistently being duplicated in the subject field across all institutions. Answers: username_1: After doing some investigation, I think the problem may in the Marmot load. I harvested a random 250 records from Plains2Peaks and it was only in those records from Marmot institutions that have subjects that are identical or strongly similar to the title. Checking the Marmot JSON feed (https://titan.marmot.org/API/ArchiveAPI?method=getDPLAFeed&page=1&pageSize=10), you'll see that the title is being duplicated in the subjects field that they are providing. If we don't want this duplication, we'll need to have Marmot adjust their feed or adjust the custom bibcat Marmot ingester. Could you check to see if there are other objects that have this title duplication from non-Marmot libraries? Otherwise, I'll close this issue. username_0: Hi, Jeremy - It does indeed appear to only be happening with Marmot libraries. You can go ahead and close the case. I will chat with Leigh about how she wants to address it with Marmot. Thanks, Amy <NAME> Collaborative Programming Coordinator Colorado State Library, Networking and Resource Sharing Web: CSL <http://www.cde.state.co.us/cdelib> | Colorado Virtual Library <http://www.coloradovirtuallibrary.org/> Twitter: @COStateLibrary <https://twitter.com/COStateLibrary> | @hitchlib <https://twitter.com/hitchlib> Facebook: Colorado State Library <https://www.facebook.com/coloradostatelibrary/> Status: Issue closed
childe/hangout
175693233
Title: 从kafka拉数据到es,启动的时候报错老是: bin/hangout -f test.yml Question: username_0: Exception in thread "main" java.lang.NullPointerException at scala.collection.convert.Wrappers$JMapWrapperLike$$anon$2.<init>(Wrappers.scala:265) at scala.collection.convert.Wrappers$JMapWrapperLike$class.iterator(Wrappers.scala:264) at scala.collection.convert.Wrappers$JMapWrapper.iterator(Wrappers.scala:275) at scala.collection.IterableLike$class.foreach(IterableLike.scala:72) at scala.collection.AbstractIterable.foreach(Iterable.scala:54) at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:144) at scala.collection.AbstractTraversable.foldLeft(Traversable.scala:105) at scala.collection.TraversableOnce$class.$div$colon(TraversableOnce.scala:138) at scala.collection.AbstractTraversable.$div$colon(Traversable.scala:105) at scala.collection.immutable.MapLike$class.$plus$plus(MapLike.scala:87) at scala.collection.immutable.AbstractMap.$plus$plus(Map.scala:187) at kafka.javaapi.consumer.ZookeeperConsumerConnector.createMessageStreams(ZookeeperConsumerConnector.scala:83) at kafka.javaapi.consumer.ZookeeperConsumerConnector.createMessageStreams(ZookeeperConsumerConnector.scala:97) at com.ctrip.ops.sysdev.inputs.Kafka.emit(Kafka.java:128) at com.ctrip.ops.sysdev.Main.main(Main.java:109) 我的配置 inputs: - Kafka: codec: json topic: wiseweb_crawler_webpage: 4 consumer_settings: group.id: hangouts # socket.receive.buffer.bytes:"1048576" # fetch.message.max.bytes:"1048576" zookeeper.connect: 10.165.65.14:2181,10.171.100.247:2181,10.251.5.35:2181 auto.commit.interval.ms: "5000" 这是输入 Answers: username_1: 你使用的zk, kafka, hangout版本是什么? username_0: 好了,我之前用hadoop用户启动老是报错,今天用root启动了下,可以了,数据到es了,谢谢 Status: Issue closed
plk/biblatex-apa
278986969
Title: Journal articles in special issues Question: username_0: When citing articles in special or supplementary issues, for which the `biblatex` manual suggests to use the `issue` field (implied on p. 20), the issue number appears in the text citation, but not in the references (I expected the reverse outcome). It works when using the `number` field instead, but Zotero BetterBiblatex will not let me export non-numeric issues to the `number` field. In any case, given the `biblatex` recommendation, the correct field should be `issue`. As per the [APA-style blog](http://blog.apastyle.org/apastyle/2012/05/citing-a-special-issue-or-special-section-in-apa-style.html), articles from supplementary issues should be cited and referenced as any other journal article, although necessarily tagged with the correct issue ("Suppl. 1" below). Hence, in the MWE below, I expect the reference entry to be `<NAME>. & <NAME>. (2013). On the Theory of Ethnic Conflict. Journal of the European Economic Association, 11(Suppl. 1).` , but I get `<NAME>. & <NAME>. (2013). On the Theory of Ethnic Conflict. Journal of the European Economic Association, 11.` , and I expect the citation `(Caselli & C<NAME>, 2013)` , but I get `Caselli & <NAME>, 2013, Suppl. 1` ``` \documentclass{article} \usepackage{filecontents} \begin{filecontents}{\jobname.bib} @article{Caselli2013, title = {On the {{Theory}} of {{Ethnic Conflict}}}, volume = {11}, doi = {10.1111/j.1542-4774.2012.01103.x}, issue = {Suppl. 1}, journaltitle = {Journal of the European Economic Association}, author = {Caselli, <NAME> <NAME>, <NAME>}, date = {2013}, } \end{filecontents} % Setup that I use, although not all options necessary here \usepackage[ backend=biber, apamaxprtauth=30, style=apa, natbib=true, doi=false, eprint=false, labeldate=year]{biblatex} \addbibresource{\jobname.bib} \begin{document} \citep{Caselli2013} \printbibliography \end{document} ``` Answers: username_1: In many cases you don't actually want `issue`, you want `number` instead. If you change `issue` to `number` in your MWE, things looks better. username_1: See also ```latex \documentclass{article} \usepackage[backend=biber, style=apa]{biblatex} % \usepackage{filecontents} \begin{filecontents}{\jobname.bib} @article{fogliano_vitro_2011, langid = {english}, title = {In Vitro Bioaccessibility and Gut Biotransformation of Polyphenols Present in the Water-Insoluble Cocoa Fraction}, volume = {55}, issn = {1613-4133}, doi = {10.1002/mnfr.201000360}, issue = {S1}, journaltitle = {Molecular Nutrition \& Food Research}, date = {2011-05-01}, pages = {S44--S55}, author = {<NAME> Corollaro, <NAME> and <NAME> and Napolitano, Aurora and Ferracane, Rosalia and Travaglia, Fabiano and Arlorio, Marco and Costabile, Adele and Klinder, Annett and Gibson, Glenn}, } \end{filecontents} \addbibresource{\jobname.bib} \begin{document} \parencite{fogliano_vitro_2011} \printbibliography \end{document} ``` adapted from https://tex.stackexchange.com/q/414495/35864 It seems to me that the `issue` field should not pop up in citations and should instead be printed in the bibliography. I expect that `issue` should either behave as `number` or as in the standard styles. username_2: According to the current APA style, use `TITLEADDON` for this, not `ISSUE`. See the comments in the example .bib files that come with the style. username_1: makes me think that it would be most appropriate to simply go with `number`, but then the problem is that one would put non-integer values there (which is actively discouraged). username_2: Ok, I have moved such cases to be periodicals, that's more sensible. For some reason I can't remember, periodical wasn't used at all. Status: Issue closed username_1: Just to reiterate and for the benefit of the people coming here via their favourite search engine: `number` is the most appropriate field here, `issue` gives sub-par results. There is an issue at the `biblatex` bugtracker (https://github.com/username_2/biblatex/issues/726) that discusses changes to the documentation and data model to officially endorse using `number` instead of `issue` here.
shfshanyue/Daily-Question
903088617
Title: 【Q543】什么是事件冒泡和事件捕获 Answers: username_1: 个人记录 [事件冒泡与事件捕获](https://github.com/username_1/javaScript-learning/blob/main/interviewJs/BrowserDom/%E4%BA%8B%E4%BB%B6%E5%86%92%E6%B3%A1%E4%BA%8B%E4%BB%B6%E6%8D%95%E8%8E%B7%E4%BA%8B%E4%BB%B6%E5%A7%94%E6%89%98.md)
space-wizards/space-station-14
738950288
Title: Block Game™️ code improvements Question: username_0: As discussed with @username_1 - Messages like `BlockGameVisualUpdateMessage` that have an internal enum to determine the purpose of their data, done to avoid having 1 C# type per variant - This is a rather baseless fear, there's nothing bad about having lots of classes for messages (just stick them in a `.messages` sub namespace if you're worried about pollution) - The message size increases to be able to hold the enum, thankfully with byte enums this shouldn't be too bad (The enums in Block Game™️ aren't byte though!), but it's still unnecessary size - The "variant"/"purpose" of the data can be determined through the type of the message if each message were its own concrete C# type -- this is effectively free as the type of the message is already sent over the network - State changes everywhere - The current piece can be moved in multiple "ticking" methods such as ProcessInput - Makes it hard to determine what the state of the game is - Bad for prediction - There's no 2d grid - Block Game™️ is a 2d rectangle grid game, the simplest storage form for that would be a 2d array - the 2d array can be easily stored as a 1d array - Positions of blocks no longer need to be stored, if a block's x and y need to be looked up from an index, it's as simple as: ```cs var x = i % widthOfGrid; var y = i / widthOfGrid; ``` - Blocks can be looked up via their x/y coords still ```cs var i = widthOfGrid * y + x; ``` - 2d grid would allow for easy removal of the `BlockGameBlock` struct, simply have the grid be an array of `BlockGameBlock` enum, eg: ```cs public enum BlockGameBlock : Byte { Red, Blue, ..., Empty } ``` - Collisions against the grid can simply be done by looking up if an x y index isn't `BlockGameBlock.Empty` Answers: username_1: (Maybe) also make certain elements (UI, Game) more portable to allow for an easy implementation of versus. oh and also do versus someday. Also some thought could be brought into maybe doing prediction while im at it refactoring the whole thing username_2: I feel like there are better uses of time than adding prediction to the Nanotrasen Block Game. username_3: 🤔 People work on what they want to work on, don't bitch at them for it. Block Game is basically unplayable anyway at high ping, so it's needed. username_2: Fair, but thats just my opinion on it
maxplanck-ie/snakepipes
325212907
Title: Streamline installation Question: username_0: In the future we should create a channel for this such that one can `conda install -c mpiie snakepipes` into a conda environment and that will bring in snakemake as a requirement. Then we can have an `setup` command within that which can create the workflow environments as needed. This will be after everything else is done in the hackathon and will then allow us to remove `module load snakemake/5.1.2` from the scripts. Answers: username_0: Currently the root environment requires the following: - snakemake - graphviz (for the `--DAG` option, though I expect 99.99% of servers already have this) username_1: @username_0 and yaml as well username_0: unless you explicitly installed it then it was pulled in as a dependency. username_1: We can also have a discussion about whether we need **snakediff** as separate package at all or simply move all functions under one folder called Rscripts and call it done.. My main reason to make it a package was to install all R dependencies when we install it, but now there's conda to deal with that..
naser44/1
99346268
Title: كوب من البردقوش صباحا ومساء يخلصك من الوزن الزائد Question: username_0: <a href="http://ift.tt/1E9CSRA">&#1603;&#1608;&#1576; &#1605;&#1606; &#1575;&#1604;&#1576;&#1585;&#1583;&#1602;&#1608;&#1588; &#1589;&#1576;&#1575;&#1581;&#1575; &#1608;&#1605;&#1587;&#1575;&#1569;&#1611; &#1610;&#1582;&#1604;&#1589;&#1603; &#1605;&#1606; &#1575;&#1604;&#1608;&#1586;&#1606; &#1575;&#1604;&#1586;&#1575;&#1574;&#1583;</a>
PiranhaCMS/piranha.core
344447468
Title: Missing content type Question: username_0: It seems piranha is not providing content type in response. In general maybe no big issue, but if you run the application with the following security header: `<add` name="X-Content-Type-Options" value="nosniff" /> The browsers (or other applications) are ask to ignore file extensions or guessing using the content. What is the result or the problem: You can see it very easily using the sitemap, the sitemap is displayed in the browser as plaintext, not as xml where you can collapse the nodes etc. Another application i am using for seo analysis thinks the sitemap is a normal page and gave me the advice to add description, headline etc. I have no idea what the result is for a crawler or so. So my request is: Please add the content type header depending on the resource (xml for the sitemap ;) ) I did a quick check, the mvc is working correcty, if i request the response from the controller directly i see the correct content type (for example json) `public IActionResult Details() { return Json(""); }` Answers: username_1: Just to clarify, I assume this error only exists for the Sitemap middleware as everything else is standard MVC? username_0: Well, additional comment. I think you can ignore the cms part. The content type is only missing in case of response type "304". I guess this is fine, i checked with other website and there the content type is also included in "304" case. But as i said, i see no error case. So fixing the sitemap.xml would be enough and great. Sorry for the confusion. Thanks. username_1: I've checked pages coming from the CMS and they return `text/html; charset=utf-8` both on **localhost** and from **azure**. Like you say, for `304` I don't see any problem with it, but I'll make sure to add `application/xml; charset=utf-8` for the sitemap Status: Issue closed username_1: This has now been released and is available in `Piranha.AspNetCore` 5.1.2
pop-os/pop
467195475
Title: Title bar menu won't stay open Question: username_0: <!-- If this is a bug, please use the template below. If this is a question or general discussion topic, please start a conversation in our chat https://chat.pop-os.org/ or post on our subreddit https://reddit.com/r/pop_os - as those are the proper forums for that type of discussion. --> **Distribution POP OS 19.04 Nvidia** **Related Application and/or Package Version N/A** **Issue/Bug Description:** Title bar menu on any window won't always stay open **Steps to reproduce (if you know):** Right-click on title bar in various locations. Sometimes it will stay open and other times it instantly closes. Just depends on where you right-click **Expected behavior:** Title bar menu should stay open no matter where you click on the title bar. **Other Notes:** I can right-click and hold and select a menu option, but I shouldn't have to. Thanks
haikuports/haikuports
1128575361
Title: Inkscape segment violation on launch Question: username_0: Per topic, Inkscape crashes immediately upon being opened under a fully-updated fresh installation of R1B3. Here are my computer's specs: Processor: AMD Ryzen 5 5600X Motherboard: MSI B550-A PRO RAM: 16 GB DDR4-3600 CL18 SSD 1: 1 TB WD Blue NVMe SSD (btrfs-formatted; contains Garuda Linux) SSD 2: 480 GB WD Blue SATA SSD (BeFS-formatted; contains Haiku) HDD: 8 TB Samsung HDD (exFAT-formatted; contains misc. data) GPU: MSI Radeon RX 570 w/ 8 GB VRAM and here's the debug report: ``` Debug information for team /boot/system/apps/Inkscape (136463): CPU(s): 12x AMD Ryzen 5 5600X Memory: 15.93 GiB total, 2.57 GiB used Haiku revision: hrev55181+63 Feb 7 2022 08:08: (x86_64) Active Threads: thread 136464: Xlibe BApplication thread 136468: pthread func thread 136469: pthread func thread 136470: pthread func thread 136471: pthread func thread 136472: pthread func thread 136473: pthread func thread 136474: pthread func thread 136475: pthread func thread 136476: pthread func thread 136477: pthread func thread 136478: pthread func thread 136479: pthread func thread 136480: pthread func thread 136481: team 136463 debug task thread 136463: Inkscape (main) state: Exception (Segment violation) Frame IP Function Name ----------------------------------------------- 0x7f596d7e7b20 0x1a801a31675 g_build_path_va + 0x85 Disassembly: g_build_path_va: 0x000001a801a315f0: 55 push %rbp 0x000001a801a315f1: 4889e5 mov %rsp, %rbp 0x000001a801a315f4: 4157 push %r15 0x000001a801a315f6: 4156 push %r14 0x000001a801a315f8: 4989fe mov %rdi, %r14 0x000001a801a315fb: 4155 push %r13 0x000001a801a315fd: 4989f5 mov %rsi, %r13 0x000001a801a31600: 4154 push %r12 0x000001a801a31602: 53 push %rbx 0x000001a801a31603: 4889cb mov %rcx, %rbx 0x000001a801a31606: 4883ec58 sub $0x58, %rsp 0x000001a801a3160a: 48895598 mov %rdx, -0x68(%rbp) 0x000001a801a3160e: 48894db0 mov %rcx, -0x50(%rbp) 0x000001a801a31612: e8d90ffeff call 0x1a801a125f0 0x000001a801a31617: 31ff xor %edi, %edi 0x000001a801a31619: 8945a4 mov %eax, -0x5c(%rbp) 0x000001a801a3161c: e82f990300 call 0x1a801a6af50 0x000001a801a31621: c745bc00000000 mov $0x0, -0x44(%rbp) [Truncated] 13170976 0 118266 width buffer 13173762 0 118272 some BLocker 13173775 0 118881 LocaleRosterData 13174432 0 127502 font list 13174433 0 127502 BLooperList lock 13174446 0 126566 Catalog 13174457 0 135868 some BBlockCache lock 13174462 0 127502 XEvents 13174471 0 118272 BMessageQueue Lock 13228451 0 118266 AppLooperPort 13474789 0 122622 AppServerLink_sLock 13476166 0 118272 some BLocker 13476329 0 127502 ***** 13477191 0 127502 token space 13477280 0 118315 some BLocker 13477407 0 127502 BMessageQueue Lock 13477589 0 129888 screen list 13477654 0 126566 clipboard ``` Answers: username_1: Could you try with a current nightly? username_2: Application launch PASSES on Haiku revision hrev55874 (x86_64) username_3: Added the details markdown, makes scrolling unnecessary once read username_0: @username_1 Will report back soon. @username_3 Thanks. I didn't know that existed 😓
uOttawaDrone/drone-fall-2019
525261016
Title: Lab Documentation link Problems Question: username_0: # Lab Documentation link Problems * Encountering issues with opening image links in the lab documentation. * Links redirect to back to documentation rather than the image. * Issue could possibly be related to markdown. Answers: username_1: This can be solved by removing the %'s in the template ;). Sorry I should have made that more clear! For example: ``` ![Power Supply Board](https://github.com/uOttawaDrone/drone-fall-2019/blob/master/docs/Lab%20Picture/power%20supply%20board.jpg "PSB picture") ``` ![Power Supply Board](https://github.com/uOttawaDrone/drone-fall-2019/blob/master/docs/Lab%20Picture/power%20supply%20board.jpg "PSB picture") <img width="928" alt="Screen Shot 2019-11-19 at 3 57 42 PM" src="https://user-images.githubusercontent.com/31967593/69185915-9ca3f200-0ae5-11ea-8150-7dd8be46d4e7.png"> Status: Issue closed
gladguys/aluco
554518090
Title: Status da chamada dos alunos na turma fica salvo para outras datas Question: username_0: Cenario: Entro pra fazer chamada para o dia 23/01/2020 ![image](https://user-images.githubusercontent.com/5836109/73041138-079ddc00-3e3b-11ea-81b1-b3336365a00c.png) Depois ao mudar para outra data: 27/01/2020 os status das chamadas do dia atual (23/01) ficam salvos nos inputs. ![image](https://user-images.githubusercontent.com/5836109/73041114-f48b0c00-3e3a-11ea-8f52-7628ef51cb98.png)<issue_closed> Status: Issue closed
getsentry/sentry-javascript
381458791
Title: Unclear which package to use for server-side / isomorphic react app Question: username_0: The @sentry/browser package is said to include React integrations, and you also have a Node package - but what do you use for a server rendered react app? I was expecting a @sentry/isomorphic or similar. Status: Issue closed Answers: username_1: Duplicate of https://github.com/getsentry/sentry-javascript/issues/951 with some ideas in there. username_2: @username_1 you closed #951 as it was too generic. Is there an open issue that tracks this? An example of implementation would be @bugsnag/js. username_2: Looks like this may be a workaround: https://github.com/syginc/isomorphic-sentry username_3: This has been closed for a good while, so quite possibly you all are aware of this already, but if you're using nextjs for your isomorphic app, we now have an SDK for that (docs [here](https://docs.sentry.io/platforms/javascript/guides/nextjs/)), and if not, it might nonetheless give you some pointers on how to work around these issues. Cheers!
reallyenglish/ansible-role-uchiwa
255512531
Title: error log does not get logged in log file Question: username_0: ##### ISSUE TYPE - Bug Report ##### ROLE VERSION ``` 1.2.0 ``` ##### CONFIGURATION n/a ##### OS / ENVIRONMENT FreeBSD ##### SUMMARY the startup script of the FreeBSD package has a bug and the error log is not logged in the log file. ##### STEPS TO REPRODUCE * run the example playbook * login to the host * run `service uchiwa restart` wait until the process logs error which should not be logged to stderr of the console. ##### EXPECTED RESULTS errors are logged in the log file. ##### ACTUAL RESULTS they are logged to stderr. Answers: username_0: this diff should fix the issue. ```diff --- /usr/local/etc/rc.d/uchiwa.orig 2017-09-06 07:20:52.821846000 +0000 +++ /usr/local/etc/rc.d/uchiwa 2017-09-06 07:21:09.181532000 +0000 @@ -29,7 +29,7 @@ pidfile="${uchiwa_rundir}/${name}.pid" command="/usr/sbin/daemon" -command_args="-P $pidfile -c /usr/local/bin/uchiwa -c ${uchiwa_config} -p ${uchiwa_publicdir} >> ${uchiwa_logfile}" +command_args="-P $pidfile -c /usr/local/bin/uchiwa -c ${uchiwa_config} -p ${uchiwa_publicdir} >> ${uchiwa_logfile} 2>>1" unset uchiwa_flags start_precmd="install -d -o ${uchiwa_user} -g ${uchiwa_group} ${pidfile%/*}" ``` username_0: the diff breaks the log file which is supposed to be JSON lines. Status: Issue closed
domoticz/domoticz-android
503911458
Title: Change the login check api method Question: username_0: Requested on slack: Gizmocuz 3:39 PM I am changing the way the logincheck is being called. This used to be a 'GET' command, but this is being changed to a POST... Is it possible for the app developers to implement this as well ? Currently the old method is still available... As example, check the new code in the web login page or the settings page Thanks in advance!!<issue_closed> Status: Issue closed
clay/clay-kiln
272591475
Title: Bug: Glitch in editing multi-graf clay-paragraph component Question: username_0: A user reported this problem. A paragraph component had this text data: ``` I felt powerless through it all​. B​eing young and pregnant ​meant a lot of societal judgment. I wasn’t in high school, but I wasn’t married, and the feeling was very much: If you’re young and pregnant, you’re stupid. That’s how we look at young parents. I don’t advocate for teenagers getting pregnant, but using teen pregnancy to determine how well a society is doing — that’s evidence of the stigma. I just wish I’d had someone to cheerlead for me, to tell me that I was strong enough to do it. Instead, I was lonely. Still, I never thought I made the wrong decision.<br /><br /><br /><strong>On the upside of a car accident. </strong>My maternity leave was only six weeks, so I had to go back to work when my daughter was 5 weeks old. With my commute I’d be gone 10, 12 hours — the whole breastfeeding thing didn’t last very long. ​So not only did I feel guilty about being away all day, but already I hadn’t been able to fulfill what I considered an important “good mom” goal. ``` When attempting to edit the graf starting with `On the upside of a car accident...` by clicking it, that graf would disappear from the UI, and not be editable, though you could still edit the _first_ graf. When clicking out of the component, the second graf would reappear. Answers: username_1: Hmm. Where were they pasting from? Those three `<br>` tags should have triggered new paragraph creation, since the paragraphs are using `multi-component` wysiwyg username_0: @username_1 [This Google Doc](https://docs.google.com/document/d/1uWgzRv5P0quTIssCOJB6eU2s2b5vlf3Z7ANYeaAqz3M/edit?usp=sharing) username_1: That doc pastes correctly for me in Chrome on Kiln 5.0.1 ![screen shot 2017-11-09 at 11 23 49 am](https://user-images.githubusercontent.com/447522/32616560-9b800b4e-c540-11e7-9247-fa1197fbb25b.png) Status: Issue closed username_0: @username_1 I'm still seeing this on Kiln 5.0.2. Here are exact steps to reproduce: 1. Create a new article. 2. Copy and paste complete Google Doc text into first clay-paragraph component. 3. Mouse over to the graf starting with "On moving forward with relationship." (Notice that it is within the same clay-paragraph component as the above graph.) 4. Click that graf, edit it in any way, and then click out. 5. Click the graf again, and it disappears. It's impossible to edit. Click out, and it reappears. username_1: Ah, thanks for the detailed steps. It looks like the issue is that the line breaks are interspersed with formatting, in a weird way. ![](http://up.keats.me/3a3m3R1F2U1F/Screen%20Shot%202017-11-09%20at%201.18.26%20PM.png) username_1: I can't really see a way for us to handle this, since that's technically valid code. Got any ideas? username_2: Won't fix, will re-open if this happens again. Status: Issue closed username_1: A user reported this problem. A paragraph component had this text data: ` I felt powerless through it all​. B​eing young and pregnant ​meant a lot of societal judgment. I wasn’t in high school, but I wasn’t married, and the feeling was very much: If you’re young and pregnant, you’re stupid. That’s how we look at young parents. I don’t advocate for teenagers getting pregnant, but using teen pregnancy to determine how well a society is doing — that’s evidence of the stigma. I just wish I’d had someone to cheerlead for me, to tell me that I was strong enough to do it. Instead, I was lonely. Still, I never thought I made the wrong decision.<br /><br /><br /><strong>On the upside of a car accident. </strong>My maternity leave was only six weeks, so I had to go back to work when my daughter was 5 weeks old. With my commute I’d be gone 10, 12 hours — the whole breastfeeding thing didn’t last very long. ​So not only did I feel guilty about being away all day, but already I hadn’t been able to fulfill what I considered an important “good mom” goal. ` When attempting to edit the graf starting with `On the upside of a car accident...` by clicking it, that graf would disappear from the UI, and not be editable, though you could still edit the _first_ graf. When clicking out of the component, the second graf would reappear. The user said this graf was likely pasted in from a Google Doc. username_1: Looks like this isn't a one-off. @username_2 @amycheng any ideas how we should handle it? username_2: Can't figure out a solution for this, but it's infrequent enough that we'll fix it manually when it comes up. Status: Issue closed
ros-industrial/staubli_experimental
145138953
Title: gazebo: give 'remap' launch arg a more descriptive name Question: username_0: The `*_gazebo.launch` files currently expose a `remap` argument that, when `true`, remap the default topics from the `gazebo_ros` plugin (and friends) to the ROS-Industrial driver spec compliant ones. `remap` is too generic a name, and should be changed into something which more clearly conveys its purpose / effect. Answers: username_0: Launch files in `rx160_gazebo` and the to be introduced `tx90_gazebo` packages (see #6) are affected. username_1: Good idea. Besides renaming "remap" to something more meaningful, I think using the ROS-Industrial driver spec compliant topics by default would be a good idea. Since this is a ROS-Industrial repo, it makes a lot of sense to work with the industrial convention in gazebo by default. username_1: What about extending the argument "remap" to something more meaningful like "ros_remapping", "ros_remap", "ros_i_remapping" or "ros_i_remap" and set the default value to true. username_1: So if I did a PR for this, would you consider merging it? Or is there a reason not to do these changes?
dotnet/runtime
908522355
Title: WASM JS->C# byte[] transfer doesn't work Question: username_0: On JS, I'm using `js_typed_array_to_array(Uint8Array)` to get the pointer to send to C# as a method parameter. Expected: C# receives a byte[] Actual: C# receives a int ptr / `m_value` This may be a lower priority issue as there's a potential work around: - JS->C# notify byte array available - C#->JS InvokeUnmarshalled<byte[]>("someFunction") - The byte array is returned. Answers: username_0: cc/ @SteveSandersonMS @username_1 Status: Issue closed username_1: @username_0 I already filed an issue for this. I'm going to close this one. username_1: This is a dupe of https://github.com/dotnet/runtime/issues/53378
aptible/docker-postgresql
403351263
Title: Documentation of available extensions Question: username_0: Is this documented anywhere? I need to know if I can use `uuid-ossp` on the default image. I can't find the string anywhere in the entire repo. Status: Issue closed Answers: username_1: The default images use the extensions that Postgres itself ships with — that includes `uuid-ossp` . That said, if you'd like to know _exactly_ what is available, you can run: `SELECT * FROM pg_available_extensions;` through e.g. a DB tunnel and get the full list.
molashaonian/article
314877862
Title: Java 调用 groovy 脚本文件,groovy 访问 MongoDB Question: username_0: groovy 访问 MongoDB 示例: shell.groovy ``` package db import com.gmongo.GMongoClient import com.mongodb.BasicDBObject import com.mongodb.MongoCredential import com.mongodb.ServerAddress /** * 本地无密 mongodb 数据库连接 */ def connect() { GMongoClient client = new GMongoClient(new ServerAddress('127.0.0.1',50513)) return client } /** * SSH mongodb 数据库连接 */ def connectSSH() { //以下这两行是针对包含用户名和密码配置的数据库的。 MongoCredential credentials = MongoCredential.createMongoCRCredential('root','xy', 'pass' as char[]) //MongoClientOptions options = MongoClientOptions.builder().connectTimeout(1000) //创建一个Client连接,如果是认证的则使用下面的这一行 GMongoClient client = new GMongoClient(new ServerAddress('10.101.114.108',22), [credentials]) return client } def topics(map) { def DB = connect().getDB('xy') /*def sl = map['$gte'] def el = map['$lte']*/ //查询条件 BasicDBObject object = new BasicDBObject('timestamp',new BasicDBObject(map)) println(object.toString()) println(DB.getCollection('topics').count(object)) //println(DB.getCollection('topics').count()) } def topic_tip(map) { def DB = connect().getDB('xy') //查询条件 BasicDBObject object1 = new BasicDBObject('$match',new BasicDBObject(map)) BasicDBObject object2 = new BasicDBObject('$group',new BasicDBObject('_id',null).append('num_tutorial',new BasicDBObject('$sum','$kd_money'))) //聚合查询 //db.getCollection('topic_tip').aggregate([{$match:{ 'timestamp' : { '$gte' : 1376065109781 , '$lte' : 1576065109781}}},{$group : {_id : null, num_tutorial : {$sum : '$kd_money'}}}]) println(DB.getCollection('topic_tip').aggregate(object1,object2)) [Truncated] Map<String,Long> params = new HashMap<>(); params.put("$gte",start.getTime()); params.put("$lte",end.getTime()); // topics 是 shell.groovy 中的方法名,params 是传给 topics 的参数,执行下面语句完成 topics 方法脚本的调用 object.invokeMethod("topics",params); } /*GroovyScriptEngine engine = new GroovyScriptEngine(Shell.class.getClassLoader().getResource("db").getPath()); Binding binding = new Binding(); binding.setVariable("language","Groovy"); engine.run("studenttopicdata.groovy",binding);*/ } catch (Exception e) { e.printStackTrace(); System.out.println("Exception e = " + e.toString()); } } } ```
rcgsheffield/sheffield_hpc
397476339
Title: No RTD builds have been triggered for ~3 weeks... Question: username_0: ...even though PRs have been merged: https://readthedocs.org/projects/iceberg/builds/ We currently rely on a [GitHub Service](https://developer.github.com/changes/2018-04-25-github-services-deprecation/) for the GitHub<->ReadTheDocs integration, which seems to be silently failing. I've enabled [GitHub Webhook](https://developer.github.com/webhooks/) for triggering RTD builds as per https://docs.readthedocs.io/en/latest/webhooks.html to see if this helps; if so, then the RTD Service for this GitHub repo can be deleted. This is a sensible move anyway as GitHub Services are being deprecated across the board (see 1st link above). Status: Issue closed Answers: username_0: RTD building again. RTD Service removed from this GitHub repo now that RTD webhook has been enabled.
thoughtworks/talisman
409332804
Title: Regression caused by removing support for .talismanignore Question: username_0: Removing support for .talismanignore means that new users are now seeing errors on files which old users don't. This means the tool no longer provides consistent behaviour for different users. There is no migration path to .talismanrc as it does not support the behaviour of .talismanignore. Answers: username_1: @username_0 You are right. We hoped to tackle this with a more verbose output explaining the steps to perform. We chose to allow users to see the output on the console itself which can be copied-pasted to .talismanrc directly, if all looks right. This will be a one-time effort, and things can move smoothly from there on. Do you still see this to be a blocking problem? Status: Issue closed
CLOSER-Cohorts/archivist
1186270677
Title: REACT: Some instruments are not loading the view page (including map) Question: username_0: There are several instruments on the live and staging which are not loading the document view or the map view. The construct view is loading. This doesn't seem to be related to whether the instrument is signed off or not, but this shouldn't effect whether the doc view is visible anyway. Examples include https://closer-archivist-staging.herokuapp.com/instruments/alspac_96_spq/ https://closer-archivist-staging.herokuapp.com/instruments/alspac_00_msdh/
selectize/selectize.js
215382942
Title: Broken table markup in usage doc Question: username_0: https://github.com/selectize/selectize.js/blob/master/docs/api.md API doc also broken Answers: username_1: I noticed that, too. But I think the documentation itself is not broken, but Github seems to have changed something again. If I paste the RAW into Atom and let it render the Markdown file, it works fine ... username_2: see #1265 Status: Issue closed username_0: https://github.com/selectize/selectize.js/blob/master/docs/api.md API doc also broken
bobmagicii/vscode-auto-fold
309151383
Title: Feature Requesr: Fold regions only instead of levels Question: username_0: It would be nice to have a setting to fold only specified regions in a file. Given that recent changes made in VSCode (https://code.visualstudio.com/updates/v1_17#_folding-regions) now allow specifying custom regions on a per-language basis, it would be great to see this integrated into auto-fold. The general idea is instead setting the preferred level of folding, one can set a bool flag of `foldRegions` instead which would only fold regions in a file. Answers: username_1: not a fan of the comments super cluttering all the code blocks myself, but i'll look into it. there is another fix i need to get in too. username_0: This would've been straightforward if `Fold All Regions` worked, doesn't seem to be so I've logged an issue on vscode. username_1: if that get that working i will mod it so autofold.default, autofold.types[].level, and, and vscode-fold can accept the string "region" username_2: If I could tell this extension to automatically fold javascript imports, that'd be really great! username_3: @username_2 I realize this is quite late after you originally posted - but vscode seems has this setting natively now. If you search in settings "Editor Fold" there is _Folding Imports By Default_ (Note: for myself, this only works when _Folding Strategy_ is set to **auto** not indentation)
backend-br/vagas
788433743
Title: [remoto] Analista Desenvolvedor(a) Oracle Sênior @ Supero Question: username_0: Sobre nós: Na Supero somos movidos por desafios, valorizamos a alta performance e quem pega junto com a equipe pra fazer acontecer! Estamos há 17 anos no mercado, realizando projetos personalizados de acordo com a necessidade de cada cliente. Nosso grande diferencial é trabalhar sempre com as melhores tecnologias e um elevado nível de resolutividade. Buscamos um perfil #Supers para agregar ao time. Você busca evolução contínua? Está a procura de novos desafios? Então confere essa oportunidade! O que você irá fazer: Aqui você vai atuar em um grande player multinacional, na área de sustentação aos processos de negócio na comercialização de energia elétrica. Buscamos um profissional autodidata, pró-ativo, organizado, focado, com visão de dono e bom relacionamento interpessoal! Para isso, você precisa de: Conhecimento em Oracle SQL e PL/SQL, Oracle Forms e Reports; Experiência com desenvolvimento, análise de sistemas, análise de negócio e arquitetura de sofware. Será um diferencial: Conhecimento em testes unitários e automatizados; Conhecimento em padrões de desenvolvimento SOLID e GoF. O que oferecemos: Ginástica Laboral; Horários flexíveis; Convênios e descontos; Workchopps; Treinamentos; No dress code; Programa indique um amigo; Cultura que valoriza feedbacks e a diversidade; Oportunidades profissionais a todo momento! Eai? Quer ser um #Supers? Então corre e se inscreve ;) Candidate-se em: http://bit.ly/3qtICEu #### Nível - Sênior<issue_closed><issue_closed> Status: Issue closed
osum-vertnet/osum-mammals
187553151
Title: Monthly VertNet data use report for 2016-10, resource osum_mammals Question: username_0: Your monthly VertNet data use report is ready! You can see the HTML rendered version of the reports with this link: http://tools-usagestats.vertnet-portal.appspot.com/reports/78dcfcbd-03c4-4d47-8618-830aac6f2ee5/201610/ Raw text and JSON-formatted versions of the report are also available for download from this link. In addition, a copy of the text version has been uploaded to your GitHub repository, under the "Reports" folder. Also, a full list of all reports can be accessed here: http://tools-usagestats.vertnet-portal.appspot.com/reports/78dcfcbd-03c4-4d47-8618-830aac6f2ee5/ You can find more information on the reporting system, along with an explanation of each metric, here: http://www.vertnet.org/resources/usagereportingguide.html Please post any comments or questions to: http://www.vertnet.org/feedback/contact.html Thank you for being a part of VertNet.
firebase/firebase-tools
293514174
Title: Are these npm errors normal? Seems I can't start emulator. Question: username_0: ### Version info $ uname -a Linux home 4.13.0-26-generic #29~16.04.2-Ubuntu SMP Tue Jan 9 22:00:44 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux $ node -v v8.9.4 $ npm -v 5.6.0 $ firebase --version 3.17.4 ### Steps to reproduce Install firebase-tools with: sudo npm i -g firebase-tools ### Expected behavior Get no errors, start emulator successfully. ### Actual behavior $ sudo npm i -g firebase-tools npm WARN deprecated [email protected]: Use uuid module instead /usr/bin/firebase -> /usr/lib/node_modules/firebase-tools/bin/firebase [email protected] install /usr/lib/node_modules/firebase-tools/node_modules/grpc node-pre-gyp install --fallback-to-build --library=static_library node-pre-gyp ERR! Tried to download(undefined): https://storage.googleapis.com/grpc-precompiled-binaries/node/grpc/v1.4.1/node-v57-linux-x64.tar.gz node-pre-gyp ERR! Pre-built binaries not found for [email protected] and [email protected] (node-v57 ABI) (falling back to source compile with node-gyp) gyp ERR! configure error gyp ERR! stack Error: EACCES: permission denied, mkdir '/usr/lib/node_modules/firebase-tools/node_modules/grpc/build' gyp ERR! System Linux 4.13.0-26-generic gyp ERR! command "/usr/bin/node" "/usr/lib/node_modules/npm/node_modules/node-gyp/bin/node-gyp.js" "configure" "--fallback-to-build" "--library=static_library" "--module=/usr/lib/node_modules/firebase-tools/node_modules/grpc/src/node/extension_binary/grpc_node.node" "--module_name=grpc_node" "--module_path=/usr/lib/node_modules/firebase-tools/node_modules/grpc/src/node/extension_binary" gyp ERR! cwd /usr/lib/node_modules/firebase-tools/node_modules/grpc gyp ERR! node -v v8.9.4 gyp ERR! node-gyp -v v3.6.2 gyp ERR! not ok node-pre-gyp ERR! build error node-pre-gyp ERR! stack Error: Failed to execute '/usr/bin/node /usr/lib/node_modules/npm/node_modules/node-gyp/bin/node-gyp.js configure --fallback-to-build --library=static_library --module=/usr/lib/node_modules/firebase-tools/node_modules/grpc/src/node/extension_binary/grpc_node.node --module_name=grpc_node --module_path=/usr/lib/node_modules/firebase-tools/node_modules/grpc/src/node/extension_binary' (1) node-pre-gyp ERR! stack at ChildProcess.<anonymous> (/usr/lib/node_modules/firebase-tools/node_modules/grpc/node_modules/node-pre-gyp/lib/util/compile.js:83:29) node-pre-gyp ERR! stack at emitTwo (events.js:126:13) node-pre-gyp ERR! stack at ChildProcess.emit (events.js:214:7) node-pre-gyp ERR! stack at maybeClose (internal/child_process.js:925:16) node-pre-gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:209:5) node-pre-gyp ERR! System Linux 4.13.0-26-generic node-pre-gyp ERR! command "/usr/bin/node" "/usr/lib/node_modules/firebase-tools/node_modules/grpc/node_modules/.bin/node-pre-gyp" "install" "--fallback-to-build" "--library=static_library" node-pre-gyp ERR! cwd /usr/lib/node_modules/firebase-tools/node_modules/grpc node-pre-gyp ERR! node -v v8.9.4 node-pre-gyp ERR! node-pre-gyp -v v0.6.36 node-pre-gyp ERR! not ok Failed to execute '/usr/bin/node /usr/lib/node_modules/npm/node_modules/node-gyp/bin/node-gyp.js configure --fallback-to-build --library=static_library --module=/usr/lib/node_modules/firebase-tools/node_modules/grpc/src/node/extension_binary/grpc_node.node --module_name=grpc_node --module_path=/usr/lib/node_modules/firebase-tools/node_modules/grpc/src/node/extension_binary' (1) npm WARN optional SKIPPING OPTIONAL DEPENDENCY: [email protected] (node_modules/firebase-tools/node_modules/grpc): npm WARN optional SKIPPING OPTIONAL DEPENDENCY: [email protected] install: `node-pre-gyp install --fallback-to-build --library=static_library` npm WARN optional SKIPPING OPTIONAL DEPENDENCY: Exit status 1 [email protected] added 502 packages and updated 1 package in 83.934s Answers: username_1: GRPC installation issues are consolidated at https://github.com/grpc/grpc-node/issues/121 Status: Issue closed username_1: Also, I recommend installing without "sudo", that may be causing the issue, and if you are unable to install without sudo, then you should fix your npm permissions: https://docs.npmjs.com/getting-started/fixing-npm-permissions
liuchengxu/vim-clap
742649143
Title: Cannot create new file with `Clap filer` Question: username_0: **Environment (please complete the following information):** - OS: macOS - (Neo)Vim version: VIM - Vi IMproved 8.2 (2019 Dec 12, compiled Nov 4 2020 11:27:02) macOS version Included patches: 1-1950 Compiled by Homebrew - vim-clap version: 45a91bf706e9dcf85172c126bbac8964e209a4ae - Have you reproduced with a minimal vimrc: Yes - Have you updated to the latest plugin version: Yes - Have you upgraded to/compiled the latest Rust binary: Yes - Tried run `:Clap install-binary!` in vim **Describe the bug** A clear and concise description of what the bug is. **Clap debug** <!-- Paste the output of :Clap debug here, or try :Clap debug+. --> ``` has cargo: 1 has maple: /Users/phong/.vim/plugged/vim-clap/target/release/maple maple info: version 0.1.21, built for x86_64-apple-darwin by rustc 1.47.0 (18bf6b4f0 2020-10-07). has +python3: 1 has py dynamic module: 1 has ctags: ctags with JSON output support Current FileType: Third Party Providers: [] Global Options: let g:clap#autoload_dir = '/Users/phong/.vim/plugged/vim-clap/autoload' let g:clap#popup#display = {'shrink': function('51'), 'open': function('<SNR>27_create_display'), 'width': 69, 'shrink_if_undersize': function('50')} let g:clap#popup#preview = {'line_count': function('<SNR>26__line_count'), 'show': function('52'), 'hide': function('53'), 'bufnr': 6, 'add_highlight': function('54'), 'get_lines': function('<SNR>26__get_lines'), 'getbufvar': function('<SNR>26__getbufvar'), 'setbufvar_batch': function('<SNR>26__setbufvar_batch'), 'setbufvar': function('<SNR>26__setbufvar'), 'winid': 1008, 'win_is_valid': function('<SNR>26__win_is_valid'), 'goto_win': function('<SNR>26__goto_win'), 'set_syntax': function('55')} let g:clap#provider_alias = {'gfiles': 'git_files', 'hist:': 'command_history', 'hist/': 'search_history'} let g:clap_background_shadow_blend = 50 let g:clap_disable_bottom_top = 0 let g:clap_disable_matches_indicator = v:false let g:clap_disable_optional_async = v:false let g:clap_disable_run_rooter = v:false let g:clap_enable_background_shadow = v:true let g:clap_enable_debug = v:false let g:clap_enable_icon = 0 let g:clap_forerunner_status_sign = {'done': '•', 'running': '!', 'using_cache': '*'} let g:clap_indicator_winid = 1009 let g:clap_insert_mode_only = v:false let g:clap_multi_selection_warning_silent = 0 let g:clap_no_matches_msg = 'NO MATCHES FOUND' let g:clap_open_action = {'ctrl-v': 'vsplit', 'ctrl-x': 'split', 'ctrl-t': 'tab split'} let g:clap_popup_border = 'rounded' let g:clap_preview_size = 5 let g:clap_providers_relaunch_code = '@@' let g:clap_search_box_border_style = 'nil' let g:clap_search_box_border_symbols = {'nil': ['', ''], 'curve': ['', ''], 'arrow': ['', '']} let g:clap_spinner_winid = 1011 Provider Variables: [] ``` **To Reproduce** Steps to reproduce the behavior: [Truncated] filetype plugin indent on ``` 2. Start (neo)vim with command: `vim -u min.vim` 3. Type `Clap filer` and input inexistent path and press Enter `<CR>` 4. See error: `No such file or directory (os error 2)` **Expected behavior** I can create new file on pressing Enter **Screenshots** <img width="1792" alt="Screen Shot 2020-11-14 at 12 35 48 AM" src="https://user-images.githubusercontent.com/438791/99102929-59edfa80-2611-11eb-8508-ffc4b8d1a111.png"> <img width="1792" alt="Screen Shot 2020-11-14 at 12 41 58 AM" src="https://user-images.githubusercontent.com/438791/99103489-37a8ac80-2612-11eb-9c08-f0322f99e255.png"> Could you help check this issue? Thanks. Answers: username_1: Can reproduce, will take a more detailed look later. username_2: @username_0 Can you see if this patch works? ```diff diff --git a/autoload/clap/provider/filer.vim b/autoload/clap/provider/filer.vim index cdc06a8..caf8811 100644 --- a/autoload/clap/provider/filer.vim +++ b/autoload/clap/provider/filer.vim @@ -218,7 +218,7 @@ function! s:cr_action() abort return endif - if exists('g:__clap_has_no_matches') && g:__clap_has_no_matches + if g:clap.display.line_count() == 1 && g:clap.display.get_lines()[0] =~# s:CREATE_FILE " Create file if it doesn't exist stopinsert call clap#handler#sink() ``` Since we injected the `query` line into the candidates, `g:__clap_has_no_matches` will never be true then. @username_1 https://github.com/username_2/vim-clap/blob/5d2a04ca8c33f4031cd89a012663ca34ae854b8f/autoload/clap/provider/filer.vim#L143-L144 username_0: Thanks @username_2. It works for these case: - creating new file from current folder - creating new file from non-empty folder <img width="713" alt="Screen Shot 2020-11-14 at 10 32 07 AM" src="https://user-images.githubusercontent.com/438791/99138671-df4eca80-2664-11eb-9939-5e16d0ebb2be.png"> <img width="386" alt="Screen Shot 2020-11-14 at 10 32 15 AM" src="https://user-images.githubusercontent.com/438791/99138673-e37ae800-2664-11eb-9790-a47abda60d10.png"> <img width="667" alt="Screen Shot 2020-11-14 at 10 32 47 AM" src="https://user-images.githubusercontent.com/438791/99138676-ebd32300-2664-11eb-8ac8-d488a968f012.png"> <img width="469" alt="Screen Shot 2020-11-14 at 10 32 53 AM" src="https://user-images.githubusercontent.com/438791/99138677-efff4080-2664-11eb-88f3-aa5ce734f977.png"> It does not work for the case - empty folder. Please check screenshots: <img width="651" alt="Screen Shot 2020-11-14 at 10 33 23 AM" src="https://user-images.githubusercontent.com/438791/99138688-fd1c2f80-2664-11eb-854b-d7eb916abf6c.png"> <img width="457" alt="Screen Shot 2020-11-14 at 10 33 29 AM" src="https://user-images.githubusercontent.com/438791/99138686-fbeb0280-2664-11eb-8492-b57923d3b327.png"> Status: Issue closed username_1: Indeed, I've applied your patch in #590. @username_0 Please update to latest master and reopen if it happens again :] username_2: @username_0 The issue of creating a new file under empty folder has been resovled in https://github.com/username_2/vim-clap/commit/b2fe93fbee23c3dba24439e0e0bf8c5b77eb5447, please have a try. username_0: Thanks @username_1. I fixed problem for creating new file. @username_2 I tried to test two cases: 1. I typed `emptydir/new.txt`, and pressed Enter, it created `new.txt` correctly ![Screen Shot 2020-11-16 at 10 42 21 AM](https://user-images.githubusercontent.com/438791/99210626-e4448300-27f8-11eb-88f3-77fc611fc7a4.png) ![Screen Shot 2020-11-16 at 10 42 35 AM](https://user-images.githubusercontent.com/438791/99210642-ed355480-27f8-11eb-94aa-7c54332cc73c.png) 2. I typed `emp`, and press `TAB` to go to `emptydir`, then I typed `new.txt`, it did not show option to create new file. I'm not sure this case is a bug or not. ![Screen Shot 2020-11-16 at 10 44 14 AM](https://user-images.githubusercontent.com/438791/99210648-f2929f00-27f8-11eb-89c7-d8ebf46e59e2.png) ![Screen Shot 2020-11-16 at 10 44 18 AM](https://user-images.githubusercontent.com/438791/99210654-f58d8f80-27f8-11eb-91b4-a37486799160.png) ![Screen Shot 2020-11-16 at 10 44 25 AM](https://user-images.githubusercontent.com/438791/99210658-f8888000-27f8-11eb-961b-902b8a9095d6.png) username_2: @username_0 You can still create the new file as expected, but the prompt message `Directory is empty` can definitely be improved. username_0: @username_2 Just tried again, press `ENTER` on `new.txt` in last screenshot. It works. Thank you.
dart-lang/co19
983703632
Title: Failures on [co19] Roll co19 to cc4df792d0f02f15bfe04911c1f856b90cb7cbf4 Question: username_0: There are new test failures on [[co19] Roll co19 to cc4df792d0f02f15bfe04911c1f856b90cb7cbf4](https://dart-review.googlesource.com/c/sdk/+/211903). The tests ``` co19/LanguageFeatures/Constructor-tear-offs/ambiguities_A15_t01 MissingCompileTimeError (expected Pass) co19/LanguageFeatures/Constructor-tear-offs/goal_A01_t03 RuntimeError (expected Pass) co19/LanguageFeatures/Constructor-tear-offs/named_constructor_A02_t02 RuntimeError (expected Pass) co19/LanguageFeatures/Constructor-tear-offs/named_constructor_A03_t02 RuntimeError (expected Pass) co19/LanguageFeatures/Constructor-tear-offs/named_constructor_A03_t03 RuntimeError (expected Pass) co19/LanguageFeatures/Constructor-tear-offs/named_constructor_A07_t07 RuntimeError (expected Pass) co19/LanguageFeatures/Constructor-tear-offs/named_constructor_A08_t01 RuntimeError (expected Pass) co19/LanguageFeatures/Constructor-tear-offs/named_constructor_A08_t02 RuntimeError (expected Pass) co19/LanguageFeatures/Constructor-tear-offs/summary_A02_t01 RuntimeError (expected Pass) co19/LanguageFeatures/Constructor-tear-offs/syntax_t01/none RuntimeError (expected Pass) co19/LanguageFeatures/regression/33786_t01 RuntimeError (expected Pass) co19/LanguageFeatures/regression/33786_t01 CompileTimeError (expected Pass) co19/LanguageFeatures/regression/35114_35115_t01 CompileTimeError (expected Pass) ``` are failing on configurations ``` analyzer-asserts-strong-linux analyzer-asserts-weak-linux cfe-strong-linux cfe-weak-linux dart2js-hostasserts-strong-linux-x64-chrome dart2js-hostasserts-weak-linux-x64-chrome dartdevk-strong-linux-release-chrome dartdevk-weak-linux-release-chrome dartk-strong-linux-debug-x64 dartk-strong-linux-release-x64 dartk-weak-asserts-linux-debug-x64 dartk-weak-asserts-linux-release-x64 dartkp-strong-linux-release-x64 dartkp-weak-asserts-linux-release-x64 ``` Answers: username_1: Triaged dartdevk... |   -- | -- CompileTimeError -> RuntimeError (expected Pass) |   co19/LanguageFeatures/Constructor-tear-offs/goal_A01_t03 | UnimplementedError co19/LanguageFeatures/Constructor-tear-offs/named_constructor_A03_t02 | UnimplementedError co19/LanguageFeatures/Constructor-tear-offs/named_constructor_A03_t03 | UnimplementedError co19/LanguageFeatures/Constructor-tear-offs/named_constructor_A07_t07 | UnimplementedError co19/LanguageFeatures/Constructor-tear-offs/named_constructor_A08_t01 | UnimplementedError co19/LanguageFeatures/Constructor-tear-offs/named_constructor_A08_t02 | UnimplementedError co19/LanguageFeatures/Constructor-tear-offs/summary_A02_t01 | UnimplementedError co19/LanguageFeatures/Constructor-tear-offs/syntax_t01/none | UnimplementedError new test -> RuntimeError (expected Pass) |   co19/LanguageFeatures/regression/33786_t01 | Issue https://github.com/dart-lang/sdk/issues/33786 dartk... |   new test -> CompileTimeError (expected Pass) |   co19/LanguageFeatures/regression/33786_t01 | Issue https://github.com/dart-lang/sdk/issues/33786   |   analyzer...cfe...dart2js...dartdevk...dartk...dartkp... |   new test -> CompileTimeError (expected Pass) |   co19/LanguageFeatures/regression/35114_35115_t01 | Issues https://github.com/dart-lang/sdk/issues/35114, https://github.com/dart-lang/sdk/issues/35115 cfe... |   Pass -> MissingCompileTimeError (expected Pass) |   co19/LanguageFeatures/Constructor-tear-offs/ambiguities_A15_t01 | Issue https://github.com/dart-lang/sdk/issues/47060 Status: Issue closed
Brandply/publicissues
132408160
Title: IOS / Android My Favorite Offers redeem via instore bug Question: username_0: Platform: IOS, Android Browser: Safari, Chrome URL of page: https://brandply.com Organization or Member Login User: archer Password: <PASSWORD> Description of Problem: Many offers displayed under My Favorite Offers are not showing link to redeem when clicking on Redeem via Instore, it can appear if clicking on Redeem via Instore Print and switching on Redeem via Instore again <img width="199" alt="brandply0" src="https://cloud.githubusercontent.com/assets/13362917/12916177/b83a44dc-cf30-11e5-89a2-80b659b4b3a1.png"> ======= <img width="201" alt="brandply" src="https://cloud.githubusercontent.com/assets/13362917/12916182/bdd7306c-cf30-11e5-91a8-a1902b02ea7f.png"> ======= <img width="196" alt="brandply2" src="https://cloud.githubusercontent.com/assets/13362917/12916186/c17d1cea-cf30-11e5-94d3-5b2e8b3394ae.png"> Here are the steps that reproduce the problem: 1) Go to https://brandply.com in your browser 2) Enter ZipCode when prompted: 84062 3) Log into Brandply site u:archer p:1234567 4) Click on My Favorite Offers 5) Choose some of offers with Redeem via Instore and Redeem via Instore Print and click on Redeem via Instore Status: Issue closed Answers: username_1: This issue was moved to Brandply/brandply#359
andmorefine/since-co
842741626
Title: お問い合わせ | andmorefine Question: username_0: Morning Defrost frozen foods in minutes safely and naturally with our THAW KING™. 50% OFF for the next 24 Hours ONLY + FREE Worldwide Shipping for a LIMITED time Buy now: thawking.online Cheers, Brendan お問い合わせ | username_0
MaryKattyVO/lim20181-Track-FE-markdown-list
356610041
Title: Feedback Técnico Question: username_0: ## Cosas que hiciste bien. - Recorrer archivo, extraer links y texto de links en un archivo markdown. - Recorrer carpetas, extraer links y texto de links en un archivo markdown. - Determinar el status de respuesta HTTP de los links encontrados. - Sacar el total de links que se encuentran en la ruta escrita, total de links únicos y total links rotos. - Trabajaste con issues y milestone. ## Cosas por mejorar. - La función mdLinks debería retornar una promesa donde tú imprimas el resultado de esa promesa. https://github.com/username_1/lim20181-Track-FE-markdown-list/blob/281799b40928cdb4421958a18b1b2bcd16337b79/cli.js#L24 donde esto debería ser así: ` mdlinks(file, options).then(response => console.log(response))` - Cada iteración que haces para sacar los resultados de las opciones escritas en CLI (`--stats , --validate`) lo haces con un `console.log() cuando deberías imprimirlo en el resultado de la promesa en la función mdLinks` - Cuando no escribes ninguna opción en CLI debería retornar *link* *texto* y *ruta*. - Trata de mejorar tus consultas según las opciones, que solo retornen lo que el usuario requiere ![image](https://user-images.githubusercontent.com/15807118/45002419-f16c5300-af9b-11e8-98e2-2dfc02ba018b.png) este ejemplo de tu readme lo describe. ## Cosas por hacer - Te recomiendo que vayas actualizando tu readme describiendo el uso de tu librería. - Subas tu package en npm. - Empieza a hacer tus test desde ahora para que puedas corregir a tiempo tus funciones. Guíate de aquí: https://jestjs.io/docs/en/asynchronous . - En tu proyecto tienes carpetas y archivos que no deberían estar agrégalos en tu `.gitignore` ![image](https://user-images.githubusercontent.com/15807118/45002567-a4897c00-af9d-11e8-8255-3fa4dc27fbd0.png) Mucha suerte! Answers: username_1: Gracias Anaflavia por tu feedback, voy a mejorarlo. Status: Issue closed
keptn/community
1082355566
Title: Add @oleg-nenashev to keptn-contrib Question: username_0: It would be great to be added to https://github.com/keptn-contrib as organization member so that my reviews could be requested and so that I could probably help with GitHub organization management. ### GitHub Username @username_0 ### Membership level Membership level according to https://github.com/keptn/community/blob/master/COMMUNITY_MEMBERSHIP.md `member` ### Requirements (for member) - [x] I have reviewed the community membership guidelines (https://github.com/keptn/community/blob/master/COMMUNITY_MEMBERSHIP.md) - [x] I have enabled 2FA on my GitHub account. See https://github.com/settings/security - [x] I have subscribed to the [Keptn Slack channel](http://slack.keptn.sh/) - [x] I am actively contributing to 1 or more Keptn subprojects - [x] I have two sponsors that meet the sponsor requirements listed in the community membership guidelines. Among other requirements, sponsors must be approvers or maintainers of at least one repository in the organization and not both affiliated with the same company - [x] I have spoken to 2 sponsors (approvers or maintainers) ahead of this application, and they have agreed to sponsor my application - [x] I have filed a PR to add myself as [project member](https://github.com/keptn/keptn/blob/master/MAINTAINERS) and referenced this issue in the PR ### Sponsors - @grabnerandi - @username_2 - @username_1 Answers: username_1: I support username_2: I support username_0: Thanks! Now I need someone to actually add me there username_3: I will handle that. username_0: I believe it is fixed now. Thanks @username_3 ! Status: Issue closed
openshift/origin-web-console
267755848
Title: how can i skip auth on webconsole Question: username_0: i'm doing `oc proxy` which already exposes api on my host without oauth. how can disable auth on origin-web-console? Answers: username_1: The console requires you to be logged in as a user, its always going to redirect to the oauth server if it doesn't find a token in LocalStorage. username_0: thanks @username_1 i will try with hard coding LocalStorage token Status: Issue closed
inaturalist/iNaturalistAndroid
953112765
Title: NullPointerException in LocationDetailsActivity.onOptionsItemSelected Question: username_0: https://console.firebase.google.com/u/2/project/inaturalist-ios/crashlytics/app/android:org.inaturalist.android/issues/a710dee3dda8cc3b3a87b5b5b47d0067 ``` Fatal Exception: java.lang.NullPointerException: Attempt to invoke virtual method 'double java.lang.Double.doubleValue()' on a null object reference at org.inaturalist.android.LocationDetailsActivity.onOptionsItemSelected(LocationDetailsActivity.java:188) at android.app.Activity.onMenuItemSelected(Activity.java:3544) ```
ruuvi/ruuvi.gateway_esp.c
809174674
Title: Gateway is restarted by "assert" after connecting Ethernet Question: username_0: ``` I (34441) ruuvi_gateway: [monitoring_task] free heap: 139172 I (36421) eth: Ethernet Link Up I (36421) eth: Ethernet HW Addr 24:0A:C4:E2:67:AB I (36421) ruuvi_gateway: [sys_evt] Ethernet connection established I (37721) tcpip_adapter: eth ip: 192.168.1.108, mask: 255.255.255.0, gw: 192.168.1.1 I (37731) eth: Ethernet Got IP Address I (37741) eth: ~~~~~~~~~~~ I (37741) eth: ETHIP:192.168.1.108 I (37741) eth: ETHMASK:255.255.255.0 I (37751) eth: ETHGW:192.168.1.1 I (37751) eth: ~~~~~~~~~~~ I (37761) ruuvi_gateway: [sys_evt] Ethernet connected I (37761) wifi_manager: [sys_evt] wifi_manager_stop I (37771) wifi:state: run -> init (0) I (37771) wifi:pm stop, total sleep time: 0 us / 31925162 us I (37781) wifi:new:<3,0>, old:<3,1>, ap:<255,255>, sta:<3,1>, prof:1 I (37781) wifi:flush txq I (37781) wifi:stop sw txq I (37791) wifi:lmac stop hw txq I (37791) wifi:Deinit lldesc rx mblock:10 I (37801) wifi_manager: [sys_evt] wifi_manager_destroy I (37801) TIME: [time_task] Activate SNTP time synchronization I (38151) http: [adv_post_task] HTTP POST: { "data": { ... /home/alexey/dev/esp-idf-v4.0/components/freertos/event_groups.c:498 (xEventGroupClearBits)- assert failed! abort() was called at PC 0x40096ef0 on core 0 ELF file SHA256: c1a3f71b46f2ff0a Backtrace: 0x4009093d:0x3ffd7bc0 0x40090cbd:0x3ffd7be0 0x40096ef0:0x3ffd7c00 0x400eaeb8:0x3ffd7c20 0x400e9d2d:0x3ffd7c40 0x400f4fd3:0x3ffd7c80 0x40093c35:0x3ffd7ca0 Rebooting... ```<issue_closed> Status: Issue closed
danielgindi/Charts
175442114
Title: I can't see the zero(float Num) on ChartYAxis Question: username_0: the lastest verson has bug?? ![f685879e-39c6-446a-b6f3-a52ed7daf428](https://cloud.githubusercontent.com/assets/20593532/18304904/94745e1a-7517-11e6-91c7-a8f4758f60d2.png) Answers: username_1: do check `NSNumberFormatter` APIs Status: Issue closed username_0: thanks a lot!
mrchypark/sendgridr
462361278
Title: How can I use "new line" in content? Question: username_0: Thanks for your package. I have a question. I want to use **new line** in content. But `content("aaa\n\naa)` dosen't work. How can I use? Answers: username_1: Now package only support html style. So please use <br> tag. I'll trying to find out how to use that syntax. Thank you for asking. Status: Issue closed username_1: @all-contributors please add @username_0 for ideas
powerline/powerline
88200226
Title: ERROR:tmus:uptime:Exception while computing segment: 'module' object has no attribute 'BOOT_TIME' Question: username_0: I'm running a fresh version of OS X 10.10.3. After installing powerline, in the bottom right corner I get the following error message: ``` ERROR:tmus:uptime:Exception while computing segment: 'module' object has no attribute 'BOOT_TIME' ``` Does this have something to do with psutil? Answers: username_1: Yes, it means that `psutil` module is missing BOOT_TIME attribute. You need to disable uptime segment then or find `psutil` module that has this attribute if there is such a thing. username_1: It appears that new psutil has `boot_time` function in place of `BOOT_TIME` attribute. username_1: psutil documentation states that old constants are still available, though should produce DeprecationWarning. So it is probably not your problem. username_1: Apparently this is only the case for psutil-2.x. psutil-3.x does not have this attribute. Status: Issue closed username_2: I am seeing this error message as well, since I updated via `pip install powerline-status --upgrade`. Before the upgrade, the `uptime`-module worked without a glitch. ```pip freeze | grep powerline powerline-status==2.1.4 ``` I understand that a fix has been merged 12 days ago, is this supposed to be inside the pip package, or should I rather install from head? username_1: It is currently not in release. username_1: Now it is in release 2.2.
tivadarbocz/spa_backend
174705796
Title: Liquibase Status: Issue closed Question: username_0: With yaml format https://devcenter.heroku.com/articles/running-database-migrations-for-java-apps#using-liquibase http://www.baeldung.com/liquibase-refactor-schema-of-java-app Refactor to xml format Answers: username_0: With yaml format https://devcenter.heroku.com/articles/running-database-migrations-for-java-apps#using-liquibase http://www.baeldung.com/liquibase-refactor-schema-of-java-app Refactor to xml format
thiagocoutinhor/burning-books
595988429
Title: Exibir o Application_Id ao se connectar Question: username_0: Seria interessante a exibição do Application_id ao se conectar o notebook, talvez posterior ao botão de conectado. Isso se mostra util uma vez que é uma atividade corriqueira a checagem da aplicação na spark-ui, checagem de log ou matar o processo , e para isso o application_id é necessário. Answers: username_1: ![](https://github.trello.services/images/mini-trello-icon.png) [Exibir o Application_Id ao se connectar](https://trello.com/c/7rsnnPaG/48-exibir-o-applicationid-ao-se-connectar) Status: Issue closed
metanorma/metanorma-standoc
507758848
Title: bibdata/abstract maintaining blocks Question: username_0: bibdata/abstract is meant to be a formatted string. Documents that have abstracts with Metanorma block structure are moving that structure into bibdata/abstract, which gives syntax errors. We should either: * strip blocks in the bibdata/abstract. and leave just inline makup; or * allow blocks in the syntax of bibdata/abstract. Status: Issue closed Answers: username_0: Addressed by stripping attributes from abstract during parsing; abstracts are xs:any blocks, and they validate fine.
junegunn/fzf
320717704
Title: FZF breaks down when terminal column width more than 181 chars. Question: username_0: <!-- Check all that apply [x] --> - Category - [x] fzf binary - [ ] fzf-tmux script - [ ] Key bindings - [ ] Completion - [ ] Vim - [ ] Neovim - [ ] Etc. - OS - [x] Linux - [ ] Mac OS X - [] Windows - [ ] Windows Subsystem for Linux - [ ] Etc. - Shell - [x] bash - [ ] zsh - [ ] fish I am using Cmder with bash configuration and I realized after ssh to Linux server (CentOS) that after invoking fzf and searching each line is sort of interlaced with empty line and after few key strokes the search breaks apart. Comparing to other terminal utilities which worked fine I found that the issue is with setting terminal width which was originally 188 columns. After changing it to 181 or less (stty cols 181) it worked. Also changing this option in other terminals broke it apart. So is there any hardcoded limit? Could it be fixed? BTW I love this utility :) Answers: username_1: Hi @username_0, could you let us know if this is still an issue? username_0: Hi @username_1 unfortunately it is still an issue. I was able to reproduce it just now. username_2: Do you mean that you manually set the width with `stty cols` command?
RGLab/flowStats
100521431
Title: warpSet: check grouping suggestion Question: username_0: I think grouping could be checked at the beginning of the function. This would avoid getting error after some computation started. Lines 657-659 if(!is.null(grouping)){ if(!grouping %in% names(pData(x))) stop("'", grouping, "' is not a phenoData variable.") could be placed around line 593. Best.<issue_closed> Status: Issue closed
nltk/nltk
194835322
Title: How to use Dictionary as training data Question: username_0: How can i train classifier, so that if i will classify `atms` it should return me `atm fees` for nsf -> overfraft fees food -> restaurant how to solve this problem ? Answers: username_1: @username_0: sorry, this is not a valid issue. You can post to nltk-users or stackoverflow. Status: Issue closed
saltstack/salt
126472728
Title: Add failhard arg to cmd_batch LocalClient Question: username_0: Currently I'm writing a custom runner that calls a LocalClient cmd_batch function to run a command in batches. This custom runner basically deploys code to a server farm and if any of the targetted nodes in the batch fails to deploy the code, I want to stop the deployment and do a rollback. Without the failhard option present in the LocalClient cmd_batch function I can't stop the batch execution upon the failure. Answers: username_1: @username_0, thanks for the feature request. username_0: Thanks @username_1... I'm currently testing if setting the LocalClient mopts with failhard=True will do the trick. Once I have some news I'll post it in here username_0: Nope, it does not change anything... The --failhard for the Salt CLI is present in 'cli/salt.py' while when running the LocalClient cmd_batch, the api calls the 'client/__init__.py' directly that does not handle the failhard option at all. username_0: I found another way to do this after checking that the return sent by cmd_batch is a generator object I can control when to not proceed raising a StopIteration exception.
Klonan/belt_buffer
638236571
Title: Doesn't work on v18 Question: username_0: First off, it's incompatible with v18. Once the stuff in info.json is bumped to 18, the first startup error is this: ![Factorio 0 18 2 6_13_2020 23_03_29](https://user-images.githubusercontent.com/7498356/84579264-b9c74780-adcc-11ea-849b-90fd3a2362de.png) After bumping the `active_energy_usage` to 0.5J, the following error happens: ![Factorio 0 18 2 6_13_2020 23_04_53](https://user-images.githubusercontent.com/7498356/84579284-d6fc1600-adcc-11ea-8094-e3f04815b2e7.png) adding the `icon_mipmaps = 0` doesn't resolve the issue, but it seems to work fine when the buffer-icon.png is rescaled to 64x64. Answers: username_0: I'll consider making a PR that supports all 3 belt types if this ends up working and does what i think/hope it does. username_0: When actually fixed to run, this is what it looks like: ![wtf](https://user-images.githubusercontent.com/7498356/84579571-0744b400-adcf-11ea-9f8d-4fd88c98bb30.png) and does not actually let items through itself at all. even items manually put into it can't get down the belt. username_1: Yea, I don't plan to support 0.18, sorry, you can fork the mod and try to update yourself Status: Issue closed
DatacomRD/dtc-fhir
240825899
Title: JAXB 在處理 base64Binary 類型資料時的問題 Question: username_0: 這是在開發 TWR 時發現的 存放於 `Extension.valueSignature.blob` 欄位,型態為 `Base64Binary`, 實際從 FHIR Server 取回的 xml 中,存放的值為 `/9j/4AAQSkZJRgABAQEAYABgAAD/....` 這樣的 Base64 String 然而,從 Java class 取出來時是 `byte[]` 型態的資料, 必須使用 `Base64.encodeBase64()` 來還原回 xml 原本的值, 推測是 JAXB 在 unmarshal 時,先進行了 `Base64.decodeBase64()` 所導致。 至於究竟是: JAXB 不應該在 unmarshal 時先進行 decodeBase64(@shvoidlee 的論點), 還是 xml 存放了不正確的資料(我的論點),還有待釐清。
DarkoPendragon/discord.js-musicbot-addon
325286633
Title: Prefix longer than one character Question: username_0: It just doesn't work, but when I give only one character it works. Please fix, it's really simple to fix. Answers: username_1: The prefix I have always used for a testing bot is longer than one character. What are you setting the prefix as? username_0: 'Aye ' username_1: Sorry for such a long delay, I forget about issues on the repo alot. I think I saw an issue similar to this on the old bot (the one that this one was originally based off of). When you're passing the prefix, pass it as `"Aye "` instead of `'Aye '`. Example: ![screenshot_1](https://user-images.githubusercontent.com/28911975/40880295-02044d4e-667c-11e8-9f79-9dc86939e308.png) Status: Issue closed username_1: Closing due to inactivity, and solution found (it's just the way strings are handled/processed).
envoyproxy/envoy
571718742
Title: Route actions prefix_rewrite and regex_rewrite should be a oneof Question: username_0: *Title*: Route actions `prefix_rewrite` and `regex_rewrite` should be a `oneof` *Description*: In the discussion for pull request #10050, it was suggested that the `prefix_rewrite` and `regex_rewrite` route actions be put under a `oneof` in the next api version, v4. However, it's too early to mark the proto files with that annotation now, so the code change for #10050 didn't add the annotation. This issue is to remember that this should be considered when the time is appropriate.
AlexProgrammerDE/balena-minecraft-server
569600654
Title: Performance issues Question: username_0: Hi Alex, Just thought I'd try and raise this here as a topic. I noticed your gif showing the chunk loading and unfortunately cannot yield the same result despite using 2GB on my RPi 4 (4GB). Do you have any advice or tips on optimizing the pi to ensure it can get the same performance yours has? Currently my console is littered with things like: `Can't keep up! Is the server overloaded? Running 73827ms or 1476 ticks behind` Answers: username_1: @username_0 you are right with this error, but even on my gaming rig is the server sometimes overloaded. The problem is that the server is not using all of your RAM and CPU cores. Just a part of it. You maybe had tried to fly into not generated chunks (https://minecraft.gamepedia.com/Chunk) In my gif i flied into already generated chunks. That is faster. But i think if you don't fly very fast around the world, can't this issue occur. You can also try to set a lower render distance. username_0: I’m guessing its not possible to push up the cpu and ram allocation a little further? That’s fair I guess, performance is a little better when taking rendered chunks into account. Other behaviours that occurred also when I tried to stay in the same chunk were the occasional block lag and slow mob AI too. Sometimes it got too much for it and took 10 seconds to catch up also. username_1: @username_0 i just saw how big your lags are. I had only one with 500ms. Can you say me what happens if you connect to `mc.hypixel.net`? Are you experiencing lags? username_0: @username_1 I'm not encountering the lag on servers. I launched a server on my own computer also and can see it is running smoothly with mods while playing. username_0: I have decided to invest in a fan for the Pi unit as I had a sneaking feeling the processor was throttling due to the temperature. The room it sits in is relatively hot and even while idle was reaching 80-90°C after longer periods of time. Now with the fan installed I may be correct as performance is noticeably better and little to no tick errors. username_1: @username_0 thats great news. My Pi has also a fan. I forgot to try it without one. Seems like a temperature issue. By the way you are using forge? Thats cool, but it uses lots of resources. I think i will recommend in the README a fan. username_0: I'd definitely recommend adding fan info to the README, it's made quite the difference to the performance overall. I'm stilling using the default Paper jar, but I do plan on finding where the limits are over time, I'll keep this issue posted with my findings. username_1: @username_0 do you have any more questions or shall i close the issue? username_0: I'm all good here, If notice anything else I'll reopen Status: Issue closed
TangleSpace/hotstepper
823532426
Title: More worked examples Question: username_0: Provide more guidance on step by step (hehe) work flow using HotStepper. The ability to explore and analyse data and then directly using the core data model in any machine learning packages that accept Numpy based data needs further elaboration
giesselmann/STRique
468720817
Title: repeat number 0 Question: username_0: Hi, I am trying to quantify repeat number in a large insertion of unknown (potentially varying) size. The alignment is very poor because this insert is not in the reference. When calling repeat number with STRique I am getting a lot of reads that have counts of 0 but when I look at the fast5 there is definitely a repeat present. Could this be a result of the poor mapping? Count distribution: ![Screen Shot 2019-07-16 at 10 55 37 AM](https://user-images.githubusercontent.com/47872912/61307453-5b5ca800-a7bc-11e9-9a9d-dd4d9694b2ad.png) example of read with count of 0: ![Screen Shot 2019-07-16 at 11 19 54 AM](https://user-images.githubusercontent.com/47872912/61307483-70393b80-a7bc-11e9-8222-c3d83b73cdc4.png) Answers: username_1: Hi, The repeat count 0 is given for reads where for instance the signal alignment of prefix and suffix failed. These reads can be filtered out, I updated our documentation accordingly. The quality of the sequence alignment is not impacting the repeat counting. On our targets we observe, that for most reads either the prefix or suffix maps with large soft-clippings on one read side. As long as a read can be located to span a region of interest, STRique will try to evaluate it. The count distribution is -given the length of the expansion- from my experience still very nice. The example signal plot shows a shift of the beginning of the repeat signal. The longer the repeat, the harder the normalization for us, as mean and stdv are impacted by the monotonic signal. For the given signal I would assume, that the mapping of the prefix signal didn't work. username_0: Thank you! I've looked at the signal for a few more of these and see that some don't extend through both flanking regions so I will just filter out the 0s.