repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
ngmansion/FE8N
752854721
Title: ユニットが移動力の最大で移動した時にフリーズ(極まれ) Question: username_0: ユニットが移動力の最大で移動した時にフリーズ(極まれ)します。 移動7のユニットが7マス移動を指定して移動した瞬間に画面が乱れ異音を吐きます。 症状としては、ユニット命令でマップ外に移動するように指定した時みたいなかんじ? 発生日は今年の9月から。 リセットすると発動しないのでセーブデータは用意できていません。 良かったらありえそうな箇所をもう一度みてもらえないでしょうか? Status: Issue closed Answers: username_1: たぶん直しました username_0: バグがまた起こったのを確認しました。 症状としてはやはり、最大移動でフリーズ。ユニット操作の矢印がムチャクチャになる等です。 username_1: ユニットが移動力の最大で移動した時にフリーズ(極まれ)します。 移動7のユニットが7マス移動を指定して移動した瞬間に画面が乱れ異音を吐きます。 症状としては、ユニット命令でマップ外に移動するように指定した時みたいなかんじ? 発生日は今年の9月から。 リセットすると発動しないのでセーブデータは用意できていません。 良かったらありえそうな箇所をもう一度みてもらえないでしょうか? username_1: @username_0 通知があった事に気づいたのですが、直りませんでした? とはいえ、他に原因は思いつかないのですが。 username_0: すいません、こことは別の所に要因を見つけたので解決しました。 お手数おかけしました。 Status: Issue closed username_1: 解決したようなのでclose
AbnormalSec/darkbox
351742394
Title: New tool: timestomp Question: username_0: Similar to this: https://www.offensive-security.com/metasploit-unleashed/TimeStomp/ Answers: username_1: this is something I've been looking into implementing lately, although some issues have arisen. TL;DR Linux is a complex ecosystem and OS agnostic code is hard af. So windows and other OS's are pretty straight forward. I need to test with metasploit to see how it handles linux (TODO) but from all my reading: The linux kernel does not directly support file creation times. It's left up to the filesystem. the field used in windows for creation (ctime) is usually used as "change time" which is for metadata changes to the file, such as changing any other timestamp, renaming or copying, or other metadata changes. If the file has never undergone metadata changes, then it will be the creation date which is technically the first metadata change, but this is unreliable. some filesystems do support creation time, but some extremely common ones such as XFS, EXT3, and others don't. bottom line: this tool requires some weird choices to be made, makes a point for why we need CI testing and likely wont be completed until I've tested meterpreter's implementation.
cozy/cozy-ui
395511089
Title: Icon sprite generation Question: username_0: Currently, during the generation of the sprite, we keep `id` and `use` provided by each SVG. But sometimes, several SVG can have the same `id` and `use`. We have two solutions : - Editing manually the SVG we want to add by editing the `id` and `use` attributes with an unique id - During the generation, editing SVG by adding unique id Answers: username_1: I'm not sure I understand the issue here 😕 username_2: We could use the filename for this right ? An automic id like random hash would be difficult to debug. username_0: @username_1 Let say I want to add this SVG (info.svg) : ```svg <svg xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" viewBox="0 0 16 16"><defs><path id="a" d="M8 16A8 8 0 1 1 8 0a8 8 0 0 1 0 16zM7 4a1 1 0 1 0 2 0 1 1 0 1 0-2 0zm1 2H6v2h1v4a1 1 0 0 0 1 1h2v-2H9V7a1 1 0 0 0-1-1z"/></defs><use fill-rule="evenodd" xlink:href="#a"/></svg> ``` The sprite will be generated and I think I'll have my `info` svg ready to use. But when I'm using it, it'll display (in this case), the help icon since this SVG define the same `id = a` path (I don't know how the browser manage that case, but it seems to take the first id) ```svg <svg width="16" height="16" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink"><defs><path d="M8 16A8 8 0 1 1 8 0a8 8 0 0 1 0 16zm1-7.198C10.122 8.355 11 7.21 11 6c0-1.552-1.448-3-3-3S5 4.448 5 6h2c0-.448.552-1 1-1 .448 0 1 .552 1 1 0 .448-.552 1-1 1a1 1 0 0 0-1 1v2h2V8.802zM7 11v2h2v-2H7z" id="a"/></defs><use xlink:href="#a" fill-rule="evenodd"/></svg> ``` @username_2 Yep, I think we should prepend the id / use with the icon name.
dadhi/FastExpressionCompiler
464824247
Title: Feature toggles to disable the not used features which affect performance Question: username_0: For example, DryIoc does not use Label, GoTo and TryCatchFinally, but to support them FEC adds the memory (fields in `ClosureInfo`) and processing overhead. May be we can use compile time vars to switch some features off if needed. Cause I want to proceed to use FEC as-is in DryIoc, but want a maximum performamce for my use case - which was the original goal anyway. Status: Issue closed Answers: username_0: on hold
maarksman/Eulib
293002726
Title: Search should separate suggestion name and value searched Question: username_0: -type article name -click suggestion -click search -if suggestion-text is different from desired article title, search will fail need to make so that search suggestions have title, type, subtype, and field, but search does not fail when clicking the suggestion
probot/create-probot-app
789422564
Title: CPA cannot find NPM module when running via NPX Question: username_0: npx: installed 231 in 12.637s Cannot find module 'npm' ``` Answers: username_1: it looks like starting with https://github.com/probot/create-probot-app/pull/316 we depend on the `npm` package, but we missed to add it to dependencies: https://github.com/probot/create-probot-app/pull/316/files#diff-7ae45ad102eab3b6d7e7896acd08c427a9b25b346470d7bc6507b6481575d519R30-R45 Status: Issue closed
theniceboy/nvim
822956287
Title: 从B站来,git clone你的neovim报错 Question: username_0: mac环境下,按照视频配置后报了错误,大都是文件找不到 Answers: username_1: 我看到了这个你的提问,一下子就知道问题出在了哪里因为我精确地猜出了你的mac,nvim版本号,并且看到了完整报错信息。 解决方案:把找不到的文件找到。 Status: Issue closed username_2: mac brew的nvim版本不对 得自己去github下载relase过得 username_3: 您好!我也是直接克隆你的neovim配置的,总是遇到各种问题,您能出一期视频从零开始使用的自己的neovim配置吗?克隆配置后的一些配置,应该怎么去配置
Azure/azure-iot-sdk-c
255236160
Title: Azure IoT SDK C - Issue during reconnection attempt when utilizing openSSL and amqp over websockets Question: username_0: Hello, I have encountered an issue during the reconnection attempt when using openSSL and amqpWS as Protocol. (Windows) <!-- Hi there! thank you for discovering and submitting an issue! Please first tell us a little bit about the environment you're running: --> - **OS and version used:** Windows 7 - **SDK version used:** 1.1.21 - Dev environment set up according to : https://github.com/Azure/azure-iot-sdk-c/blob/master/doc/devbox_setup.md#windows - Visual Studio 2015 - openSSL 1.0.2j or 1.0.2l (should also be reproducible with 1.0.2.k) - target platform x86 # Description of the issue: Within c-utility\src\tlsio_openssl.c, the tlsio_openssl_setoption function expects the value as pointer to int during evaluation of the tls_version option. This works fine when invoking IoTHubClient_SetOption. However, on a reconnection attempt the tlsio_openssl_setoption function will be invoked like this: exampled.exe!tlsio_openssl_setoption(void * tls_io, const char * optionName, const void * value) Line 1548 C exampled.exe!OptionHandler_FeedOptions(OPTIONHANDLER_HANDLE_DATA_TAG * handle, void * destinationHandle) Line 236 C exampled.exe!xio_setoption(XIO_INSTANCE_TAG * xio, const char * optionName, const void * value) Line 184 C exampled.exe!OptionHandler_FeedOptions(OPTIONHANDLER_HANDLE_DATA_TAG * handle, void * destinationHandle) Line 236 C exampled.exe!uws_client_set_option(UWS_CLIENT_INSTANCE_TAG * uws_client, const char * option_name, const void * value) Line 1841 C exampled.exe!OptionHandler_FeedOptions(OPTIONHANDLER_HANDLE_DATA_TAG * handle, void * destinationHandle) Line 236 C exampled.exe!wsio_setoption(void * ws_io, const char * optionName, const void * value) Line 670 C exampled.exe!OptionHandler_FeedOptions(OPTIONHANDLER_HANDLE_DATA_TAG * handle, void * destinationHandle) Line 236 C exampled.exe!xio_setoption(XIO_INSTANCE_TAG * xio, const char * optionName, const void * value) Line 184 C exampled.exe!OptionHandler_FeedOptions(OPTIONHANDLER_HANDLE_DATA_TAG * handle, void * destinationHandle) Line 236 C exampled.exe!restore_underlying_io_transport_options(AMQP_TRANSPORT_INSTANCE_TAG * transport_instance, XIO_INSTANCE_TAG * xio_handle) Line 613 C exampled.exe!get_new_underlying_io_transport(AMQP_TRANSPORT_INSTANCE_TAG * transport_instance, XIO_INSTANCE_TAG * * xio_handle) Line 659 C exampled.exe!establish_amqp_connection(AMQP_TRANSPORT_INSTANCE_TAG * transport_instance) Line 716 C When invoked like this, the transport_instance->saved_tls_options OPTIONHANDLER_HANDLE gets evaluated which instead of containing a pointer to the value, contains the value itself. The issue occurs in c-utility\src\tlsio_openssl.c on lines 1548ff, when the passed void pointer gets cast and accessed. # Code sample exhibiting the issue: ` const int* version_option = value;` `if ((int)(intptr_t)*version_option == 0 || (int)(intptr_t)*version_option == 10)` ` {` ` tls_io_instance->tls_version = VERSION_1_0;` ` }` # Console log of the issue: Exception thrown: read access violation. version_option was 0x2. Answers: username_1: @username_0 - Thanks for filing this issue. We're investigating it. username_2: Hi @username_0 , a fix was checked in and is now available on the latest code in master branch. If possible, please validate it on your side and let us know the results. Thanks for contributing to the Azure IoT SDK! username_2: https://github.com/Azure/azure-c-shared-utility/commit/68470a382a3be1e64ace3082f1a6eb6dedfc4f6c https://github.com/Azure/azure-c-shared-utility/commit/71ff9b4e900e6f5ecf2195d1f238441463b69449 Status: Issue closed username_2: Hi @username_0 , this issue will be closed, but please feel free to reopen it if you need to follow up on the issue. Thanks, Azure IoT SDK Team
dotnet/cli
185108582
Title: 'dotnet restore3' failing with "Unable to resolve 'Microsoft.NET.Sdk.Web'" Question: username_0: ## Steps to reproduce Install SDK version `1.0.0-preview3-003892` and create an "empty" web project: `mkdir test` `cd test` `dotnet new -t web` `dotnet restore3` ## Expected behavior - restoration of all NuGet packages ## Actual behavior `restore3` fails with the following message: `C:\Program Files\dotnet\sdk\1.0.0-preview3-003892\NuGet.targets(70,5): error : Unable to resolve 'Microsoft.NET.Sdk.Web (>= 1.0.0-alpha-20161019-1)' for '.NETCoreApp,Version=v1.0'.` ## Environment data `dotnet --info` output: .NET Command Line Tools (1.0.0-preview3-003892) Product Information: Version: 1.0.0-preview3-003892 Commit SHA-1 hash: 5301f08423 Runtime Environment: OS Name: Windows OS Version: 10.0.14393 OS Platform: Windows RID: win10-x64 Answers: username_1: Right now these packages are not published to NuGet.org yet, so you need a NuGet.Config pointing to a feed that contains the packages. You can look for the one we have in the CLI as an example: https://github.com/dotnet/cli/blob/rel/1.0.0/NuGet.Config Status: Issue closed username_0: @username_1 thanks for the reply, I added the NuGet.Config file as described but I was still getting the error: `Unable to resolve 'Microsoft.NET.Sdk.Web (>= 1.0.0-alpha-20161019-1)' for '.NETCoreApp,Version=v1.0'.` So I changed the reference to the package in `test.csproj` to: <PackageReference Include="Microsoft.NET.Sdk.Web"> <Version>1.0.0-alpha-*</Version> </PackageReference> in order to get the latest version of the package and `restore3` was able to resolve it, but now it's failing for another package: `Unable to resolve 'Microsoft.NET.Sdk.Publish (>= 1.0.0-alpha-20161006)' for '.NETCoreApp,Version=v1.0'.` I can't find a reference to `Microsoft.NET.Sdk.Publish` in any of the project files, so I'm at a loss regarding how to solve this. Any help is greatly appreciated. Thanks in advance!
mull-project/mull-jit-lab
310539317
Title: Objective-C: Implement missing property attributes. Question: username_0: ``` property const mull::objc::objc_property64 * 0x1047ea3a0 0x00000001047ea3a0 name const char * "hash" 0x00000001047e9d17 attributes const char * "TQ,R" 0x00000001047e9d1c ``` Answers: username_1: ``` Property: searchList token:T@"NSArray" token:N token:C Assertion failed: (false && "Unknown type"), function parsePropertyAttributes, file /opt/mull-jit-lab/lab-jit-objc/llvm-jit-lab/src/ObjCRuntime.mm, line 342. ```
stanford-oval/genie-toolkit
662144054
Title: Heuristics for slot-filling measures & currencies Question: username_0: If the dialog agent asks for a currency or measure and the user provides a number, we should heuristically convert it. Potentially, this could include a translatable annotation `#_[default_unit]` associated with the parameter indicating the default unit to use (which would be locale-dependent). Answers: username_0: This issue was transferred from almond-dialog-agent. It should be implemented in terms of templates: we change the templates such that filters and imprecise answers can use numbers where measurements are expected, and generate the right measurement code in the neural network.
sebasmonia/panda
474716180
Title: Auto adjust column width Question: username_0: Make it an option (there's a chance that lists too big make it unusable). Sample of `tabulated-list-mode` with adjustable cols: https://stackoverflow.com/a/28488084 Answers: username_0: A newish binding in tablist mode uses `{` and `}` to modify column sizes, which makes this obsolete. Status: Issue closed
alphagov/govuk-prototype-kit
348729386
Title: Make it easier to do branching Question: username_0: For example, we could let people add URLs to radio buttons. If a user selects that radio button, it automatically redirects to that URL, with no need for back end coding. Answers: username_1: Has this moved at all? be good to have this feature moving forward... username_0: On potential solution is to significantly improve the documentation and examples for branching username_2: One product that does this really well is, https://github.com/DEFRA/digital-form-builder, it's a UI which decreases any learning curve, but we could take inspiration from the way they do branches and implement it into the prototype kit. Initial suggestion is have a file specifically for adding branches, that file needs to be changed and everything else will work. A really simple, rushed design for that could be something like ``` const redirects = branchRedirect([ { fieldName: 'eligibilityQuestion', route: '/question-asked-here', values: { yes: '/go-here', no: '/go-there', 'dont-know: 'go-anywhere', }, } ]); module.exports = redirects; ``` ``` /question-asked-here.html {{ govukRadios({ classes: "govuk-radios--inline", idPrefix: "eligibilityQuestion", name: "eligibilityQuestion", fieldset: { legend: { text: "Are you the person paying?", isPageHeading: true, classes: "govuk-fieldset__legend--xl" } }, items: [ { value: "yes", text: "Yes" }, { value: "no", text: "No" }, { value: "dont-know", text: "I don't know" } ] }) }} ``` username_0: Another risk that would be good to get more research on: if we do radio button branching, is it then confusing to have 2 different ways to do branching? (radios vs routes)
mrmike/Ok2Curl
311104739
Title: Android Studio 3 : Failed to resolve "com.github.mrmike:Ok2Curl:0.4.5" Question: username_0: as described in README, added it to dependencies `compile "com.github.username_1:Ok2Curl:0.4.5"` jcenter() is there in the repository list and also in the buildScript list Answers: username_1: Hi @username_0 I think it's happening because of the typo in README file. I've already prepared a fix for that: https://github.com/username_1/Ok2Curl/pull/46. Cause library is available on jcenter(as you can see [here](https://jcenter.bintray.com/com/github/username_1/ok2curl/0.4.5/)). Instead of using `com.github.username_1:Ok2Curl:0.4.5` use `com.github.username_1:ok2curl:0.4.5` (no uppercased letters). Thanks Status: Issue closed
inkle/ink
165648799
Title: Manipulate variables outside of Ink script. Question: username_0: I'd to be able to modify the value of a global variable (ie. set hasSpoken = true) that was declared in Ink from an external C# script. It would help with being able to control the flow of the story and identify speakers that are currently holding a conversation. Is there any functionality built in that supports this? Looking through the available functions, the closest I could get was from observing the current variables. I appreciate any and all help! Answers: username_1: It sounds like you're after the Setting/Getting variables logic: https://github.com/inkle/ink/blob/master/Documentation/RunningYourInk.md#settinggetting-ink-variables unless I'm misunderstanding what it is you're aiming to do. username_0: Ah, that's it. Thanks! Status: Issue closed
marko-js/marko
103744364
Title: Please add for loop iteration with `index` Question: username_0: Would really like if we had the ability to use the index on iteration somehow... Here using the imaginary `index='pos'` `for` attribute (default `index='index') ```jade ui-grid(ui='inverted') ui-row(ui='green' list='data.users.header') for(each='item in data.users.list' index='pos') ui-row(ui='silver' attrs='item' if='pos % 2 == 0') ui-row(ui='blue' attrs='item' if='pos % 2 == 0') ``` Answers: username_0: Sweet :) Learning every day... Status: Issue closed
cms-sw/cmssw
295937495
Title: Avoid calling TClass::GetClass for all setRefCoreStreamer calls Question: username_0: The clall to `TClass::GetClass` in `setRefCoreStreamer` is meant to do a lazy evaluation for putting our own custom Streamers into ROOT. At present, the custom streamers are just there to customize the writing. The reading is customized by an IO rule. Setting the Streamer at read is probably historic, from the time before the IO rules. It would probably be possible to move the actually change of the Streamers to the PoolOutputModule's constructor (since the custom Streamers are only needed for writing, not reading). The call to `EDProductGetter::switchProductGetter(...)` would have to remain to keep the IO rule functional. Answers: username_0: assign core username_1: In the discussion with Philippe, the only case where we may not want the streamer set globally if a particular object would need to be serialized differently on output versus input (for example, if we wanted to use a custom streamer for `edm::Ref` on read but write out `edm::Ref` in a ntuple that can be read by bare ROOT. We didn't think this applied to CMSSW, but felt it was reasonable to point out. username_2: PR #22249 should address this issue and when/if merged allow us to close it. username_2: PR #22249 was merged and this issue should be closed. Status: Issue closed username_0: +1
facebook/react
639671702
Title: hooks Question: username_0: How can I make the input display and get the focus of the input after the setstate value? import React, { useState, useRef } from "react"; import "./styles.css"; export default function App() { const [state, setState] = useState(false); const refEl = useRef(null); const clickBtn = () => { setState(true); refEl.current.focus(); }; return ( <div className="App"> <button onClick={clickBtn}>showIbout</button> {state && <input ref={refEl} />} </div> ); } Answers: username_1: Support requests filed as GitHub issues often go unanswered. We want you to find the answer you're looking for, so we suggest the following alternatives: ##### Coding Questions If you have a coding question related to React and React DOM, it might be better suited for Stack Overflow. It's a great place to browse through frequent questions about using React, as well as ask for help with specific questions. [https://stackoverflow.com/questions/tagged/react](https://stackoverflow.com/questions/tagged/react) ##### Talk to other React developers There are many online forums which are a great place for discussion about best practices and application architecture as well as the future of React. [https://reactjs.org/community/support.html](https://reactjs.org/community/support.html#popular-discussion-forums) Status: Issue closed
sigp/lighthouse
560575706
Title: Docker container doesn't exit gracefully Question: username_0: ## Description After bringing up the lighthouse docker container using `docker-compose up`, killing the container using `Ctrl + C` causes the `beacon_node` process within the container to exit with exit status 137 (out of memory). Since the process doesn't exit gracefully, the beacon node doesn't get persisted and subsequently, bringing up the container again using `docker-compose up` errors with ```Failed to start Lighthouse: Failed to start beacon node: No persisted beacon chain found in store``` ## Expected Behaviour Killing the container with `Ctrl + C` should exit gracefully after persisting the beacon chain.<issue_closed> Status: Issue closed
Zrips/CMI
758011564
Title: Homes at deleted worlds not visible in /homes GUI Question: username_0: **Description of issue:** Homes directing to worlds that don't exist anymore aren't shown in the /homes GUI. This leads to a missleading GUI title (It says "7/10" but I only see 5 icons because 2 homes point to a deleted world) and rubbish in the cmi_users db. --- **Cmi Version (using`/cmi version`):** 172.16.31.10 **Server Type (Spigot/Paperspigot/etc):** Paperspigot **Server Version (using `/ver`):** Paper#294 1.16.4-R0.1-SNAPSHOT Answers: username_1: I think it's good that they aren't seen there, but they should be deleted probably. username_0: Well of course, but it's lot easier to implement when you just show all icons. So if a player tries to tp to such a home, he'll already receive an appropiate error message so he will delete his home on his own. You could also delete a home instantly when a player tries to use it the first time. username_2: Confirmed for version 192.168.127.12. Status: Issue closed username_3: Issue with auto removal would be the cases when world is simply not loaded but home location should remain. Like some plugins will load world which is meant to be for specific player, if world was not loaded when player opened UI then his home would get removed which is no go. What i can do is to actually include those homes into UI but add extra indication that location is in none existing world, just to be more convenient.
Heather/gentoo-gnome
270300643
Title: app-misc/tracker has some conflicts! Question: username_0: app-misc/tracker:0 (app-misc/tracker-2.0.1:0/100::gnome-next, ebuild scheduled for merge) conflicts with <=app-misc/tracker-1.99.0 required by (gnome-extra/nautilus-tracker-tags-1.13.0:0/0::gnome-next, installed) ^^ ^^^^^^ <=app-misc/tracker-1.99:= required by (gnome-base/nautilus-3.26.0:0/0::gnome-next, installed) ^^ ^^^^ <=app-misc/tracker-1.99:0/100= required by (gnome-base/nautilus-3.26.0:0/0::gnome-next, installed) ^^ ^^^^ ~app-misc/tracker-1.13.0 required by (gnome-extra/nautilus-tracker-tags-1.13.0:0/0::gnome-next, installed) ^ ^^^^^^ Answers: username_1: @username_0 it's because I don't know how to make new tracker ebuild with multilib support :woman_shrugging: username_2: Greetings. So what can I do to upgrade my gnome, get rid of nautilus? username_1: remove `gnome-extra/nautilus-tracker-tags` username_2: @username_1 just unmerged it but now gnome-online-miners and folks won't merge. Don't sure if I need them though... username_1: @username_2 I don't know what's that, try to get rid of it too or check why you need it with `equery d gnome-online-miners` username_2: @username_1 I ended up disabling gnome-tracker globally. Thanks for your help. username_1: I need to recheck whether is issue is still being actual username_2: Sorry cannot check if it's still present, I'm not happy gentoo user anymore. Status: Issue closed
stripe/stripe-python
1069075725
Title: Feature Request: Please Add Type-hint Question: username_0: Hi, is there any plan for adding type-hint for Stripe? It's hard for new comer like me to write correct method or variable type without type-hints. Thanks. Status: Issue closed Answers: username_1: @username_0 It's something we want to do in the future but no short term plans for this. Adding type hints will require typing most objects which in turns requiring "pinning" those objects to a specific version of our API, so that you have the right exact list of properties or parameters. This comes at a cost in ease of adoption of the library too and would require major reworks too We know we will get to this in the future but it's not something we're working on at the moment unfortunately. I'm going to close as we already have https://github.com/stripe/stripe-python/issues/650 for this but thanks for the feedback, we definitely hear it!
Hejsil/zig-clap
314517765
Title: Sub commands Question: username_0: Allow the clap to parse subcommands. ``` const parser = comptime Clap(Options).Builder .init( Options { .print_values = false, .a = 0, .b = 0, .c = 0, } ) .command( Command.Builder .init("com") .arguments( []Argument { Argument.Builder .init("a") .help("Set the a field of Option.") .short('a') .takesValue(true) .build(), } ) .subCommands( []Command { Command.Builder .init("sub-com") .arguments( []Argument { Argument.Builder .init("b") .help("Set the a field of Option.") .short('b') .takesValue(true) .build(), } ) .build() } ) .build() ) .build(); ``` Subcommands have their own options, which cannot be accessed by their parent. `com -a 1 sub-com -b 1` works, but `com -a 1 -b 1` does not. Questions: * Should sub commands be able to access the options of its parent? * `com sub-com -a 1 -b 1` * We could have an option that allows/disallows this. * It would probably be useful to be able to tell from the resulting struct that a subcommand was parse. * Should it just set some boolean field? Maybe subcommands will have sub results, which could be their own struct (or union, which would allow for all subcommands to return different tags of the same union). * Should subcommands have an option to be required? Answers: username_0: This can be implemented as a wrapper on top of `StreamingClap`. Status: Issue closed username_0: Out of scope. This is a small package for `-` and `--` style argument parsing only. One could implement a package on top of `zig-clap` that does this.
ampproject/amphtml
255347305
Title: doubleclick Error executing onCreativeRender: G(a.ownerDocument.defaultView).loadExtension is not a function. (In 'G(a.ownerDocument.defaultView).loadExtension("amp-analytics")', 'G(a.ownerDocument.defaultView).loadExtension' is undefined) Question: username_0: doubleclick Error executing onCreativeRender: G(a.ownerDocument.defaultView).loadExtension is not a function. (In 'G(a.ownerDocument.defaultView).loadExtension("amp-analytics")', 'G(a.ownerDocument.defaultView).loadExtension' is undefined) user agent: Mozilla/5.0 (iPhone; CPU iPhone OS 10_2_1 like Mac OS X) AppleWebKit/602.1.50 (KHTML, like Gecko) GSA/33.0.164895372 Mobile/14D27 Safari/602.1 might possibly just be in canary right now. I don't have much other information but please reach out to me so we can dig for more if needed. @username_2 @username_1 do you guys mind triaging this? Thanks Answers: username_1: Not currently able to repro this, even in canary. Do you have more detailed instructions for doing so? username_2: +@username_3 This appears to line up with https://github.com/ampproject/amphtml/blame/master/extensions/amp-ad-network-doubleclick-impl/0.1/amp-ad-network-doubleclick-impl.js#L490. Looking at code history the code was changed as part of #10992 which changed from this.extensions_./*OK*/loadExtension('amp-analytics'); to this.extensions_./*OK*/installExtensionForDoc(this.getAmpDoc(), 'amp-analytics'); Could this be a version sync issue where the extension service changed and older code is consuming the new service? username_3: @username_0 Is this on Canary? The only way I see this happening is via version mismatch. username_3: Setting to P0 until we confirm whether this is canary issue or PROD, at the very least. I don't see how this would happen for the specified URL: this is RTV serving path. username_2: This appears to be a canary issue (at least I can confirm that ad request is sending art=2 which indicates canary). We are not seeing any alerts associated with production traffic. username_0: @username_3 i can only see canary right now with very low occurrences. double checking right now username_3: Still, in RTV path we can never have this kind of issues, right? username_0: @username_3 yes. is this code path not usually reached? im just curious about the low occurrence. username_3: In this case it'd never be reached. Right? username_3: In other words, we can never have RTV version mismatched. We can't have it in non-RTV either, but definitely not in RTV. username_0: some addt'l info. notice that version ends with 313 (current canary) but rtv is 635 (current prod) Query string: v = 001504119607313 noAmp = 0 m = doubleclick Error executing onCreativeRender: G(a.ownerDocument.defaultView).loadExtension is not a function. (In 'G(a.ownerDocument.defaultView).loadExtension("amp-analytics")', 'G(a.ownerDocument.defaultView).loadExtension' is undefined) a = 0 rt = 1p ca = 1 vs = channelOpen rvu = https://https%253A%252F%252Fwww.google.co.jp%252F/amp/s/gamp-ameblo-jp.cdn.ampproject.org/c/s/gamp.ameblo.jp/rkyhlove1207/entry-12198146478.html mso = https://www.google.co.jp/ jse = Safari exps = canary=1,expAdsenseA4A=0,expDoubleclickA4A=0,expAdSenseFFDR=0,a4aProfilingRate=1,dbclk_a4a_viz_change=0,ad-type-custom=1,ios-embed-wrapper=1,amp-apester-media=1,amp-playbuzz=1,chunked-amp=1,pump-early-frame=1,sticky-ad-early-load=1,amp-auto-ads=1,amp-auto-ads-adsense-holdout=0,slidescroll-disable-css-snap=1,visibility-v3=1,version-locking=1,as-use-attr-for-format=1,a4aFastFetchDoubleclickLaunched=0,a4aFastFetchAdSenseLaunched=0,a4a-measure-get-ad-urls=1,ad-loader-v2=1,3p-use-ampcontext=0,amp-animation=1,amp-sidebar toolbar=1 el = u s = sd@https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:115:483 onCreativeRender@https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:135:415 https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:65:117 Yc@https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:93:392 Xc@https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:89:85 https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:82:105 promiseReactionJob@[native code] r = ae = doubleclick Error executing SHORT = 1 Fragment: username_4: Seems like canary runtime is loading prod version of ad extension script. As talked with @username_0 offline: - It's not likely for canary amp runtime to know the current prod RTV and compose a wrong extension URL. More likely it's a environment problem - This failure is before ad request, so it couldn't be the version from the ad either. - We don't allow amp-ad in AMP ad, so it shouldn't be amp-inabox I'm not very familiar what service-worker is doing for caching. Does it rewrite URL to full RTV version? Do we use service-worker on CDN in any case? @jridgewell Also, worth to mention that this bug might exist for a while, just recently revealed because of the method signature change. username_3: It looks like service workers are not currently involved. I really don't see how RTV can ever be wrong - they are always explicitly selected. The only possibility I see now is that Cache intermittently inserts wrong RTV. But not sure if that's really possible or the case here. I also don't think Cache ever inserts `amp-ad-network-doubleclick-impl-0.1.js`, but I could be wrong. username_0: i also don't believe cache inserts `amp-ad-network-doubleclick-impl-0.1.js` from what @username_4 told me that is dynamically fetched (correct me if i'm wrong @username_4) username_0: @username_3 is there any additional logging we can do to give us more information? username_3: It is definitely dynamically fetched, but we are definitely considering to add them in Cache. Nor sure where that project is now. /cc @codewiz username_3: @username_0 To answer your question about logging. The current logging is pretty clear and points to a problem very well. We just don't know how we ever arrive at it. username_1: The cache insertion project is still in the works and definitely hasn't been deployed to production yet. username_3: @username_1 This is canary error. HAs it been deployed to canary? username_1: No. Also, this is (will be) a feature of the Google AMP Cache; it's entirely separate from the AMP runtime release process. username_3: @username_1 Definitely a Cache thing, I get this. The thing is, however, AMP Runtime has non way of knowing RTVs - it only knows one - it's own. While Cache knows them all. So, I just don't see it'd be possible to get a mismatch via runtime. username_0: we just promoted 313 to prod and the issue occurred: Query string: v = 011504040004635 noAmp = 0 m = doubleclick Error executing onCreativeRender: G(...).loadExtension is not a function a = 0 rt = 1p jse = IE exps = canary=0,expAdsenseA4A=0,expDoubleclickA4A=0,expAdSenseFFDR=0,dbclk_a4a_viz_change=0,a4aProfilingRate=0,ad-type-custom=1,ios-embed-wrapper=1,amp-apester-media=1,amp-playbuzz=1,chunked-amp=1,sticky-ad-early-load=1,amp-auto-ads=1,amp-auto-ads-adsense-holdout=0,slidescroll-disable-css-snap=1,version-locking=1,visibility-v3=1,a4aFastFetchDoubleclickLaunched=0,a4aFastFetchAdSenseLaunched=0,pump-early-frame=1,a4a-measure-get-ad-urls=0,ad-loader-v2=1,3p-use-ampcontext=0,amp-animation=1 el = u s = amp-a4a: doubleclick Error executing onCreativeRender: G(...).loadExtension is not a function at sd (https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:115:470) at Ad.push.f.h.onCreativeRender (https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:135:413) at https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:65:112 at Yc (https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:93:392) at Xc (https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:89:83) at https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:82:103 at <anonymous> ae = doubleclick Error executing onCreativeRender: G(...).loadExtension is not a function fr = Fragment: username_0: we've more or less ruled this out as a client side issue. we've filed an issue at google internally to look into this (b/65461170) username_0: well i haven't seen this in a while so Im going to close it. Status: Issue closed
bbc/simorgh
489729501
Title: Improve setupTests.js Question: username_0: **Is your feature request related to a problem? Please describe.** When an error is thrown it stops running other tests. **Describe the solution you'd like** Update setupTests.js to fix this. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Testing notes** [Tester to complete] Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc - [ ] This feature is expected to need manual testing. **Additional context** Add any other context or screenshots about the feature request here.<issue_closed> Status: Issue closed
magefree/mage
686452420
Title: Historic matches not shown in the list Question: username_0: When I create a Historic match it doesn't appear in the matches list, so other players can't join. ![Historic game](https://user-images.githubusercontent.com/36141078/91329816-d9dc9500-e785-11ea-9ee4-7cc0ba26eea8.png) Answers: username_1: Fixed via 40fe3b137e52805dce19f8ae7db6167feb134c8d, will be available with next hotfix/release (soon). Status: Issue closed
expo/sentry-expo
831736001
Title: Native Client is not available error when debug is set to false in Sentry.init call Question: username_0: ### Summary I set up expo sentry in our app. At first, everything seemed to be working, but then I tried to disable debug in development and I now I cannot start the app, because of this error: SentryError: Native Client is not available, can't start on native. This is my config. When I set `debug` to `true` it works again. ``` Sentry.init({ dsn: SENTRY_DSN, enableInExpoDevelopment: false, debug: false, }); ``` ### Managed or bare workflow? If you have `ios/` or `android/` directories in your project, the answer is bare! managed ### What platform(s) does this occur on? iOS ### SDK Version (managed workflow only) 39 ### Environment Expo CLI 4.3.0 environment info: System: OS: macOS 11.2.3 Shell: 5.7.1 - /usr/local/bin/zsh Binaries: Node: 14.15.4 - ~/.nvm/versions/node/v14.15.4/bin/node Yarn: 1.22.4 - ~/.yarn/bin/yarn npm: 6.14.10 - ~/.nvm/versions/node/v14.15.4/bin/npm Watchman: 4.9.0 - /usr/local/bin/watchman Managers: CocoaPods: 1.8.4 - /Users/username_0/.rvm/rubies/ruby-2.6.3/bin/pod SDKs: iOS SDK: Platforms: iOS 14.4, DriverKit 20.2, macOS 11.1, tvOS 14.3, watchOS 7.2 Android SDK: API Levels: 23, 25, 28, 29 Build Tools: 28.0.3, 29.0.2 System Images: android-23 | Google APIs Intel x86 Atom, android-28 | Intel x86 Atom_64, android-28 | Google APIs Intel x86 Atom, android-28 | Google Play Intel x86 Atom, android-29 | Google APIs Intel x86 Atom IDEs: Android Studio: 4.0 AI-193.6911.18.40.6514223 Xcode: 12.4/12D4e - /usr/bin/xcodebuild npmPackages: expo: ^39.0.4 => 39.0.4 react: 16.13.1 => 16.13.1 react-dom: 16.13.1 => 16.13.1 react-native: https://github.com/expo/react-native/archive/sdk-39.0.4.tar.gz => 0.63.2 Expo Workflow: managed ### Reproducible demo or steps to reproduce from a blank project Just start sentry in App.js with disabled debug mode: ``` Sentry.init({ dsn: SENTRY_DSN, enableInExpoDevelopment: false, debug: false, }); ``` Answers: username_1: What version of `sentry-expo`? This could be a duplicate of [any of these issues](https://github.com/expo/sentry-expo/search?q=Native+Client+is+not+available%2C+can%27t+start+on+native&type=issues) username_0: Hi @username_1. Sorry I forgot to mention `sentry-expo` version. I am using `v 3.0.5`. I've seen the issues you linked above, unfortunately they are all closed without any resolution that would work for me. As for the repro of the issue: I believe it happens as soon as you set `debug` to `false`. I am not sure what else to add to the repro steps but please let me know if there's anything more I can do to help you find the issue. username_2: Same problem as this issue: https://github.com/expo/sentry-expo/issues/178 With this configuration and latest release of sentry-expo : ``` Sentry.init({ dsn: "blablabla", enableNative: Device.isDevice, enableInExpoDevelopment: Device.isDevice, release: version, environment: env, integrations: (integrations) => (isDev ? [...integrations, new Debug() as Integration] : integrations), debug: env !== "production", enabled: !isDev, }); ``` username_1: @username_2 you have `enableNative: true`, which is why it's trying to initialize the native client. Be sure you: - are in the bare workflow - have `@sentry/react-native` in your dependencies - run `npx pod-install` for iOS username_2: @username_1 Oh. I'm using the managed workflow but it was working before. Removed `enableNative` and it's working as expected now. Thanks ! username_1: Closing since the above issue is resolved, and still haven't been able to repro @username_0's original issue Status: Issue closed
joaodrp/gelf-formatter
901831843
Title: gelf-formatter induces logging error in django when uncaptured exception is raised Question: username_0: **Solution:** This can be fixed by converting value to string: `log_record[_prefix(key)] = str(value)` While this breaks deep encoding json dicts, this is fine for my use case (hopefully for other users, too) Status: Issue closed Answers: username_1: Fixed with #18
uhlik/bpy
460345638
Title: PCV, 2.80, points somehow lost depth sorting, in fact, looks inverted now Question: username_0: blender-2.80.0-git20190625.c0c1b4542f39 25.6.2019 Answers: username_0: blender-2.80.0-git20190625.c0c1b4542f39 25.6.2019 username_1: I can confirm that the depth sorting appears to be awry on the latest 2.8: **Blender Version** Broken: version: 2.80 (sub 74), branch: master, commit date: 2019-06-25 23:26, hash: `rBabaf6b106d3a` Worked: (optional) **But** it appears to work fine on this older build of 2.8, so maybe its an API compatibility issue? **Blender Version** Broken: version: 2.80 (sub 74), branch: blender2.7, commit date: 2019-06-17 21:07, hash: `rB741641f4c3b8` Worked: (optional) username_0: see here: https://developer.blender.org/T66148 will update it soon.. Status: Issue closed username_0: fixed
dubocr/homebridge-tahoma
244801649
Title: Error with more 10 blinds Question: username_0: Hello, I've an error when I start a scene with more 10 blinds, is it possible to control more ? This limit is in the plugin or in homekit? Best Regards Christopher Answers: username_1: Hi Christopher, What kind of error do you have ? Have you log trace ? I have no knowledge of restriction but if this come from Tahoma you will have some error log in homebridge : `sudo journalctl | grep homebridge` Best username_0: Hello, the error in hombridge: [2017-07-25 08:31:07] [Tahoma] Error 400 Execution queue is full (soft limit: 10) (EXEC_QUEUE_FULL) username_1: Hi, Seem this is a TaHoma API restriction so if you want to control more than 10 device with a scene, you have to create a TaHoma scene which will appear as on/off switch in homebridge. username_0: Hi, ok, the scene create in TaHoma don't appear in home... I will test to create an automation with a fake button who control many scene. Status: Issue closed username_1: Hi, Sorry, to expose TaHoma scenes as on/off switch you have to add the folowing line in your config file : “exposeScenarios“: true username_2: While this might be a limit of the Somfy API, I believe that homebridge-tahoma needs to implement some sort of fix to work around it. As a user, I should be able to make a HomeKit Scene that consists of more than 10 devices without thinking about whatever API restrictions lie behind. Cheers, username_1: Hi, Yes, that’s true homebridge-tahoma needs to implement a work around. Best, username_3: Hi, I have the same problem with more than 10 blinds. And when I tried to expose the TaHoma scenes I get the following error: ReferenceError: Service is not defined at new ScenarioAccessory (/usr/local/lib/node_modules/homebridge-tahoma/index.js:226:17) at /usr/local/lib/node_modules/homebridge-tahoma/index.js:200:32 at /usr/local/lib/node_modules/homebridge-tahoma/overkiz-api.js:193:5 at Request.authCallback [as _callback] (/usr/local/lib/node_modules/homebridge-tahoma/overkiz-api.js:216:17) at Request.self.callback (/usr/local/lib/node_modules/homebridge-tahoma/node_modules/request/request.js:185:22) at Request.emit (events.js:196:13) at Request.<anonymous> (/usr/local/lib/node_modules/homebridge-tahoma/node_modules/request/request.js:1161:10) at Request.emit (events.js:196:13) at IncomingMessage.<anonymous> (/usr/local/lib/node_modules/homebridge-tahoma/node_modules/request/request.js:1083:12) at Object.onceWrapper (events.js:284:20) at IncomingMessage.emit (events.js:201:15) at endReadableNT (_stream_readable.js:1130:12) at processTicksAndRejections (internal/process/task_queues.js:84:17) What could be the problem? Homebridge is running on my iMac: +-- [email protected] +-- [email protected] Best regards Joel username_1: Hi, Issue with Scénario must be fixed in last release 0.2.3. May I’ll find a workaround for API limit to 10. Stay tuned username_3: Thanks! My Scenarios from the TaHoma-Box now show up! :-) username_1: Small improvement made. Command over quota (10) are postpone to be executed later (10 sec).
ktorio/ktor
483305218
Title: ktor-client hangs forever with empty Input in formData Question: username_0: **To Reproduce** ``` input = Files.newInputStream(emptyFile) ... httpClient.execute(HttpRequestBuilder().apply { method = HttpMethod.Post url( ... ) body = MultiPartFormDataContent( formData { ... appendInput("file", Headers.build { append(HttpHeaders.ContentDisposition, "filename=$fileName") append(HttpHeaders.ContentType, contentType) } ) { input } } ) header(HttpHeaders.IfMatch, etag0) }).use {response -> ... } ``` Perhaps this is the same or related to [#1290 ](https://github.com/ktorio/ktor/issues/1290) **Expected behavior** 1. size check should be added to Input.copyTo ``` val size = [email protected](buffer) if (size < 0) break; written(size) ``` 2. When exception occurs inside ktor-client it should not cause the call to hang<issue_closed> Status: Issue closed
spring-projects/spring-framework
838740672
Title: Autowiring HttpServletResponse in the request flow to set a header Question: username_0: Hi Colleagues, I wanted confirm on one particular flow. If there is a spring controller for REST and we auto-wire HttpServletResponse along the flow in any of the underlying classes. I mean not in @Controller, but any other @Service class. Still would it be RequestScoped? Please help. Answers: username_1: Thanks for getting in touch, but it feels like this is a question that would be better suited to [Stack Overflow](http://stackoverflow.com/). As mentioned in [the guidelines for contributing](https://github.com/spring-projects/spring-framework/blob/master/CONTRIBUTING.md#discuss), we prefer to use the issue tracker only for bugs and enhancements. Feel free to update this issue with a link to the re-posted question (so that other people can find it) or add some more details if you feel this is a genuine bug. I'm not sure I understand your question, but indeed [injecting a scoped-component](https://docs.spring.io/spring-framework/docs/current/reference/html/core.html#beans-factory-scopes) (you wouldn't inject the request itself). I think that in general it's a good idea to use `@Controller` and other MVC infrastructure classes to act as an interface between the web and the application; extracting information from the request thanks to controller flexible signatures, or providing that information so that a shared component can set it as a header, is a better idea. In your case, asking a question on StackOverflow and describing 1) what you're trying to achieve (the use case) 2) what you've tried so far and 3) why it's not working out for you so far - will help you more than questions here. Thanks! Status: Issue closed
rust-lang/rust
631480463
Title: ICE E0431.rs with -Zsave-analysis Question: username_0: <!-- Thank you for finding an Internal Compiler Error! 🧊 If possible, try to provide a minimal verifiable example. You can read "Rust Bug Minimization Patterns" for how to create smaller examples. http://blog.pnkfx.org/blog/2019/11/18/rust-bug-minimization-patterns/ --> ### Code code from `./src/test/ui/error-codes/E0431.rs` ```Rust use {self}; //~ ERROR E0431 fn main () { } ``` ### Meta <!-- If you're using the stable version of the compiler, you should also check if the bug also exists in the beta or nightly versions. --> repo is @ 219380d840877de97b0e17674be7212ebc929d51 ### Error output `build/x86_64-unknown-linux-gnu/stage2/bin/rustc ./src/test/ui/error-codes/E0431.rs -Zsave-analysis` ``` thread 'rustc' panicked at 'index 18446744073709551615 out of range for slice of length 0', src/librustc_save_analysis/dump_visitor.rs:213:21 note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace error: internal compiler error: unexpected panic note: the compiler unexpectedly panicked. this is a bug. ``` <!-- Include a backtrace in the code block by setting `RUST_BACKTRACE=1` in your environment. E.g. `RUST_BACKTRACE=1 cargo build`. --> <details><summary><strong>Backtrace</strong></summary> <p> ``` error[E0431]: `self` import can only appear in an import list with a non-empty prefix --> ./src/test/ui/error-codes/E0431.rs:1:6 | 1 | use {self}; //~ ERROR E0431 | ^^^^ can only appear in an import list with a non-empty prefix warning: unused import: `self` --> ./src/test/ui/error-codes/E0431.rs:1:6 | 1 | use {self}; //~ ERROR E0431 | ^^^^ | = note: `#[warn(unused_imports)]` on by default [Truncated] error: internal compiler error: unexpected panic note: the compiler unexpectedly panicked. this is a bug. note: we would appreciate a bug report: https://github.com/rust-lang/rust/blob/master/CONTRIBUTING.md#bug-reports note: rustc 1.45.0-dev running on x86_64-unknown-linux-gnu note: compiler flags: -Z save-analysis query stack during panic: end of query stack error: aborting due to previous error; 1 warning emitted For more information about this error, try `rustc --explain E0431`. ``` </p> </details> Answers: username_0: another one: `./src/test/ui/issues/issue-47673.rs` ````rust // check-pass #![allow(unused_imports)] use {{}, {}}; fn main() {} ```` username_0: another one: `./src/test/ui/issues/issue-10806.rs` ````rust // run-pass #![allow(unused_imports)] // pretty-expanded FIXME #23616 pub fn foo() -> isize { 3 } pub fn bar() -> isize { 4 } pub mod baz { use {foo, bar}; pub fn quux() -> isize { foo() + bar() } } pub mod grault { use {foo}; pub fn garply() -> isize { foo() } } pub mod waldo { use {}; pub fn plugh() -> isize { 0 } } pub fn main() { let _x = baz::quux(); let _y = grault::garply(); let _z = waldo::plugh(); } ```` username_1: I am on it username_2: Same issue happened after upgraded to the latest version today. username_1: Yes the fix #73046 has not landed yet. Status: Issue closed
haslo/lists_juggle_browser
216217379
Title: Icons for ships are not correct in Microsoft Edge 38.14393.0.0 Question: username_0: Was checking out the meta-wing page on my surface and noticed that the icons for many of the ships are not correct. Screenshot attached. ![2017-03-22 1](https://cloud.githubusercontent.com/assets/5503381/24222009/9893aaf4-0f1e-11e7-81f4-a7aec5e608da.png) Answers: username_1: Gee, thanks. Looks like there's other cross-browser problems, with a variety of browsers asking before any page change. Gonna have to fix this. username_1: I'm not sure whether I can fix the icons themselves; Edge breaks them in the original repository as well: https://geordanr.github.io/xwing-miniatures-font/ I guess we should go file a bug there, too. Or with Microsoft, preferrably. username_1: FWIW, I've fixed the nagging "do you really want to change page" bug in a variety of browsers. Not the actual bug you've reported though, so I'm leaving this open. username_0: Thanks for doing that... that one was really getting to me too :) Good job!
aquasecurity/tracee
979857390
Title: container_id initialization regex has to be fixed Question: username_0: When running in an Ubuntu environment with LXD, the regex used to recognize running containers, and populate the map in charge for tracking existing containers + pids, will have false positives like: ``` ./fs/cgroup/systemd/user.slice/user-1000.slice/[email protected]/app.slice/snap.lxd.lxc.4b3296a2-c139-4da5-b782-006c431cca74.scope/tasks ``` ``` $ sudo ./dist/tracee-ebpf --debug --trace event=mount BTF: bpfenv = false, btfenv = false, vmlinux = true BPF: using embedded BPF object unpacked CO:RE bpf object file into memory TIME UID COMM PID TID RET EVENT ARGS Running container = vice/app.slice/snap.lxd.lxc.4b3296a2-c139-4da5-b782-006c431cca74 pid = 2558259 ``` We should not have "**UUID.scope/tasks**" regex as a 'generic' regex for now IMO.<issue_closed> Status: Issue closed
tummychow/git-absorb
748583606
Title: Getting "old path already occupied" Question: username_0: Great tool! Unfortunately, I was trying to absorb some changes today and got this error: ``` Nov 23 09:17:36.878 CRIT absorb failed, err: old path already occupied ``` It is not clear to me why am I getting this error, what it means, or how to solve it. :( Answers: username_1: i can't diagnose this issue without more information. that error comes from my patch parsing code, so i need to see the diffs of your git index (`git diff --cached`) and the commits in your git absorb stack username_0: I cannot share the exact diff (it's proprietary source code), but I could locate the issue: I had a symbolic link to a file that I replaced with an actual file. This gives a diff similar to this one: ```diff diff --git a/some-file.ext b/some-file.ext new file mode 100644 index 0000000000..7cb58a6452 --- /dev/null +++ b/some-file.ext @@ -0,0 +1,135 @@ +# File contents following here ``` username_1: ah... i can see how that would cause this problem ``` diff --git a/bar b/bar deleted file mode 120000 index 1910281..0000000 --- a/bar +++ /dev/null @@ -1 +0,0 @@ -foo \ No newline at end of file diff --git a/bar b/bar new file mode 100644 index 0000000..e69de29 ``` looks like link and non-link changes get different hunks. that'll require a special case in the parser username_0: I've found another case not involving symbolic links where I'm getting this error: ```patch diff --git a/some_path/pvm.spec b/some_path/pvm.spec index aa58341137..943bacfa2c 100644 --- a/some_path/pvm.spec +++ b/some_path/pvm.spec @@ -58,7 +58,7 @@ tar -xzf %{SOURCE0} cd pvm3 tar -xzf %{SOURCE1} %patch0 -p0 -#patch1 -p1 +%patch1 -p1 %patch2 -p1 %patch3 -p1 %patch4 -p1 ``` username_1: is that the entire index? this bug requires a diff with two patches that both have the same filename. your example only has one patch in it username_0: Yes, that is the entire index. username_1: yeah that sounds likely, every commit has to get parsed the same way as the index. probably one of them has the same issue where a symlink is being turned into a regular file or vice versa
BEEmod/BEE2-items
213603354
Title: Glass Item causes "Failure To Compile" on P1 Style Question: username_0: VBSP failed. [INFO] vbsp.main(): BEE VBSP hook initiallised. [INFO] cond.core.import_conditions(): Imported all conditions modules! [INFO] vbsp.main(): Map path is "E:/Steam/steamapps/common/Portal 2/sdk_content\maps/preview.vmf" [INFO] vbsp.main(): New path: "E:/Steam/steamapps/common/Portal 2/sdk_content\maps\styled\preview.vmf" [INFO] vbsp.main(): Clearing inject/ directory.. [INFO] vbsp.main(): PeTI map detected! [INFO] vbsp.main(): Loading settings... [WARNING] vbsp_options.load(): Extra config options: {'bottomless_pit': '0', 'fizz_visiblity': '0'} [WARNING] instanceLocs.get_subitems(): Invalid custom instance name - "piston_grate_low" for <item_rex_pellet_destroyer> (Valid: {'head': 'instances/bee2/clean/items/rex/pellet_dest/head.vmf', 'piston_logic_0': 'instances/bee2/clean/items/rex/pellet_dest/logic_0.vmf', 'piston_logic_1': 'instances/bee2/clean/items/rex/pellet_dest/logic_1.vmf', 'piston_logic_2': 'instances/bee2/clean/items/rex/pellet_dest/logic_2.vmf', 'piston_logic_3': 'instances/bee2/clean/items/rex/pellet_dest/logic_3.vmf', 'common': 'instances/bee2/p1/items/rex/pellet_dest/common.vmf', 'piston_bottom_0': 'instances/bee2/p1/items/rex/pellet_dest/pist_0.vmf', 'piston_bottom_1': 'instances/bee2/p1/items/rex/pellet_dest/pist_1.vmf', 'piston_bottom_2': 'instances/bee2/p1/items/rex/pellet_dest/pist_2.vmf', 'piston_bottom_3': 'instances/bee2/p1/items/rex/pellet_dest/pist_3.vmf', 'piston_static_0': 'instances/bee2/p1/items/rex/pellet_dest/static_0.vmf', 'piston_static_1': 'instances/bee2/p1/items/rex/pellet_dest/static_1.vmf', 'piston_static_2': 'instances/bee2/p1/items/rex/pellet_dest/static_2.vmf', 'piston_static_3': 'instances/bee2/p1/items/rex/pellet_dest/static_3.vmf', 'piston_static_4': 'instances/bee2/p1/items/rex/pellet_dest/static_4.vmf'}) [WARNING] instanceLocs.get_subitems(): Invalid custom instance name - "piston_grate_high" for <item_rex_pellet_destroyer> (Valid: {'head': 'instances/bee2/clean/items/rex/pellet_dest/head.vmf', 'piston_logic_0': 'instances/bee2/clean/items/rex/pellet_dest/logic_0.vmf', 'piston_logic_1': 'instances/bee2/clean/items/rex/pellet_dest/logic_1.vmf', 'piston_logic_2': 'instances/bee2/clean/items/rex/pellet_dest/logic_2.vmf', 'piston_logic_3': 'instances/bee2/clean/items/rex/pellet_dest/logic_3.vmf', 'common': 'instances/bee2/p1/items/rex/pellet_dest/common.vmf', 'piston_bottom_0': 'instances/bee2/p1/items/rex/pellet_dest/pist_0.vmf', 'piston_bottom_1': 'instances/bee2/p1/items/rex/pellet_dest/pist_1.vmf', 'piston_bottom_2': 'instances/bee2/p1/items/rex/pellet_dest/pist_2.vmf', 'piston_bottom_3': 'instances/bee2/p1/items/rex/pellet_dest/pist_3.vmf', 'piston_static_0': 'instances/bee2/p1/items/rex/pellet_dest/static_0.vmf', 'piston_static_1': 'instances/bee2/p1/items/rex/pellet_dest/static_1.vmf', 'piston_static_2': 'instances/bee2/p1/items/rex/pellet_dest/static_2.vmf', 'piston_static_3': 'instances/bee2/p1/items/rex/pellet_dest/static_3.vmf', 'piston_static_4': 'instances/bee2/p1/items/rex/pellet_dest/static_4.vmf'}) [INFO] vbsp.load_settings(): Settings Loaded! [INFO] vbsp.load_map(): Parsing Map... [INFO] vbsp.load_map(): Reading Map... [INFO] vbsp.load_map(): Loading complete! [INFO] vbsp.mod_entryexit(): Setting SP Entry to 6 [INFO] vbsp.mod_entryexit(): Setting SP Exit to 3 [INFO] vbsp.get_map_info(): Game Mode: SP [INFO] vbsp.get_map_info(): Is Preview: True [INFO] brushLoc.read_from_map(): Analysed map, filling air... (21 starting positions..) [INFO] brushLoc.read_from_map(): Air filled! [INFO] cond.core.check_all(): Checking Conditions... [INFO] cond.fizzler.fizzler_out_relay(): Fizzler data: {'marker_inst': [], 'pos_to_name': {}, 'fizz_by_name': {}, 'model': 'instances/bee2/p1/items/tdb/csfi_model.vmf', 'base': 'instances/bee2/p1/items/tdb/csfi_base.vmf', 'fizz_id': 'item_tdb_csfi', 'fizz_bases': {'instances/bee2/p1/items/fizzler/fizzler_base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/clean/items/paint_fizz/base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/p1/items/fr/csf_base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/p1/items/fr/fdf_base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/p1/items/fr/mif_base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/p1/items/death_fizz/base.vmf': ItemConnections(in_act=('branch_fizz', 'ToggleTest'), in_deact=('branch_las', 'ToggleTest'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/p1/items/physler/base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/clean/items/rex/pellet_field_base.vmf': ItemConnections(in_act=(None, ''), in_deact=(None, ''), out_act=('out', 'OnUser2'), out_deact=('out', 'OnUser1')), 'instances/bee2/p1/items/tdb/csfi_base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, ''))}, 'fizz_models': {'instances/bee2/p1/items/physler/model.vmf', 'instances/bee2/p1/items/tdb/csfi_model.vmf', 'instances/bee2/p1/items/fizzler/fizzler_model.vmf', 'instances/bee2/p1/items/fr/mif_model.vmf', 'instances/bee2/p1/items/fr/csf_model.vmf', 'instances/bee2/clea[INFO] cond.catwalks.res_make_catwalk(): Starting catwalk generator... [INFO] cond.scaffold.res_unst_scaffold(): Running Scaffold Generator (BEE2_SCAFFOLD)... [INFO] cond.scaffold.res_unst_scaffold(): Finished Scaffold generation! [INFO] cond.vactubes.res_make_vactubes(): Running Generator (BEE2_VACTUBES)... [INFO] vbsp.set_elev_videos(): Elevator type: NONE [INFO] voiceLine.add_voice(): Adding Voice Lines! [INFO] voiceLine.add_voice(): Quote events: [] [INFO] voiceLine.add_voice(): No responses data.. [INFO] voiceLine.add_voice(): 0 Mid quotes [INFO] voiceLine.add_voice(): Done! [INFO] cond.monitor.mon_camera_link(): Bullseye defaultdict(<class 'int'>, {}) [INFO] vbsp.set_player_portalgun(): Setting Portalgun: [INFO] vbsp.set_player_portalgun(): Blue: Y, Orange: Y [INFO] vbsp.set_player_portalgun(): Done! [INFO] vbsp.add_screenshot_logic(): Added Screenshot Logic [INFO] cond.core.check_all(): Map has attributes: ['pack_p1_radio_clean', 'spawn_dual'] [INFO] cond.core.check_all(): instanceLocs cache: CacheInfo(hits=5318, misses=335, maxsize=256, currsize=256) [INFO] cond.core.check_all(): Style Vars: {'blockplayergelbomb': True, 'doorcameras': True, 'p1_ported_pgun': True, 'restartonexit': False, 'multiversecave': False, 'fixfizzlerbump': False, 'nomidvoices': False, 'unlockdefault': False, 'allowgoomist': False, 'funnelallowswitchedlights': True, 'enableshapesignageframe': True} [INFO] cond.core.check_all(): Global instances: {'instances/BEE2/p1/set_portalgun.vmf', 'instances/BEE2/p1/global_ents.vmf'} [INFO] vbsp.add_extra_ents(): Adding Music... [INFO] vbsp.add_extra_ents(): Adding Global PTI Ents [INFO] vbsp.change_ents(): Editing Other Entities... [INFO] vbsp.fixup_goo_sides(): Changing goo sides... [INFO] vbsp.fixup_goo_sides(): Done! [INFO] vbsp.change_brush(): Editing Brushes... [INFO] vbsp.change_brush(): Make Bottomless Pit: False [INFO] vbsp.change_brush(): Goo heights: 0 <- Counter() [INFO] vbsp.change_brush(): Adding Goo Mist... [INFO] vbsp.change_brush(): Done! [INFO] vbsp.change_overlays(): Editing Overlays... [INFO] vbsp.change_trig(): Editing Triggers... [INFO] vbsp.collapse_goo_trig(): Collapsing goo triggers... [INFO] vbsp.collapse_goo_trig(): Done! [INFO] vbsp.change_func_brush(): Editing Brush Entities... [ERROR] (unknown file).(unknown function)(): Uncaught Exception: Traceback (most recent call last): File "<frozen importlib._bootstrap>", line 961, in _find_and_load [Truncated] File "<frozen importlib._bootstrap>", line 646, in _load_unlocked File "<frozen importlib._bootstrap>", line 616, in _load_backward_compatible File "C:\Users\Lautaro\AppData\Local\Programs\Python\Python36\lib\site-packages\cx_Freeze\initscripts\__startup__.py", line 12, in <module> __import__(name + "__init__") File "<frozen importlib._bootstrap>", line 961, in _find_and_load File "<frozen importlib._bootstrap>", line 950, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 646, in _load_unlocked File "<frozen importlib._bootstrap>", line 616, in _load_backward_compatible File "C:\Users\Lautaro\AppData\Local\Programs\Python\Python36\lib\site-packages\cx_Freeze\initscripts\Console.py", line 24, in <module> exec(code, m.__dict__) File "vbsp_launch.py", line 7, in <module> File "C:\Users\Lautaro\Documents\BEEMOD 2.4 GHITUB\BEE2.4\src\vbsp.py", line 3583, in main change_func_brush() File "C:\Users\Lautaro\Documents\BEEMOD 2.4 GHITUB\BEE2.4\src\vbsp.py", line 2829, in change_func_brush grate_temp.apply(side, change_mat=False) File "C:\Users\Lautaro\Documents\BEEMOD 2.4 GHITUB\BEE2.4\src\template_brush.py", line 288, in apply mat, face.uaxis, face.vaxis, face.ham_rot = self[face.normal().as_tuple()] File "C:\Users\Lautaro\Documents\BEEMOD 2.4 GHITUB\BEE2.4\src\template_brush.py", line 272, in __getitem__ mat, axis_u, axis_v, rotation = self._axes[tuple(normal)] ValueError: not enough values to unpack (expected 4, got 3) Status: Issue closed Answers: username_1: Fixed.
greenboxal/dns-heaven
400218760
Title: kubectl throws error related to Google OAuth when dns-heaven is up Question: username_0: Thanks for creating this tool, it temporarily fixed the problem with kube dns. But after some time kubectl started to give this error ``` Unable to connect to the server: failed to refresh token: Post https://oauth2.googleapis.com/token: dial tcp: lookup oauth2.googleapis.com on 127.0.0.1:53: read udp 127.0.0.1:59449->127.0.0.1:53: i/o timeout ``` when trying to execute any command. e.g. ``` kubectl version Client Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.2", GitCommit:"<PASSWORD>", GitTreeState:"clean", BuildDate:"2018-08-07T23:17:28Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"darwin/amd64"} Unable to connect to the server: failed to refresh token: Post https://oauth2.googleapis.com/token: dial tcp: lookup oauth2.googleapis.com on 127.0.0.1:53: read udp 127.0.0.1:59449->127.0.0.1:53: i/o timeout ``` If I kill the dns-heaven process, kubectl works as before (still giving error related to dns) ``` kubectl version Client Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.2", GitCommit:"<PASSWORD>", GitTreeState:"clean", BuildDate:"2018-08-07T23:17:28Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"darwin/amd64"} The connection to the server ... was refused - did you specify the right host or port? ``` Answers: username_0: Found this log in dns-heaven output ``` ERRO[0005] error resolving query error="read udp 192.168.0.65:56801->172.16.58.3:53: i/o timeout" ns="172.16.58.3:53" qname=oauth2.googleapis.com. ERRO[0005] error resolving query error="read udp 192.168.0.65:62111->172.16.58.3:53: i/o timeout" ns="172.16.58.3:53" qname=oauth2.googleapis.com. ``` username_1: I think the problem is with your DNS server: `172.16.58.3` It may be blocking me (being in Russia and me in the USA) but I get a timeout whenever using that DNS server. username_2: I have the same trouble but with terraform aws provider. How did you get dns-heaven output?
dotnet/roslyn
863008039
Title: Need a way to detect Power Save mode Question: username_0: Currently there is no way to detect Power Save mode. Since features may operate more slowly or with reduced functionality in this mode, we need a way to uniquely detect a Power Save request from other requests (e.g. #29123).
amcharts/amcharts5
1075379355
Title: Custom Data in Tooltip With Drill Down Map Chart Question: username_0: Issue: When using custom data for continents in Map chart, the continent will become unclickable (click event will not fire). **_My Scenario_** While trying to build a drilldown map with: - Tooltip on continents will show calculated value: "{name} : {value}". - Clicking on continent will zoom-in to the continent and show the countries map. If Asia will have a value = 150, the tooltip will show "Asia : 150". ``` let continentSeries = chart.series.push(am5map.MapPolygonSeries.new(root, { geoJSON: am5geodata_continentsLow, exclude: ["antarctica"] })); continentSeries.data.setAll([{ id: "asia", name: "Asia", value: 150 }]); continentSeries.mapPolygons.template.setAll({ tooltipText: "{name}: {value}", interactive: true, }); ``` Now, when clicking Asia (or any other continent), we will zoom in and show it's contries: ``` continentSeries.mapPolygons.template.events.on("click", function (ev) { continentSeries.zoomToDataItem(ev.target.dataItem as am5.DataItem<am5map.IMapPolygonSeriesDataItem>); continentSeries.hide(); countrySeries.show(); homeButton.show(); }); ``` For each continent in the `continentSeries.data.setAll([...]);`, the click event will not fire. Thanks for reading 👍 Answers: username_1: Try moving `setData()` call **after** everything else: ```JavaScript let continentSeries = chart.series.push(am5map.MapPolygonSeries.new(root, { geoJSON: am5geodata_continentsLow, exclude: ["antarctica"] })); continentSeries.mapPolygons.template.setAll({ tooltipText: "{name}: {value}", interactive: true, }); continentSeries.mapPolygons.template.events.on("click", function (ev) { continentSeries.zoomToDataItem(ev.target.dataItem as am5.DataItem<am5map.IMapPolygonSeriesDataItem>); continentSeries.hide(); countrySeries.show(); homeButton.show(); }); continentSeries.data.setAll([{ id: "asia", name: "Asia", value: 150 }]); ``` username_0: Moving setAll call after defining click fixed the issue. Thanks! Status: Issue closed
MaelSemler/mini-cap
576427390
Title: DEV-9 Autocomplete on search fields Answers: username_1: **UI Mockup** Portrait: <img width="271" alt="Capture d’écran 2020-03-05 à 20 13 56" src="https://user-images.githubusercontent.com/35980232/76040640-eb06c080-5f1d-11ea-8bae-e5bae68d1742.png"> Landscape: <img width="731" alt="Capture d’écran 2020-03-05 à 20 14 45" src="https://user-images.githubusercontent.com/35980232/76040677-feb22700-5f1d-11ea-9b12-d56b177f27a1.png"> username_2: Acceptance Criteria - [ ] The feature must complete a user's entry - [ ] The feature must suggest all nearby places according to the user's entry - [ ] The feature must display an empty "Heart" icon (non-functional) username_2: https://github.com/username_0/mini-cap/issues/80 Status: Issue closed
bbcarchdev/spindle
142327742
Title: Add standalone spindle-migrate utility Question: username_0: This utility will perform two functions: 1. Ensure that the database schema is up to date (analogous to `twine -S` with the Spindle modules loaded) 2. Perform any one-time data migrations which aren't suitable for inclusion in the normal schema update process, because they may be long-running Answers: username_0: Sanity check after migrating proxies into `state` table is required: the number of rows in the `proxy` table should match the number of rows in the `state` table. Status: Issue closed
wso2/product-is
530218106
Title: Tenant supported claims not saved to the user profile during self registration Question: username_0: When trying the self-registration feature of Identity Server with a secondary userstore, tenant supported claims are not saved only super tenant supported claims are saved. **Steps to reproduce** 1) Create a tenant in the Identity Server 5.8.0 2) Enable self-registration with account confirmation for this tenant [1]. 3) Create a new claim and make this supported by default for the tenant. 4) Go to the dashboard app and click on Register Now. 5) In the username option provide the username as "userstoredomain/username@tenantdomain" 6) Fill in the values on the form and click on submit. 7) Login to the management console and view the user profile. You can see that the new claim value is not set. **Affected Versions** wso2is-5.8.0 [1] - https://docs.wso2.com/display/IS580/Self-Registration+and+Account+Confirmation<issue_closed> Status: Issue closed
dotnet/roslyn
416433980
Title: Suggestion mode does not work in C# Interactive view with async completion Question: username_0: **Version Used**: 2773f5e The command is not available. We did not have any tests that relied on in working (they should be added), but several tests assumed the command was available. Answers: username_1: @username_2 is this going to be fixed for 16.0/16.1 ? @ivanbasov to followup and add test our end. username_2: yes it will be fixed in 16.0
matthew-brett/delocate
1002943992
Title: Code cleanup suggestions. Question: username_0: Some ideas that could make things easier for me and future maintainers: I'd be nice to setup auto formatters like Black and isort. It's tedious to format all code manually. [doc/devel/make_release.rst](https://github.com/username_1/delocate/blob/dbb9367bcf2b8519199587b5336bcbd2d12e3c82/doc/devel/make_release.rst) should have been in a `CONTRIBUTING` file. I missed it entirely when I made the `0.9.1` release. A `CONTRIBUTING` file could also be used to explain how to add any test data. In #93 I attempted to fully type-hint the package. I never got that merged but it showed some areas that are causing issues such as `InWheelCtx.__enter__` and `back_tick`. I think `InWheelCtx` needs to be refactored or removed (or finish #46 and fix things there,) and all calls to `back_tick` should be replaced with `subprocess.run` now that the code base is on Python 3. Testing has some issues. Everything from `pytest_tools.py` breaks the pytest debugger so they'll all eventually need to be replaced with plain asserts. It'd be nice to replace `scriptrunner.py` with a more standard tool like [pytest-console-scripts](https://pypi.org/project/pytest-console-scripts/). Issue and PR templates might be a good idea. It sucks that most of the current bug reports don't mention which version of decloate they used and don't provide a wheel that can be checked. Answers: username_1: All good suggestions. Would you like me to review #93? I suppose we should keep / deprecate `backticks` for a bit, just in case, but yes, all uses should be replaced with `subprocess.run`.
quasarframework/quasar
1082612329
Title: Update event for q-toggle passes different targets Question: username_0: Greetings! I've been trying to capture status changes for q-toggle, using @update:model-value, so I can take actions with a custom function. The problem is that the custom function receives the JS event object, but the 'target' references the toggle 'thumb', rather than the toggle itself. In the custom function, I need to be able to retrieve the toggle name or id so I know what to do with it. Code here: https://jsfiddle.net/username_0/nL98x5zk/22/ Steps to reproduce the behavior: 1. Open the jsfiddle page 2. Open the browser console (not the console in jsfiddle, as it outputs too much text) 3. Flip the toggle ON & OFF to run the custom function on status change 4. Look in the console: evt.target points to the toggle 'thumb'. But in the DOM, the toggle id 'someId' and name 'someName' are set at the parent.parent and sibling level: <div class="q-toggle cursor-pointer no-outline row inline no-wrap items-center" tabindex="0" role="checkbox" aria-checked="true" id="**someId**"> <div class="q-toggle__inner relative-position non-selectable q-toggle__inner--truthy"> <input class="hidden q-toggle__native absolute q-ma-none q-pa-none" type="checkbox" name="**someName**" value="true"> <div class="q-toggle__track"> </div> <div class="**q-toggle__thumb** absolute flex flex-center no-wrap"> </div> </div> </div> **Expected behavior** Shouldn't the evt.target always point to the top element of the toggle where the id is set? **Platform (please complete the following information):** Quasar Version: @quasar/cl v1.2.1 @quasar/app Version: 3.1.10 Quasar mode: - [ X ] SPA - [ ] SSR - [ ] PWA - [ ] Electron - [ ] Cordova - [ ] Capacitor - [ ] BEX - [ ] Vite Plugin Additonal Info: - [ ] Typescript - [ ] SFC usage with `<script setup>` - [ ] Vuex - [ ] Prettier (eslint) - [ ] Standard (eslint) Tested on: - [ X ] SPA - [ ] SSR - [ ] PWA - [ ] Electron - [ ] Cordova - [ ] Capacitor - [ ] BEX OS: MacOS Node: NPM: Yarn: Browsers: Safari, Chrome iOS: Android: Electron: Answers: username_1: To put it in apple's terms, you are holding it wrong :) https://jsfiddle.net/a3suj25f/ The event is the DOM event that triggered the change. If you want to do something you don't need the DOM, you have the model. Status: Issue closed username_0: @username_1 Thank you! That makes total sense.
ipython/ipython
59180393
Title: Horse in tab completion. Question: username_0: In notebook. ``` object.__dict__.__.<tab><delay1><c><delay2><tab> ``` will complete to `copy` `copyopy` `<someting else weird>` depending on the value of `<delay1>` and `<delay2>` especially reproducible if `<delay1>` ~1s and `<delay2>` ~ (fast enough but not too much). Discover trying to fix #4777 that might be related. Answers: username_1: Should this have been closed by #7889, or is there still some issue? username_1: @username_0 shouldn't this have been closed by #7889? Status: Issue closed username_0: Oh, yes, it should I guess github does not parse the AND.
bevyengine/bevy
684252353
Title: Scale for OrthographicProjection Question: username_0: Right now, the `OrthographicProjection` seems to have the units in pixels, which makes any 3d renders very small compared to the same render using `PerspectiveProjection`. Perhaps there should be a `Scale` field added to the `OrthographicProjection` and used [here](https://github.com/bevyengine/bevy/blob/7b6fbcb932f240e7789bb4ea292cff7eb3682cc9/crates/bevy_render/src/camera/projection.rs#L79) to allow one to shrink the very wide field of view. Answers: username_1: Just ran into this too. It will use a scale attached to the entity, and `Camera2dComponents` adds a uniform scale, but since it's a *uniform* scale it scales in XYZ, which messes up the depth (in a 2D game, if scale <1.0, then sprites don't draw at all because they're outside of zmin..zmax). At minimum it needs a nonuniform scale, which means you can't use `Camera2dComponents` (or you need to manually remove the uniform scale that it creates.) Ideally I think `OrthographicProjection` should replace `window_origin` with a more explicit enum that covers more options, something like: - pixels with (0,0) in lower-left, Y up - pixels with (0,0) in top-left, Y down - pixels with (0,0) at center, Y up - H units tall and/or W units wide, with different aspect ratios handled as: (1) H exact, adjust W to preserve aspect ratio (2) W exact, adjust H to preserve aspect ratio (3) minimum of (1) or (2) (4) maximum of (1) or (2) (5) exactly H*W, so don't preserve 1:1 aspect ratio Status: Issue closed username_2: I believe this was solved by #400 getting merged :)
wellprog/TaskManagetV2
394878318
Title: Создать модуль контактов Question: username_0: Создать модуль контактов, оформить его в виде сервиса (так же как модуль таймеров) В модуле контактов должно быть поле (День рождения) который связан с модулем таймер. Все записи которые добавляются в модуль таймер в нем не должны отображаться, однако должны работать. При удалении контакта так же удаляются и таймеры.
biggora/caminte
425332368
Title: Any plan to support GraphQL? Question: username_0: The idea would be as follow GraphQL Schema to CarminteJS Shema (GraphQL -> [AST](https://www.contentful.com/blog/2018/07/04/graphql-abstract-syntax-tree-new-schema/) -> Carminte) Then support [OpenCRUD](https://www.opencrud.org/) for the queries. This is what [prisma](https://www.prisma.io/) is doing right now and as GraphQL seems to be the next generic query language, it might worth taking a look =) Answers: username_1: I just wanted to add a bit.. TypeORM and caminte both are doing most of the things in common. I guess caminte is sleek though, If you guys can add GraphQL support you guys would become awesome. username_0: In the meantime, I discovered Graphql-box https://www.npmjs.com/package/graphql-box I still did not have time to play with it but it might be a good alternative… at least for the time to have an ORM which would compile the queries.
kubernetes-sigs/cluster-api-provider-aws
1124631292
Title: Failing to build release-notes binary Question: username_0: Looks like a version incompatibility issue. ``` ➜ cluster-api-provider-aws git:(main) make release make -C hack/tools bin/release-notes mkdir -p bin go build -tags tools -o bin/release-notes k8s.io/release/cmd/release-notes go: downloading cloud.google.com/go v0.99.0 go: downloading github.com/shirou/gopsutil/v3 v3.21.12 go: downloading github.com/BurntSushi/toml v1.0.0 go: downloading golang.org/x/sys v0.0.0-20211216021012-1d35b9e2eb4e go: downloading golang.org/x/tools v0.1.9-0.20211228192929-ee1ca4ffc4da go: downloading golang.org/x/crypto v0.0.0-20220112180741-5e0467b6c7ce go: downloading golang.org/x/net v0.0.0-20211112202133-69e39bad7dc2 go: downloading google.golang.org/api v0.63.0 go: downloading google.golang.org/grpc v1.43.0 # cloud.google.com/go/storage /Users/ssavas/go/pkg/mod/cloud.google.com/go/[email protected]/storage.go:1416:53: o.GetCustomerEncryption().GetKeySha256 undefined (type *"google.golang.org/genproto/googleapis/storage/v2".Object_CustomerEncryption has no field or method GetKeySha256) make[1]: *** [bin/release-notes] Error 2 make: *** [hack/tools/bin/release-notes] Error 2 ``` /kind bug Answers: username_1: I've been wondering for a while if we should change to use the GH release notes generation instead. username_2: /assign username_0: I am using GH release to double check we are not missing any changes. Only thing is it is missing categorization, we need to go over each of them to decide bug/cleanup etc. We can possibly do something similar to what [cluster-api has ](https://github.com/kubernetes-sigs/cluster-api/tree/main/hack/tools/release)using GH release notes. username_1: I have done this with a different project, we can probably get away without custom code. You can customize the GH release notes generation using a config file placed in `.github/release.yml`. [This](https://github.com/weaveworks/flintlock/blob/main/.github/release.yml) is an example. The configuration options can be found [here](https://docs.github.com/en/repositories/releasing-projects-on-github/automatically-generated-release-notes#configuration-options). I like with the GitHub release notes generation that it also highlights first time contributors. @username_0 - i will add it to the office hours call this evening. @username_2 - if we change the focus of this to user the GitHub release notes generation is this still something you'd want to work on? username_2: Yes, I would like to work on it. username_1: Added #3170 with the details of implementing github release notes. username_0: Closing this issue as it will not be a problem once we move to GH release notes. #3170 Status: Issue closed
Nordstrom/chefdk_bootstrap
110354097
Title: Buff::Errors::InvalidConfig error when running ChefDK_bootstrap script (Berkshelf) Question: username_0: ## Error When running the boostrap script you may encounter an error like the following: ``` Installing ChefDK... C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/buff-config-1.0.1/lib/buff/config/ruby.rb:30:in `rescue in initializ e': uninitialized constant Buff::Config::Ruby::Evaluator::Chef (Buff::Errors::InvalidConfig) from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/buff-config-1.0.1/lib/buff/config/ruby.rb:25:in `initia lize' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/buff-config-1.0.1/lib/buff/config/ruby.rb:17:in `new' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/buff-config-1.0.1/lib/buff/config/ruby.rb:17:in `parse' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/buff-config-1.0.1/lib/buff/config/ruby.rb:119:in `from_ ruby' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/buff-config-1.0.1/lib/buff/config/ruby.rb:112:in `initi alize' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/ridley-4.2.0/lib/ridley/chef/config.rb:83:in `initializ e' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf.rb:90:in `new' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf.rb:90:in `chef_config' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/config.rb:80:in `<class:C onfig>' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/config.rb:4:in `<module:B erkshelf>' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/config.rb:3:in `<top (req uired)>' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/cookbook_generator.rb:3:i n `require_relative' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/cookbook_generator.rb:3:i n `<class:CookbookGenerator>' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/cookbook_generator.rb:2:i n `<module:Berkshelf>' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/cookbook_generator.rb:1:i n `<top (required)>' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf.rb:197:in `require_relati ve' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf.rb:197:in `<top (required )>' from C:/opscode/chefdk/embedded/lib/ruby/site_ruby/2.1.0/rubygems/core_ext/kernel_require.rb:54:in `require' from C:/opscode/chefdk/embedded/lib/ruby/site_ruby/2.1.0/rubygems/core_ext/kernel_require.rb:54:in `require' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/cli.rb:1:in `<top (requir ed)>' from C:/opscode/chefdk/embedded/lib/ruby/site_ruby/2.1.0/rubygems/core_ext/kernel_require.rb:54:in `require' from C:/opscode/chefdk/embedded/lib/ruby/site_ruby/2.1.0/rubygems/core_ext/kernel_require.rb:54:in `require' from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/bin/berks:3:in `<top (required)>' from C:/opscode/chefdk/bin/berks:43:in `load' from C:/opscode/chefdk/bin/berks:43:in `<main>' Starting Chef Client, version 12.4.4 resolving cookbooks for run list: ["chefdk_bootstrap"] ================================================================================ Error Resolving Cookbooks for Run List: ================================================================================ Missing Cookbooks: ------------------ No such cookbook: chefdk_bootstrap Expanded Run List: ------------------ [Truncated] ## Cause This is this [issue in Berkshelf](https://github.com/berkshelf/berkshelf/issues/965). The issue is caused by having a line like this in your `knife.rb`: `Chef::Config.from_file(knife_override) if File.exist?(knife_override)` ## Solution Replace the offending line in your knife.rb with this code: ```ruby # See this Berkshelf issue for why we can't call Chef::Config.from_file() # https://github.com/berkshelf/berkshelf/issues/965 # This code is what Chef::Config.from_file() does. if File.exists?(org_override) && File.readable?(org_override) self.instance_eval(IO.read(org_override), org_override, 1) end ``` Status: Issue closed Answers: username_0: This is not actually a bug in chefdk_bootstrap. Closing.
rapidsai/cusignal
705143365
Title: [FEA] Reevaluate Peak Finding module for performance improvements Question: username_0: The Peak Finding module need to be reevaluated to find areas of further performance improvements. Answers: username_1: Note: This includes similar investigations as explored in https://github.com/rapidsai/cusignal/pull/235. Developers should be examining: - Use of numpy over cupy for small-scale calculations like overlap size, FFT size, and other generic parameters that don't require GPU compute - Elementwise kernels, where appropriate (either raw cupy, as denoted in the above PR, or Numba) - Caching of FFT plans Status: Issue closed username_0: Closed by #267
NPellet/visualizer
160879140
Title: pasting png file and clipboard content Question: username_0: I did a "private" view (in Test flavor) : Test png: https://my.cheminfo.org/?viewURL=https%3A%2F%2Fmydb.cheminfo.org%2Fdb%2Fvisualizer%2Fentry%2Fb1d9fbc6670f5236f9ff2a00d0386f34%2Fview.json In the module drag / drop there is a "Read type" that may be Array buffer. Is this expected to work and how ? Would it be possible to use the PNGreader library in this code to check the png content of the image. If I have many "flavor" in the clipboard I would have a pop-up window ? Still looking for the meta data of the image ... It "should" be there somewhere ... Maybe there is a difference from chemdraw when you save as png or if you copy / paste the structure directly from the clipboard ... ![image](https://cloud.githubusercontent.com/assets/1484241/16150717/160a5eda-3499-11e6-9991-bf12b870ccfe.png) Answers: username_1: ArrayBuffer works only if you export the parent object (not content) otherwise it is converted to a DataObject Status: Issue closed username_0: Apparently even if you clipboard contains many flavor of the data it seems that the browser will filter out only the one that can be paste.
Submitty/Submitty
466664661
Title: Long course titles cause display issues on mobile Question: username_0: **Describe the bug** If a course has a long name the course list page looks weird. The blue buttons around the title don't read which causes have the title to be unreadable. There's also a horizontal scroll wheel which probably shouldn't be there on mobile **Expected behavior** I'm not sure whats a good way to handle this but my thoughts are are to wrap the text around. I think scaling things down might make it too hard to read. **To Reproduce** Steps to reproduce the behavior: 1. Login on mobile 4. See error **Configuration** - Android chrome/firefox **Screenshots** If applicable, add screenshots to help explain your problem. <img src= "https://user-images.githubusercontent.com/12129065/61023301-5f7d6580-a377-11e9-856d-7bbf3bc3afc3.png" height = 800 /> Horizontal scroll: <img src ="https://user-images.githubusercontent.com/12129065/61023300-5db3a200-a377-11e9-9313-5e7822cb7191.png" height = 800 /> **Additional context** Add any other context about the problem here.<issue_closed> Status: Issue closed
OBOFoundry/OBOFoundry.github.io
229084707
Title: Visibility of 'replaced by' on Foundry page Question: username_0: On the OBO Foundry page there are a number of ontologies listed as obsolete. Right now, if there is a suitable other ontology in the same domain, that information is buried one click away under 'details'. I think it would be useful to make this directly visible. Answers: username_1: I can see the argument for that; on the other hand, "details" reveals a bunch of other fields, and the replacement ontology is just one of those. username_0: True, though one could argue that--for obsolete ontologies--what other ontology would be of most interest is probably the KEY piece of information. I don't feel strongly about it, though I suspect if this was directly displayed we would probably have more cases of replacement ontologies indicated. username_1: Is this still relevant and desired? username_2: I will put it on the list for @ddooley team - this should be easy to resolve. (I am thinking mouseover tooltip type solution).
Thom1729/Sublime-JS-Custom
553484969
Title: Highlight function calls Question: username_0: I've been using JavaScript Next for ages, and have really got used to it. But it's old, and now doesn't support things like arrow function parameter default values which are function calls amongst [other things](https://github.com/Benvie/JavaScriptNext.tmLanguage/issues/4). So I moved to JS Custom, but there is one thing that niggles me—I'm really used to function _calls_ highlighted as function names. Is it possible to fix with a local customisation? Or does it need a change to the syntax file itself? Compare JavaScript Next on left vs JS Custom on right. The `thing` function definition and parameters look great, but I'd like the calls to also be green... ![image](https://user-images.githubusercontent.com/900542/72890987-1be2bb80-3d0b-11ea-90f2-1f1962295ef5.png) Answers: username_1: The function names in the calls are scoped `variable.function.js`, which should be colored by most schemes. Your color scheme looks like Monokai, but I've just checked and the version of Monokai bundled with Sublime does highlight the function calls. Are you using a third-party color scheme? username_0: I have both Monokai and Monokai (SL) installed—not sure if either came with JS Next. But both those schemes colour function calls in JS Next and don't in JS Custom. I've removed JS Next, cleared the \Cache fodler and restarted ST, but still no coloring of function calls. Then I installed Monokai++, and function call coloring is there. So I suspect I just have a random collection of old theme files which are out of date :). Sorry for the interruption! Still no idea why JS Next worked, unless it patched the color scheme when active... Anyway, problem solved. Status: Issue closed
mike-lischke/antlr4-c3
250013972
Title: Some further usage examples Question: username_0: Not a bug, a request. This project looks very hopeful for our usage, thanks. To be really useful it would help to have some more concrete use cases on how to retrieve the incomplete symbol and its context, for example (from my experiments, where I'm not sure what to do. I'm writing a simple filter grammar (imagine something grepping through logfiles) and want to auto-complete on something like: fu = "incomplete string When finding possible completions, I'd like to be able to have the context that I'm comparing a field called 'fu' (and thus be able to query all known values of 'fu') and that the token I'm trying to complete on is "incomplete string Similarly we might not have the double-quotes on the incomplete string if we want the user to be able to type in a string without quotes... An example or test case tying together the grammar with the symbol table and extracting the parse context + the half-complete symbol would be really helpful, I'm a little stuck here... As an aside, if interested I've ported the completion core and tests to Java and may be able to release them back to this project. If so please context me. Answers: username_1: An incomplete string is a syntax errror and the walker might not find the path to the invocation point for candidate collection. But I haven't tested this scenario yet, as you usually don't have completion for strings (and literals in general). Have to think a bit how to get a context for that symbol. Regarding the Java port: yes, sure, I'm interested in that. There is already a Kotlin port of the code (and I've written myself a C++ variant). I could publish all that together in this repo. username_0: Hi Mike, Would it be possible for the walker to record the matching text and/or the set of tokens matching a given preferred rule, which would give the ability for the code trying to do completion to get not just the most recent token but the preferred rule’s full string? In the case I describe this would mean that we could have a rule called ‘partial_string_value’ matching an optional quote then a white-spaced string, and this would be a preferred rule which I could use in auto-complete and I’d be able to get the value. Today I can’t get the value since I can’t just use the last token, I need all the tokens that matched the ‘partial_string_value’ rule. I guess that information must be available during the processRule evaluation? I’ve started getting approvals to give back the java port, will keep you informed… I guess we could then move each port into separate sub-directories, I could prepare a cloned repo with a merge request once approvals are in place, which hopefully should be pretty quick. Nick username_0: I believe I've found out how to extract the multi-token rule-string value as required to be able to do completions on things which span more than one token such as the given example. It's done as an extension to collectCandidates() and a new field in the candidates object which maps a given candidate rule to its string fragment. In my java implementation it looks like this: ~~~ this.processRule(this.atn.ruleToStartState[startRule], 0, callStack, "\n"); // now post-process the rule candidates and their last occurrences to find the rule strings for (int ruleId : candidates.rules.keySet()) { Map<Integer, Set<Integer>> shortcut = shortcutMap.get(ruleId); if (shortcut == null) { continue; } // select the right-most occurrence - the highest entry int startToken = Collections.max(shortcut.keySet()); Set<Integer> endSet = shortcut.get(startToken); int endToken = endSet.isEmpty() ? caretTokenIndex - currentIndex : Collections.max(shortcut.get(startToken)); // extract the matching text candidates.ruleStrings.put(ruleId, tokenStream.getText(tokenStream.LT(startToken), tokenStream.LT(endToken))); } ~~~ then for each candidate rule I have the last string in the parsed input which matched that rule. HTH Nick username_1: If I understand this right you want to extract the longest text that was matched when walking a candidate rule. What are you going to do with that text? It's not something you would show during code completion. Also, often candidate rules are not completely executed (since they are candidates, the caret position is somewhere in the text covered by those rules) and hence have no entry in the shortcutMap. username_1: Regarding the contribution of your Java code: glad to see the approval process has started. I'm still thinking about the repo structure and where to place the ports, since we must ensure we don't break the Node.js module setup. Maybe we just add a new folder called `ports` or `contributions` and exclude that in the package.json file? username_0: Hi Mike, The idea is to extract the most recent (closest to the caret position) and longest text that matches when walking a candidate rule, so that completions of multi-token rules can have the partially-completed text… I’m still experimenting with this… it’s not yet seeming to get all the string values I’d like, maybe your comment identifies the problem. Here are a couple of use-cases for a technical-english grammar which performs filtering of log lines: Imagine a log line like this: Aug 19 00:07:00 info nicks-mbp cron[17522]: no path for address 0x10fab9000 And a filter like this: log text contains no path for We might want to detect that the field being filtered on is the ‘log text’ field (which contains the full line) and the value to filter on is ‘no path for’ which in itself is a legal filter but which could be completed by scanning logs and finding that it could be completed to become ‘no path for address 0x10fab9000’ (or other addresses) based upon frequency of log messages. To do this requires finding the ‘no path for’ text for the ‘value’ rule, and indeed also finding the ‘log text’ for the ‘field’ rule to give context for the search. Similarly if typing: log t I’d like to complete with ‘log text’ and ‘log timestamp’ And typing: log priority is I’d like to propose completion with the values from the ‘priority’ field of log files, based upon their occurrence in the logs (e.g. ‘info’ in the above line). Hope this is clear! Concerning where to put in the source code base, I agree that separate subdirectories would be good, up to you if you want to create a ‘ports’ subdirectory, let me know… Thx Nick Hi Mike, The idea is to extract the most recent (closest to the caret position) and longest text that matches when walking a candidate rule, so that completions of multi-token rules can have the partially-completed text… I’m still experimenting with this… it’s not yet seeming to get all the string values I’d like, maybe your comment identifies the problem. Here are a couple of use-cases for a technical-english grammar which performs filtering of log lines: Imagine a log line like this: Aug 19 00:07:00 info nicks-mbp cron[17522]: no path for address 0x10fab9000 And a filter like this: log text contains no path for We might want to detect that the field being filtered on is the ‘log text’ field (which contains the full line) and the value to filter on is ‘no path for’ which in itself is a legal filter but which could be completed by scanning logs and finding that it could be completed to become ‘no path for address 0x10fab9000’ (or other addresses) based upon frequency of log messages. To do this requires finding the ‘no path for’ text for the ‘value’ rule, and indeed also finding the ‘log text’ for the ‘field’ rule to give context for the search. Similarly if typing: log t [Truncated] And typing: log priority is I’d like to propose completion with the values from the ‘priority’ field of log files, based upon their occurrence in the logs (e.g. ‘info’ in the above line). Hope this is clear! Concerning where to put in the source code base, I agree that separate subdirectories would be good, up to you if you want to create a ‘ports’ subdirectory, let me know… Thx Nick On 19/08/2017, 11:50, "<NAME>" <<EMAIL><mailto:<EMAIL>>> wrote: If I understand this right you want to extract the longest text that was matched when walking a candidate rule. What are you going to do with that text? It's not something you would show during code completion. Also, often candidate rules are not completely executed (since they are candidates, the caret position is somewhere in the text covered by those rules) and hence have no entry in the shortcutMap. username_1: This looks very much like completing keywords. Why do you want to scan your input first before offering `text` and `timestamp` as candidates, if you know these are valid values? username_0: Let's close the issue. Thanks! Status: Issue closed username_2: Very appreciate about the work! I came across a problem and believe walker should have ability to go further even if there is syntax error before caret position. Suppose I have defined mysql grammar. And editing the following sentence, ```sql select from | ``` where `'|'` represents caret position and it will get nothing by calling `collectCandidates`. Well, I can still do a little work to make it work. But It is also reasonable the walker keep walking if there is potential road to current request position. And I believe this is achievable by search algorithm like A-star. When walker walked to `select` and find no bridge to `from`, it search all neighbor bridges until it finally reach `from`. And this search should be constrained because it maybe very time-consuming. Is it make sense? Thanks! username_1: @username_2 Can you please open a new request instead of hijacking a closed one?
stackblitz/core
807118431
Title: Github angular project import results in mismatching devDependencies Question: username_0: When importing an angular project via the github import integration, devDependencies are added to the package.json file, which mismatch the given angular version. Input package.json file from our project (https://github.com/dynatrace-oss/barista-examples/tree/master/examples/DtExampleAlertDark) ``` { "name": "DtExampleAlertDark", "dependencies": { "@angular/animations": "^11.0.5", "@angular/cdk": "^11.0.3", "@angular/common": "^11.0.5", "@angular/compiler": "^11.0.5", "@angular/core": "^11.0.5", "@angular/forms": "^11.0.5", "@angular/localize": "^11.0.5", "@angular/router": "^11.0.5", "@angular/platform-browser": "^11.0.5", "@angular/platform-browser-dynamic": "^11.0.5", "core-js": "^3.8.3", "d3-scale": "^3.0.0", "d3-shape": "^1.3.5", "rxjs": "^6.6.3", "tslib": "^2.0.0", "zone.js": "~0.10.2", "@dynatrace/barista-components": "latest", "@dynatrace/barista-fonts": "latest", "@dynatrace/barista-icons": "latest", "highcharts": "^7.2.1", "lodash-es": "^4.17.15" }, "devDependencies": { "@angular-devkit/build-angular": "0.1100.5", "@angular/cli": "^11.0.5", "@angular/compiler-cli": "^11.0.5" } } ``` When importing this via the github import url structure: https://stackblitz.com/github/dynatrace-oss/barista-examples/tree/master/examples/DtExampleAlertDark/ the resulting package.json lists angular 9 devDependencies instead of the given angular 11 ones. This mismatch leads to broken projects. When creating a new project in angular via the UI, the dev dependencies are set correctly to angular 11 ones. ``` { "name": "aoniryjnz.github", "version": "0.0.0", "private": true, "dependencies": { "rxjs": "6.6.3", "tslib": "2.1.0", "core-js": "3.8.3", "zone.js": "0.10.3", "d3-scale": "3.2.3", "d3-shape": "1.3.7", "lodash-es": "4.17.20", "highcharts": "7.2.2", "@angular/cdk": "11.2.0", "@angular/core": "11.2.0", [Truncated] "@angular/compiler-cli": "~9.1.1", "@angular/language-service": "~9.1.1", "@types/node": "^12.11.1", "@types/jasmine": "~3.5.0", "@types/jasminewd2": "~2.0.3", "codelyzer": "^5.1.2", "jasmine-core": "~3.5.0", "jasmine-spec-reporter": "~4.2.1", "karma": "~4.4.1", "karma-chrome-launcher": "~3.1.0", "karma-coverage-istanbul-reporter": "~2.1.0", "karma-jasmine": "~3.0.1", "karma-jasmine-html-reporter": "^1.4.2", "protractor": "~5.4.3", "ts-node": "~8.3.0", "tslint": "~6.1.0", "typescript": "~3.8.3" } } ```
OpenMined/PySyft
684082237
Title: Add torch.Tensor.q_per_channel_scales to allowlist and test suite Question: username_0: # Description This issue is a part of Syft 0.3.0 Epic 2: https://github.com/OpenMined/PySyft/issues/3696 In this issue, you will be adding support for remote execution of the torch.Tensor.q_per_channel_scales method or property. This might be a really small project (literally a one-liner) or it might require adding significant functionality to PySyft OR to the testing suite in order to make sure the feature is both functional and tested. ## Step 0: Run tests and ./scripts/pre_commit.sh Before you get started with this project, let's make sure you have everything building and testing correctly. Clone the codebase and run: ```pip uninstall syft``` followed by ```pip install -e .``` Then run the pre-commit file (which will also run the tests) ```./scripts/pre_commit.sh``` If all of these tests pass, continue on. If not, make sure you have all the dependencies in requirements.txt installed, etc. ## Step 1: Uncomment your method in the allowlist.py file Inside [allowlist.py](https://github.com/OpenMined/PySyft/blob/syft_0.3.0/src/syft/lib/torch/allowlist.py) you will find a huge dictionary of methods. Find your method and uncomment the line its on. At the time of writing this Issue (WARNING: THIS MAY HAVE CHANGED) the dictionary maps from the string name of the method (in your case 'torch.Tensor.q_per_channel_scales') to the string representation of the type the method returns. ## Step 2: Run Unit Tests Run the following: ```python setup.py test``` And wait to see if some of the tests fail. Why might the tests fail now? I'm so glad you asked! https://github.com/OpenMined/PySyft/blob/syft_0.3.0/tests/syft/lib/torch/tensor/tensor_remote_method_api_suite_test.py In this file you'll find the torch method test suite. It AUTOMATICALLY loads all methods from the allowlist.py file you modified in the previous step. It attempts to test them. # Step 3: If you get a Failing Test If you get a failing test, this could be for one of a few reasons: ### Reason 1 - The testing suite passed in non-compatible arguments The testing suite is pretty dumb. It literally just has a permutation of possible arguments to pass into every method on torch tensors. So, if one of those permutations doesn't work for your method (aka... perhaps it tries to call your method without any arguments but torch.Tensor.q_per_channel_scales actually requires some) then the test will fail if the error hasn't been seen before. If this happens - don't worry! Just look inside the only test in that file and look [Truncated] pointer objects to very many remote object types. So, if your method returns anything other than a single tensor, you probably need to add support for the type it returns (Such as a bool, None, int, or other types). *IMPORTANT:* do NOT return the value itself to the end user!!! Return a pointer object to that type! *NOTE:* at the time of writing - there are several core pieces of Syft not yet working to allow you to return any type other than a torch tensor. If you're not comfortable investigating what those might be - skip this issue and try again later once someone else has solved these issues. ### Reason 3 - There's something else broken Chase those stack traces! Talk to friends in Slack. Look at how other methods are supported. This is a challenging project in a fast moving codebase! And don't forget - if this project seems to complex - there are plenty of others that might be easier.<issue_closed> Status: Issue closed
vmware-tanzu/velero
945898591
Title: Got "velero container not found in velero deployment" error when trying to add plugin Question: username_0: I found the following issue during running the E2E test: I installed velero on TKG which created two deployments under the namespace `velero`: `velero` and `backup-driver` And when I tried to add another plugin, got the error: `velero container not found in velero deployment` I think the root cause should be here: https://github.com/vmware-tanzu/velero/blob/main/pkg/cmd/cli/plugin/add.go#L100 https://github.com/vmware-tanzu/velero/blob/main/pkg/cmd/cli/plugin/helpers.go#L49 velero cli tries to get the `velero` deployment by label selector `"component": "velero"`, but in this case both `velero` and `backup-driver` are labeled with `"component": "velero"`, this cuases the `backup-driver` is returned rather than `velero` If this is the case, maybe we can filter the velero deployment by both label and name? **Vote on this issue!** This is an invitation to the Velero community to vote on issues, you can see the project's [top voted issues listed here](https://github.com/vmware-tanzu/velero/issues?q=is%3Aissue+is%3Aopen+sort%3Areactions-%2B1-desc). Use the "reaction smiley face" up to the right of this comment to vote. - :+1: for "I would like to see this bug fixed as soon as possible" - :-1: for "There are more important bugs to focus on right now" Answers: username_0: Seems that comparing the name isn't the correct way either. The naming comparing logic is removed by the https://github.com/vmware-tanzu/velero/pull/3447 because the name of velero installed by helm chart can be customized. Status: Issue closed
robot-lab/judyst-main-web-service
371738007
Title: Настройки для организации Question: username_0: # Task request ## Цель задачи Дать возможность менять настройки организации. На основе feature #55 ## Решение задачи Представление которое обрабатывает следующие запросы: изменение имени организации, удаление членов организации, удаление доступных файлов, удаление организации. ## Дополнительный контекст или ссылки на связанные с данной задачей issues
Wynncraft/Issues
220491655
Title: The Envoy Part I Question: username_0: Heading through an area following Maxie I got stuck between two walls of barrier blocks. This is while heading to her where the Quest Book updates to "Stop the mechs from taking over the diner.". This is not the first quest that has given us a lot of issues due to barrier blocks. I have tried /c;lass and coming back and switching the server. Answers: username_1: Try doing /kill with 5 or more SP username_0: I did /kill. That is not the point, I should not have to keep killing myself because of constant issues in quests with the over use of barrier blocks. This has happened on a lot of quests and it is becoming ridiculous now.
libtcod/libtcod
548934343
Title: Remove C++ dependencies. Question: username_0: As much as I liked using C++ over C it has made the code less portable and made the C ABI harder to maintain. I now plan on rewriting the code in C and porting to C++ instead of the other way around. The following needs to be rewritten in C: - [ ] New tileset objects. - [ ] Tileset font loader. - [ ] TrueType font loader. - [ ] SDL2 renderer. - [ ] OpenGL2 renderer. - [ ] Console UTF-8 functions. - [ ] Console drawing functions. - [ ] Console initialization. - [ ] Image functions. Answers: username_1: This would definitely help in supporting emscripten/wasm too. What's your ETA on this? username_0: I thought Emscripten could compile C++. If anything currently prevents an emscripten/wasm build then feel free to create an issue for it. Some of the refactoring is hard to do since a lot of the rendering functions depend on each other, so those all have to be done at once before I can test and upload it. No real ETA, but I'm trying to get as much as I can done before the February 7DRL. This _might_ be done before the end of January, but there will still be a lot of compiled C++ code left over after this task is done. username_1: Emscripten can compile C++ but it’s a bit of a mess when there are external libraries and C++14 involved. Some standard libraries like zlib and SDL2 have already been ported but others like GLAD and probably others would need to be ported manually. The best would be to reduce the number of external dependencies to the bare minimum. On top of that, a port to pure C would be nice for people like me that usually avoid C++ username_0: GLAD is for the desktop builds, and it keeps the function names standard so that I can switch between OpenGL and OpenGLES easier. I just need to add the preprocessor code that actually does that. Status: Issue closed username_0: It's done, all the major components compile and run without the C++ sources. Right now it's only python-tcod which is setup to compile without C++. I want to test Emscripten but I've been unable to setup the tool-chain so far. I expect it to fail, and that I'll need to switch the headers to use GLES. That kind of thing will be tracked in issue #41 if I make more progress. username_1: Nice! ;)
flutter/flutter
1023535710
Title: Flutter error "type 'bool' is not a subtype of type 'List<dynamic>' " Question: username_0: I am getting while running flutter/dart program "type 'bool' is not a subtype of type 'List<dynamic>' ". The code is as follows: import 'package:flutter/material.dart'; import 'package:http/http.dart' as http; import 'dart:convert'; void main() => runApp(MainEntry()); class MainEntry extends StatelessWidget { // const MainEntry({Key? key}) : super(key: key); @override Widget build(BuildContext context) { return MaterialApp( theme: ThemeData( backgroundColor: Colors.white70, primarySwatch: Colors.blue, brightness: Brightness.light, ), debugShowCheckedModeBanner: false, title: 'flutter', home: Entry(), ); } } class Entry extends StatefulWidget { @override _EntryState createState() => _EntryState(); } class _EntryState extends State<Entry> { getMethod() async { String theUrl = 'http://ananyananu.000webhostapp.com/getData.php'; var res = await http.get(Uri.parse(theUrl), headers: {"Accept":"application/json"}); var responsBody = json.decode(res.body); print ('I reached here'); print('responsBody'); return responsBody; } @override Widget build(BuildContext context) { return Scaffold( backgroundColor: Colors.grey[200], appBar: AppBar( title: Text('Connect to Local Host') , centerTitle: true, backgroundColor: Colors.redAccent, ), body: FutureBuilder( future: getMethod(), builder: (BuildContext context, AsyncSnapshot snapshot) { if (snapshot.connectionState == ConnectionState.waiting) { [Truncated] return Center( child: Text('Error Fetching Data'), ); } List snap = snapshot.data; return ListView.builder( itemCount: snap.length, itemBuilder: (context, index) { return ListTile( title: Text("head: ${snap[index]['heading']}"), subtitle: Text("body ${snap[index] ['body']}"), ); }, ); }), ); } }
jlippold/tweakCompatible
483395718
Title: `Amber` working on iOS 12.4 Question: username_0: ``` { "packageId": "com.ps.amber", "action": "working", "userInfo": { "arch32": false, "packageId": "com.ps.amber", "deviceId": "iPhone8,4", "url": "http://cydia.saurik.com/package/com.ps.amber/", "iOSVersion": "12.4", "packageVersionIndexed": true, "packageName": "Amber", "category": "Tweaks", "repository": "A Main Repo", "name": "Amber", "installed": "0.0.2.9", "packageIndexed": true, "packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.", "id": "com.ps.amber", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.1.5", "shortDescription": "Amber-ify your LED torch.", "latest": "0.0.2.9", "author": "PoomSmart", "packageStatus": "Unknown" }, "base64": "<KEY>", "chosenStatus": "working", "notes": "" } ```<issue_closed> Status: Issue closed
hail-is/hail
526125283
Title: breaking hl.rand_unif via CSE Question: username_0: shoudn't be terribly surprising that `rand_unif` has weird behavior, but here's a case that is definitely The Wrong Thing: ``` Python 3.6.0 |Continuum Analytics, Inc.| (default, Dec 23 2016, 13:19:00) Type 'copyright', 'credits' or 'license' for more information IPython 7.6.1 -- An enhanced Interactive Python. Type '?' for help. In [1]: import hail as hl In [2]: r = hl.rand_unif(0, 1) In [3]: hl.eval(r) Out[3]: 0.5387579341676381 In [4]: hl.eval(hl.tuple([r, r])) Out[4]: (0.5387579341676381, 0.5387579341676381) ``` okay, this makes sense becuase they have the same seed: ``` In [5]: print(hl.tuple([r, r])._ir) (MakeTuple (0 1) (ApplySeeded rand_unif 806694938962853089 Float64 (Apply toFloat64 Float64 (I32 0)) (Apply toFloat64 Float64 (I32 1))) (ApplySeeded rand_unif 806694938962853089 Float64 (Apply toFloat64 Float64 (I32 0)) (Apply toFloat64 Float64 (I32 1)))) ``` how about this: ``` In [6]: hl.eval(hl.range(2).map(lambda x: r)) Out[6]: [0.5387579341676381, 0.9394799645512691] ``` odd. but maybe rand_unif inside an iteration has some semantics for advancing the RNG (like an aggregation). ``` In [7]: p = 1 - r In [8]: hl.eval(hl.range(2).map(lambda x: p)) Out[8]: [0.46124206583236194, 0.06052003544873086] ``` ok... ``` In [9]: hl.eval((p, hl.range(2).map(lambda x: p))) Out[9]: (0.46124206583236194, [0.46124206583236194, 0.46124206583236194]) ``` wtf? if you look in the logs, its explained by the fact that only the final IR triggers CSE: ``` (Let __cse_1 (ApplyBinaryPrimOp Subtract (ApplyIR toFloat64 Float64 (I32 1)) (ApplySeeded rand_unif 806694938962853089 Float64 (ApplyIR toFloat64 Float64 (I32 0)) (ApplyIR toFloat64 Float64 (I32 1)))) (MakeTuple (0 1) (Ref __cse_1) (ArrayMap __uid_5 (ArrayRange (I32 0) (I32 2) (I32 1)) (Ref __cse_1)))) ``` Answers: username_1: I think I'd flip the logic. I'm not sure if this one is wrong: ``` In [9]: hl.eval((p, hl.range(2).map(lambda x: p))) Out[9]: (0.46124206583236194, [0.46124206583236194, 0.46124206583236194]) ``` But if it's right, clearly this one is wrong: ``` In [7]: p = 1 - r In [8]: hl.eval(hl.range(2).map(lambda x: p)) Out[8]: [0.46124206583236194, 0.06052003544873086] ``` username_2: The latter seems wrong. It violates the identity of `p`. I prefer the semantics of the former. username_1: I agree. It's extremely hard for us to provide these semantics though -- we'd need to introspect on the Python AST of the lambda username_2: It seems the right answer is that CSE cannot lift side-effecting operations outside of `map`. username_3: so just as an aside, the behavior of all the random functions is actually documented in the hail docs: https://hail.is/docs/0.2/functions/random.html The one thing that I apparently didn't write up is how it's supposed to behave in an array context. The way it's currently intended to work is as executed in: ``` In [8]: hl.eval(hl.range(2).map(lambda x: p)) Out[8]: [0.46124206583236194, 0.06052003544873086] ``` since, as Tim points out, it's kind of hard to differentiate the scope of `hl.range(2).map(lambda x: p)` and `hl.range(2).map(lambda x: hl.rand_unif(0, 1))` if p is only used once. (This behavior is also pretty consistent with treating iteration through an array value as the same as iteration through an axis of a table or a matrix table, since that's exactly what happens there.) The way you'd use the same value of a random number in an array map, previously, was with bind: ``` hl.bind(lambda p: hl.range(2).map(lambda x: p), hl.rand_unif(0, 1)) ``` I'm not sure what a bind-free answer to this would be short of implementing a way of differentiating between: ``` p = hl.rand_unif(0, 1) hl.range(2).map(lambda x: p) ``` and ``` hl.range(2).map(lambda x: hl.rand_unif(0, 1)) ``` which would be nice, but also definitely a breaking change. Status: Issue closed
CompuCell3D/CompuCell3D
676471958
Title: FPP links rendering Question: username_0: The links are NOT drawn properly if you chance the view to 3D or the plane of view (they are always rendered as if in x-y view no matter what) Answers: username_0: Many people noticed odd behavior with FPP in CC3D during the class. FPP edges weren’t being displayed properly in player, the edges seem to come and go randomly and other odd behaviors. I have a simulation now that appears to be changing the membership in the cell list by type data. For example; for cell1 in self.cell_list_by_type(self.ACELL): print("\t\t\t cell1.id,",cell1.id) for cell2 in self.cell_list_by_type(self.BCELL): d=cell_cell_dist(cell1,cell2) self.focalPointPlasticityPlugin.createFocalPointPlasticityLink(cell1, cell2, 10, d, 500) print("\t\t\t cell2.id",cell2.id) # calculate the distance between two cells, cell1 and cell2 def cell_cell_dist(cell1,cell2): import math d=math.sqrt((cell1.xCOM-cell2.xCOM)**2+(cell1.yCOM-cell2.yCOM)**2+(cell1.zCOM-cell2.zCOM)**2) return(d) Gives as output: Len acell and bcell lists: 4 5 acell lists: 1 2 5 8 bcell lists: 3 4 6 7 9 cell1.id, 1 cell2.id 3 cell2.id 4 cell2.id 6 cell2.id 7 cell2.id 9 [Truncated] cell2.id 7 cell2.id 9 Len acell and bcell lists: 4 5 acell lists: 1 2 5 8 bcell lists: 3 4 6 7 9 Notice the highlighted values are not cell.ids that are on the respective list and a value is repeated. The above behavior changes every run (even though I have an explicit seed set in the Potts section). username_0: [FPP_test_Aug_2020b.zip](https://github.com/username_0/username_0/files/5053923/FPP_test_Aug_2020b.zip) username_1: This might not be an FPP problem. It looks like CC3D doesn't properly process cell lists by type whe you embed them in cascade loops: (pseudo code) ``` for cells of type a for cells of type b [gives a corrupted iteration list for the type a cells] ```
FAForever/downlords-faf-client
556412515
Title: add mapping rules and validation tool as link to upload process Question: username_0: Hello, I'm missing a link to validation tool (https://api.faforever.com/maps/validate) in addition to the already added mapping rules (https://forums.faforever.com/viewtopic.php?f=2&t=17873) in the upload window/process. It helps you check your map name and content before upload. Currently we are enforcing the rules while uploading a map and return these error messages. But the validation tool is not linked. https://cdn.discordapp.com/attachments/622773120377421851/671783149092601875/unknown.png https://cdn.discordapp.com/attachments/622773120377421851/671783474407014411/unknown.png ### **My Dream Upload** I believe the best error free upload can be achieved when we add the validation tool or its features to the upload process. Here are some ideas.. **improve pre upload process** 0)hit `check map` button ![image](https://user-images.githubusercontent.com/28671151/73296196-81eaa980-4209-11ea-99d0-1814b145c1a3.png) A.1) open link to validation tool: https://api.faforever.com/maps/validate A.2) manual user input or B.1) select map B.2) auto-fill validation tool and show results **improve upload process** 0)hit `upload map` button 1) select map (add buttons or autodecide if the map source (folder, files, path) has `.v000x` in files and folder -> this way removing `.v0001` before upload is not a thing anymore) 2) auto-fill validation tool and show results 3) if rules are met the upload is possible, otherwise errors are shown Answers: username_1: Imo with the new (?) restrictions on naming file names should also be renamed to `-save` `-scenario` `-script` to be consistent username_2: I added a button for now, a test within the client would be the best solution, but I advocate against that. It'll take way too much time away from more important developement tasks. username_0: thx for the button. what exactly are your referring to? "a test within the client would be the best solution, but I advocate against that." username_2: A system within the client that checks what you have selected and tells you what's wrong would be the best solution. But this takes too much time to be worth it for the small amount of people that use the map upload. Status: Issue closed username_2: Hello, we are missing a link to the validation tool (https://api.faforever.com/maps/validate) in addition to the already added mapping rules (https://forums.faforever.com/viewtopic.php?f=2&t=17873) in the upload window/process. ![image](https://user-images.githubusercontent.com/28671151/73301193-9f704100-4212-11ea-9332-167350476ef2.png) The validation tool helps you check and validate your map name / content before upload. It also provides the right file names and pathing. Currently we are enforcing the rules while uploading a map and return these error messages. But the validation tool is not linked and some people don't know why they get this issue or how to solve it. ![name_restriction](https://user-images.githubusercontent.com/28671151/73297157-58cb1880-420b-11ea-9c43-a13562b260f4.png) ![name_okay_names_given](https://user-images.githubusercontent.com/28671151/73297160-5a94dc00-420b-11ea-85b1-ae4e4fc40a2a.png) Morax also supports the idea of the link. ### **My Dream Upload** While writing this I like to bring up a long desired dream of mine ;) I believe the best error free upload can be achieved when we add the validation tool or its features to the upload process. Here are some ideas.. **improve pre upload process** Encourage people to validate the map (name, scenario) before upload... Option A: 1) hit `validate/check map` button (needs to be added next to the `upload map` button) ![image](https://user-images.githubusercontent.com/28671151/73296196-81eaa980-4209-11ea-99d0-1814b145c1a3.png) 2) open link to validation tool: https://api.faforever.com/maps/validate 3) manual user input Option B: 1) hit `validate/check map` button (needs to be added next to the `upload map` button) 2) select map 3) auto-fill validation tool and show results online **improve upload process** 1) hit `upload map` button 2) select map 3) auto-fill validation tool and show results in upload window 4) if rules are met and no errors remain the upload is possible, otherwise errors are displayed and explained 4.1) server version: complete rewrite of the map files names (folder, scmap, save, script, options, tables, ...) and path (scmap, scenario, script,...) based on validated `name = "map name"` from scenario.lua -> this way removing `.v0001` (from folder and in files) before upload is not a thing anymore Status: Issue closed
gpe-mw-ansible-org/rh-sso-multi-realm
455941912
Title: new kie-realm users and groups Question: username_0: -u 'Super' -p '<PASSWORD>!' -g 'kie-server,admin,Administrators' -u 'Ava' -p '<PASSWORD>!' -g 'kie-server,applicant' u 'Bill' -p '<PASSWORD>!' -g 'kie-server,applicant' -u 'Tina' -p '<PASSWORD>!' -g 'kie-server,talent-acquisition' -u 'Tom' -p '<PASSWORD>!' -g 'kie-server,talent-acquisition' -u 'Eve' -p '<PASSWORD>!' -g 'kie-server,interviewer' -u 'Ann' -p '<PASSWORD>!' -g 'kie-server,interviewer' -u 'Bob' -p '<PASSWORD>!' -g 'kie-server,interviewer'`
BitBagCommerce/SyliusElasticsearchPlugin
611943378
Title: Unable to use checkbox filter Question: username_0: On product, I have many checkbox attributes. When I check "1" in the filter and I click on "Filter" button, I have this error : ``` An exception has been thrown during the rendering of a template ("failed to create query: { "bool": { "must": [ { "term": { "enabled": { "value": true, "boost": 1.0 } } }, { "terms": { "channels": [ "sylius" ], "boost": 1.0 } }, { "terms": { "product_taxons": [ "01" ], "boost": 1.0 } }, { "range": { "price_sylius": { "from": 0, "to": 9223372036854775807, "include_lower": true, "include_upper": true, "boost": 1.0 } } }, { "bool": { "should": [ { "term": { "attribute_promotion": { "value": "1", "boost": 1.0 } } } ], "adjust_pure_negative": true, "boost": 1.0 } }, { "bool": { "should": [ [Truncated] "term": { "attribute_news_auto": { "value": "9a9bf86a_8e14_11ea_ad31_0242ac120005", "boost": 1.0 } } } ], "adjust_pure_negative": true, "boost": 1.0 } } ], "adjust_pure_negative": true, "boost": 1.0 } } [index: bitbag_shop_products_dev] [reason: all shards failed]"). ``` When I replace the "1" value by `true` in the query string, the filter work fine but the filter form is not checked. Answers: username_1: Here's a quick POC in case someone needs to implement this feature https://github.com/username_1/SyliusElasticsearchPlugin/commit/d006d35d530e81c59fab45e065efc7122f50b422 username_2: Issued a PR to fix this issue
jgromes/RadioLib
451856260
Title: RFM95 aka SX127x not detected Question: username_0: Using [SparkFun LoRa Gateway - 1-Channel (ESP32)](https://www.sparkfun.com/products/15006) NSS = 16, DIO0 = 26, DIO1 = 33, DIO2 = 32 according to [schematic](https://cdn.sparkfun.com/assets/5/3/1/9/c/ESP32_LoRa_1_Channel_Gateway.pdf) Setting up the connection to the RFM95 module: ```cpp RFM95 lora = new Module(16, 26, 33, 32); void setup() { int state = lora.begin(868.0, 125.0, 9, 5, 0x3444, 22, 8); if (state == ERR_NONE) { Serial.println("success!"); } else { Serial.print("failed: "); printError(state); return false; } ... } ``` fails with ```log ===================================== RFM95 sender test ===================================== DeviceID: 84 0D 8E 0C 56 3C 00 00 Initialize RFM95 SX127x not found! (1 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011 SX127x not found! (2 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011 SX127x not found! (3 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011 SX127x not found! (4 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011 SX127x not found! (5 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011 SX127x not found! (6 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011 SX127x not found! (7 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011 SX127x not found! (8 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011 SX127x not found! (9 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011 SX127x not found! (10 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011 No SX127x found! failed: SX127x not found ``` Using the abandoned library [RadioHead](https://github.com/PaulStoffregen/RadioHead) Setup with ```cpp // Singleton instance of the radio driver RH_RF95 rf95(16, 26); // 16 = NSS, 26 = INT void setup() { // Init the RFM95 module if (!rf95.init()) { Serial.println("[RFM95] init failed"); delay(500); } } ``` works perfect, log output: ```log RH_RF95::init() started RH_RF95::init() success ``` Any idea? Answers: username_1: During initialization, RadioLib checks the version register of RFM95. It should read 0x11, yours returns 0x00. That usually indicates faulty SPI bus or damaged device. It looks like RadioHead performs no such check, does the device actually work? username_0: Yes the device works perfectly. The counterpart is [SparkFun Pro RF - LoRa, 915MHz (SAMD21)](https://www.sparkfun.com/products/14916) and the two communicate without problems with each other. But...... Found the problem. I am using an ESP32 which has two SPI, one called HSPI and a second one VSPI The RF96 is connected to HSPI, but it seems the ESP32 defaults SPI to VSPI. So the SPI is initialized on the wrong pins. That's why it doesn't work. But no idea why it works with the RadioHead library. As a quick test I hard-coded the pins into Module::init() ```cpp _spi->end(); _spi->begin(14, 12, 13, 16); ``` Then it works. Now I need to figure out how to set the SPI to the correct pins and give it as argument to ``` RFM96 lora = new Module(16, 26, 33, 32, mySPI); ``` Sorry for bothering you with this. It's not a bug, but an ESP32 specific problem. Status: Issue closed
arangodb/arangodb-php
184836377
Title: Undefined variable $revision in DocumentHandler.php line 705 Question: username_0: https://github.com/arangodb/arangodb-php/blob/master/lib/triagens/ArangoDb/DocumentHandler.php#L705 adding $revision = `$document->getRevision();` before this line fixed the problem. Answers: username_1: Thanks for taking the time to report. This one was already fixed in in devel but did not make it to master yet. Commit: <PASSWORD>ef<PASSWORD> I will hopefully release a new 3.0.x version within the next couple of days, which will include this and other fixes and improvements. Closing this for now. If there is anything else concerning this, feel free to comment. PS: Release 3.1 of the driver will finally have the very old deprecated stuff removed, which unfortunately wasn't removed in 2.x and 3.0. This in turn will allow for some refactoring and improvements further down the road. Status: Issue closed
infinitered/solidarity
803330298
Title: High severity vulnerability in gluegun dependency (apisauce) Question: username_0: Hello :) Solidarity currently depends on [email protected] which depends on a version of axios with the high severity vulnerability https://www.npmjs.com/advisories/1594. Can we please upgrade gluegun to latest version ? Answers: username_1: Sounds good. Do you want to do a PR? If not I can put this on a TODO list. username_0: Hi @username_1 Sorry, didn't see your reply. I made a local PR but I'm waiting for this [MR](https://github.com/SimonJang/gluegun/commit/a411ecbf3805a57c8f5abeb69ef5450b155c329a) to be merged :) username_1: I merged that, now we'll just need a new version pushed and then an update. Thanks! Status: Issue closed
Aldrin-John-Olaer-Manalansan/DOTA-2-MOD-Master
1122570347
Title: ANTI VAC Question: username_0: Can anyone share a fix for this? I really missed playing with great looking skins, Answers: username_1: Hi, there is a fix, that disables vac message, but it requires special gameinfo.gi, that made specially for D2MO (Dota 2 Mod Organizer, and you don't need to use it), i'm playing with mods for almost month, and everything is working. You can download it here: https://www.mediafire.com/file/0fe3799qegsag0o/gameinfo.gi/file . Also make sure you'll instal D2MO, you don't need to install every mod through this app, but it will make base folder for mods in your dota directory. Status: Issue closed
apollographql/apollo-feature-requests
657463159
Title: Validate Apollo Client typePolicies `keyFields` with Apollo CLI Question: username_0: Apollo Client 3 support's a new cache configuration mechanic with typePolicies, specifically allowing us to define a canonical id for a type via `keyFields`. I've run into several cases where I add a `keyFields` value and then get runtime errors for not fetching that field. This seems like a great place to integrate with the Apollo CLI to verify that `keyFields` definitions in `typePolicies` always have the fields required in all queries.
HackerHappyHour/impact
138858879
Title: Auto creation of team sub-contexts via integrations Question: username_0: new threads/channels/groups should be able to automatically be created for any integration *impact* is linked to, such that a new sub-context will become available inside of a chat context, when a new context is created within the integration. *sub-context*: a channel or group, existing within an *impact* team. Let's take github for example: You should be able to associate an *impact* team to either a github organization, or a github team. - Whenever a new *github team* is added to a *github organization* that is associated to an `impact team`, it should automatically create a new sub-context inside the `impact team - Whenever a new *github repository* is added to a *github team* associated with an *impact team*, it should create a new sub-context ### Considerations - Privacy/visibility settings of the new context item from the integration (new *github team/repo*), should carry over, or sync via LDAP, to the access and visibility of the sub-context in *impact*. - The privacy/visibiltiy settings could determine *what type* of sub-context is created. For example, if a github team is publicly visible, perhaps a channel is added to the *impact* team. However, if the github team is not publicly visible, then it creates a private group inside the *impact* team.
jlippold/tweakCompatible
339998215
Title: `HapticFeedback` working on iOS 11.2.6 Question: username_0: ``` { "packageId": "com.ziph0n.hapticfeedback", "action": "working", "userInfo": { "arch32": false, "packageId": "com.ziph0n.hapticfeedback", "deviceId": "iPhone9,4", "url": "http://cydia.saurik.com/package/com.ziph0n.hapticfeedback/", "iOSVersion": "11.2.6", "packageVersionIndexed": true, "packageName": "HapticFeedback", "category": "Tweaks", "repository": "BigBoss", "name": "HapticFeedback", "packageIndexed": true, "packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 100% with 2 working reports.", "id": "com.ziph0n.hapticfeedback", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.0.7", "shortDescription": "Get an haptic feedback when pushing buttons!", "latest": "1.2-1", "author": "Ziph0n", "packageStatus": "Working" }, "base64": "<KEY> "chosenStatus": "working", "notes": "" } ```
clearlinux/distribution
927921048
Title: conda bundle Module Not Found Error: No module named 'conda_package_handling' Question: username_0: Severity: major System: 34770 Description: After initializtion using `conda init` and session restart, conda still cannot be used. conda command such as `list`, `install` or `create` shows the following error: Traceback (most recent call last): File "/usr/lib/python3.9/site-packages/conda/exceptions.py", line 1079, in __call__ return func(*args, **kwargs) File "/usr/lib/python3.9/site-packages/conda/cli/main.py", line 84, in _main exit_code = do_call(args, p) File "/usr/lib/python3.9/site-packages/conda/cli/conda_argparse.py", line 83, in do_call return getattr(module, func_name)(args, parser) File "/usr/lib/python3.9/site-packages/conda/cli/main_create.py", line 41, in execute install(args, parser, 'create') File "/usr/lib/python3.9/site-packages/conda/cli/install.py", line 116, in install if context.use_only_tar_bz2: File "/usr/lib/python3.9/site-packages/conda/base/context.py", line 734, in use_only_tar_bz2 import conda_package_handling.api ModuleNotFoundError: No module named 'conda_package_handling' ...... Answers: username_1: Upstream bug https://github.com/conda/conda/issues/9510
JacksonBates/fcc-scraper
272741965
Title: Fix regex in the challenge name parser Question: username_0: Spaces are being dealt with inconsistently. See line 5 of controllers/getSolution.js Answers: username_1: Hi Jackson, Im Matt, I'm a freecodecamper from Melbourne, seen you a few times ages ago at FCC Melb. Do you mind if have a go at my first pull requests here with you? I'm fairly sure I've sorted this issue, the replace function is missing the first space. line 5 of controllers/getSolution.js was removing the '%' symbol thus leaving just a '20' and the subsequent replace function in writeFile.js was ignoring that space... I've worked out what the solution is, but am just trying to figure out best way to make a pull request, never done one before! username_0: Hi Matt! You're more than welcome to have a crack at it. I actually made a video walkthrough for pull requests last year which might help. The gist of it is to fork the repo, check out a new branch for the fix, make the change, commit the change, push the commit and then open up a pull request across branches on my original repo. Sounds complicated, but this should help: https://www.youtube.com/watch?v=dSl_qnWO104&t=104s When you test it - test against the camper **portablestick**. He has some older solutions that include odd characters that broke the old regex I had. Fixing that introduced the current problem, so ensure your fix doesn't undo that fix :) Once you PR, I'll review your code and walk you through any changes they might need. username_1: Thanks Jackson. I totally thought this was an open source thing and here I am crashing your lil project haha! Your vid is excellent by the way. If you could leave those simpler bugs to me, I'll get on to them. PR coming through! username_0: If it's on GH it's an open source thing! Always happy to have contributors :) username_0: fixed by #16 Status: Issue closed
AppMetrics/AppMetrics
425342973
Title: RegisterFromAssembly Get Error "Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information." Question: username_0: Hi , when I was use health part in 3.0 version ,I got an error "Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information." And the stacktrace is: at System.Reflection.RuntimeModule.GetTypes(RuntimeModule module) at System.Reflection.RuntimeAssembly.get_DefinedTypes() at System.Linq.Enumerable.SelectManySingleSelectorIterator`2.MoveNext() at System.Linq.Enumerable.WhereSelectEnumerableIterator`2.MoveNext() at App.Metrics.Health.Extensions.DependencyInjection.Internal.HealthChecksAsServices.AddHealthChecksAsServices(IServiceCollection services, IEnumerable`1 types) in C:\projects\microsoftextensions\src\App.Metrics.Health.Extensions.DependencyInjection\Internal\HealthChecksAsServices.cs:line 27 at Microsoft.Extensions.DependencyInjection.ServiceCollectionHealthCheckBuilderExtensions.RegisterFromAssembly(IHealthCheckBuilder healthCheckBuilder, IServiceCollection services, DependencyContext dependencyContext) in C:\projects\microsoftextensions\src\App.Metrics.Health.Extensions.DependencyInjection\ServiceCollectionHealthCheckBuilderExtensions.cs:line 34 at Demo.Startup.ConfigureServices(IServiceCollection services) in D:\下载\AppMetricsDemo-master\AppMetricsDemo\Demo\Startup.cs:line 82 So I debug with the source code, and find out that, is was caused by the class "DefaultHealthCheckTypeProvider" in line 25,and the reason is that,the Assembly in StaticHealthAssemblyProvider include "App.Metrics.Health.Abstractions", but at the run time,it`s "DefinedTypes" is null. I think ,we should exclude “App.Metrics.Health.Abstractions” assembly. Answers: username_1: Refer to the [sample application](https://github.com/AppMetrics/Samples.V2/tree/master/AspNetCore2.Health.Api.QuickStart) Status: Issue closed
bvaisvil/zenith
582718614
Title: Panic on WSL writing db Question: username_0: `thread '<unnamed>' panicked at 'Couldn't open database.: Io(Custom { kind: Other, error: "Os { code: 38, kind: Other, message: \"Function not implemented\" }" })', src/metrics.rs:248:17 thread '<unnamed>' panicked at 'Couldn't open database.: Io(Custom { kind: Other, error: "Os { code: 38, kind: Other, message: \"Function not implemented\" }" })', src/metrics.rs:248:17 thread panicked while panicking. aborting. Illegal instruction (core dumped)` Seems to happen at the right write in the sled db. Answers: username_0: perhaps related to https://github.com/spacejam/sled/issues/958 username_1: Just curious, is this happening in WSL or WSL2? username_0: Believe it is WSL as ` cmd.exe /c "systeminfo" | grep "^OS Version"` gives me `OS Version: 10.0.18363 N/A Build 18363` and WSL2 is in build >= 18917 ? username_1: I see, yeah, I think so. `wsl -l -v` is another way to explicitly check if it's version 2. FWIW wsl2 is actually a stock linux kernel running in a VM, instead of a reimplementation of Linux APIs on Windows, so it's much less likely to have incompatibilities. If you can provide a binary (I don't see a windows executable for the latest release) I can test to see if this also happens on WSL2. No worries if that's not convenient for you though. username_1: Ok, it seems to work fine on WSL2! username_0: That's good. Thanks very much for testing that! username_0: Latest version has no issues in WSL for me. Feel free to re-open this if you find anything. Thanks! Status: Issue closed
artsy/force
173220895
Title: Partner Article Redirect Question: username_0: Partner articles should redirect from /article/ to their profile slug at /:id/article/:slug Answers: username_0: Based off: https://artsy.slack.com/archives/publishing-support/p1471872867000277 Actually @owendodd do we want articles for non-2.0 partners to have the partner header too? Because this redirect will require both Partner and Partner_2 to have /:id/article/:slug support. Currently only Partner_2 has this. Status: Issue closed
Jessecar96/SteamDesktopAuthenticator
947075850
Title: Invalid length for a Base-64 array or string ERROR Question: username_0: ### **When i launch SDA:** ![image](https://user-images.githubusercontent.com/68283261/126075553-26ce0d83-cb32-4be9-a988-05bbba97a3a9.png) ### **If I click on continue both of my accounts disappear** ![image](https://user-images.githubusercontent.com/68283261/126075602-55616c2a-e3f5-4c45-be3e-aac8942c9898.png) **anyone had this issue before?** Answers: username_1: Are you using the same phone no /sim dongle/desktop as you've always used ? I hot a similar error because i forgot i had swapped the 3g dongle i was using for internet to wi-fi fibre broadband. username_1: I had to plug in my 3g dongle and use that to do confirmations.
test-repo-billy/sdk-release-request
618891205
Title: [data-plane] JS: 'Test' Question: username_0: **Link**: [test](test) **Tag**: test **Release date**: 2020-05-29 **Contacts**: [<NAME> <<EMAIL>>](mailto:<EMAIL>); @username_0 , [<NAME> <<EMAIL>>](mailto:<EMAIL>); @zhenglaizhang **Comments**: Answers: username_0: Test add comment
coredns/coredns
287816559
Title: RRSIG caching issue Question: username_0: After 8 days of operation (with file and without changing the zone file) the cached signatures are still served (and they are only signed for 8 days). [dnssec.go#L40](https://github.com/coredns/coredns/blob/master/plugin/dnssec/dnssec.go#L40) states that "Signatures will be cached for a short while.". But I can't find any code where expired or old signatures are dropped. Answers: username_1: hmmm, me neither :( I 100% sure I thought about, but somehow that didn't translated in to code. username_1: If you change `eightDays = 8 * 24 * time.Hour` into something smaller like `1 * 24 * time.Hour` you should see failures earlier. Could you easily try #1385 ? username_0: I'll try and report. But if I understand the code correctly we will run into the following problem: We have an RRSIG in the coredns cache which is valid now and not yet expired. However, given the TTL, the signature will expire while still in the cache of a resolver. And even if we subtract the TTL for the validity check, we don't consider a possible clock drift between the server and a validator. username_1: You're right we need to do this at 3/4 of the validity period and then regen. Status: Issue closed
google/dagger
533294964
Title: Hello. I am having trouble using dagger. Question: username_0: This is my Application.kt ` class SellerHubApplication : Application(), HasAndroidInjector { @Inject lateinit var activityDispatchingAndroidInjector: AndroidInjector<Any> override fun androidInjector(): AndroidInjector<Any> = activityDispatchingAndroidInjector override fun onCreate() { super.onCreate() DaggerAppComponent.builder() .application(this) .build() .inject(this) Fresco.initialize(applicationContext) } }` and this is my AppComponent `@Singleton @Component(modules = [AndroidInjectionModule::class, ActivityBuilder::class, AppModule::class]) interface AppComponent { @Component.Builder interface Builder { @BindsInstance fun application(app: Application): Builder fun build(): AppComponent } fun inject(app: SellerHubApplication) }` and my issue is that "dagger.android.AndroidInjector<java.lang.Object> cannot be provided without an @Provides-annotated method.". if i change "fun inject(app: SellerHubApplication)" to "fun inject(app: Application)". It works but in runtime, it shows "activityDispatchingAndroidInjector has not been initialized". Please help me.<issue_closed> Status: Issue closed
geneontology/go-annotation
326398065
Title: IPR019404 mapped to a merged term Question: username_0: Hello InterPro, IPR019404 is mapped to GO:0001104 RNA polymerase II transcription cofactor activity which now has as a primary ID: GO:0003712 Do you automatically migrate merge terms ? If not we'll make sure to notify you ahead of time. Thanks, Pascale Answers: username_1: Hi Pascale, Thank you for this. We do have a system (script) to migrate these terms. We usually do so as soon as we receive the report from GO. However, as you may be aware, these results are only visible after our release. In this case, we changed these terms in May, so the new ID 'GO:0003712' will be visible in our next release (v.69) in June. Therefore, it's always good to get your report as early as possible to synchronise the InterPro and GO websites. :) Thank you, Hsin-Yu username_0: Thanks @username_1 Can you specify which report you are talking about ? Some email with errors ? username_1: Hi Pascale, Sorry for the delayed reply. I just double checked with Amaia. We usually pick up the data from GOA and run a sanity check in InterPro. We then migrate the terms using SQL. Regards, Hsin-Yu Status: Issue closed
BluSunrize/ImmersiveEngineering
464874219
Title: [1.12.2] Bug inside Engineers workbench Question: username_0: ### Description of the issue: Recipes where the wirecutter is involved can only give you 1 result item at a time. After that you can't take another result item out. Removing and adding an ingrediënt resets it. ### Crashlog: no crash. ### Versions & Modlist forge -14.23.5.28.38 No other mods, tested it with only IE<issue_closed> Status: Issue closed
RPi-Distro/raspi-config
900187263
Title: Disable the new GL driver on Pi 4, now see error on boot? Question: username_0: Disable the new GL driver on Pi 4, now see error on boot? `# systemctl status gldriver-test ● gldriver-test.service - Check for v3d driver Loaded: loaded (/lib/systemd/system/gldriver-test.service; enabled; vendor preset: enabled) Active: failed (Result: exit-code) since Tue 2021-05-25 00:33:41 UTC; 6min ago Main PID: 416 (code=exited, status=1/FAILURE) May 25 00:33:41 localhost systemd[1]: Starting Check for v3d driver... May 25 00:33:41 localhost gldriver_test.sh[416]: /usr/lib/systemd/scripts/gldriver_test.sh: line 12: /usr/share/X11/xorg.conf.d/99-fbturbo.conf: No such file or directory May 25 00:33:41 localhost systemd[1]: gldriver-test.service: Main process exited, code=exited, status=1/FAILURE May 25 00:33:41 localhost systemd[1]: gldriver-test.service: Failed with result 'exit-code'. May 25 00:33:41 localhost systemd[1]: Failed to start Check for v3d driver. ` Raspberry Pi OS lite installed on Pi4. The GL driver is applicable only to X11 environment right? This error should not be occurring thus? Answers: username_1: @spl237 The `[Unit]` section of `gldriver-test`'s `gldriver-test.service` should probably have `ConditionPathIsDirectory=/usr/share/X11/xorg.conf.d`? username_0: I will check,.. Nope, or I should say, yeah, maybe that is the fix, the unit file is pretty basic... `[Unit] Description=Check for v3d driver [Service] Type=oneshot ExecStart=/usr/lib/systemd/scripts/gldriver_test.sh RemainAfterExit=yes [Install] WantedBy=multi-user.target ` Since I am running lite, none of the is applicable per se. Seems lite something that was missed when the lite image was established. For now I disabled the service to avoid the error, which as far as I can see, just informational if you are not using the GL driver in the first place. The actuall script executed... `#!/bin/bash if grep -s -q okay /proc/device-tree/soc/v3d@7ec00000/status \ /proc/device-tree/soc/firmwarekms@7e600000/status \ /proc/device-tree/v3dbus/v3d@7ec04000/status || [ "$(arch)" == aarch64 ]; then if [ -e /usr/share/X11/xorg.conf.d/99-fbturbo.conf ] ; then rm /usr/share/X11/xorg.conf.d/99-fbturbo.conf fi else if ! [ -e /usr/share/X11/xorg.conf.d/99-fbturbo.conf ] ; then cat > /usr/share/X11/xorg.conf.d/99-fbturbo.conf << EOF # This is a minimal sample config file, which can be copied to # /etc/X11/xorg.conf in order to make the Xorg server pick up # and load xf86-video-fbturbo driver installed in the system. # # When troubleshooting, check /var/log/Xorg.0.log for the debugging # output and error messages. # # Run "man fbturbo" to get additional information about the extra # configuration options for tuning the driver. Section "Device" Identifier "Allwinner A10/A13 FBDEV" Driver "fbturbo" Option "fbdev" "/dev/fb0" Option "SwapbuffersWait" "true" EndSection EOF fi fi ` Running the script it fails with... '/usr/lib/systemd/scripts/gldriver_test.sh: line 12: /usr/share/X11/xorg.conf.d/99-fbturbo.conf: No such file or directory' For Pi OS lite image.
NizarSemlali/INF3172-TP2
220358444
Title: Questions à poser Lundi : Question: username_0: - Si l'exécutable tsh doit être dans le bin ou simplement dans le répertoire TP2 . - Valider l'arborescence du programme . -Si le fichier à afficher avec fin est vide, qu'est ce qu'on affiche ? -.DS_Store , es ce que le fichier doit être considéré dans nos calculs ou pas ?<issue_closed> Status: Issue closed
rclone/rclone
606600006
Title: (req) search box Question: username_0: ![image](https://user-images.githubusercontent.com/19517680/80259142-1baddf80-86af-11ea-9a7f-aeeb3412d759.png) Answers: username_1: You can use the forums for any site feedback rather than raising an issue. https://forums.rclone.org Thanks! Status: Issue closed
ripperhe/Bob
737383063
Title: 百度翻译使用自动检测可能出现无法准确翻译问题 Question: username_0: <!-- ⚠️ 反馈前请确保已阅读 ⚠️ 反馈前请确保已阅读 ⚠️ 反馈前请确保已阅读 1. 请确保你已经认真阅读了 README 文件,可能你的问题不是「问题」。 2. 请在 issues 页面搜索你的问题,很可能已被解决。 3. 如果仍旧有问题,请填写模板描述问题,以便大家理解、定位和解决问题。 --> <!-- 这是隐藏的信息 --> <!-- 👆👆👆这样括起来的信息将被隐藏,填写时注意不要写在里面。 --> <!-- 点击编辑器上方的 preview 可预览你填写的效果 --> ## 运行环境 ``` ❯ neofetch 'c. <EMAIL> ,xNMM. ---------------------- .OMMMMo OS: macOS Catalina 10.15.7 19H2 x86_64 OMMM0, Host: MacBookPro11,5 .;loddo:' loolloddol;. Kernel: 19.6.0 cKMMMMMMMMMMNWMMMMMMMMMM0: Uptime: 2 days, 13 hours, 55 mins .KMMMMMMMMMMMMMMMMMMMMMMMWd. Packages: 156 (brew) XMMMMMMMMMMMMMMMMMMMMMMMX. Shell: zsh 5.7.1 ;MMMMMMMMMMMMMMMMMMMMMMMM: Resolution: 1920x1080 :MMMMMMMMMMMMMMMMMMMMMMMM: DE: Aqua .MMMMMMMMMMMMMMMMMMMMMMMMX. WM: Quartz Compositor kMMMMMMMMMMMMMMMMMMMMMMMMWd. WM Theme: Red (Dark) .XMMMMMMMMMMMMMMMMMMMMMMMMMMk Terminal: iTerm2 .XMMMMMMMMMMMMMMMMMMMMMMMMK. Terminal Font: MesloLGS-NF-Regular 13 kMMMMMMMMMMMMMMMMMMMMMMd CPU: Intel i7-4870HQ (8) @ 2.50GHz ;KMMMMMMMWXXWMMMMMMMk. GPU: AMD Radeon R9 M370X, Intel Iris Pro .cooc,. .,coo:. Memory: 9661MiB / 16384MiB ``` ## 描述问题 在百度私有api中测试 使用自动检测 ![image](https://user-images.githubusercontent.com/15372793/98314525-1a238380-2011-11eb-8cc9-c8cd37e083b5.png) 指定语言 ![image](https://user-images.githubusercontent.com/15372793/98314549-2c052680-2011-11eb-9bd4-4e3a9b3d9222.png) Answers: username_1: 了解了,感谢反馈 🤝 username_2: 更新了以后用上百度的 识别经常失败 不如上一版有道 Status: Issue closed username_1: 新版加了语种检查按钮,如果不幸检查错误,可以手动点击这个按钮切换下语言 ![image](https://user-images.githubusercontent.com/13943595/143744719-1a6fe914-05fd-4de4-a80a-24b505458ad3.png)
ssborbis/ContextSearch-web-ext
285073295
Title: possible quick menu style change? Question: username_0: Quick Menu Style change? select text and mouse left-click( not hold) Quick Context Search https://addons.mozilla.org/ko/firefox/addon/quickcontextsearch/ Answers: username_1: I'll see about adding that. In the meantime, you might like the latest betas where I added an option for opening the quick menu automatically after selecting text. https://addons.mozilla.org/en-US/firefox/addon/contextsearch-web-ext/versions/beta username_1: If you're wanting the grid in the context menu like in the add-on you linked to, it's not possible with the new web extensions api username_0: thanks. Status: Issue closed
everpeace/banditsbook-scala
208867080
Title: Hedge Algorithm run results in a NotANumberException Question: username_0: [info] Running com.github.everpeace.banditsbook.algorithm.hedge.TestHedge ------------------------------- Hedge Algorithm ------------------------------- arms = (μ=0.1), (μ=0.9), (μ=0.1), (μ=0.1), (μ=0.1) (Best Arm = 1) horizon = 500 nSims = 5000 η = (0.2,0.4,0.8,0.9,0.95,1.0) starts simulation on η=0.2. `results are written to output/test-hedge-results.csv` `[error] (run-main-0) org.apache.commons.math3.exception.NotANumberException: NaN is not allowed org.apache.commons.math3.exception.NotANumberException: NaN is not allowed at org.apache.commons.math3.distribution.EnumeratedDistribution.<init>(EnumeratedDistribution.java:121) at org.apache.commons.math3.distribution.EnumeratedIntegerDistribution.<init>(EnumeratedIntegerDistribution.java:100) at org.apache.commons.math3.distribution.EnumeratedIntegerDistribution.<init>(EnumeratedIntegerDistribution.java:68) at com.github.everpeace.banditsbook.algorithm.package$$anon$1.<init>(package.scala:34) at com.github.everpeace.banditsbook.algorithm.package$.CategoricalDistribution(package.scala:32) at com.github.everpeace.banditsbook.algorithm.hedge.Hedge$$anon$1.selectArm(Hedge.scala:55) at com.github.everpeace.banditsbook.algorithm.hedge.Hedge$$anon$1.selectArm(Hedge.scala:45) at com.github.everpeace.banditsbook.algorithm.Algorithm$$anonfun$selectArm$1.apply(Algorithm.scala:42) at com.github.everpeace.banditsbook.algorithm.Algorithm$$anonfun$selectArm$1.apply(Algorithm.scala:40) at cats.data.StateFunctions$$anonfun$inspect$2.apply(StateT.scala:184) at cats.data.StateFunctions$$anonfun$inspect$2.apply(StateT.scala:184) at cats.data.StateFunctions$$anonfun$apply$10.apply(StateT.scala:169) at cats.data.StateFunctions$$anonfun$apply$10.apply(StateT.scala:169) at cats.data.StateT$$anonfun$run$1.apply(StateT.scala:36) at cats.data.StateT$$anonfun$run$1.apply(StateT.scala:36) at cats.Eval$$anon$5$$anonfun$1$$anon$6$$anonfun$2.apply(Eval.scala:81) at cats.Eval$$anon$5$$anonfun$1$$anon$6$$anonfun$2.apply(Eval.scala:81) at cats.Eval$Compute.loop$1(Eval.scala:276) at cats.Eval$Compute.value(Eval.scala:290) at com.github.everpeace.banditsbook.algorithm.TracedAlgorithmDriver$$anonfun$2.apply(TracedAlgorithmDriver.scala:60) at com.github.everpeace.banditsbook.algorithm.TracedAlgorithmDriver$$anonfun$2.apply(TracedAlgorithmDriver.scala:59) at cats.data.StateT$$anonfun$map$1.apply(StateT.scala:30) at cats.data.StateT$$anonfun$map$1.apply(StateT.scala:30) at cats.data.StateT$$anonfun$transform$1$$anonfun$apply$6.apply(StateT.scala:72) at cats.data.StateT$$anonfun$transform$1$$anonfun$apply$6.apply(StateT.scala:72) at cats.Eval$$anonfun$map$1.apply(Eval.scala:58) at cats.Eval$$anonfun$map$1.apply(Eval.scala:58) at cats.Eval$Compute.loop$1(Eval.scala:286) at cats.Eval$Compute.value(Eval.scala:290) at com.github.everpeace.banditsbook.algorithm.TracedAlgorithmDriver.runFrom(TracedAlgorithmDriver.scala:95) at com.github.everpeace.banditsbook.algorithm.TracedAlgorithmDriver.run(TracedAlgorithmDriver.scala:81) at com.github.everpeace.banditsbook.algorithm.TracedAlgorithmDriver.run(TracedAlgorithmDriver.scala:88) at com.github.everpeace.banditsbook.testing_framework.TestRunner$$anonfun$run$1.apply$mcVI$sp(TestRunner.scala:84) at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160) at com.github.everpeace.banditsbook.testing_framework.TestRunner$.run(TestRunner.scala:79) at com.github.everpeace.banditsbook.algorithm.hedge._TestHedge$$anonfun$1.apply$mcVD$sp(TestHedge.scala:69) at com.github.everpeace.banditsbook.algorithm.hedge._TestHedge$$anonfun$1.apply(TestHedge.scala:64) at com.github.everpeace.banditsbook.algorithm.hedge._TestHedge$$anonfun$1.apply(TestHedge.scala:64) at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234) at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) at scala.collection.TraversableLike$class.map(TraversableLike.scala:234) at scala.collection.AbstractTraversable.map(Traversable.scala:104) at com.github.everpeace.banditsbook.algorithm.hedge._TestHedge$class.run(TestHedge.scala:64) at com.github.everpeace.banditsbook.algorithm.hedge.TestHedge$.run(TestHedge.scala:35) [Truncated] at com.github.everpeace.banditsbook.algorithm.hedge.TestHedge$delayedInit$body.apply(TestHedge.scala:35) at scala.Function0$class.apply$mcV$sp(Function0.scala:34) at scala.runtime.AbstractFunction0.apply$mcV$sp(AbstractFunction0.scala:12) at scala.App$$anonfun$main$1.apply(App.scala:76) at scala.App$$anonfun$main$1.apply(App.scala:76) at scala.collection.immutable.List.foreach(List.scala:381) at scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:35) at scala.App$class.main(App.scala:76) at com.github.everpeace.banditsbook.algorithm.hedge.TestHedge$.main(TestHedge.scala:35) at com.github.everpeace.banditsbook.algorithm.hedge.TestHedge.main(TestHedge.scala) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) [trace] Stack trace suppressed: run last compile:runMain for the full output. java.lang.RuntimeException: Nonzero exit code: 1 at scala.sys.package$.error(package.scala:27) [trace] Stack trace suppressed: run last compile:runMain for the full output. [error] (compile:runMain) Nonzero exit code: 1 [error] Total time: 1 s, completed Feb 20, 2017 1:31:39 PM`
Purii/react-native-tableview-simple
200316689
Title: CustomCell doesn't render > Question: username_0: Hi there, Is there a reacon I don't see ">" when I use CustomCell? Also, how can I use the ">" in my code? How is it rendered? Thanks. Answers: username_0: Found it: it's this: <View style={styles.accessory_disclosureIndicator} /> Status: Issue closed
enzoampil/fastquant
678490010
Title: [FEATURE] How to control the size plots when backtrader plots are embedded into notebook output Question: username_0: Currently, it's common to have workflows where backtrader plots are rendered straight from a jupyter or colab notebook. The issue is that the size of the plot is quite small by default, which makes it difficult to see through the busy contents of the backtesting results. It would be great if `backtest` could have a `figsize` argument similar to `matplotlib.pyplot`, where the default is much bigger than the current setup. E.g. `figsize=(12, 7)` could be a good default. Answers: username_0: Sample default size: ![image](https://user-images.githubusercontent.com/39557688/90148298-c4df2980-ddb5-11ea-9166-76305fcde6f4.png) username_1: I looked into this problem and found a solution below. We just need to add these lines on top of the jupyter notebook. ```python import matplotlib.pyplot as pl pl.style.use("default") #ggplot is also fine pl.rcParams["figure.figsize"] = (15,12) #all styles: print(pl.style.available) ``` Documentation: https://matplotlib.org/3.3.0/tutorials/introductory/customizing.html We may be able to set this global config automatically if code is being run in jupyter but there may be [caveats](https://stackoverflow.com/questions/15411967/how-can-i-check-if-code-is-executed-in-the-ipython-notebook). username_0: Understand @username_1 . Thanks will use this as quick fix. I think this should be automated as much as possible straight from `backtest`. Yeah, think it makes sense to start chipping away at these plotting issues. Really affects the UX of the tool. username_0: Also, great to refer these two relevant threads on backtrader docs: [link1](https://community.backtrader.com/topic/291/ipython-notebook-plot-sizing/5), [link2](https://community.backtrader.com/topic/22/plot-maximize-windows/2) Status: Issue closed username_2: I added support for `figsize()` parameter in the `backtest` function in https://github.com/username_0/fastquant/pull/286/files#diff-9081e26fa869af389e67c6f9e0872cb6eb190052979673832a9b2f4a6215343eR67 username_0: Thanks @username_2 ! :smile:
jenkins-x/enhancements
584994339
Title: Enhance JX UI Question: username_0: JX UI looks great and it gives a high-level view of the current status of the builds, environments, projects and so on. As user, it would be really great to be able to: - Configure namespaces (at the moment 'jx' is hardcoded in some places and it is not possible to configure it) - Support GHE or a configurable wat for Git Providers/Services (e.g. links/images seem to point towards Github only - so they are broken) - Have covered edge-cases in builds' completion/status (collected builds, old ones, stuck, etc..) - Check the Codebase for potential contributions Answers: username_1: we've an OSS plugin for octant now: https://github.com/jenkins-x/octant-jx which hopefully has the features you need; if not we could create the necessary issues there & hack on the code to improve it? username_0: thank you very much @username_1! I have just tried it and it looks great (and those features seem implemented there)! what about users not able to access the cluster resources directly and not able to use octant? (e.g. a developer who is taking care of a project managed by jx - or non-technical users such as Scrum Master or Product Owner)
PAhelper/PhoenixAdult.bundle
993763209
Title: HotCrazyMess [Nubiles] Error Loading Actor Question: username_0: **Describe the bug** When matching or refreshing metadata for any scene from HotCrazyMess the actor fails to load with an error in the logs. **Include** 1. All Scenes from https://hotcrazymess.com/video/ 2. HotCrazyMess - Scene Title ``` 2021-09-11 02:09:15,492 (15bc) : INFO (logkit:16) - ******UPDATE CALLED******* 2021-09-11 02:09:15,492 (15bc) : INFO (logkit:16) - SiteNum: 543 2021-09-11 02:09:15,492 (15bc) : INFO (logkit:16) - Provider: networkNubiles 2021-09-11 02:09:15,493 (15bc) : DEBUG (preferences:258) - Loaded preferences from DefaultPrefs.json 2021-09-11 02:09:15,494 (15bc) : DEBUG (preferences:178) - Loaded the user preferences for com.plexapp.agents.phoenixadult 2021-09-11 02:09:15,496 (15bc) : INFO (logkit:16) - Requesting GET "https://hotcrazymess.com/video/watch/72198" 2021-09-11 02:09:15,844 (15bc) : INFO (logkit:16) - Requesting GET "https://hotcrazymess.com/model/profile/10220/chloe-temple" 2021-09-11 02:09:16,118 (15bc) : CRITICAL (core:574) - Exception in the update function of agent named 'PhoenixAdult', called with guid 'com.plexapp.agents.phoenixadult://72198|543?lang=en' (most recent call last): File "C:\Program Files (x86)\Plex\Plex Media Server\Resources\Plug-ins-2b1b51db9\Framework.bundle\Contents\Resources\Versions\2\Python\Framework\api\agentkit.py", line 1093, in _update agent.update(obj, media, lang, **kwargs) File "C:\Plex\Plex Media Server\Plug-ins\PhoenixAdult.bundle\Contents\Code\__init__.py", line 136, in update provider.update(metadata, lang, siteNum, movieGenres, movieActors) File "C:\Plex\Plex Media Server\Plug-ins\PhoenixAdult.bundle\Contents\Code\networkNubiles.py", line 138, in update galleryURL = PAsearchSites.getSearchBaseURL(siteNum) + detailsPageElements.xpath('//div[contains(@class, "content-pane-related-links")]/a[contains(., "Pic")]/@href')[0] File "C:\Program Files (x86)\Plex\Plex Media Server\Resources\Plug-ins-2b1b51db9\Framework.bundle\Contents\Resources\Versions\2\Python\Framework\code\sandbox.py", line 108, in <lambda> _getitem_ = lambda x, y: x.__getitem__(y), IndexError: list index out of range ```<issue_closed> Status: Issue closed
kapi2289/fb-anti-unsend
894236183
Title: No license Question: username_0: EN: This software isn't licensed. This means that nobody can use this software without permission. Please add license file. I recommend MIT license for this software. You can find another license at https://choosealicense.com/ If you choose GNU GPL as your license, read their guidelines for using the GPL at https://www.gnu.org/licenses/gpl-howto.html.en or read guidelines at the end of gpl text. If you choose GPL, please add a note about using newer versions of GNU GPL. PL: Te oprogramowanie nie jest udostępnione z licencją. To oznacza, że nikt nie może go używać bez zgody. Proszę, dodaj licencję. Polecam licencję MIT dla tego programu. Jeśli wolisz, możesz znaleźć inną licencję na https://choosealicense.com/ . Jeżeli wybierzesz GNU GPL jako licencję, przeczytaj ich wskazówki używania GPL na https://www.gnu.org/licenses/gpl-howto.html.en lub na końcu tekstu GPL. Dodatkowo: jeśli wybierzesz GPL, dodaj informacje o używaniu nowszych wersji GNU GPL. Answers: username_1: Yeah, you're right. It's old, small and not used anymore because of problems with Facebook project so I haven't thought of any license yet. I'll add it, thanks! Status: Issue closed