repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
ngmansion/FE8N | 752854721 | Title: ユニットが移動力の最大で移動した時にフリーズ(極まれ)
Question:
username_0: ユニットが移動力の最大で移動した時にフリーズ(極まれ)します。
移動7のユニットが7マス移動を指定して移動した瞬間に画面が乱れ異音を吐きます。
症状としては、ユニット命令でマップ外に移動するように指定した時みたいなかんじ?
発生日は今年の9月から。
リセットすると発動しないのでセーブデータは用意できていません。
良かったらありえそうな箇所をもう一度みてもらえないでしょうか?
Status: Issue closed
Answers:
username_1: たぶん直しました
username_0: バグがまた起こったのを確認しました。
症状としてはやはり、最大移動でフリーズ。ユニット操作の矢印がムチャクチャになる等です。
username_1: ユニットが移動力の最大で移動した時にフリーズ(極まれ)します。
移動7のユニットが7マス移動を指定して移動した瞬間に画面が乱れ異音を吐きます。
症状としては、ユニット命令でマップ外に移動するように指定した時みたいなかんじ?
発生日は今年の9月から。
リセットすると発動しないのでセーブデータは用意できていません。
良かったらありえそうな箇所をもう一度みてもらえないでしょうか?
username_1: @username_0
通知があった事に気づいたのですが、直りませんでした?
とはいえ、他に原因は思いつかないのですが。
username_0: すいません、こことは別の所に要因を見つけたので解決しました。
お手数おかけしました。
Status: Issue closed
username_1: 解決したようなのでclose |
AbnormalSec/darkbox | 351742394 | Title: New tool: timestomp
Question:
username_0: Similar to this: https://www.offensive-security.com/metasploit-unleashed/TimeStomp/
Answers:
username_1: this is something I've been looking into implementing lately, although some issues have arisen.
TL;DR Linux is a complex ecosystem and OS agnostic code is hard af.
So windows and other OS's are pretty straight forward. I need to test with metasploit to see how it handles linux (TODO) but from all my reading:
The linux kernel does not directly support file creation times. It's left up to the filesystem. the field used in windows for creation (ctime) is usually used as "change time" which is for metadata changes to the file, such as changing any other timestamp, renaming or copying, or other metadata changes. If the file has never undergone metadata changes, then it will be the creation date which is technically the first metadata change, but this is unreliable.
some filesystems do support creation time, but some extremely common ones such as XFS, EXT3, and others don't.
bottom line: this tool requires some weird choices to be made, makes a point for why we need CI testing and likely wont be completed until I've tested meterpreter's implementation. |
cozy/cozy-ui | 395511089 | Title: Icon sprite generation
Question:
username_0: Currently, during the generation of the sprite, we keep `id` and `use` provided by each SVG. But sometimes, several SVG can have the same `id` and `use`.
We have two solutions :
- Editing manually the SVG we want to add by editing the `id` and `use` attributes with an unique id
- During the generation, editing SVG by adding unique id
Answers:
username_1: I'm not sure I understand the issue here 😕
username_2: We could use the filename for this right ? An automic id like random hash would be difficult to debug.
username_0: @username_1 Let say I want to add this SVG (info.svg) :
```svg
<svg xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" viewBox="0 0 16 16"><defs><path id="a" d="M8 16A8 8 0 1 1 8 0a8 8 0 0 1 0 16zM7 4a1 1 0 1 0 2 0 1 1 0 1 0-2 0zm1 2H6v2h1v4a1 1 0 0 0 1 1h2v-2H9V7a1 1 0 0 0-1-1z"/></defs><use fill-rule="evenodd" xlink:href="#a"/></svg>
```
The sprite will be generated and I think I'll have my `info` svg ready to use. But when I'm using it, it'll display (in this case), the help icon since this SVG define the same `id = a` path (I don't know how the browser manage that case, but it seems to take the first id)
```svg
<svg width="16" height="16" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink"><defs><path d="M8 16A8 8 0 1 1 8 0a8 8 0 0 1 0 16zm1-7.198C10.122 8.355 11 7.21 11 6c0-1.552-1.448-3-3-3S5 4.448 5 6h2c0-.448.552-1 1-1 .448 0 1 .552 1 1 0 .448-.552 1-1 1a1 1 0 0 0-1 1v2h2V8.802zM7 11v2h2v-2H7z" id="a"/></defs><use xlink:href="#a" fill-rule="evenodd"/></svg>
```
@username_2 Yep, I think we should prepend the id / use with the icon name. |
dadhi/FastExpressionCompiler | 464824247 | Title: Feature toggles to disable the not used features which affect performance
Question:
username_0: For example, DryIoc does not use Label, GoTo and TryCatchFinally, but to support them FEC adds the memory (fields in `ClosureInfo`) and processing overhead.
May be we can use compile time vars to switch some features off if needed.
Cause I want to proceed to use FEC as-is in DryIoc, but want a maximum performamce for my use case - which was the original goal anyway.
Status: Issue closed
Answers:
username_0: on hold |
maarksman/Eulib | 293002726 | Title: Search should separate suggestion name and value searched
Question:
username_0: -type article name
-click suggestion
-click search
-if suggestion-text is different from desired article title, search will fail
need to make so that search suggestions have title, type, subtype, and field, but search does not fail when clicking the suggestion |
probot/create-probot-app | 789422564 | Title: CPA cannot find NPM module when running via NPX
Question:
username_0: npx: installed 231 in 12.637s
Cannot find module 'npm'
```
Answers:
username_1: it looks like starting with https://github.com/probot/create-probot-app/pull/316 we depend on the `npm` package, but we missed to add it to dependencies:
https://github.com/probot/create-probot-app/pull/316/files#diff-7ae45ad102eab3b6d7e7896acd08c427a9b25b346470d7bc6507b6481575d519R30-R45
Status: Issue closed
|
theniceboy/nvim | 822956287 | Title: 从B站来,git clone你的neovim报错
Question:
username_0: mac环境下,按照视频配置后报了错误,大都是文件找不到
Answers:
username_1: 我看到了这个你的提问,一下子就知道问题出在了哪里因为我精确地猜出了你的mac,nvim版本号,并且看到了完整报错信息。
解决方案:把找不到的文件找到。
Status: Issue closed
username_2: mac brew的nvim版本不对 得自己去github下载relase过得
username_3: 您好!我也是直接克隆你的neovim配置的,总是遇到各种问题,您能出一期视频从零开始使用的自己的neovim配置吗?克隆配置后的一些配置,应该怎么去配置 |
Azure/azure-iot-sdk-c | 255236160 | Title: Azure IoT SDK C - Issue during reconnection attempt when utilizing openSSL and amqp over websockets
Question:
username_0: Hello,
I have encountered an issue during the reconnection attempt when using openSSL and amqpWS as Protocol. (Windows)
<!--
Hi there! thank you for discovering and submitting an issue!
Please first tell us a little bit about the environment you're running:
-->
- **OS and version used:** Windows 7
- **SDK version used:** 1.1.21
- Dev environment set up according to : https://github.com/Azure/azure-iot-sdk-c/blob/master/doc/devbox_setup.md#windows
- Visual Studio 2015
- openSSL 1.0.2j or 1.0.2l (should also be reproducible with 1.0.2.k)
- target platform x86
# Description of the issue:
Within c-utility\src\tlsio_openssl.c, the tlsio_openssl_setoption function expects the value as pointer to int during evaluation of the tls_version option. This works fine when invoking IoTHubClient_SetOption.
However, on a reconnection attempt the tlsio_openssl_setoption function will be invoked like this:
exampled.exe!tlsio_openssl_setoption(void * tls_io, const char * optionName, const void * value) Line 1548 C
exampled.exe!OptionHandler_FeedOptions(OPTIONHANDLER_HANDLE_DATA_TAG * handle, void * destinationHandle) Line 236 C
exampled.exe!xio_setoption(XIO_INSTANCE_TAG * xio, const char * optionName, const void * value) Line 184 C
exampled.exe!OptionHandler_FeedOptions(OPTIONHANDLER_HANDLE_DATA_TAG * handle, void * destinationHandle) Line 236 C
exampled.exe!uws_client_set_option(UWS_CLIENT_INSTANCE_TAG * uws_client, const char * option_name, const void * value) Line 1841 C
exampled.exe!OptionHandler_FeedOptions(OPTIONHANDLER_HANDLE_DATA_TAG * handle, void * destinationHandle) Line 236 C
exampled.exe!wsio_setoption(void * ws_io, const char * optionName, const void * value) Line 670 C
exampled.exe!OptionHandler_FeedOptions(OPTIONHANDLER_HANDLE_DATA_TAG * handle, void * destinationHandle) Line 236 C
exampled.exe!xio_setoption(XIO_INSTANCE_TAG * xio, const char * optionName, const void * value) Line 184 C
exampled.exe!OptionHandler_FeedOptions(OPTIONHANDLER_HANDLE_DATA_TAG * handle, void * destinationHandle) Line 236 C
exampled.exe!restore_underlying_io_transport_options(AMQP_TRANSPORT_INSTANCE_TAG * transport_instance, XIO_INSTANCE_TAG * xio_handle) Line 613 C
exampled.exe!get_new_underlying_io_transport(AMQP_TRANSPORT_INSTANCE_TAG * transport_instance, XIO_INSTANCE_TAG * * xio_handle) Line 659 C
exampled.exe!establish_amqp_connection(AMQP_TRANSPORT_INSTANCE_TAG * transport_instance) Line 716 C
When invoked like this, the transport_instance->saved_tls_options OPTIONHANDLER_HANDLE gets evaluated which instead of containing a pointer to the value, contains the value itself.
The issue occurs in c-utility\src\tlsio_openssl.c on lines 1548ff, when the passed void pointer gets cast and accessed.
# Code sample exhibiting the issue:
` const int* version_option = value;`
`if ((int)(intptr_t)*version_option == 0 || (int)(intptr_t)*version_option == 10)`
` {`
` tls_io_instance->tls_version = VERSION_1_0;`
` }`
# Console log of the issue:
Exception thrown: read access violation.
version_option was 0x2.
Answers:
username_1: @username_0 - Thanks for filing this issue. We're investigating it.
username_2: Hi @username_0 ,
a fix was checked in and is now available on the latest code in master branch.
If possible, please validate it on your side and let us know the results.
Thanks for contributing to the Azure IoT SDK!
username_2: https://github.com/Azure/azure-c-shared-utility/commit/68470a382a3be1e64ace3082f1a6eb6dedfc4f6c
https://github.com/Azure/azure-c-shared-utility/commit/71ff9b4e900e6f5ecf2195d1f238441463b69449
Status: Issue closed
username_2: Hi @username_0 ,
this issue will be closed, but please feel free to reopen it if you need to follow up on the issue.
Thanks,
Azure IoT SDK Team |
dotnet/cli | 185108582 | Title: 'dotnet restore3' failing with "Unable to resolve 'Microsoft.NET.Sdk.Web'"
Question:
username_0: ## Steps to reproduce
Install SDK version `1.0.0-preview3-003892` and create an "empty" web project:
`mkdir test`
`cd test`
`dotnet new -t web`
`dotnet restore3`
## Expected behavior
- restoration of all NuGet packages
## Actual behavior
`restore3` fails with the following message:
`C:\Program Files\dotnet\sdk\1.0.0-preview3-003892\NuGet.targets(70,5): error : Unable to resolve 'Microsoft.NET.Sdk.Web (>= 1.0.0-alpha-20161019-1)' for '.NETCoreApp,Version=v1.0'.`
## Environment data
`dotnet --info` output:
.NET Command Line Tools (1.0.0-preview3-003892)
Product Information:
Version: 1.0.0-preview3-003892
Commit SHA-1 hash: 5301f08423
Runtime Environment:
OS Name: Windows
OS Version: 10.0.14393
OS Platform: Windows
RID: win10-x64
Answers:
username_1: Right now these packages are not published to NuGet.org yet, so you need a NuGet.Config pointing to a feed that contains the packages.
You can look for the one we have in the CLI as an example: https://github.com/dotnet/cli/blob/rel/1.0.0/NuGet.Config
Status: Issue closed
username_0: @username_1 thanks for the reply, I added the NuGet.Config file as described but I was still getting the error:
`Unable to resolve 'Microsoft.NET.Sdk.Web (>= 1.0.0-alpha-20161019-1)' for '.NETCoreApp,Version=v1.0'.`
So I changed the reference to the package in `test.csproj` to:
<PackageReference Include="Microsoft.NET.Sdk.Web">
<Version>1.0.0-alpha-*</Version>
</PackageReference>
in order to get the latest version of the package and `restore3` was able to resolve it, but now it's failing for another package:
`Unable to resolve 'Microsoft.NET.Sdk.Publish (>= 1.0.0-alpha-20161006)' for '.NETCoreApp,Version=v1.0'.`
I can't find a reference to `Microsoft.NET.Sdk.Publish` in any of the project files, so I'm at a loss regarding how to solve this. Any help is greatly appreciated.
Thanks in advance! |
mull-project/mull-jit-lab | 310539317 | Title: Objective-C: Implement missing property attributes.
Question:
username_0: ```
property const mull::objc::objc_property64 * 0x1047ea3a0 0x00000001047ea3a0
name const char * "hash" 0x00000001047e9d17
attributes const char * "TQ,R" 0x00000001047e9d1c
```
Answers:
username_1: ```
Property: searchList
token:T@"NSArray"
token:N
token:C
Assertion failed: (false && "Unknown type"), function parsePropertyAttributes, file /opt/mull-jit-lab/lab-jit-objc/llvm-jit-lab/src/ObjCRuntime.mm, line 342.
``` |
stanford-oval/genie-toolkit | 662144054 | Title: Heuristics for slot-filling measures & currencies
Question:
username_0: If the dialog agent asks for a currency or measure and the user provides a number, we should heuristically convert it.
Potentially, this could include a translatable annotation `#_[default_unit]` associated with the parameter indicating the default unit to use (which would be locale-dependent).
Answers:
username_0: This issue was transferred from almond-dialog-agent. It should be implemented in terms of templates: we change the templates such that filters and imprecise answers can use numbers where measurements are expected, and generate the right measurement code in the neural network. |
sebasmonia/panda | 474716180 | Title: Auto adjust column width
Question:
username_0: Make it an option (there's a chance that lists too big make it unusable).
Sample of `tabulated-list-mode` with adjustable cols: https://stackoverflow.com/a/28488084
Answers:
username_0: A newish binding in tablist mode uses `{` and `}` to modify column sizes, which makes this obsolete.
Status: Issue closed
|
alphagov/govuk-prototype-kit | 348729386 | Title: Make it easier to do branching
Question:
username_0: For example, we could let people add URLs to radio buttons. If a user selects that radio button, it automatically redirects to that URL, with no need for back end coding.
Answers:
username_1: Has this moved at all? be good to have this feature moving forward...
username_0: On potential solution is to significantly improve the documentation and examples for branching
username_2: One product that does this really well is, https://github.com/DEFRA/digital-form-builder, it's a UI which decreases any learning curve, but we could take inspiration from the way they do branches and implement it into the prototype kit.
Initial suggestion is have a file specifically for adding branches, that file needs to be changed and everything else will work.
A really simple, rushed design for that could be something like
```
const redirects = branchRedirect([
{
fieldName: 'eligibilityQuestion',
route: '/question-asked-here',
values: {
yes: '/go-here',
no: '/go-there',
'dont-know: 'go-anywhere',
},
}
]);
module.exports = redirects;
```
```
/question-asked-here.html
{{ govukRadios({
classes: "govuk-radios--inline",
idPrefix: "eligibilityQuestion",
name: "eligibilityQuestion",
fieldset: {
legend: {
text: "Are you the person paying?",
isPageHeading: true,
classes: "govuk-fieldset__legend--xl"
}
},
items: [
{
value: "yes",
text: "Yes"
},
{
value: "no",
text: "No"
},
{
value: "dont-know",
text: "I don't know"
}
]
}) }}
```
username_0: Another risk that would be good to get more research on: if we do radio button branching, is it then confusing to have 2 different ways to do branching? (radios vs routes) |
mrmike/Ok2Curl | 311104739 | Title: Android Studio 3 : Failed to resolve "com.github.mrmike:Ok2Curl:0.4.5"
Question:
username_0: as described in README,
added it to dependencies
`compile "com.github.username_1:Ok2Curl:0.4.5"`
jcenter() is there in the repository list and also in the buildScript list
Answers:
username_1: Hi @username_0
I think it's happening because of the typo in README file. I've already prepared a fix for that: https://github.com/username_1/Ok2Curl/pull/46. Cause library is available on jcenter(as you can see [here](https://jcenter.bintray.com/com/github/username_1/ok2curl/0.4.5/)).
Instead of using `com.github.username_1:Ok2Curl:0.4.5` use `com.github.username_1:ok2curl:0.4.5` (no uppercased letters).
Thanks
Status: Issue closed
|
inkle/ink | 165648799 | Title: Manipulate variables outside of Ink script.
Question:
username_0: I'd to be able to modify the value of a global variable (ie. set hasSpoken = true) that was declared in Ink from an external C# script. It would help with being able to control the flow of the story and identify speakers that are currently holding a conversation.
Is there any functionality built in that supports this? Looking through the available functions, the closest I could get was from observing the current variables.
I appreciate any and all help!
Answers:
username_1: It sounds like you're after the Setting/Getting variables logic: https://github.com/inkle/ink/blob/master/Documentation/RunningYourInk.md#settinggetting-ink-variables
unless I'm misunderstanding what it is you're aiming to do.
username_0: Ah, that's it. Thanks!
Status: Issue closed
|
marko-js/marko | 103744364 | Title: Please add for loop iteration with `index`
Question:
username_0: Would really like if we had the ability to use the index on iteration somehow...
Here using the imaginary `index='pos'` `for` attribute (default `index='index')
```jade
ui-grid(ui='inverted')
ui-row(ui='green' list='data.users.header')
for(each='item in data.users.list' index='pos')
ui-row(ui='silver' attrs='item' if='pos % 2 == 0')
ui-row(ui='blue' attrs='item' if='pos % 2 == 0')
```
Answers:
username_0: Sweet :) Learning every day...
Status: Issue closed
|
cms-sw/cmssw | 295937495 | Title: Avoid calling TClass::GetClass for all setRefCoreStreamer calls
Question:
username_0: The clall to `TClass::GetClass` in `setRefCoreStreamer` is meant to do a lazy evaluation for putting our own custom Streamers into ROOT. At present, the custom streamers are just there to customize the writing. The reading is customized by an IO rule. Setting the Streamer at read is probably historic, from the time before the IO rules.
It would probably be possible to move the actually change of the Streamers to the PoolOutputModule's constructor (since the custom Streamers are only needed for writing, not reading). The call to `EDProductGetter::switchProductGetter(...)` would have to remain to keep the IO rule functional.
Answers:
username_0: assign core
username_1: In the discussion with Philippe, the only case where we may not want the streamer set globally if a particular object would need to be serialized differently on output versus input (for example, if we wanted to use a custom streamer for `edm::Ref` on read but write out `edm::Ref` in a ntuple that can be read by bare ROOT.
We didn't think this applied to CMSSW, but felt it was reasonable to point out.
username_2: PR #22249 should address this issue and when/if merged allow us to close it.
username_2: PR #22249 was merged and this issue should be closed.
Status: Issue closed
username_0: +1 |
facebook/react | 639671702 | Title: hooks
Question:
username_0: How can I make the input display and get the focus of the input after the setstate value?
import React, { useState, useRef } from "react";
import "./styles.css";
export default function App() {
const [state, setState] = useState(false);
const refEl = useRef(null);
const clickBtn = () => {
setState(true);
refEl.current.focus();
};
return (
<div className="App">
<button onClick={clickBtn}>showIbout</button>
{state && <input ref={refEl} />}
</div>
);
}
Answers:
username_1: Support requests filed as GitHub issues often go unanswered. We want you to find the answer you're looking for, so we suggest the following alternatives:
##### Coding Questions
If you have a coding question related to React and React DOM, it might be better suited for Stack Overflow. It's a great place to browse through frequent questions about using React, as well as ask for help with specific questions.
[https://stackoverflow.com/questions/tagged/react](https://stackoverflow.com/questions/tagged/react)
##### Talk to other React developers
There are many online forums which are a great place for discussion about best practices and application architecture as well as the future of React.
[https://reactjs.org/community/support.html](https://reactjs.org/community/support.html#popular-discussion-forums)
Status: Issue closed
|
sigp/lighthouse | 560575706 | Title: Docker container doesn't exit gracefully
Question:
username_0: ## Description
After bringing up the lighthouse docker container using `docker-compose up`, killing the container using `Ctrl + C` causes the `beacon_node` process within the container to exit with exit status 137 (out of memory). Since the process doesn't exit gracefully, the beacon node doesn't get persisted and subsequently, bringing up the container again using `docker-compose up` errors with
```Failed to start Lighthouse: Failed to start beacon node: No persisted beacon chain found in store```
## Expected Behaviour
Killing the container with `Ctrl + C` should exit gracefully after persisting the beacon chain.<issue_closed>
Status: Issue closed |
Zrips/CMI | 758011564 | Title: Homes at deleted worlds not visible in /homes GUI
Question:
username_0: **Description of issue:**
Homes directing to worlds that don't exist anymore aren't shown in the /homes GUI.
This leads to a missleading GUI title (It says "7/10" but I only see 5 icons because 2 homes point to a deleted world) and rubbish in the cmi_users db.
---
**Cmi Version (using`/cmi version`):**
172.16.31.10
**Server Type (Spigot/Paperspigot/etc):**
Paperspigot
**Server Version (using `/ver`):**
Paper#294 1.16.4-R0.1-SNAPSHOT
Answers:
username_1: I think it's good that they aren't seen there, but they should be deleted probably.
username_0: Well of course, but it's lot easier to implement when you just show all icons. So if a player tries to tp to such a home, he'll already receive an appropiate error message so he will delete his home on his own.
You could also delete a home instantly when a player tries to use it the first time.
username_2: Confirmed for version 192.168.127.12.
Status: Issue closed
username_3: Issue with auto removal would be the cases when world is simply not loaded but home location should remain. Like some plugins will load world which is meant to be for specific player, if world was not loaded when player opened UI then his home would get removed which is no go. What i can do is to actually include those homes into UI but add extra indication that location is in none existing world, just to be more convenient. |
Heather/gentoo-gnome | 270300643 | Title: app-misc/tracker has some conflicts!
Question:
username_0: app-misc/tracker:0
(app-misc/tracker-2.0.1:0/100::gnome-next, ebuild scheduled for merge) conflicts with
<=app-misc/tracker-1.99.0 required by (gnome-extra/nautilus-tracker-tags-1.13.0:0/0::gnome-next, installed)
^^ ^^^^^^
<=app-misc/tracker-1.99:= required by (gnome-base/nautilus-3.26.0:0/0::gnome-next, installed)
^^ ^^^^
<=app-misc/tracker-1.99:0/100= required by (gnome-base/nautilus-3.26.0:0/0::gnome-next, installed)
^^ ^^^^
~app-misc/tracker-1.13.0 required by (gnome-extra/nautilus-tracker-tags-1.13.0:0/0::gnome-next, installed)
^ ^^^^^^
Answers:
username_1: @username_0 it's because I don't know how to make new tracker ebuild with multilib support :woman_shrugging:
username_2: Greetings. So what can I do to upgrade my gnome, get rid of nautilus?
username_1: remove `gnome-extra/nautilus-tracker-tags`
username_2: @username_1 just unmerged it but now gnome-online-miners and folks won't merge. Don't sure if I need them though...
username_1: @username_2 I don't know what's that, try to get rid of it too or check why you need it with `equery d gnome-online-miners`
username_2: @username_1 I ended up disabling gnome-tracker globally. Thanks for your help.
username_1: I need to recheck whether is issue is still being actual
username_2: Sorry cannot check if it's still present, I'm not happy gentoo user anymore.
Status: Issue closed
|
stripe/stripe-python | 1069075725 | Title: Feature Request: Please Add Type-hint
Question:
username_0: Hi, is there any plan for adding type-hint for Stripe?
It's hard for new comer like me to write correct method or variable type without type-hints.
Thanks.
Status: Issue closed
Answers:
username_1: @username_0 It's something we want to do in the future but no short term plans for this. Adding type hints will require typing most objects which in turns requiring "pinning" those objects to a specific version of our API, so that you have the right exact list of properties or parameters. This comes at a cost in ease of adoption of the library too and would require major reworks too
We know we will get to this in the future but it's not something we're working on at the moment unfortunately. I'm going to close as we already have https://github.com/stripe/stripe-python/issues/650 for this but thanks for the feedback, we definitely hear it! |
Hejsil/zig-clap | 314517765 | Title: Sub commands
Question:
username_0: Allow the clap to parse subcommands.
```
const parser = comptime Clap(Options).Builder
.init(
Options {
.print_values = false,
.a = 0,
.b = 0,
.c = 0,
}
)
.command(
Command.Builder
.init("com")
.arguments(
[]Argument {
Argument.Builder
.init("a")
.help("Set the a field of Option.")
.short('a')
.takesValue(true)
.build(),
}
)
.subCommands(
[]Command {
Command.Builder
.init("sub-com")
.arguments(
[]Argument {
Argument.Builder
.init("b")
.help("Set the a field of Option.")
.short('b')
.takesValue(true)
.build(),
}
)
.build()
}
)
.build()
)
.build();
```
Subcommands have their own options, which cannot be accessed by their parent. `com -a 1 sub-com -b 1` works, but `com -a 1 -b 1` does not.
Questions:
* Should sub commands be able to access the options of its parent?
* `com sub-com -a 1 -b 1`
* We could have an option that allows/disallows this.
* It would probably be useful to be able to tell from the resulting struct that a subcommand was parse.
* Should it just set some boolean field? Maybe subcommands will have sub results, which could be their own struct (or union, which would allow for all subcommands to return different tags of the same union).
* Should subcommands have an option to be required?
Answers:
username_0: This can be implemented as a wrapper on top of `StreamingClap`.
Status: Issue closed
username_0: Out of scope. This is a small package for `-` and `--` style argument parsing only. One could implement a package on top of `zig-clap` that does this. |
ampproject/amphtml | 255347305 | Title: doubleclick Error executing onCreativeRender: G(a.ownerDocument.defaultView).loadExtension is not a function. (In 'G(a.ownerDocument.defaultView).loadExtension("amp-analytics")', 'G(a.ownerDocument.defaultView).loadExtension' is undefined)
Question:
username_0: doubleclick Error executing onCreativeRender: G(a.ownerDocument.defaultView).loadExtension is not a function. (In 'G(a.ownerDocument.defaultView).loadExtension("amp-analytics")', 'G(a.ownerDocument.defaultView).loadExtension' is undefined)
user agent: Mozilla/5.0 (iPhone; CPU iPhone OS 10_2_1 like Mac OS X) AppleWebKit/602.1.50 (KHTML, like Gecko) GSA/33.0.164895372 Mobile/14D27 Safari/602.1
might possibly just be in canary right now. I don't have much other information but please reach out to me so we can dig for more if needed.
@username_2 @username_1 do you guys mind triaging this? Thanks
Answers:
username_1: Not currently able to repro this, even in canary. Do you have more detailed instructions for doing so?
username_2: +@username_3
This appears to line up with https://github.com/ampproject/amphtml/blame/master/extensions/amp-ad-network-doubleclick-impl/0.1/amp-ad-network-doubleclick-impl.js#L490. Looking at code history the code was changed as part of #10992 which changed from this.extensions_./*OK*/loadExtension('amp-analytics'); to this.extensions_./*OK*/installExtensionForDoc(this.getAmpDoc(), 'amp-analytics');
Could this be a version sync issue where the extension service changed and older code is consuming the new service?
username_3: @username_0 Is this on Canary?
The only way I see this happening is via version mismatch.
username_3: Setting to P0 until we confirm whether this is canary issue or PROD, at the very least.
I don't see how this would happen for the specified URL: this is RTV serving path.
username_2: This appears to be a canary issue (at least I can confirm that ad request is sending art=2 which indicates canary). We are not seeing any alerts associated with production traffic.
username_0: @username_3 i can only see canary right now with very low occurrences. double checking right now
username_3: Still, in RTV path we can never have this kind of issues, right?
username_0: @username_3 yes. is this code path not usually reached? im just curious about the low occurrence.
username_3: In this case it'd never be reached. Right?
username_3: In other words, we can never have RTV version mismatched. We can't have it in non-RTV either, but definitely not in RTV.
username_0: some addt'l info. notice that version ends with 313 (current canary) but rtv is 635 (current prod)
Query string:
v = 001504119607313
noAmp = 0
m = doubleclick Error executing onCreativeRender: G(a.ownerDocument.defaultView).loadExtension is not a function. (In 'G(a.ownerDocument.defaultView).loadExtension("amp-analytics")', 'G(a.ownerDocument.defaultView).loadExtension' is undefined)
a = 0
rt = 1p
ca = 1
vs = channelOpen
rvu = https://https%253A%252F%252Fwww.google.co.jp%252F/amp/s/gamp-ameblo-jp.cdn.ampproject.org/c/s/gamp.ameblo.jp/rkyhlove1207/entry-12198146478.html
mso = https://www.google.co.jp/
jse = Safari
exps = canary=1,expAdsenseA4A=0,expDoubleclickA4A=0,expAdSenseFFDR=0,a4aProfilingRate=1,dbclk_a4a_viz_change=0,ad-type-custom=1,ios-embed-wrapper=1,amp-apester-media=1,amp-playbuzz=1,chunked-amp=1,pump-early-frame=1,sticky-ad-early-load=1,amp-auto-ads=1,amp-auto-ads-adsense-holdout=0,slidescroll-disable-css-snap=1,visibility-v3=1,version-locking=1,as-use-attr-for-format=1,a4aFastFetchDoubleclickLaunched=0,a4aFastFetchAdSenseLaunched=0,a4a-measure-get-ad-urls=1,ad-loader-v2=1,3p-use-ampcontext=0,amp-animation=1,amp-sidebar toolbar=1
el = u
s = sd@https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:115:483
onCreativeRender@https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:135:415
https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:65:117
Yc@https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:93:392
Xc@https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:89:85
https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:82:105
promiseReactionJob@[native code]
r =
ae = doubleclick Error executing
SHORT = 1
Fragment:
username_4: Seems like canary runtime is loading prod version of ad extension script. As talked with @username_0 offline:
- It's not likely for canary amp runtime to know the current prod RTV and compose a wrong extension URL. More likely it's a environment problem
- This failure is before ad request, so it couldn't be the version from the ad either.
- We don't allow amp-ad in AMP ad, so it shouldn't be amp-inabox
I'm not very familiar what service-worker is doing for caching. Does it rewrite URL to full RTV version? Do we use service-worker on CDN in any case? @jridgewell
Also, worth to mention that this bug might exist for a while, just recently revealed because of the method signature change.
username_3: It looks like service workers are not currently involved. I really don't see how RTV can ever be wrong - they are always explicitly selected. The only possibility I see now is that Cache intermittently inserts wrong RTV. But not sure if that's really possible or the case here. I also don't think Cache ever inserts `amp-ad-network-doubleclick-impl-0.1.js`, but I could be wrong.
username_0: i also don't believe cache inserts `amp-ad-network-doubleclick-impl-0.1.js` from what @username_4 told me that is dynamically fetched (correct me if i'm wrong @username_4)
username_0: @username_3 is there any additional logging we can do to give us more information?
username_3: It is definitely dynamically fetched, but we are definitely considering to add them in Cache. Nor sure where that project is now.
/cc @codewiz
username_3: @username_0 To answer your question about logging. The current logging is pretty clear and points to a problem very well. We just don't know how we ever arrive at it.
username_1: The cache insertion project is still in the works and definitely hasn't been deployed to production yet.
username_3: @username_1 This is canary error. HAs it been deployed to canary?
username_1: No. Also, this is (will be) a feature of the Google AMP Cache; it's entirely separate from the AMP runtime release process.
username_3: @username_1 Definitely a Cache thing, I get this. The thing is, however, AMP Runtime has non way of knowing RTVs - it only knows one - it's own. While Cache knows them all. So, I just don't see it'd be possible to get a mismatch via runtime.
username_0: we just promoted 313 to prod and the issue occurred:
Query string:
v = 011504040004635
noAmp = 0
m = doubleclick Error executing onCreativeRender: G(...).loadExtension is not a function
a = 0
rt = 1p
jse = IE
exps = canary=0,expAdsenseA4A=0,expDoubleclickA4A=0,expAdSenseFFDR=0,dbclk_a4a_viz_change=0,a4aProfilingRate=0,ad-type-custom=1,ios-embed-wrapper=1,amp-apester-media=1,amp-playbuzz=1,chunked-amp=1,sticky-ad-early-load=1,amp-auto-ads=1,amp-auto-ads-adsense-holdout=0,slidescroll-disable-css-snap=1,version-locking=1,visibility-v3=1,a4aFastFetchDoubleclickLaunched=0,a4aFastFetchAdSenseLaunched=0,pump-early-frame=1,a4a-measure-get-ad-urls=0,ad-loader-v2=1,3p-use-ampcontext=0,amp-animation=1
el = u
s = amp-a4a: doubleclick Error executing onCreativeRender: G(...).loadExtension is not a function
at sd (https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:115:470)
at Ad.push.f.h.onCreativeRender (https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:135:413)
at https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:65:112
at Yc (https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:93:392)
at Xc (https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:89:83)
at https://cdn.ampproject.org/rtv/011504040004635/v0/amp-ad-network-doubleclick-impl-0.1.js:82:103
at <anonymous>
ae = doubleclick Error executing onCreativeRender: G(...).loadExtension is not a function
fr =
Fragment:
username_0: we've more or less ruled this out as a client side issue.
we've filed an issue at google internally to look into this (b/65461170)
username_0: well i haven't seen this in a while so Im going to close it.
Status: Issue closed
|
bbc/simorgh | 489729501 | Title: Improve setupTests.js
Question:
username_0: **Is your feature request related to a problem? Please describe.**
When an error is thrown it stops running other tests.
**Describe the solution you'd like**
Update setupTests.js to fix this.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Testing notes**
[Tester to complete]
Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc
- [ ] This feature is expected to need manual testing.
**Additional context**
Add any other context or screenshots about the feature request here.<issue_closed>
Status: Issue closed |
magefree/mage | 686452420 | Title: Historic matches not shown in the list
Question:
username_0: When I create a Historic match it doesn't appear in the matches list, so other players can't join.

Answers:
username_1: Fixed via 40fe3b137e52805dce19f8ae7db6167feb134c8d, will be available with next hotfix/release (soon).
Status: Issue closed
|
expo/sentry-expo | 831736001 | Title: Native Client is not available error when debug is set to false in Sentry.init call
Question:
username_0: ### Summary
I set up expo sentry in our app. At first, everything seemed to be working, but then I tried to disable debug in development and I now I cannot start the app, because of this error: SentryError: Native Client is not available, can't start on native.
This is my config. When I set `debug` to `true` it works again.
```
Sentry.init({
dsn: SENTRY_DSN,
enableInExpoDevelopment: false,
debug: false,
});
```
### Managed or bare workflow? If you have `ios/` or `android/` directories in your project, the answer is bare!
managed
### What platform(s) does this occur on?
iOS
### SDK Version (managed workflow only)
39
### Environment
Expo CLI 4.3.0 environment info:
System:
OS: macOS 11.2.3
Shell: 5.7.1 - /usr/local/bin/zsh
Binaries:
Node: 14.15.4 - ~/.nvm/versions/node/v14.15.4/bin/node
Yarn: 1.22.4 - ~/.yarn/bin/yarn
npm: 6.14.10 - ~/.nvm/versions/node/v14.15.4/bin/npm
Watchman: 4.9.0 - /usr/local/bin/watchman
Managers:
CocoaPods: 1.8.4 - /Users/username_0/.rvm/rubies/ruby-2.6.3/bin/pod
SDKs:
iOS SDK:
Platforms: iOS 14.4, DriverKit 20.2, macOS 11.1, tvOS 14.3, watchOS 7.2
Android SDK:
API Levels: 23, 25, 28, 29
Build Tools: 28.0.3, 29.0.2
System Images: android-23 | Google APIs Intel x86 Atom, android-28 | Intel x86 Atom_64, android-28 | Google APIs Intel x86 Atom, android-28 | Google Play Intel x86 Atom, android-29 | Google APIs Intel x86 Atom
IDEs:
Android Studio: 4.0 AI-193.6911.18.40.6514223
Xcode: 12.4/12D4e - /usr/bin/xcodebuild
npmPackages:
expo: ^39.0.4 => 39.0.4
react: 16.13.1 => 16.13.1
react-dom: 16.13.1 => 16.13.1
react-native: https://github.com/expo/react-native/archive/sdk-39.0.4.tar.gz => 0.63.2
Expo Workflow: managed
### Reproducible demo or steps to reproduce from a blank project
Just start sentry in App.js with disabled debug mode:
```
Sentry.init({
dsn: SENTRY_DSN,
enableInExpoDevelopment: false,
debug: false,
});
```
Answers:
username_1: What version of `sentry-expo`? This could be a duplicate of [any of these issues](https://github.com/expo/sentry-expo/search?q=Native+Client+is+not+available%2C+can%27t+start+on+native&type=issues)
username_0: Hi @username_1. Sorry I forgot to mention `sentry-expo` version. I am using `v 3.0.5`. I've seen the issues you linked above, unfortunately they are all closed without any resolution that would work for me.
As for the repro of the issue: I believe it happens as soon as you set `debug` to `false`. I am not sure what else to add to the repro steps but please let me know if there's anything more I can do to help you find the issue.
username_2: Same problem as this issue: https://github.com/expo/sentry-expo/issues/178
With this configuration and latest release of sentry-expo :
```
Sentry.init({
dsn: "blablabla",
enableNative: Device.isDevice,
enableInExpoDevelopment: Device.isDevice,
release: version,
environment: env,
integrations: (integrations) => (isDev ? [...integrations, new Debug() as Integration] : integrations),
debug: env !== "production",
enabled: !isDev,
});
```
username_1: @username_2 you have `enableNative: true`, which is why it's trying to initialize the native client. Be sure you:
- are in the bare workflow
- have `@sentry/react-native` in your dependencies
- run `npx pod-install` for iOS
username_2: @username_1 Oh. I'm using the managed workflow but it was working before. Removed `enableNative` and it's working as expected now. Thanks !
username_1: Closing since the above issue is resolved, and still haven't been able to repro @username_0's original issue
Status: Issue closed
|
joaodrp/gelf-formatter | 901831843 | Title: gelf-formatter induces logging error in django when uncaptured exception is raised
Question:
username_0: **Solution:**
This can be fixed by converting value to string:
`log_record[_prefix(key)] = str(value)`
While this breaks deep encoding json dicts,
this is fine for my use case (hopefully for other users, too)
Status: Issue closed
Answers:
username_1: Fixed with #18 |
uhlik/bpy | 460345638 | Title: PCV, 2.80, points somehow lost depth sorting, in fact, looks inverted now
Question:
username_0: blender-2.80.0-git20190625.c0c1b4542f39 25.6.2019
Answers:
username_0: blender-2.80.0-git20190625.c0c1b4542f39 25.6.2019
username_1: I can confirm that the depth sorting appears to be awry on the latest 2.8:
**Blender Version**
Broken: version: 2.80 (sub 74), branch: master, commit date: 2019-06-25 23:26, hash: `rBabaf6b106d3a`
Worked: (optional)
**But** it appears to work fine on this older build of 2.8, so maybe its an API compatibility issue?
**Blender Version**
Broken: version: 2.80 (sub 74), branch: blender2.7, commit date: 2019-06-17 21:07, hash: `rB741641f4c3b8`
Worked: (optional)
username_0: see here: https://developer.blender.org/T66148
will update it soon..
Status: Issue closed
username_0: fixed |
dubocr/homebridge-tahoma | 244801649 | Title: Error with more 10 blinds
Question:
username_0: Hello,
I've an error when I start a scene with more 10 blinds, is it possible to control more ?
This limit is in the plugin or in homekit?
Best Regards
Christopher
Answers:
username_1: Hi Christopher,
What kind of error do you have ? Have you log trace ? I have no knowledge of restriction but if this come from Tahoma you will have some error log in homebridge :
`sudo journalctl | grep homebridge`
Best
username_0: Hello,
the error in hombridge:
[2017-07-25 08:31:07] [Tahoma] Error 400 Execution queue is full (soft limit: 10) (EXEC_QUEUE_FULL)
username_1: Hi,
Seem this is a TaHoma API restriction so if you want to control more than 10 device with a scene, you have to create a TaHoma scene which will appear as on/off switch in homebridge.
username_0: Hi,
ok, the scene create in TaHoma don't appear in home...
I will test to create an automation with a fake button who control many scene.
Status: Issue closed
username_1: Hi,
Sorry, to expose TaHoma scenes as on/off switch you have to add the folowing line in your config file :
“exposeScenarios“: true
username_2: While this might be a limit of the Somfy API, I believe that homebridge-tahoma needs to implement some sort of fix to work around it. As a user, I should be able to make a HomeKit Scene that consists of more than 10 devices without thinking about whatever API restrictions lie behind.
Cheers,
username_1: Hi,
Yes, that’s true homebridge-tahoma needs to implement a work around.
Best,
username_3: Hi,
I have the same problem with more than 10 blinds. And when I tried to expose the TaHoma scenes I get the following error:
ReferenceError: Service is not defined
at new ScenarioAccessory (/usr/local/lib/node_modules/homebridge-tahoma/index.js:226:17)
at /usr/local/lib/node_modules/homebridge-tahoma/index.js:200:32
at /usr/local/lib/node_modules/homebridge-tahoma/overkiz-api.js:193:5
at Request.authCallback [as _callback] (/usr/local/lib/node_modules/homebridge-tahoma/overkiz-api.js:216:17)
at Request.self.callback (/usr/local/lib/node_modules/homebridge-tahoma/node_modules/request/request.js:185:22)
at Request.emit (events.js:196:13)
at Request.<anonymous> (/usr/local/lib/node_modules/homebridge-tahoma/node_modules/request/request.js:1161:10)
at Request.emit (events.js:196:13)
at IncomingMessage.<anonymous> (/usr/local/lib/node_modules/homebridge-tahoma/node_modules/request/request.js:1083:12)
at Object.onceWrapper (events.js:284:20)
at IncomingMessage.emit (events.js:201:15)
at endReadableNT (_stream_readable.js:1130:12)
at processTicksAndRejections (internal/process/task_queues.js:84:17)
What could be the problem?
Homebridge is running on my iMac:
+-- [email protected]
+-- [email protected]
Best regards
Joel
username_1: Hi,
Issue with Scénario must be fixed in last release 0.2.3.
May I’ll find a workaround for API limit to 10. Stay tuned
username_3: Thanks! My Scenarios from the TaHoma-Box now show up! :-)
username_1: Small improvement made. Command over quota (10) are postpone to be executed later (10 sec). |
ktorio/ktor | 483305218 | Title: ktor-client hangs forever with empty Input in formData
Question:
username_0: **To Reproduce**
```
input = Files.newInputStream(emptyFile)
...
httpClient.execute(HttpRequestBuilder().apply {
method = HttpMethod.Post
url(
...
)
body = MultiPartFormDataContent(
formData {
...
appendInput("file",
Headers.build {
append(HttpHeaders.ContentDisposition, "filename=$fileName")
append(HttpHeaders.ContentType, contentType)
}
) { input }
}
)
header(HttpHeaders.IfMatch, etag0)
}).use {response ->
...
}
```
Perhaps this is the same or related to [#1290 ](https://github.com/ktorio/ktor/issues/1290)
**Expected behavior**
1. size check should be added to Input.copyTo
```
val size = [email protected](buffer)
if (size < 0)
break;
written(size)
```
2. When exception occurs inside ktor-client it should not cause the call to hang<issue_closed>
Status: Issue closed |
spring-projects/spring-framework | 838740672 | Title: Autowiring HttpServletResponse in the request flow to set a header
Question:
username_0: Hi Colleagues,
I wanted confirm on one particular flow.
If there is a spring controller for REST and we auto-wire HttpServletResponse along the flow in any of the underlying classes. I mean not in @Controller, but any other @Service class. Still would it be RequestScoped?
Please help.
Answers:
username_1: Thanks for getting in touch, but it feels like this is a question that would be better suited to [Stack Overflow](http://stackoverflow.com/). As mentioned in [the guidelines for contributing](https://github.com/spring-projects/spring-framework/blob/master/CONTRIBUTING.md#discuss), we prefer to use the issue tracker only for bugs and enhancements. Feel free to update this issue with a link to the re-posted question (so that other people can find it) or add some more details if you feel this is a genuine bug.
I'm not sure I understand your question, but indeed [injecting a scoped-component](https://docs.spring.io/spring-framework/docs/current/reference/html/core.html#beans-factory-scopes) (you wouldn't inject the request itself).
I think that in general it's a good idea to use `@Controller` and other MVC infrastructure classes to act as an interface between the web and the application; extracting information from the request thanks to controller flexible signatures, or providing that information so that a shared component can set it as a header, is a better idea.
In your case, asking a question on StackOverflow and describing 1) what you're trying to achieve (the use case) 2) what you've tried so far and 3) why it's not working out for you so far - will help you more than questions here.
Thanks!
Status: Issue closed
|
rust-lang/rust | 631480463 | Title: ICE E0431.rs with -Zsave-analysis
Question:
username_0: <!--
Thank you for finding an Internal Compiler Error! 🧊 If possible, try to provide
a minimal verifiable example. You can read "Rust Bug Minimization Patterns" for
how to create smaller examples.
http://blog.pnkfx.org/blog/2019/11/18/rust-bug-minimization-patterns/
-->
### Code
code from `./src/test/ui/error-codes/E0431.rs`
```Rust
use {self}; //~ ERROR E0431
fn main () {
}
```
### Meta
<!--
If you're using the stable version of the compiler, you should also check if the
bug also exists in the beta or nightly versions.
-->
repo is @ 219380d840877de97b0e17674be7212ebc929d51
### Error output
`build/x86_64-unknown-linux-gnu/stage2/bin/rustc ./src/test/ui/error-codes/E0431.rs -Zsave-analysis`
```
thread 'rustc' panicked at 'index 18446744073709551615 out of range for slice of length 0', src/librustc_save_analysis/dump_visitor.rs:213:21
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
error: internal compiler error: unexpected panic
note: the compiler unexpectedly panicked. this is a bug.
```
<!--
Include a backtrace in the code block by setting `RUST_BACKTRACE=1` in your
environment. E.g. `RUST_BACKTRACE=1 cargo build`.
-->
<details><summary><strong>Backtrace</strong></summary>
<p>
```
error[E0431]: `self` import can only appear in an import list with a non-empty prefix
--> ./src/test/ui/error-codes/E0431.rs:1:6
|
1 | use {self}; //~ ERROR E0431
| ^^^^ can only appear in an import list with a non-empty prefix
warning: unused import: `self`
--> ./src/test/ui/error-codes/E0431.rs:1:6
|
1 | use {self}; //~ ERROR E0431
| ^^^^
|
= note: `#[warn(unused_imports)]` on by default
[Truncated]
error: internal compiler error: unexpected panic
note: the compiler unexpectedly panicked. this is a bug.
note: we would appreciate a bug report: https://github.com/rust-lang/rust/blob/master/CONTRIBUTING.md#bug-reports
note: rustc 1.45.0-dev running on x86_64-unknown-linux-gnu
note: compiler flags: -Z save-analysis
query stack during panic:
end of query stack
error: aborting due to previous error; 1 warning emitted
For more information about this error, try `rustc --explain E0431`.
```
</p>
</details>
Answers:
username_0: another one: `./src/test/ui/issues/issue-47673.rs`
````rust
// check-pass
#![allow(unused_imports)]
use {{}, {}};
fn main() {}
````
username_0: another one: `./src/test/ui/issues/issue-10806.rs`
````rust
// run-pass
#![allow(unused_imports)]
// pretty-expanded FIXME #23616
pub fn foo() -> isize {
3
}
pub fn bar() -> isize {
4
}
pub mod baz {
use {foo, bar};
pub fn quux() -> isize {
foo() + bar()
}
}
pub mod grault {
use {foo};
pub fn garply() -> isize {
foo()
}
}
pub mod waldo {
use {};
pub fn plugh() -> isize {
0
}
}
pub fn main() {
let _x = baz::quux();
let _y = grault::garply();
let _z = waldo::plugh();
}
````
username_1: I am on it
username_2: Same issue happened after upgraded to the latest version today.
username_1: Yes the fix #73046 has not landed yet.
Status: Issue closed
|
haslo/lists_juggle_browser | 216217379 | Title: Icons for ships are not correct in Microsoft Edge 38.14393.0.0
Question:
username_0: Was checking out the meta-wing page on my surface and noticed that the icons for many of the ships are not correct. Screenshot attached.

Answers:
username_1: Gee, thanks. Looks like there's other cross-browser problems, with a variety of browsers asking before any page change. Gonna have to fix this.
username_1: I'm not sure whether I can fix the icons themselves; Edge breaks them in the original repository as well:
https://geordanr.github.io/xwing-miniatures-font/
I guess we should go file a bug there, too. Or with Microsoft, preferrably.
username_1: FWIW, I've fixed the nagging "do you really want to change page" bug in a variety of browsers. Not the actual bug you've reported though, so I'm leaving this open.
username_0: Thanks for doing that... that one was really getting to me too :) Good job! |
aquasecurity/tracee | 979857390 | Title: container_id initialization regex has to be fixed
Question:
username_0: When running in an Ubuntu environment with LXD, the regex used to recognize running containers, and populate the map in charge for tracking existing containers + pids, will have false positives like:
```
./fs/cgroup/systemd/user.slice/user-1000.slice/[email protected]/app.slice/snap.lxd.lxc.4b3296a2-c139-4da5-b782-006c431cca74.scope/tasks
```
```
$ sudo ./dist/tracee-ebpf --debug --trace event=mount
BTF: bpfenv = false, btfenv = false, vmlinux = true
BPF: using embedded BPF object
unpacked CO:RE bpf object file into memory
TIME UID COMM PID TID RET EVENT ARGS
Running container = vice/app.slice/snap.lxd.lxc.4b3296a2-c139-4da5-b782-006c431cca74 pid = 2558259
```
We should not have "**UUID.scope/tasks**" regex as a 'generic' regex for now IMO.<issue_closed>
Status: Issue closed |
tummychow/git-absorb | 748583606 | Title: Getting "old path already occupied"
Question:
username_0: Great tool! Unfortunately, I was trying to absorb some changes today and got this error:
```
Nov 23 09:17:36.878 CRIT absorb failed, err: old path already occupied
```
It is not clear to me why am I getting this error, what it means, or how to solve it. :(
Answers:
username_1: i can't diagnose this issue without more information. that error comes from my patch parsing code, so i need to see the diffs of your git index (`git diff --cached`) and the commits in your git absorb stack
username_0: I cannot share the exact diff (it's proprietary source code), but I could locate the issue: I had a symbolic link to a file that I replaced with an actual file. This gives a diff similar to this one:
```diff
diff --git a/some-file.ext b/some-file.ext
new file mode 100644
index 0000000000..7cb58a6452
--- /dev/null
+++ b/some-file.ext
@@ -0,0 +1,135 @@
+# File contents following here
```
username_1: ah... i can see how that would cause this problem
```
diff --git a/bar b/bar
deleted file mode 120000
index 1910281..0000000
--- a/bar
+++ /dev/null
@@ -1 +0,0 @@
-foo
\ No newline at end of file
diff --git a/bar b/bar
new file mode 100644
index 0000000..e69de29
```
looks like link and non-link changes get different hunks. that'll require a special case in the parser
username_0: I've found another case not involving symbolic links where I'm getting this error:
```patch
diff --git a/some_path/pvm.spec b/some_path/pvm.spec
index aa58341137..943bacfa2c 100644
--- a/some_path/pvm.spec
+++ b/some_path/pvm.spec
@@ -58,7 +58,7 @@ tar -xzf %{SOURCE0}
cd pvm3
tar -xzf %{SOURCE1}
%patch0 -p0
-#patch1 -p1
+%patch1 -p1
%patch2 -p1
%patch3 -p1
%patch4 -p1
```
username_1: is that the entire index? this bug requires a diff with two patches that both have the same filename. your example only has one patch in it
username_0: Yes, that is the entire index.
username_1: yeah that sounds likely, every commit has to get parsed the same way as the index. probably one of them has the same issue where a symlink is being turned into a regular file or vice versa |
BEEmod/BEE2-items | 213603354 | Title: Glass Item causes "Failure To Compile" on P1 Style
Question:
username_0: VBSP failed.
[INFO] vbsp.main(): BEE VBSP hook initiallised.
[INFO] cond.core.import_conditions(): Imported all conditions modules!
[INFO] vbsp.main(): Map path is "E:/Steam/steamapps/common/Portal 2/sdk_content\maps/preview.vmf"
[INFO] vbsp.main(): New path: "E:/Steam/steamapps/common/Portal 2/sdk_content\maps\styled\preview.vmf"
[INFO] vbsp.main(): Clearing inject/ directory..
[INFO] vbsp.main(): PeTI map detected!
[INFO] vbsp.main(): Loading settings...
[WARNING] vbsp_options.load(): Extra config options: {'bottomless_pit': '0', 'fizz_visiblity': '0'}
[WARNING] instanceLocs.get_subitems(): Invalid custom instance name - "piston_grate_low" for <item_rex_pellet_destroyer> (Valid: {'head': 'instances/bee2/clean/items/rex/pellet_dest/head.vmf', 'piston_logic_0': 'instances/bee2/clean/items/rex/pellet_dest/logic_0.vmf', 'piston_logic_1': 'instances/bee2/clean/items/rex/pellet_dest/logic_1.vmf', 'piston_logic_2': 'instances/bee2/clean/items/rex/pellet_dest/logic_2.vmf', 'piston_logic_3': 'instances/bee2/clean/items/rex/pellet_dest/logic_3.vmf', 'common': 'instances/bee2/p1/items/rex/pellet_dest/common.vmf', 'piston_bottom_0': 'instances/bee2/p1/items/rex/pellet_dest/pist_0.vmf', 'piston_bottom_1': 'instances/bee2/p1/items/rex/pellet_dest/pist_1.vmf', 'piston_bottom_2': 'instances/bee2/p1/items/rex/pellet_dest/pist_2.vmf', 'piston_bottom_3': 'instances/bee2/p1/items/rex/pellet_dest/pist_3.vmf', 'piston_static_0': 'instances/bee2/p1/items/rex/pellet_dest/static_0.vmf', 'piston_static_1': 'instances/bee2/p1/items/rex/pellet_dest/static_1.vmf', 'piston_static_2': 'instances/bee2/p1/items/rex/pellet_dest/static_2.vmf', 'piston_static_3': 'instances/bee2/p1/items/rex/pellet_dest/static_3.vmf', 'piston_static_4': 'instances/bee2/p1/items/rex/pellet_dest/static_4.vmf'})
[WARNING] instanceLocs.get_subitems(): Invalid custom instance name - "piston_grate_high" for <item_rex_pellet_destroyer> (Valid: {'head': 'instances/bee2/clean/items/rex/pellet_dest/head.vmf', 'piston_logic_0': 'instances/bee2/clean/items/rex/pellet_dest/logic_0.vmf', 'piston_logic_1': 'instances/bee2/clean/items/rex/pellet_dest/logic_1.vmf', 'piston_logic_2': 'instances/bee2/clean/items/rex/pellet_dest/logic_2.vmf', 'piston_logic_3': 'instances/bee2/clean/items/rex/pellet_dest/logic_3.vmf', 'common': 'instances/bee2/p1/items/rex/pellet_dest/common.vmf', 'piston_bottom_0': 'instances/bee2/p1/items/rex/pellet_dest/pist_0.vmf', 'piston_bottom_1': 'instances/bee2/p1/items/rex/pellet_dest/pist_1.vmf', 'piston_bottom_2': 'instances/bee2/p1/items/rex/pellet_dest/pist_2.vmf', 'piston_bottom_3': 'instances/bee2/p1/items/rex/pellet_dest/pist_3.vmf', 'piston_static_0': 'instances/bee2/p1/items/rex/pellet_dest/static_0.vmf', 'piston_static_1': 'instances/bee2/p1/items/rex/pellet_dest/static_1.vmf', 'piston_static_2': 'instances/bee2/p1/items/rex/pellet_dest/static_2.vmf', 'piston_static_3': 'instances/bee2/p1/items/rex/pellet_dest/static_3.vmf', 'piston_static_4': 'instances/bee2/p1/items/rex/pellet_dest/static_4.vmf'})
[INFO] vbsp.load_settings(): Settings Loaded!
[INFO] vbsp.load_map(): Parsing Map...
[INFO] vbsp.load_map(): Reading Map...
[INFO] vbsp.load_map(): Loading complete!
[INFO] vbsp.mod_entryexit(): Setting SP Entry to 6
[INFO] vbsp.mod_entryexit(): Setting SP Exit to 3
[INFO] vbsp.get_map_info(): Game Mode: SP
[INFO] vbsp.get_map_info(): Is Preview: True
[INFO] brushLoc.read_from_map(): Analysed map, filling air... (21 starting positions..)
[INFO] brushLoc.read_from_map(): Air filled!
[INFO] cond.core.check_all(): Checking Conditions...
[INFO] cond.fizzler.fizzler_out_relay(): Fizzler data: {'marker_inst': [], 'pos_to_name': {}, 'fizz_by_name': {}, 'model': 'instances/bee2/p1/items/tdb/csfi_model.vmf', 'base': 'instances/bee2/p1/items/tdb/csfi_base.vmf', 'fizz_id': 'item_tdb_csfi', 'fizz_bases': {'instances/bee2/p1/items/fizzler/fizzler_base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/clean/items/paint_fizz/base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/p1/items/fr/csf_base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/p1/items/fr/fdf_base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/p1/items/fr/mif_base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/p1/items/death_fizz/base.vmf': ItemConnections(in_act=('branch_fizz', 'ToggleTest'), in_deact=('branch_las', 'ToggleTest'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/p1/items/physler/base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, '')), 'instances/bee2/clean/items/rex/pellet_field_base.vmf': ItemConnections(in_act=(None, ''), in_deact=(None, ''), out_act=('out', 'OnUser2'), out_deact=('out', 'OnUser1')), 'instances/bee2/p1/items/tdb/csfi_base.vmf': ItemConnections(in_act=('counter', 'Add'), in_deact=('counter', 'Subtract'), out_act=(None, ''), out_deact=(None, ''))}, 'fizz_models': {'instances/bee2/p1/items/physler/model.vmf', 'instances/bee2/p1/items/tdb/csfi_model.vmf', 'instances/bee2/p1/items/fizzler/fizzler_model.vmf', 'instances/bee2/p1/items/fr/mif_model.vmf', 'instances/bee2/p1/items/fr/csf_model.vmf', 'instances/bee2/clea[INFO] cond.catwalks.res_make_catwalk(): Starting catwalk generator...
[INFO] cond.scaffold.res_unst_scaffold(): Running Scaffold Generator (BEE2_SCAFFOLD)...
[INFO] cond.scaffold.res_unst_scaffold(): Finished Scaffold generation!
[INFO] cond.vactubes.res_make_vactubes(): Running Generator (BEE2_VACTUBES)...
[INFO] vbsp.set_elev_videos(): Elevator type: NONE
[INFO] voiceLine.add_voice(): Adding Voice Lines!
[INFO] voiceLine.add_voice(): Quote events: []
[INFO] voiceLine.add_voice(): No responses data..
[INFO] voiceLine.add_voice(): 0 Mid quotes
[INFO] voiceLine.add_voice(): Done!
[INFO] cond.monitor.mon_camera_link(): Bullseye defaultdict(<class 'int'>, {})
[INFO] vbsp.set_player_portalgun(): Setting Portalgun:
[INFO] vbsp.set_player_portalgun(): Blue: Y, Orange: Y
[INFO] vbsp.set_player_portalgun(): Done!
[INFO] vbsp.add_screenshot_logic(): Added Screenshot Logic
[INFO] cond.core.check_all(): Map has attributes: ['pack_p1_radio_clean', 'spawn_dual']
[INFO] cond.core.check_all(): instanceLocs cache: CacheInfo(hits=5318, misses=335, maxsize=256, currsize=256)
[INFO] cond.core.check_all(): Style Vars: {'blockplayergelbomb': True, 'doorcameras': True, 'p1_ported_pgun': True, 'restartonexit': False, 'multiversecave': False, 'fixfizzlerbump': False, 'nomidvoices': False, 'unlockdefault': False, 'allowgoomist': False, 'funnelallowswitchedlights': True, 'enableshapesignageframe': True}
[INFO] cond.core.check_all(): Global instances: {'instances/BEE2/p1/set_portalgun.vmf', 'instances/BEE2/p1/global_ents.vmf'}
[INFO] vbsp.add_extra_ents(): Adding Music...
[INFO] vbsp.add_extra_ents(): Adding Global PTI Ents
[INFO] vbsp.change_ents(): Editing Other Entities...
[INFO] vbsp.fixup_goo_sides(): Changing goo sides...
[INFO] vbsp.fixup_goo_sides(): Done!
[INFO] vbsp.change_brush(): Editing Brushes...
[INFO] vbsp.change_brush(): Make Bottomless Pit: False
[INFO] vbsp.change_brush(): Goo heights: 0 <- Counter()
[INFO] vbsp.change_brush(): Adding Goo Mist...
[INFO] vbsp.change_brush(): Done!
[INFO] vbsp.change_overlays(): Editing Overlays...
[INFO] vbsp.change_trig(): Editing Triggers...
[INFO] vbsp.collapse_goo_trig(): Collapsing goo triggers...
[INFO] vbsp.collapse_goo_trig(): Done!
[INFO] vbsp.change_func_brush(): Editing Brush Entities...
[ERROR] (unknown file).(unknown function)(): Uncaught Exception:
Traceback (most recent call last):
File "<frozen importlib._bootstrap>", line 961, in _find_and_load
[Truncated]
File "<frozen importlib._bootstrap>", line 646, in _load_unlocked
File "<frozen importlib._bootstrap>", line 616, in _load_backward_compatible
File "C:\Users\Lautaro\AppData\Local\Programs\Python\Python36\lib\site-packages\cx_Freeze\initscripts\__startup__.py", line 12, in <module>
__import__(name + "__init__")
File "<frozen importlib._bootstrap>", line 961, in _find_and_load
File "<frozen importlib._bootstrap>", line 950, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 646, in _load_unlocked
File "<frozen importlib._bootstrap>", line 616, in _load_backward_compatible
File "C:\Users\Lautaro\AppData\Local\Programs\Python\Python36\lib\site-packages\cx_Freeze\initscripts\Console.py", line 24, in <module>
exec(code, m.__dict__)
File "vbsp_launch.py", line 7, in <module>
File "C:\Users\Lautaro\Documents\BEEMOD 2.4 GHITUB\BEE2.4\src\vbsp.py", line 3583, in main
change_func_brush()
File "C:\Users\Lautaro\Documents\BEEMOD 2.4 GHITUB\BEE2.4\src\vbsp.py", line 2829, in change_func_brush
grate_temp.apply(side, change_mat=False)
File "C:\Users\Lautaro\Documents\BEEMOD 2.4 GHITUB\BEE2.4\src\template_brush.py", line 288, in apply
mat, face.uaxis, face.vaxis, face.ham_rot = self[face.normal().as_tuple()]
File "C:\Users\Lautaro\Documents\BEEMOD 2.4 GHITUB\BEE2.4\src\template_brush.py", line 272, in __getitem__
mat, axis_u, axis_v, rotation = self._axes[tuple(normal)]
ValueError: not enough values to unpack (expected 4, got 3)
Status: Issue closed
Answers:
username_1: Fixed. |
greenboxal/dns-heaven | 400218760 | Title: kubectl throws error related to Google OAuth when dns-heaven is up
Question:
username_0: Thanks for creating this tool, it temporarily fixed the problem with kube dns.
But after some time kubectl started to give this error
```
Unable to connect to the server: failed to refresh token: Post https://oauth2.googleapis.com/token: dial tcp: lookup oauth2.googleapis.com on 127.0.0.1:53: read udp 127.0.0.1:59449->127.0.0.1:53: i/o timeout
```
when trying to execute any command.
e.g.
```
kubectl version
Client Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.2", GitCommit:"<PASSWORD>", GitTreeState:"clean", BuildDate:"2018-08-07T23:17:28Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"darwin/amd64"}
Unable to connect to the server: failed to refresh token: Post https://oauth2.googleapis.com/token: dial tcp: lookup oauth2.googleapis.com on 127.0.0.1:53: read udp 127.0.0.1:59449->127.0.0.1:53: i/o timeout
```
If I kill the dns-heaven process, kubectl works as before (still giving error related to dns)
```
kubectl version
Client Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.2", GitCommit:"<PASSWORD>", GitTreeState:"clean", BuildDate:"2018-08-07T23:17:28Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"darwin/amd64"}
The connection to the server ... was refused - did you specify the right host or port?
```
Answers:
username_0: Found this log in dns-heaven output
```
ERRO[0005] error resolving query error="read udp 192.168.0.65:56801->172.16.58.3:53: i/o timeout" ns="172.16.58.3:53" qname=oauth2.googleapis.com.
ERRO[0005] error resolving query error="read udp 192.168.0.65:62111->172.16.58.3:53: i/o timeout" ns="172.16.58.3:53" qname=oauth2.googleapis.com.
```
username_1: I think the problem is with your DNS server: `172.16.58.3`
It may be blocking me (being in Russia and me in the USA) but I get a timeout whenever using that DNS server.
username_2: I have the same trouble but with terraform aws provider.
How did you get dns-heaven output? |
dotnet/roslyn | 863008039 | Title: Need a way to detect Power Save mode
Question:
username_0: Currently there is no way to detect Power Save mode. Since features may operate more slowly or with reduced functionality in this mode, we need a way to uniquely detect a Power Save request from other requests (e.g. #29123). |
amcharts/amcharts5 | 1075379355 | Title: Custom Data in Tooltip With Drill Down Map Chart
Question:
username_0: Issue:
When using custom data for continents in Map chart, the continent will become unclickable (click event will not fire).
**_My Scenario_**
While trying to build a drilldown map with:
- Tooltip on continents will show calculated value: "{name} : {value}".
- Clicking on continent will zoom-in to the continent and show the countries map.
If Asia will have a value = 150, the tooltip will show "Asia : 150".
```
let continentSeries = chart.series.push(am5map.MapPolygonSeries.new(root, {
geoJSON: am5geodata_continentsLow,
exclude: ["antarctica"]
}));
continentSeries.data.setAll([{
id: "asia",
name: "Asia",
value: 150
}]);
continentSeries.mapPolygons.template.setAll({
tooltipText: "{name}: {value}",
interactive: true,
});
```
Now, when clicking Asia (or any other continent), we will zoom in and show it's contries:
```
continentSeries.mapPolygons.template.events.on("click", function (ev) {
continentSeries.zoomToDataItem(ev.target.dataItem as am5.DataItem<am5map.IMapPolygonSeriesDataItem>);
continentSeries.hide();
countrySeries.show();
homeButton.show();
});
```
For each continent in the `continentSeries.data.setAll([...]);`, the click event will not fire.
Thanks for reading 👍
Answers:
username_1: Try moving `setData()` call **after** everything else:
```JavaScript
let continentSeries = chart.series.push(am5map.MapPolygonSeries.new(root, {
geoJSON: am5geodata_continentsLow,
exclude: ["antarctica"]
}));
continentSeries.mapPolygons.template.setAll({
tooltipText: "{name}: {value}",
interactive: true,
});
continentSeries.mapPolygons.template.events.on("click", function (ev) {
continentSeries.zoomToDataItem(ev.target.dataItem as am5.DataItem<am5map.IMapPolygonSeriesDataItem>);
continentSeries.hide();
countrySeries.show();
homeButton.show();
});
continentSeries.data.setAll([{
id: "asia",
name: "Asia",
value: 150
}]);
```
username_0: Moving setAll call after defining click fixed the issue.
Thanks!
Status: Issue closed
|
MaelSemler/mini-cap | 576427390 | Title: DEV-9 Autocomplete on search fields
Answers:
username_1: **UI Mockup**
Portrait:
<img width="271" alt="Capture d’écran 2020-03-05 à 20 13 56" src="https://user-images.githubusercontent.com/35980232/76040640-eb06c080-5f1d-11ea-8bae-e5bae68d1742.png">
Landscape:
<img width="731" alt="Capture d’écran 2020-03-05 à 20 14 45" src="https://user-images.githubusercontent.com/35980232/76040677-feb22700-5f1d-11ea-9b12-d56b177f27a1.png">
username_2: Acceptance Criteria
- [ ] The feature must complete a user's entry
- [ ] The feature must suggest all nearby places according to the user's entry
- [ ] The feature must display an empty "Heart" icon (non-functional)
username_2: https://github.com/username_0/mini-cap/issues/80
Status: Issue closed
|
bbcarchdev/spindle | 142327742 | Title: Add standalone spindle-migrate utility
Question:
username_0: This utility will perform two functions:
1. Ensure that the database schema is up to date (analogous to `twine -S` with the Spindle modules loaded)
2. Perform any one-time data migrations which aren't suitable for inclusion in the normal schema update process, because they may be long-running
Answers:
username_0: Sanity check after migrating proxies into `state` table is required: the number of rows in the `proxy` table should match the number of rows in the `state` table.
Status: Issue closed
|
wso2/product-is | 530218106 | Title: Tenant supported claims not saved to the user profile during self registration
Question:
username_0: When trying the self-registration feature of Identity Server with a secondary userstore, tenant supported claims are not saved only super tenant supported claims are saved.
**Steps to reproduce**
1) Create a tenant in the Identity Server 5.8.0
2) Enable self-registration with account confirmation for this tenant [1].
3) Create a new claim and make this supported by default for the tenant.
4) Go to the dashboard app and click on Register Now.
5) In the username option provide the username as "userstoredomain/username@tenantdomain"
6) Fill in the values on the form and click on submit.
7) Login to the management console and view the user profile. You can see that the new claim value is not set.
**Affected Versions**
wso2is-5.8.0
[1] - https://docs.wso2.com/display/IS580/Self-Registration+and+Account+Confirmation<issue_closed>
Status: Issue closed |
dotnet/roslyn | 416433980 | Title: Suggestion mode does not work in C# Interactive view with async completion
Question:
username_0: **Version Used**: 2773f5e
The command is not available. We did not have any tests that relied on in working (they should be added), but several tests assumed the command was available.
Answers:
username_1: @username_2 is this going to be fixed for 16.0/16.1 ?
@ivanbasov to followup and add test our end.
username_2: yes it will be fixed in 16.0 |
matthew-brett/delocate | 1002943992 | Title: Code cleanup suggestions.
Question:
username_0: Some ideas that could make things easier for me and future maintainers:
I'd be nice to setup auto formatters like Black and isort. It's tedious to format all code manually.
[doc/devel/make_release.rst](https://github.com/username_1/delocate/blob/dbb9367bcf2b8519199587b5336bcbd2d12e3c82/doc/devel/make_release.rst) should have been in a `CONTRIBUTING` file. I missed it entirely when I made the `0.9.1` release. A `CONTRIBUTING` file could also be used to explain how to add any test data.
In #93 I attempted to fully type-hint the package. I never got that merged but it showed some areas that are causing issues such as `InWheelCtx.__enter__` and `back_tick`. I think `InWheelCtx` needs to be refactored or removed (or finish #46 and fix things there,) and all calls to `back_tick` should be replaced with `subprocess.run` now that the code base is on Python 3.
Testing has some issues. Everything from `pytest_tools.py` breaks the pytest debugger so they'll all eventually need to be replaced with plain asserts. It'd be nice to replace `scriptrunner.py` with a more standard tool like [pytest-console-scripts](https://pypi.org/project/pytest-console-scripts/).
Issue and PR templates might be a good idea. It sucks that most of the current bug reports don't mention which version of decloate they used and don't provide a wheel that can be checked.
Answers:
username_1: All good suggestions. Would you like me to review #93? I suppose we should keep / deprecate `backticks` for a bit, just in case, but yes, all uses should be replaced with `subprocess.run`. |
quasarframework/quasar | 1082612329 | Title: Update event for q-toggle passes different targets
Question:
username_0: Greetings!
I've been trying to capture status changes for q-toggle, using @update:model-value, so I can take actions with a custom function.
The problem is that the custom function receives the JS event object, but the 'target' references the toggle 'thumb', rather than the toggle itself.
In the custom function, I need to be able to retrieve the toggle name or id so I know what to do with it.
Code here: https://jsfiddle.net/username_0/nL98x5zk/22/
Steps to reproduce the behavior:
1. Open the jsfiddle page
2. Open the browser console (not the console in jsfiddle, as it outputs too much text)
3. Flip the toggle ON & OFF to run the custom function on status change
4. Look in the console: evt.target points to the toggle 'thumb'. But in the DOM, the toggle id 'someId' and name 'someName' are set at the parent.parent and sibling level:
<div class="q-toggle cursor-pointer no-outline row inline no-wrap items-center" tabindex="0" role="checkbox" aria-checked="true" id="**someId**">
<div class="q-toggle__inner relative-position non-selectable q-toggle__inner--truthy">
<input class="hidden q-toggle__native absolute q-ma-none q-pa-none" type="checkbox" name="**someName**" value="true">
<div class="q-toggle__track">
</div>
<div class="**q-toggle__thumb** absolute flex flex-center no-wrap">
</div>
</div>
</div>
**Expected behavior**
Shouldn't the evt.target always point to the top element of the toggle where the id is set?
**Platform (please complete the following information):**
Quasar Version: @quasar/cl v1.2.1
@quasar/app Version: 3.1.10
Quasar mode:
- [ X ] SPA
- [ ] SSR
- [ ] PWA
- [ ] Electron
- [ ] Cordova
- [ ] Capacitor
- [ ] BEX
- [ ] Vite Plugin
Additonal Info:
- [ ] Typescript
- [ ] SFC usage with `<script setup>`
- [ ] Vuex
- [ ] Prettier (eslint)
- [ ] Standard (eslint)
Tested on:
- [ X ] SPA
- [ ] SSR
- [ ] PWA
- [ ] Electron
- [ ] Cordova
- [ ] Capacitor
- [ ] BEX
OS: MacOS
Node:
NPM:
Yarn:
Browsers: Safari, Chrome
iOS:
Android:
Electron:
Answers:
username_1: To put it in apple's terms, you are holding it wrong :)
https://jsfiddle.net/a3suj25f/
The event is the DOM event that triggered the change.
If you want to do something you don't need the DOM, you have the model.
Status: Issue closed
username_0: @username_1 Thank you! That makes total sense. |
ipython/ipython | 59180393 | Title: Horse in tab completion.
Question:
username_0: In notebook.
```
object.__dict__.__.<tab><delay1><c><delay2><tab>
```
will complete to `copy` `copyopy` `<someting else weird>` depending on the value of `<delay1>` and `<delay2>` especially reproducible if `<delay1>` ~1s and `<delay2>` ~ (fast enough but not too much).
Discover trying to fix #4777 that might be related.
Answers:
username_1: Should this have been closed by #7889, or is there still some issue?
username_1: @username_0 shouldn't this have been closed by #7889?
Status: Issue closed
username_0: Oh, yes, it should I guess github does not parse the AND. |
bevyengine/bevy | 684252353 | Title: Scale for OrthographicProjection
Question:
username_0: Right now, the `OrthographicProjection` seems to have the units in pixels, which makes any 3d renders very small compared to the same render using `PerspectiveProjection`. Perhaps there should be a `Scale` field added to the `OrthographicProjection` and used [here](https://github.com/bevyengine/bevy/blob/7b6fbcb932f240e7789bb4ea292cff7eb3682cc9/crates/bevy_render/src/camera/projection.rs#L79) to allow one to shrink the very wide field of view.
Answers:
username_1: Just ran into this too. It will use a scale attached to the entity, and `Camera2dComponents` adds a uniform scale, but since it's a *uniform* scale it scales in XYZ, which messes up the depth (in a 2D game, if scale <1.0, then sprites don't draw at all because they're outside of zmin..zmax).
At minimum it needs a nonuniform scale, which means you can't use `Camera2dComponents` (or you need to manually remove the uniform scale that it creates.)
Ideally I think `OrthographicProjection` should replace `window_origin` with a more explicit enum that covers more options, something like:
- pixels with (0,0) in lower-left, Y up
- pixels with (0,0) in top-left, Y down
- pixels with (0,0) at center, Y up
- H units tall and/or W units wide, with different aspect ratios handled as:
(1) H exact, adjust W to preserve aspect ratio
(2) W exact, adjust H to preserve aspect ratio
(3) minimum of (1) or (2)
(4) maximum of (1) or (2)
(5) exactly H*W, so don't preserve 1:1 aspect ratio
Status: Issue closed
username_2: I believe this was solved by #400 getting merged :) |
wellprog/TaskManagetV2 | 394878318 | Title: Создать модуль контактов
Question:
username_0: Создать модуль контактов, оформить его в виде сервиса (так же как модуль таймеров)
В модуле контактов должно быть поле (День рождения) который связан с модулем таймер.
Все записи которые добавляются в модуль таймер в нем не должны отображаться, однако должны работать.
При удалении контакта так же удаляются и таймеры. |
biggora/caminte | 425332368 | Title: Any plan to support GraphQL?
Question:
username_0: The idea would be as follow
GraphQL Schema to CarminteJS Shema (GraphQL -> [AST](https://www.contentful.com/blog/2018/07/04/graphql-abstract-syntax-tree-new-schema/) -> Carminte)
Then support [OpenCRUD](https://www.opencrud.org/) for the queries.
This is what [prisma](https://www.prisma.io/) is doing right now and as GraphQL seems to be the next generic query language, it might worth taking a look =)
Answers:
username_1: I just wanted to add a bit.. TypeORM and caminte both are doing most of the things in common. I guess caminte is sleek though, If you guys can add GraphQL support you guys would become awesome.
username_0: In the meantime, I discovered Graphql-box https://www.npmjs.com/package/graphql-box I still did not have time to play with it but it might be a good alternative… at least for the time to have an ORM which would compile the queries. |
kubernetes-sigs/cluster-api-provider-aws | 1124631292 | Title: Failing to build release-notes binary
Question:
username_0: Looks like a version incompatibility issue.
```
➜ cluster-api-provider-aws git:(main) make release
make -C hack/tools bin/release-notes
mkdir -p bin
go build -tags tools -o bin/release-notes k8s.io/release/cmd/release-notes
go: downloading cloud.google.com/go v0.99.0
go: downloading github.com/shirou/gopsutil/v3 v3.21.12
go: downloading github.com/BurntSushi/toml v1.0.0
go: downloading golang.org/x/sys v0.0.0-20211216021012-1d35b9e2eb4e
go: downloading golang.org/x/tools v0.1.9-0.20211228192929-ee1ca4ffc4da
go: downloading golang.org/x/crypto v0.0.0-20220112180741-5e0467b6c7ce
go: downloading golang.org/x/net v0.0.0-20211112202133-69e39bad7dc2
go: downloading google.golang.org/api v0.63.0
go: downloading google.golang.org/grpc v1.43.0
# cloud.google.com/go/storage
/Users/ssavas/go/pkg/mod/cloud.google.com/go/[email protected]/storage.go:1416:53: o.GetCustomerEncryption().GetKeySha256 undefined (type *"google.golang.org/genproto/googleapis/storage/v2".Object_CustomerEncryption has no field or method GetKeySha256)
make[1]: *** [bin/release-notes] Error 2
make: *** [hack/tools/bin/release-notes] Error 2
```
/kind bug
Answers:
username_1: I've been wondering for a while if we should change to use the GH release notes generation instead.
username_2: /assign
username_0: I am using GH release to double check we are not missing any changes.
Only thing is it is missing categorization, we need to go over each of them to decide bug/cleanup etc.
We can possibly do something similar to what [cluster-api has ](https://github.com/kubernetes-sigs/cluster-api/tree/main/hack/tools/release)using GH release notes.
username_1: I have done this with a different project, we can probably get away without custom code. You can customize the GH release notes generation using a config file placed in `.github/release.yml`. [This](https://github.com/weaveworks/flintlock/blob/main/.github/release.yml) is an example. The configuration options can be found [here](https://docs.github.com/en/repositories/releasing-projects-on-github/automatically-generated-release-notes#configuration-options).
I like with the GitHub release notes generation that it also highlights first time contributors.
@username_0 - i will add it to the office hours call this evening.
@username_2 - if we change the focus of this to user the GitHub release notes generation is this still something you'd want to work on?
username_2: Yes, I would like to work on it.
username_1: Added #3170 with the details of implementing github release notes.
username_0: Closing this issue as it will not be a problem once we move to GH release notes. #3170
Status: Issue closed
|
Nordstrom/chefdk_bootstrap | 110354097 | Title: Buff::Errors::InvalidConfig error when running ChefDK_bootstrap script (Berkshelf)
Question:
username_0: ## Error
When running the boostrap script you may encounter an error like the following:
```
Installing ChefDK...
C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/buff-config-1.0.1/lib/buff/config/ruby.rb:30:in `rescue in initializ
e': uninitialized constant Buff::Config::Ruby::Evaluator::Chef (Buff::Errors::InvalidConfig)
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/buff-config-1.0.1/lib/buff/config/ruby.rb:25:in `initia
lize'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/buff-config-1.0.1/lib/buff/config/ruby.rb:17:in `new'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/buff-config-1.0.1/lib/buff/config/ruby.rb:17:in `parse'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/buff-config-1.0.1/lib/buff/config/ruby.rb:119:in `from_
ruby'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/buff-config-1.0.1/lib/buff/config/ruby.rb:112:in `initi
alize'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/ridley-4.2.0/lib/ridley/chef/config.rb:83:in `initializ
e'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf.rb:90:in `new'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf.rb:90:in `chef_config'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/config.rb:80:in `<class:C
onfig>'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/config.rb:4:in `<module:B
erkshelf>'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/config.rb:3:in `<top (req
uired)>'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/cookbook_generator.rb:3:i
n `require_relative'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/cookbook_generator.rb:3:i
n `<class:CookbookGenerator>'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/cookbook_generator.rb:2:i
n `<module:Berkshelf>'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/cookbook_generator.rb:1:i
n `<top (required)>'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf.rb:197:in `require_relati
ve'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf.rb:197:in `<top (required
)>'
from C:/opscode/chefdk/embedded/lib/ruby/site_ruby/2.1.0/rubygems/core_ext/kernel_require.rb:54:in `require'
from C:/opscode/chefdk/embedded/lib/ruby/site_ruby/2.1.0/rubygems/core_ext/kernel_require.rb:54:in `require'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/lib/berkshelf/cli.rb:1:in `<top (requir
ed)>'
from C:/opscode/chefdk/embedded/lib/ruby/site_ruby/2.1.0/rubygems/core_ext/kernel_require.rb:54:in `require'
from C:/opscode/chefdk/embedded/lib/ruby/site_ruby/2.1.0/rubygems/core_ext/kernel_require.rb:54:in `require'
from C:/opscode/chefdk/embedded/lib/ruby/gems/2.1.0/gems/berkshelf-3.3.0/bin/berks:3:in `<top (required)>'
from C:/opscode/chefdk/bin/berks:43:in `load'
from C:/opscode/chefdk/bin/berks:43:in `<main>'
Starting Chef Client, version 12.4.4
resolving cookbooks for run list: ["chefdk_bootstrap"]
================================================================================
Error Resolving Cookbooks for Run List:
================================================================================
Missing Cookbooks:
------------------
No such cookbook: chefdk_bootstrap
Expanded Run List:
------------------
[Truncated]
## Cause
This is this [issue in Berkshelf](https://github.com/berkshelf/berkshelf/issues/965).
The issue is caused by having a line like this in your `knife.rb`:
`Chef::Config.from_file(knife_override) if File.exist?(knife_override)`
## Solution
Replace the offending line in your knife.rb with this code:
```ruby
# See this Berkshelf issue for why we can't call Chef::Config.from_file()
# https://github.com/berkshelf/berkshelf/issues/965
# This code is what Chef::Config.from_file() does.
if File.exists?(org_override) && File.readable?(org_override)
self.instance_eval(IO.read(org_override), org_override, 1)
end
```
Status: Issue closed
Answers:
username_0: This is not actually a bug in chefdk_bootstrap. Closing. |
rapidsai/cusignal | 705143365 | Title: [FEA] Reevaluate Peak Finding module for performance improvements
Question:
username_0: The Peak Finding module need to be reevaluated to find areas of further performance improvements.
Answers:
username_1: Note: This includes similar investigations as explored in https://github.com/rapidsai/cusignal/pull/235. Developers should be examining:
- Use of numpy over cupy for small-scale calculations like overlap size, FFT size, and other generic parameters that don't require GPU compute
- Elementwise kernels, where appropriate (either raw cupy, as denoted in the above PR, or Numba)
- Caching of FFT plans
Status: Issue closed
username_0: Closed by #267 |
NPellet/visualizer | 160879140 | Title: pasting png file and clipboard content
Question:
username_0: I did a "private" view (in Test flavor) : Test png:
https://my.cheminfo.org/?viewURL=https%3A%2F%2Fmydb.cheminfo.org%2Fdb%2Fvisualizer%2Fentry%2Fb1d9fbc6670f5236f9ff2a00d0386f34%2Fview.json
In the module drag / drop there is a "Read type" that may be Array buffer. Is this expected to work and how ?
Would it be possible to use the PNGreader library in this code to check the png content of the image.
If I have many "flavor" in the clipboard I would have a pop-up window ?
Still looking for the meta data of the image ... It "should" be there somewhere ... Maybe there is a difference from chemdraw when you save as png or if you copy / paste the structure directly from the clipboard ...

Answers:
username_1: ArrayBuffer works only if you export the parent object (not content) otherwise it is converted to a DataObject
Status: Issue closed
username_0: Apparently even if you clipboard contains many flavor of the data it seems that the browser will filter out only the one that can be paste. |
Submitty/Submitty | 466664661 | Title: Long course titles cause display issues on mobile
Question:
username_0: **Describe the bug**
If a course has a long name the course list page looks weird. The blue buttons around the title don't read which causes have the title to be unreadable. There's also a horizontal scroll wheel which probably shouldn't be there on mobile
**Expected behavior**
I'm not sure whats a good way to handle this but my thoughts are are to wrap the text around. I think scaling things down might make it too hard to read.
**To Reproduce**
Steps to reproduce the behavior:
1. Login on mobile
4. See error
**Configuration**
- Android chrome/firefox
**Screenshots**
If applicable, add screenshots to help explain your problem.
<img src= "https://user-images.githubusercontent.com/12129065/61023301-5f7d6580-a377-11e9-856d-7bbf3bc3afc3.png" height = 800 />
Horizontal scroll:
<img src ="https://user-images.githubusercontent.com/12129065/61023300-5db3a200-a377-11e9-9313-5e7822cb7191.png" height = 800 />
**Additional context**
Add any other context about the problem here.<issue_closed>
Status: Issue closed |
OBOFoundry/OBOFoundry.github.io | 229084707 | Title: Visibility of 'replaced by' on Foundry page
Question:
username_0: On the OBO Foundry page there are a number of ontologies listed as obsolete. Right now, if there is a suitable other ontology in the same domain, that information is buried one click away under 'details'. I think it would be useful to make this directly visible.
Answers:
username_1: I can see the argument for that; on the other hand, "details" reveals a bunch of other fields, and the replacement ontology is just one of those.
username_0: True, though one could argue that--for obsolete ontologies--what other ontology would be of most interest is probably the KEY piece of information. I don't feel strongly about it, though I suspect if this was directly displayed we would probably have more cases of replacement ontologies indicated.
username_1: Is this still relevant and desired?
username_2: I will put it on the list for @ddooley team - this should be easy to resolve. (I am thinking mouseover tooltip type solution). |
Thom1729/Sublime-JS-Custom | 553484969 | Title: Highlight function calls
Question:
username_0: I've been using JavaScript Next for ages, and have really got used to it. But it's old, and now doesn't support things like arrow function parameter default values which are function calls amongst [other things](https://github.com/Benvie/JavaScriptNext.tmLanguage/issues/4). So I moved to JS Custom, but there is one thing that niggles me—I'm really used to function _calls_ highlighted as function names. Is it possible to fix with a local customisation? Or does it need a change to the syntax file itself? Compare JavaScript Next on left vs JS Custom on right. The `thing` function definition and parameters look great, but I'd like the calls to also be green...

Answers:
username_1: The function names in the calls are scoped `variable.function.js`, which should be colored by most schemes. Your color scheme looks like Monokai, but I've just checked and the version of Monokai bundled with Sublime does highlight the function calls. Are you using a third-party color scheme?
username_0: I have both Monokai and Monokai (SL) installed—not sure if either came with JS Next. But both those schemes colour function calls in JS Next and don't in JS Custom.
I've removed JS Next, cleared the \Cache fodler and restarted ST, but still no coloring of function calls.
Then I installed Monokai++, and function call coloring is there. So I suspect I just have a random collection of old theme files which are out of date :). Sorry for the interruption! Still no idea why JS Next worked, unless it patched the color scheme when active... Anyway, problem solved.
Status: Issue closed
|
mike-lischke/antlr4-c3 | 250013972 | Title: Some further usage examples
Question:
username_0: Not a bug, a request. This project looks very hopeful for our usage, thanks.
To be really useful it would help to have some more concrete use cases on how to retrieve the incomplete symbol and its context, for example (from my experiments, where I'm not sure what to do.
I'm writing a simple filter grammar (imagine something grepping through logfiles) and want to auto-complete on something like:
fu = "incomplete string
When finding possible completions, I'd like to be able to have the context that I'm comparing a field called 'fu' (and thus be able to query all known values of 'fu') and that the token I'm trying to complete on is "incomplete string
Similarly we might not have the double-quotes on the incomplete string if we want the user to be able to type in a string without quotes...
An example or test case tying together the grammar with the symbol table and extracting the parse context + the half-complete symbol would be really helpful, I'm a little stuck here...
As an aside, if interested I've ported the completion core and tests to Java and may be able to release them back to this project. If so please context me.
Answers:
username_1: An incomplete string is a syntax errror and the walker might not find the path to the invocation point for candidate collection. But I haven't tested this scenario yet, as you usually don't have completion for strings (and literals in general). Have to think a bit how to get a context for that symbol.
Regarding the Java port: yes, sure, I'm interested in that. There is already a Kotlin port of the code (and I've written myself a C++ variant). I could publish all that together in this repo.
username_0: Hi Mike,
Would it be possible for the walker to record the matching text and/or the set of tokens matching a given preferred rule, which would give the ability for the code trying to do completion to get not just the most recent token but the preferred rule’s full string? In the case I describe this would mean that we could have a rule called ‘partial_string_value’ matching an optional quote then a white-spaced string, and this would be a preferred rule which I could use in auto-complete and I’d be able to get the value. Today I can’t get the value since I can’t just use the last token, I need all the tokens that matched the ‘partial_string_value’ rule. I guess that information must be available during the processRule evaluation?
I’ve started getting approvals to give back the java port, will keep you informed… I guess we could then move each port into separate sub-directories, I could prepare a cloned repo with a merge request once approvals are in place, which hopefully should be pretty quick.
Nick
username_0: I believe I've found out how to extract the multi-token rule-string value as required to be able to do completions on things which span more than one token such as the given example.
It's done as an extension to collectCandidates() and a new field in the candidates object which maps a given candidate rule to its string fragment.
In my java implementation it looks like this:
~~~
this.processRule(this.atn.ruleToStartState[startRule], 0, callStack, "\n");
// now post-process the rule candidates and their last occurrences to find the rule strings
for (int ruleId : candidates.rules.keySet()) {
Map<Integer, Set<Integer>> shortcut = shortcutMap.get(ruleId);
if (shortcut == null) {
continue;
}
// select the right-most occurrence - the highest entry
int startToken = Collections.max(shortcut.keySet());
Set<Integer> endSet = shortcut.get(startToken);
int endToken = endSet.isEmpty() ?
caretTokenIndex - currentIndex :
Collections.max(shortcut.get(startToken));
// extract the matching text
candidates.ruleStrings.put(ruleId, tokenStream.getText(tokenStream.LT(startToken), tokenStream.LT(endToken)));
}
~~~
then for each candidate rule I have the last string in the parsed input which matched that rule.
HTH
Nick
username_1: If I understand this right you want to extract the longest text that was matched when walking a candidate rule. What are you going to do with that text? It's not something you would show during code completion. Also, often candidate rules are not completely executed (since they are candidates, the caret position is somewhere in the text covered by those rules) and hence have no entry in the shortcutMap.
username_1: Regarding the contribution of your Java code: glad to see the approval process has started. I'm still thinking about the repo structure and where to place the ports, since we must ensure we don't break the Node.js module setup. Maybe we just add a new folder called `ports` or `contributions` and exclude that in the package.json file?
username_0: Hi Mike,
The idea is to extract the most recent (closest to the caret position) and longest text that matches when walking a candidate rule, so that completions of multi-token rules can have the partially-completed text… I’m still experimenting with this… it’s not yet seeming to get all the string values I’d like, maybe your comment identifies the problem.
Here are a couple of use-cases for a technical-english grammar which performs filtering of log lines:
Imagine a log line like this:
Aug 19 00:07:00 info nicks-mbp cron[17522]: no path for address 0x10fab9000
And a filter like this:
log text contains no path for
We might want to detect that the field being filtered on is the ‘log text’ field (which contains the full line) and the value to filter on is ‘no path for’ which in itself is a legal filter but which could be completed by scanning logs and finding that it could be completed to become ‘no path for address 0x10fab9000’ (or other addresses) based upon frequency of log messages.
To do this requires finding the ‘no path for’ text for the ‘value’ rule, and indeed also finding the ‘log text’ for the ‘field’ rule to give context for the search.
Similarly if typing:
log t
I’d like to complete with ‘log text’ and ‘log timestamp’
And typing:
log priority is
I’d like to propose completion with the values from the ‘priority’ field of log files, based upon their occurrence in the logs (e.g. ‘info’ in the above line).
Hope this is clear!
Concerning where to put in the source code base, I agree that separate subdirectories would be good, up to you if you want to create a ‘ports’ subdirectory, let me know…
Thx
Nick
Hi Mike,
The idea is to extract the most recent (closest to the caret position) and longest text that matches when walking a candidate rule, so that completions of multi-token rules can have the partially-completed text… I’m still experimenting with this… it’s not yet seeming to get all the string values I’d like, maybe your comment identifies the problem.
Here are a couple of use-cases for a technical-english grammar which performs filtering of log lines:
Imagine a log line like this:
Aug 19 00:07:00 info nicks-mbp cron[17522]: no path for address 0x10fab9000
And a filter like this:
log text contains no path for
We might want to detect that the field being filtered on is the ‘log text’ field (which contains the full line) and the value to filter on is ‘no path for’ which in itself is a legal filter but which could be completed by scanning logs and finding that it could be completed to become ‘no path for address 0x10fab9000’ (or other addresses) based upon frequency of log messages.
To do this requires finding the ‘no path for’ text for the ‘value’ rule, and indeed also finding the ‘log text’ for the ‘field’ rule to give context for the search.
Similarly if typing:
log t
[Truncated]
And typing:
log priority is
I’d like to propose completion with the values from the ‘priority’ field of log files, based upon their occurrence in the logs (e.g. ‘info’ in the above line).
Hope this is clear!
Concerning where to put in the source code base, I agree that separate subdirectories would be good, up to you if you want to create a ‘ports’ subdirectory, let me know…
Thx
Nick
On 19/08/2017, 11:50, "<NAME>" <<EMAIL><mailto:<EMAIL>>> wrote:
If I understand this right you want to extract the longest text that was matched when walking a candidate rule. What are you going to do with that text? It's not something you would show during code completion. Also, often candidate rules are not completely executed (since they are candidates, the caret position is somewhere in the text covered by those rules) and hence have no entry in the shortcutMap.
username_1: This looks very much like completing keywords. Why do you want to scan your input first before offering `text` and `timestamp` as candidates, if you know these are valid values?
username_0: Let's close the issue. Thanks!
Status: Issue closed
username_2: Very appreciate about the work! I came across a problem and believe walker should have ability to go further even if there is syntax error before caret position. Suppose I have defined mysql grammar. And editing the following sentence,
```sql
select from |
```
where `'|'` represents caret position and it will get nothing by calling `collectCandidates`. Well, I can still do a little work to make it work. But It is also reasonable the walker keep walking if there is potential road to current request position. And I believe this is achievable by search algorithm like A-star. When walker walked to `select` and find no bridge to `from`, it search all neighbor bridges until it finally reach `from`. And this search should be constrained because it maybe very time-consuming.
Is it make sense? Thanks!
username_1: @username_2 Can you please open a new request instead of hijacking a closed one? |
stackblitz/core | 807118431 | Title: Github angular project import results in mismatching devDependencies
Question:
username_0: When importing an angular project via the github import integration, devDependencies are added to the package.json file, which mismatch the given angular version.
Input package.json file from our project (https://github.com/dynatrace-oss/barista-examples/tree/master/examples/DtExampleAlertDark)
```
{
"name": "DtExampleAlertDark",
"dependencies": {
"@angular/animations": "^11.0.5",
"@angular/cdk": "^11.0.3",
"@angular/common": "^11.0.5",
"@angular/compiler": "^11.0.5",
"@angular/core": "^11.0.5",
"@angular/forms": "^11.0.5",
"@angular/localize": "^11.0.5",
"@angular/router": "^11.0.5",
"@angular/platform-browser": "^11.0.5",
"@angular/platform-browser-dynamic": "^11.0.5",
"core-js": "^3.8.3",
"d3-scale": "^3.0.0",
"d3-shape": "^1.3.5",
"rxjs": "^6.6.3",
"tslib": "^2.0.0",
"zone.js": "~0.10.2",
"@dynatrace/barista-components": "latest",
"@dynatrace/barista-fonts": "latest",
"@dynatrace/barista-icons": "latest",
"highcharts": "^7.2.1",
"lodash-es": "^4.17.15"
},
"devDependencies": {
"@angular-devkit/build-angular": "0.1100.5",
"@angular/cli": "^11.0.5",
"@angular/compiler-cli": "^11.0.5"
}
}
```
When importing this via the github import url structure:
https://stackblitz.com/github/dynatrace-oss/barista-examples/tree/master/examples/DtExampleAlertDark/
the resulting package.json lists angular 9 devDependencies instead of the given angular 11 ones. This mismatch leads to broken projects.
When creating a new project in angular via the UI, the dev dependencies are set correctly to angular 11 ones.
```
{
"name": "aoniryjnz.github",
"version": "0.0.0",
"private": true,
"dependencies": {
"rxjs": "6.6.3",
"tslib": "2.1.0",
"core-js": "3.8.3",
"zone.js": "0.10.3",
"d3-scale": "3.2.3",
"d3-shape": "1.3.7",
"lodash-es": "4.17.20",
"highcharts": "7.2.2",
"@angular/cdk": "11.2.0",
"@angular/core": "11.2.0",
[Truncated]
"@angular/compiler-cli": "~9.1.1",
"@angular/language-service": "~9.1.1",
"@types/node": "^12.11.1",
"@types/jasmine": "~3.5.0",
"@types/jasminewd2": "~2.0.3",
"codelyzer": "^5.1.2",
"jasmine-core": "~3.5.0",
"jasmine-spec-reporter": "~4.2.1",
"karma": "~4.4.1",
"karma-chrome-launcher": "~3.1.0",
"karma-coverage-istanbul-reporter": "~2.1.0",
"karma-jasmine": "~3.0.1",
"karma-jasmine-html-reporter": "^1.4.2",
"protractor": "~5.4.3",
"ts-node": "~8.3.0",
"tslint": "~6.1.0",
"typescript": "~3.8.3"
}
}
``` |
OpenMined/PySyft | 684082237 | Title: Add torch.Tensor.q_per_channel_scales to allowlist and test suite
Question:
username_0: # Description
This issue is a part of Syft 0.3.0 Epic 2: https://github.com/OpenMined/PySyft/issues/3696
In this issue, you will be adding support for remote execution of the torch.Tensor.q_per_channel_scales
method or property. This might be a really small project (literally a one-liner) or
it might require adding significant functionality to PySyft OR to the testing suite
in order to make sure the feature is both functional and tested.
## Step 0: Run tests and ./scripts/pre_commit.sh
Before you get started with this project, let's make sure you have everything building and testing
correctly. Clone the codebase and run:
```pip uninstall syft```
followed by
```pip install -e .```
Then run the pre-commit file (which will also run the tests)
```./scripts/pre_commit.sh```
If all of these tests pass, continue on. If not, make sure you have all the
dependencies in requirements.txt installed, etc.
## Step 1: Uncomment your method in the allowlist.py file
Inside [allowlist.py](https://github.com/OpenMined/PySyft/blob/syft_0.3.0/src/syft/lib/torch/allowlist.py) you will find a huge dictionary of methods. Find your method and uncomment the line its on. At the time
of writing this Issue (WARNING: THIS MAY HAVE CHANGED) the dictionary maps from the
string name of the method (in your case 'torch.Tensor.q_per_channel_scales') to the string representation
of the type the method returns.
## Step 2: Run Unit Tests
Run the following:
```python setup.py test```
And wait to see if some of the tests fail. Why might the tests fail now? I'm so glad you asked!
https://github.com/OpenMined/PySyft/blob/syft_0.3.0/tests/syft/lib/torch/tensor/tensor_remote_method_api_suite_test.py
In this file you'll find the torch method test suite. It AUTOMATICALLY loads all methods
from the allowlist.py file you modified in the previous step. It attempts to test them.
# Step 3: If you get a Failing Test
If you get a failing test, this could be for one of a few reasons:
### Reason 1 - The testing suite passed in non-compatible arguments
The testing suite is pretty dumb. It literally just has a permutation of possible
arguments to pass into every method on torch tensors. So, if one of those permutations
doesn't work for your method (aka... perhaps it tries to call your method without
any arguments but torch.Tensor.q_per_channel_scales actually requires some) then the test will
fail if the error hasn't been seen before.
If this happens - don't worry! Just look inside the only test in that file and look
[Truncated]
pointer objects to very many remote object types. So, if your method returns anything
other than a single tensor, you probably need to add support for the type it returns
(Such as a bool, None, int, or other types).
*IMPORTANT:* do NOT return the value itself to the end user!!! Return a pointer object
to that type!
*NOTE:* at the time of writing - there are several core pieces of Syft not yet working
to allow you to return any type other than a torch tensor. If you're not comfortable
investigating what those might be - skip this issue and try again later once
someone else has solved these issues.
### Reason 3 - There's something else broken
Chase those stack traces! Talk to friends in Slack. Look at how other methods are supported.
This is a challenging project in a fast moving codebase!
And don't forget - if this project seems to complex - there are plenty of others that
might be easier.<issue_closed>
Status: Issue closed |
vmware-tanzu/velero | 945898591 | Title: Got "velero container not found in velero deployment" error when trying to add plugin
Question:
username_0: I found the following issue during running the E2E test:
I installed velero on TKG which created two deployments under the namespace `velero`: `velero` and `backup-driver`
And when I tried to add another plugin, got the error: `velero container not found in velero deployment`
I think the root cause should be here:
https://github.com/vmware-tanzu/velero/blob/main/pkg/cmd/cli/plugin/add.go#L100
https://github.com/vmware-tanzu/velero/blob/main/pkg/cmd/cli/plugin/helpers.go#L49
velero cli tries to get the `velero` deployment by label selector `"component": "velero"`, but in this case both `velero` and `backup-driver` are labeled with `"component": "velero"`, this cuases the `backup-driver` is returned rather than `velero`
If this is the case, maybe we can filter the velero deployment by both label and name?
**Vote on this issue!**
This is an invitation to the Velero community to vote on issues, you can see the project's [top voted issues listed here](https://github.com/vmware-tanzu/velero/issues?q=is%3Aissue+is%3Aopen+sort%3Areactions-%2B1-desc).
Use the "reaction smiley face" up to the right of this comment to vote.
- :+1: for "I would like to see this bug fixed as soon as possible"
- :-1: for "There are more important bugs to focus on right now"
Answers:
username_0: Seems that comparing the name isn't the correct way either. The naming comparing logic is removed by the https://github.com/vmware-tanzu/velero/pull/3447 because the name of velero installed by helm chart can be customized.
Status: Issue closed
|
robot-lab/judyst-main-web-service | 371738007 | Title: Настройки для организации
Question:
username_0: # Task request
## Цель задачи
Дать возможность менять настройки организации.
На основе feature #55
## Решение задачи
Представление которое обрабатывает следующие запросы: изменение имени организации, удаление членов организации, удаление доступных файлов, удаление организации.
## Дополнительный контекст или ссылки на связанные с данной задачей issues |
Wynncraft/Issues | 220491655 | Title: The Envoy Part I
Question:
username_0: Heading through an area following Maxie I got stuck between two walls of barrier blocks. This is while heading to her where the Quest Book updates to "Stop the mechs from taking over the diner.". This is not the first quest that has given us a lot of issues due to barrier blocks. I have tried /c;lass and coming back and switching the server.
Answers:
username_1: Try doing /kill with 5 or more SP
username_0: I did /kill. That is not the point, I should not have to keep killing myself because of constant issues in quests with the over use of barrier blocks. This has happened on a lot of quests and it is becoming ridiculous now. |
libtcod/libtcod | 548934343 | Title: Remove C++ dependencies.
Question:
username_0: As much as I liked using C++ over C it has made the code less portable and made the C ABI harder to maintain. I now plan on rewriting the code in C and porting to C++ instead of the other way around.
The following needs to be rewritten in C:
- [ ] New tileset objects.
- [ ] Tileset font loader.
- [ ] TrueType font loader.
- [ ] SDL2 renderer.
- [ ] OpenGL2 renderer.
- [ ] Console UTF-8 functions.
- [ ] Console drawing functions.
- [ ] Console initialization.
- [ ] Image functions.
Answers:
username_1: This would definitely help in supporting emscripten/wasm too. What's your ETA on this?
username_0: I thought Emscripten could compile C++. If anything currently prevents an emscripten/wasm build then feel free to create an issue for it.
Some of the refactoring is hard to do since a lot of the rendering functions depend on each other, so those all have to be done at once before I can test and upload it.
No real ETA, but I'm trying to get as much as I can done before the February 7DRL. This _might_ be done before the end of January, but there will still be a lot of compiled C++ code left over after this task is done.
username_1: Emscripten can compile C++ but it’s a bit of a mess when there are external libraries and C++14 involved.
Some standard libraries like zlib and SDL2 have already been ported but others like GLAD and probably others would need to be ported manually.
The best would be to reduce the number of external dependencies to the bare minimum.
On top of that, a port to pure C would be nice for people like me that usually avoid C++
username_0: GLAD is for the desktop builds, and it keeps the function names standard so that I can switch between OpenGL and OpenGLES easier. I just need to add the preprocessor code that actually does that.
Status: Issue closed
username_0: It's done, all the major components compile and run without the C++ sources. Right now it's only python-tcod which is setup to compile without C++.
I want to test Emscripten but I've been unable to setup the tool-chain so far. I expect it to fail, and that I'll need to switch the headers to use GLES. That kind of thing will be tracked in issue #41 if I make more progress.
username_1: Nice! ;) |
flutter/flutter | 1023535710 | Title: Flutter error "type 'bool' is not a subtype of type 'List<dynamic>' "
Question:
username_0: I am getting while running flutter/dart program "type 'bool' is not a subtype of type 'List<dynamic>' ". The code is as follows:
import 'package:flutter/material.dart';
import 'package:http/http.dart' as http;
import 'dart:convert';
void main() => runApp(MainEntry());
class MainEntry extends StatelessWidget {
// const MainEntry({Key? key}) : super(key: key);
@override
Widget build(BuildContext context) {
return MaterialApp(
theme: ThemeData(
backgroundColor: Colors.white70,
primarySwatch: Colors.blue,
brightness: Brightness.light,
),
debugShowCheckedModeBanner: false,
title: 'flutter',
home: Entry(),
);
}
}
class Entry extends StatefulWidget {
@override
_EntryState createState() => _EntryState();
}
class _EntryState extends State<Entry> {
getMethod() async {
String theUrl = 'http://ananyananu.000webhostapp.com/getData.php';
var res = await http.get(Uri.parse(theUrl), headers: {"Accept":"application/json"});
var responsBody = json.decode(res.body);
print ('I reached here');
print('responsBody');
return responsBody;
}
@override
Widget build(BuildContext context) {
return Scaffold(
backgroundColor: Colors.grey[200],
appBar: AppBar(
title: Text('Connect to Local Host') ,
centerTitle: true,
backgroundColor: Colors.redAccent,
),
body: FutureBuilder(
future: getMethod(),
builder: (BuildContext context, AsyncSnapshot snapshot) {
if (snapshot.connectionState == ConnectionState.waiting) {
[Truncated]
return Center(
child: Text('Error Fetching Data'),
);
}
List snap = snapshot.data;
return ListView.builder(
itemCount: snap.length,
itemBuilder: (context, index) {
return ListTile(
title: Text("head: ${snap[index]['heading']}"),
subtitle: Text("body ${snap[index] ['body']}"),
);
},
);
}),
);
}
} |
jlippold/tweakCompatible | 483395718 | Title: `Amber` working on iOS 12.4
Question:
username_0: ```
{
"packageId": "com.ps.amber",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "com.ps.amber",
"deviceId": "iPhone8,4",
"url": "http://cydia.saurik.com/package/com.ps.amber/",
"iOSVersion": "12.4",
"packageVersionIndexed": true,
"packageName": "Amber",
"category": "Tweaks",
"repository": "A Main Repo",
"name": "Amber",
"installed": "0.0.2.9",
"packageIndexed": true,
"packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.",
"id": "com.ps.amber",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.1.5",
"shortDescription": "Amber-ify your LED torch.",
"latest": "0.0.2.9",
"author": "PoomSmart",
"packageStatus": "Unknown"
},
"base64": "<KEY>",
"chosenStatus": "working",
"notes": ""
}
```<issue_closed>
Status: Issue closed |
hail-is/hail | 526125283 | Title: breaking hl.rand_unif via CSE
Question:
username_0: shoudn't be terribly surprising that `rand_unif` has weird behavior, but here's a case that is definitely The Wrong Thing:
```
Python 3.6.0 |Continuum Analytics, Inc.| (default, Dec 23 2016, 13:19:00)
Type 'copyright', 'credits' or 'license' for more information
IPython 7.6.1 -- An enhanced Interactive Python. Type '?' for help.
In [1]: import hail as hl
In [2]: r = hl.rand_unif(0, 1)
In [3]: hl.eval(r)
Out[3]: 0.5387579341676381
In [4]: hl.eval(hl.tuple([r, r]))
Out[4]: (0.5387579341676381, 0.5387579341676381)
```
okay, this makes sense becuase they have the same seed:
```
In [5]: print(hl.tuple([r, r])._ir)
(MakeTuple (0 1) (ApplySeeded rand_unif 806694938962853089 Float64 (Apply toFloat64 Float64 (I32 0)) (Apply toFloat64 Float64 (I32 1))) (ApplySeeded rand_unif 806694938962853089 Float64 (Apply toFloat64 Float64 (I32 0)) (Apply toFloat64 Float64 (I32 1))))
```
how about this:
```
In [6]: hl.eval(hl.range(2).map(lambda x: r))
Out[6]: [0.5387579341676381, 0.9394799645512691]
```
odd. but maybe rand_unif inside an iteration has some semantics for advancing the RNG (like an aggregation).
```
In [7]: p = 1 - r
In [8]: hl.eval(hl.range(2).map(lambda x: p))
Out[8]: [0.46124206583236194, 0.06052003544873086]
```
ok...
```
In [9]: hl.eval((p, hl.range(2).map(lambda x: p)))
Out[9]: (0.46124206583236194, [0.46124206583236194, 0.46124206583236194])
```
wtf?
if you look in the logs, its explained by the fact that only the final IR triggers CSE:
```
(Let __cse_1
(ApplyBinaryPrimOp Subtract
(ApplyIR toFloat64 Float64
(I32 1))
(ApplySeeded rand_unif 806694938962853089 Float64
(ApplyIR toFloat64 Float64
(I32 0))
(ApplyIR toFloat64 Float64
(I32 1))))
(MakeTuple (0 1)
(Ref __cse_1)
(ArrayMap __uid_5
(ArrayRange
(I32 0)
(I32 2)
(I32 1))
(Ref __cse_1))))
```
Answers:
username_1: I think I'd flip the logic. I'm not sure if this one is wrong:
```
In [9]: hl.eval((p, hl.range(2).map(lambda x: p)))
Out[9]: (0.46124206583236194, [0.46124206583236194, 0.46124206583236194])
```
But if it's right, clearly this one is wrong:
```
In [7]: p = 1 - r
In [8]: hl.eval(hl.range(2).map(lambda x: p))
Out[8]: [0.46124206583236194, 0.06052003544873086]
```
username_2: The latter seems wrong. It violates the identity of `p`. I prefer the semantics of the former.
username_1: I agree. It's extremely hard for us to provide these semantics though -- we'd need to introspect on the Python AST of the lambda
username_2: It seems the right answer is that CSE cannot lift side-effecting operations outside of `map`.
username_3: so just as an aside, the behavior of all the random functions is actually documented in the hail docs:
https://hail.is/docs/0.2/functions/random.html
The one thing that I apparently didn't write up is how it's supposed to behave in an array context. The way it's currently intended to work is as executed in:
```
In [8]: hl.eval(hl.range(2).map(lambda x: p))
Out[8]: [0.46124206583236194, 0.06052003544873086]
```
since, as Tim points out, it's kind of hard to differentiate the scope of `hl.range(2).map(lambda x: p)` and `hl.range(2).map(lambda x: hl.rand_unif(0, 1))` if p is only used once.
(This behavior is also pretty consistent with treating iteration through an array value as the same as iteration through an axis of a table or a matrix table, since that's exactly what happens there.)
The way you'd use the same value of a random number in an array map, previously, was with bind:
```
hl.bind(lambda p: hl.range(2).map(lambda x: p), hl.rand_unif(0, 1))
```
I'm not sure what a bind-free answer to this would be short of implementing a way of differentiating between:
```
p = hl.rand_unif(0, 1)
hl.range(2).map(lambda x: p)
```
and
```
hl.range(2).map(lambda x: hl.rand_unif(0, 1))
```
which would be nice, but also definitely a breaking change.
Status: Issue closed
|
CompuCell3D/CompuCell3D | 676471958 | Title: FPP links rendering
Question:
username_0: The links are NOT drawn properly if you chance the view to 3D or the plane of view (they are always rendered as if in x-y view no matter what)
Answers:
username_0: Many people noticed odd behavior with FPP in CC3D during the class. FPP edges weren’t being displayed properly in player, the edges seem to come and go randomly and other odd behaviors.
I have a simulation now that appears to be changing the membership in the cell list by type data. For example;
for cell1 in self.cell_list_by_type(self.ACELL):
print("\t\t\t cell1.id,",cell1.id)
for cell2 in self.cell_list_by_type(self.BCELL):
d=cell_cell_dist(cell1,cell2)
self.focalPointPlasticityPlugin.createFocalPointPlasticityLink(cell1, cell2, 10, d, 500)
print("\t\t\t cell2.id",cell2.id)
# calculate the distance between two cells, cell1 and cell2
def cell_cell_dist(cell1,cell2):
import math
d=math.sqrt((cell1.xCOM-cell2.xCOM)**2+(cell1.yCOM-cell2.yCOM)**2+(cell1.zCOM-cell2.zCOM)**2)
return(d)
Gives as output:
Len acell and bcell lists: 4 5
acell lists:
1 2 5 8
bcell lists:
3 4 6 7 9
cell1.id, 1
cell2.id 3
cell2.id 4
cell2.id 6
cell2.id 7
cell2.id 9
[Truncated]
cell2.id 7
cell2.id 9
Len acell and bcell lists: 4 5
acell lists:
1 2 5 8
bcell lists:
3 4 6 7 9
Notice the highlighted values are not cell.ids that are on the respective list and a value is repeated. The above behavior changes every run (even though I have an explicit seed set in the Potts section).
username_0: [FPP_test_Aug_2020b.zip](https://github.com/username_0/username_0/files/5053923/FPP_test_Aug_2020b.zip)
username_1: This might not be an FPP problem. It looks like CC3D doesn't properly process cell lists by type whe you embed them in cascade loops: (pseudo code)
```
for cells of type a
for cells of type b
[gives a corrupted iteration list for the type a cells]
``` |
FAForever/downlords-faf-client | 556412515 | Title: add mapping rules and validation tool as link to upload process
Question:
username_0: Hello,
I'm missing a link to validation tool (https://api.faforever.com/maps/validate) in addition to the already added mapping rules (https://forums.faforever.com/viewtopic.php?f=2&t=17873) in the upload window/process. It helps you check your map name and content before upload.
Currently we are enforcing the rules while uploading a map and return these error messages.
But the validation tool is not linked.
https://cdn.discordapp.com/attachments/622773120377421851/671783149092601875/unknown.png
https://cdn.discordapp.com/attachments/622773120377421851/671783474407014411/unknown.png
### **My Dream Upload**
I believe the best error free upload can be achieved when we add the validation tool or its features to the upload process. Here are some ideas..
**improve pre upload process**
0)hit `check map` button

A.1) open link to validation tool: https://api.faforever.com/maps/validate
A.2) manual user input
or
B.1) select map
B.2) auto-fill validation tool and show results
**improve upload process**
0)hit `upload map` button
1) select map (add buttons or autodecide if the map source (folder, files, path) has `.v000x` in files and folder -> this way removing `.v0001` before upload is not a thing anymore)
2) auto-fill validation tool and show results
3) if rules are met the upload is possible, otherwise errors are shown
Answers:
username_1: Imo with the new (?) restrictions on naming file names should also be renamed to `-save` `-scenario` `-script` to be consistent
username_2: I added a button for now, a test within the client would be the best solution, but I advocate against that. It'll take way too much time away from more important developement tasks.
username_0: thx for the button.
what exactly are your referring to?
"a test within the client would be the best solution, but I advocate against that."
username_2: A system within the client that checks what you have selected and tells you what's wrong would be the best solution. But this takes too much time to be worth it for the small amount of people that use the map upload.
Status: Issue closed
username_2: Hello,
we are missing a link to the validation tool (https://api.faforever.com/maps/validate) in addition to the already added mapping rules (https://forums.faforever.com/viewtopic.php?f=2&t=17873) in the upload window/process.

The validation tool helps you check and validate your map name / content before upload. It also provides the right file names and pathing. Currently we are enforcing the rules while uploading a map and return these error messages. But the validation tool is not linked and some people don't know why they get this issue or how to solve it.


Morax also supports the idea of the link.
### **My Dream Upload**
While writing this I like to bring up a long desired dream of mine ;)
I believe the best error free upload can be achieved when we add the validation tool or its features to the upload process. Here are some ideas..
**improve pre upload process**
Encourage people to validate the map (name, scenario) before upload...
Option A:
1) hit `validate/check map` button (needs to be added next to the `upload map` button)

2) open link to validation tool: https://api.faforever.com/maps/validate
3) manual user input
Option B:
1) hit `validate/check map` button (needs to be added next to the `upload map` button)
2) select map
3) auto-fill validation tool and show results online
**improve upload process**
1) hit `upload map` button
2) select map
3) auto-fill validation tool and show results in upload window
4) if rules are met and no errors remain the upload is possible, otherwise errors are displayed and explained
4.1) server version: complete rewrite of the map files names (folder, scmap, save, script, options, tables, ...) and path (scmap, scenario, script,...) based on validated `name = "map name"` from scenario.lua -> this way removing `.v0001` (from folder and in files) before upload is not a thing anymore
Status: Issue closed
|
gpe-mw-ansible-org/rh-sso-multi-realm | 455941912 | Title: new kie-realm users and groups
Question:
username_0: -u 'Super' -p '<PASSWORD>!' -g 'kie-server,admin,Administrators'
-u 'Ava' -p '<PASSWORD>!' -g 'kie-server,applicant'
u 'Bill' -p '<PASSWORD>!' -g 'kie-server,applicant'
-u 'Tina' -p '<PASSWORD>!' -g 'kie-server,talent-acquisition'
-u 'Tom' -p '<PASSWORD>!' -g 'kie-server,talent-acquisition'
-u 'Eve' -p '<PASSWORD>!' -g 'kie-server,interviewer'
-u 'Ann' -p '<PASSWORD>!' -g 'kie-server,interviewer'
-u 'Bob' -p '<PASSWORD>!' -g 'kie-server,interviewer'` |
BitBagCommerce/SyliusElasticsearchPlugin | 611943378 | Title: Unable to use checkbox filter
Question:
username_0: On product, I have many checkbox attributes. When I check "1" in the filter and I click on "Filter" button, I have this error :
```
An exception has been thrown during the rendering of a template ("failed to create query: {
"bool": {
"must": [
{
"term": {
"enabled": {
"value": true,
"boost": 1.0
}
}
},
{
"terms": {
"channels": [
"sylius"
],
"boost": 1.0
}
},
{
"terms": {
"product_taxons": [
"01"
],
"boost": 1.0
}
},
{
"range": {
"price_sylius": {
"from": 0,
"to": 9223372036854775807,
"include_lower": true,
"include_upper": true,
"boost": 1.0
}
}
},
{
"bool": {
"should": [
{
"term": {
"attribute_promotion": {
"value": "1",
"boost": 1.0
}
}
}
],
"adjust_pure_negative": true,
"boost": 1.0
}
},
{
"bool": {
"should": [
[Truncated]
"term": {
"attribute_news_auto": {
"value": "9a9bf86a_8e14_11ea_ad31_0242ac120005",
"boost": 1.0
}
}
}
],
"adjust_pure_negative": true,
"boost": 1.0
}
}
],
"adjust_pure_negative": true,
"boost": 1.0
}
} [index: bitbag_shop_products_dev] [reason: all shards failed]").
```
When I replace the "1" value by `true` in the query string, the filter work fine but the filter form is not checked.
Answers:
username_1: Here's a quick POC in case someone needs to implement this feature https://github.com/username_1/SyliusElasticsearchPlugin/commit/d006d35d530e81c59fab45e065efc7122f50b422
username_2: Issued a PR to fix this issue |
jgromes/RadioLib | 451856260 | Title: RFM95 aka SX127x not detected
Question:
username_0: Using [SparkFun LoRa Gateway - 1-Channel (ESP32)](https://www.sparkfun.com/products/15006)
NSS = 16, DIO0 = 26, DIO1 = 33, DIO2 = 32 according to [schematic](https://cdn.sparkfun.com/assets/5/3/1/9/c/ESP32_LoRa_1_Channel_Gateway.pdf)
Setting up the connection to the RFM95 module:
```cpp
RFM95 lora = new Module(16, 26, 33, 32);
void setup() {
int state = lora.begin(868.0, 125.0, 9, 5, 0x3444, 22, 8);
if (state == ERR_NONE)
{
Serial.println("success!");
}
else
{
Serial.print("failed: ");
printError(state);
return false;
}
...
}
```
fails with
```log
=====================================
RFM95 sender test
=====================================
DeviceID: 84 0D 8E 0C 56 3C 00 00
Initialize RFM95
SX127x not found! (1 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011
SX127x not found! (2 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011
SX127x not found! (3 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011
SX127x not found! (4 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011
SX127x not found! (5 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011
SX127x not found! (6 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011
SX127x not found! (7 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011
SX127x not found! (8 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011
SX127x not found! (9 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011
SX127x not found! (10 of 10 tries) SX127X_REG_VERSION == 0x00, expected 0x0011
No SX127x found!
failed: SX127x not found
```
Using the abandoned library [RadioHead](https://github.com/PaulStoffregen/RadioHead)
Setup with
```cpp
// Singleton instance of the radio driver
RH_RF95 rf95(16, 26);
// 16 = NSS, 26 = INT
void setup() {
// Init the RFM95 module
if (!rf95.init())
{
Serial.println("[RFM95] init failed");
delay(500);
}
}
```
works perfect, log output:
```log
RH_RF95::init() started
RH_RF95::init() success
```
Any idea?
Answers:
username_1: During initialization, RadioLib checks the version register of RFM95. It should read 0x11, yours returns 0x00. That usually indicates faulty SPI bus or damaged device. It looks like RadioHead performs no such check, does the device actually work?
username_0: Yes the device works perfectly. The counterpart is [SparkFun Pro RF - LoRa, 915MHz (SAMD21)](https://www.sparkfun.com/products/14916) and the two communicate without problems with each other.
But......
Found the problem.
I am using an ESP32 which has two SPI, one called HSPI and a second one VSPI
The RF96 is connected to HSPI, but it seems the ESP32 defaults SPI to VSPI.
So the SPI is initialized on the wrong pins. That's why it doesn't work.
But no idea why it works with the RadioHead library.
As a quick test I hard-coded the pins into Module::init()
```cpp
_spi->end();
_spi->begin(14, 12, 13, 16);
```
Then it works.
Now I need to figure out how to set the SPI to the correct pins and give it as argument to
```
RFM96 lora = new Module(16, 26, 33, 32, mySPI);
```
Sorry for bothering you with this. It's not a bug, but an ESP32 specific problem.
Status: Issue closed
|
arangodb/arangodb-php | 184836377 | Title: Undefined variable $revision in DocumentHandler.php line 705
Question:
username_0: https://github.com/arangodb/arangodb-php/blob/master/lib/triagens/ArangoDb/DocumentHandler.php#L705
adding $revision = `$document->getRevision();` before this line fixed the problem.
Answers:
username_1: Thanks for taking the time to report.
This one was already fixed in in devel but did not make it to master yet.
Commit: <PASSWORD>ef<PASSWORD>
I will hopefully release a new 3.0.x version within the next couple of days, which will include this and other fixes and improvements.
Closing this for now. If there is anything else concerning this, feel free to comment.
PS: Release 3.1 of the driver will finally have the very old deprecated stuff removed, which unfortunately wasn't removed in 2.x and 3.0. This in turn will allow for some refactoring and improvements further down the road.
Status: Issue closed
|
infinitered/solidarity | 803330298 | Title: High severity vulnerability in gluegun dependency (apisauce)
Question:
username_0: Hello :)
Solidarity currently depends on [email protected] which depends on a version of axios with the high severity vulnerability https://www.npmjs.com/advisories/1594. Can we please upgrade gluegun to latest version ?
Answers:
username_1: Sounds good. Do you want to do a PR? If not I can put this on a TODO list.
username_0: Hi @username_1
Sorry, didn't see your reply. I made a local PR but I'm waiting for this [MR](https://github.com/SimonJang/gluegun/commit/a411ecbf3805a57c8f5abeb69ef5450b155c329a) to be merged :)
username_1: I merged that, now we'll just need a new version pushed and then an update. Thanks!
Status: Issue closed
|
Aldrin-John-Olaer-Manalansan/DOTA-2-MOD-Master | 1122570347 | Title: ANTI VAC
Question:
username_0: Can anyone share a fix for this? I really missed playing with great looking skins,
Answers:
username_1: Hi, there is a fix, that disables vac message, but it requires special gameinfo.gi, that made specially for D2MO (Dota 2 Mod Organizer, and you don't need to use it), i'm playing with mods for almost month, and everything is working. You can download it here: https://www.mediafire.com/file/0fe3799qegsag0o/gameinfo.gi/file . Also make sure you'll instal D2MO, you don't need to install every mod through this app, but it will make base folder for mods in your dota directory.
Status: Issue closed
|
apollographql/apollo-feature-requests | 657463159 | Title: Validate Apollo Client typePolicies `keyFields` with Apollo CLI
Question:
username_0: Apollo Client 3 support's a new cache configuration mechanic with typePolicies, specifically allowing us to define a canonical id for a type via `keyFields`. I've run into several cases where I add a `keyFields` value and then get runtime errors for not fetching that field. This seems like a great place to integrate with the Apollo CLI to verify that `keyFields` definitions in `typePolicies` always have the fields required in all queries. |
HackerHappyHour/impact | 138858879 | Title: Auto creation of team sub-contexts via integrations
Question:
username_0: new threads/channels/groups should be able to automatically be created for any integration *impact* is linked to, such that a new sub-context will become available inside of a chat context, when a new context is created within the integration.
*sub-context*: a channel or group, existing within an *impact* team.
Let's take github for example:
You should be able to associate an *impact* team to either a github organization, or a github team.
- Whenever a new *github team* is added to a *github organization* that is associated to an `impact team`, it should automatically create a new sub-context inside the `impact team
- Whenever a new *github repository* is added to a *github team* associated with an *impact team*, it should create a new sub-context
### Considerations
- Privacy/visibility settings of the new context item from the integration (new *github team/repo*), should carry over, or sync via LDAP, to the access and visibility of the sub-context in *impact*.
- The privacy/visibiltiy settings could determine *what type* of sub-context is created. For example, if a github team is publicly visible, perhaps a channel is added to the *impact* team. However, if the github team is not publicly visible, then it creates a private group inside the *impact* team. |
jlippold/tweakCompatible | 339998215 | Title: `HapticFeedback` working on iOS 11.2.6
Question:
username_0: ```
{
"packageId": "com.ziph0n.hapticfeedback",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "com.ziph0n.hapticfeedback",
"deviceId": "iPhone9,4",
"url": "http://cydia.saurik.com/package/com.ziph0n.hapticfeedback/",
"iOSVersion": "11.2.6",
"packageVersionIndexed": true,
"packageName": "HapticFeedback",
"category": "Tweaks",
"repository": "BigBoss",
"name": "HapticFeedback",
"packageIndexed": true,
"packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 100% with 2 working reports.",
"id": "com.ziph0n.hapticfeedback",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.0.7",
"shortDescription": "Get an haptic feedback when pushing buttons!",
"latest": "1.2-1",
"author": "Ziph0n",
"packageStatus": "Working"
},
"base64": "<KEY>
"chosenStatus": "working",
"notes": ""
}
``` |
clearlinux/distribution | 927921048 | Title: conda bundle Module Not Found Error: No module named 'conda_package_handling'
Question:
username_0: Severity: major
System: 34770
Description: After initializtion using `conda init` and session restart, conda still cannot be used. conda command such as `list`, `install` or `create` shows the following error:
Traceback (most recent call last):
File "/usr/lib/python3.9/site-packages/conda/exceptions.py", line 1079, in __call__
return func(*args, **kwargs)
File "/usr/lib/python3.9/site-packages/conda/cli/main.py", line 84, in _main
exit_code = do_call(args, p)
File "/usr/lib/python3.9/site-packages/conda/cli/conda_argparse.py", line 83, in do_call
return getattr(module, func_name)(args, parser)
File "/usr/lib/python3.9/site-packages/conda/cli/main_create.py", line 41, in execute
install(args, parser, 'create')
File "/usr/lib/python3.9/site-packages/conda/cli/install.py", line 116, in install
if context.use_only_tar_bz2:
File "/usr/lib/python3.9/site-packages/conda/base/context.py", line 734, in use_only_tar_bz2
import conda_package_handling.api
ModuleNotFoundError: No module named 'conda_package_handling'
......
Answers:
username_1: Upstream bug
https://github.com/conda/conda/issues/9510 |
JacksonBates/fcc-scraper | 272741965 | Title: Fix regex in the challenge name parser
Question:
username_0: Spaces are being dealt with inconsistently.
See line 5 of controllers/getSolution.js
Answers:
username_1: Hi Jackson, Im Matt, I'm a freecodecamper from Melbourne, seen you a few times ages ago at FCC Melb. Do you mind if have a go at my first pull requests here with you?
I'm fairly sure I've sorted this issue, the replace function is missing the first space.
line 5 of controllers/getSolution.js was removing the '%' symbol thus leaving just a '20' and the subsequent replace function in writeFile.js was ignoring that space...
I've worked out what the solution is, but am just trying to figure out best way to make a pull request, never done one before!
username_0: Hi Matt!
You're more than welcome to have a crack at it.
I actually made a video walkthrough for pull requests last year which might help. The gist of it is to fork the repo, check out a new branch for the fix, make the change, commit the change, push the commit and then open up a pull request across branches on my original repo. Sounds complicated, but this should help:
https://www.youtube.com/watch?v=dSl_qnWO104&t=104s
When you test it - test against the camper **portablestick**. He has some older solutions that include odd characters that broke the old regex I had. Fixing that introduced the current problem, so ensure your fix doesn't undo that fix :)
Once you PR, I'll review your code and walk you through any changes they might need.
username_1: Thanks Jackson. I totally thought this was an open source thing and here I am crashing your lil project haha!
Your vid is excellent by the way. If you could leave those simpler bugs to me, I'll get on to them.
PR coming through!
username_0: If it's on GH it's an open source thing! Always happy to have contributors :)
username_0: fixed by #16
Status: Issue closed
|
AppMetrics/AppMetrics | 425342973 | Title: RegisterFromAssembly Get Error "Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information."
Question:
username_0: Hi , when I was use health part in 3.0 version ,I got an error "Unable to load one or more of the requested types. Retrieve the LoaderExceptions property for more information."
And the stacktrace is:
at System.Reflection.RuntimeModule.GetTypes(RuntimeModule module)
at System.Reflection.RuntimeAssembly.get_DefinedTypes()
at System.Linq.Enumerable.SelectManySingleSelectorIterator`2.MoveNext()
at System.Linq.Enumerable.WhereSelectEnumerableIterator`2.MoveNext()
at App.Metrics.Health.Extensions.DependencyInjection.Internal.HealthChecksAsServices.AddHealthChecksAsServices(IServiceCollection services, IEnumerable`1 types) in C:\projects\microsoftextensions\src\App.Metrics.Health.Extensions.DependencyInjection\Internal\HealthChecksAsServices.cs:line 27
at Microsoft.Extensions.DependencyInjection.ServiceCollectionHealthCheckBuilderExtensions.RegisterFromAssembly(IHealthCheckBuilder healthCheckBuilder, IServiceCollection services, DependencyContext dependencyContext) in C:\projects\microsoftextensions\src\App.Metrics.Health.Extensions.DependencyInjection\ServiceCollectionHealthCheckBuilderExtensions.cs:line 34
at Demo.Startup.ConfigureServices(IServiceCollection services) in D:\下载\AppMetricsDemo-master\AppMetricsDemo\Demo\Startup.cs:line 82
So I debug with the source code, and find out that, is was caused by the class "DefaultHealthCheckTypeProvider" in line 25,and the reason is that,the Assembly in StaticHealthAssemblyProvider include "App.Metrics.Health.Abstractions", but at the run time,it`s "DefinedTypes" is null.
I think ,we should exclude “App.Metrics.Health.Abstractions” assembly.
Answers:
username_1: Refer to the [sample application](https://github.com/AppMetrics/Samples.V2/tree/master/AspNetCore2.Health.Api.QuickStart)
Status: Issue closed
|
bvaisvil/zenith | 582718614 | Title: Panic on WSL writing db
Question:
username_0: `thread '<unnamed>' panicked at 'Couldn't open database.: Io(Custom { kind: Other, error: "Os { code: 38, kind: Other, message: \"Function not implemented\" }" })', src/metrics.rs:248:17
thread '<unnamed>' panicked at 'Couldn't open database.: Io(Custom { kind: Other, error: "Os { code: 38, kind: Other, message: \"Function not implemented\" }" })', src/metrics.rs:248:17
thread panicked while panicking. aborting.
Illegal instruction (core dumped)`
Seems to happen at the right write in the sled db.
Answers:
username_0: perhaps related to https://github.com/spacejam/sled/issues/958
username_1: Just curious, is this happening in WSL or WSL2?
username_0: Believe it is WSL as
` cmd.exe /c "systeminfo" | grep "^OS Version"`
gives me
`OS Version: 10.0.18363 N/A Build 18363`
and WSL2 is in build >= 18917 ?
username_1: I see, yeah, I think so. `wsl -l -v` is another way to explicitly check if it's version 2.
FWIW wsl2 is actually a stock linux kernel running in a VM, instead of a reimplementation of Linux APIs on Windows, so it's much less likely to have incompatibilities.
If you can provide a binary (I don't see a windows executable for the latest release) I can test to see if this also happens on WSL2. No worries if that's not convenient for you though.
username_1: Ok, it seems to work fine on WSL2!
username_0: That's good. Thanks very much for testing that!
username_0: Latest version has no issues in WSL for me. Feel free to re-open this if you find anything. Thanks!
Status: Issue closed
|
artsy/force | 173220895 | Title: Partner Article Redirect
Question:
username_0: Partner articles should redirect from /article/ to their profile slug at /:id/article/:slug
Answers:
username_0: Based off: https://artsy.slack.com/archives/publishing-support/p1471872867000277
Actually @owendodd do we want articles for non-2.0 partners to have the partner header too? Because this redirect will require both Partner and Partner_2 to have /:id/article/:slug support. Currently only Partner_2 has this.
Status: Issue closed
|
Jessecar96/SteamDesktopAuthenticator | 947075850 | Title: Invalid length for a Base-64 array or string ERROR
Question:
username_0: ### **When i launch SDA:**

### **If I click on continue both of my accounts disappear**

**anyone had this issue before?**
Answers:
username_1: Are you using the same phone no /sim dongle/desktop as you've always used ?
I hot a similar error because i forgot i had swapped the 3g dongle i was using for internet to wi-fi fibre broadband.
username_1: I had to plug in my 3g dongle and use that to do confirmations. |
test-repo-billy/sdk-release-request | 618891205 | Title: [data-plane] JS: 'Test'
Question:
username_0: **Link**: [test](test)
**Tag**: test
**Release date**: 2020-05-29
**Contacts**:
[<NAME> <<EMAIL>>](mailto:<EMAIL>); @username_0 ,
[<NAME> <<EMAIL>>](mailto:<EMAIL>); @zhenglaizhang
**Comments**:
Answers:
username_0: Test add comment |
coredns/coredns | 287816559 | Title: RRSIG caching issue
Question:
username_0: After 8 days of operation (with file and without changing the zone file) the cached signatures are still served (and they are only signed for 8 days).
[dnssec.go#L40](https://github.com/coredns/coredns/blob/master/plugin/dnssec/dnssec.go#L40) states that "Signatures will be cached for a short while.".
But I can't find any code where expired or old signatures are dropped.
Answers:
username_1: hmmm, me neither :(
I 100% sure I thought about, but somehow that didn't translated in to code.
username_1: If you change `eightDays = 8 * 24 * time.Hour` into something smaller like `1 * 24 * time.Hour` you should see failures earlier. Could you easily try #1385 ?
username_0: I'll try and report.
But if I understand the code correctly we will run into the following problem:
We have an RRSIG in the coredns cache which is valid now and not yet expired.
However, given the TTL, the signature will expire while still in the cache of a resolver.
And even if we subtract the TTL for the validity check, we don't consider a possible clock drift between the server and a validator.
username_1: You're right we need to do this at 3/4 of the validity period and then regen.
Status: Issue closed
|
google/dagger | 533294964 | Title: Hello. I am having trouble using dagger.
Question:
username_0: This is my Application.kt
` class SellerHubApplication : Application(), HasAndroidInjector {
@Inject
lateinit var activityDispatchingAndroidInjector: AndroidInjector<Any>
override fun androidInjector(): AndroidInjector<Any> = activityDispatchingAndroidInjector
override fun onCreate() {
super.onCreate()
DaggerAppComponent.builder()
.application(this)
.build()
.inject(this)
Fresco.initialize(applicationContext)
}
}`
and this is my AppComponent
`@Singleton
@Component(modules = [AndroidInjectionModule::class, ActivityBuilder::class, AppModule::class])
interface AppComponent {
@Component.Builder
interface Builder {
@BindsInstance
fun application(app: Application): Builder
fun build(): AppComponent
}
fun inject(app: SellerHubApplication)
}`
and my issue is that "dagger.android.AndroidInjector<java.lang.Object> cannot be provided without an @Provides-annotated method.".
if i change "fun inject(app: SellerHubApplication)" to "fun inject(app: Application)". It works but in runtime, it shows "activityDispatchingAndroidInjector has not been initialized". Please help me.<issue_closed>
Status: Issue closed |
geneontology/go-annotation | 326398065 | Title: IPR019404 mapped to a merged term
Question:
username_0: Hello InterPro,
IPR019404 is mapped to
GO:0001104 RNA polymerase II transcription cofactor activity
which now has as a primary ID: GO:0003712
Do you automatically migrate merge terms ? If not we'll make sure to notify you ahead of time.
Thanks, Pascale
Answers:
username_1: Hi Pascale,
Thank you for this. We do have a system (script) to migrate these terms. We usually do so as soon as we receive the report from GO. However, as you may be aware, these results are only visible after our release. In this case, we changed these terms in May, so the new ID 'GO:0003712' will be visible in our next release (v.69) in June. Therefore, it's always good to get your report as early as possible to synchronise the InterPro and GO websites. :)
Thank you,
Hsin-Yu
username_0: Thanks @username_1
Can you specify which report you are talking about ? Some email with errors ?
username_1: Hi Pascale,
Sorry for the delayed reply. I just double checked with Amaia. We usually pick up the data from GOA and run a sanity check in InterPro. We then migrate the terms using SQL.
Regards,
Hsin-Yu
Status: Issue closed
|
BluSunrize/ImmersiveEngineering | 464874219 | Title: [1.12.2] Bug inside Engineers workbench
Question:
username_0: ### Description of the issue:
Recipes where the wirecutter is involved can only give you 1 result item at a time.
After that you can't take another result item out. Removing and adding an ingrediënt resets it.
### Crashlog:
no crash.
### Versions & Modlist
forge -14.23.5.28.38
No other mods, tested it with only IE<issue_closed>
Status: Issue closed |
RPi-Distro/raspi-config | 900187263 | Title: Disable the new GL driver on Pi 4, now see error on boot?
Question:
username_0: Disable the new GL driver on Pi 4, now see error on boot?
`# systemctl status gldriver-test
● gldriver-test.service - Check for v3d driver
Loaded: loaded (/lib/systemd/system/gldriver-test.service; enabled; vendor preset: enabled)
Active: failed (Result: exit-code) since Tue 2021-05-25 00:33:41 UTC; 6min ago
Main PID: 416 (code=exited, status=1/FAILURE)
May 25 00:33:41 localhost systemd[1]: Starting Check for v3d driver...
May 25 00:33:41 localhost gldriver_test.sh[416]: /usr/lib/systemd/scripts/gldriver_test.sh: line 12: /usr/share/X11/xorg.conf.d/99-fbturbo.conf: No such file or directory
May 25 00:33:41 localhost systemd[1]: gldriver-test.service: Main process exited, code=exited, status=1/FAILURE
May 25 00:33:41 localhost systemd[1]: gldriver-test.service: Failed with result 'exit-code'.
May 25 00:33:41 localhost systemd[1]: Failed to start Check for v3d driver.
`
Raspberry Pi OS lite installed on Pi4. The GL driver is applicable only to X11 environment right? This error should not be occurring thus?
Answers:
username_1: @spl237 The `[Unit]` section of `gldriver-test`'s `gldriver-test.service` should probably have `ConditionPathIsDirectory=/usr/share/X11/xorg.conf.d`?
username_0: I will check,.. Nope, or I should say, yeah, maybe that is the fix, the unit file is pretty basic...
`[Unit]
Description=Check for v3d driver
[Service]
Type=oneshot
ExecStart=/usr/lib/systemd/scripts/gldriver_test.sh
RemainAfterExit=yes
[Install]
WantedBy=multi-user.target
`
Since I am running lite, none of the is applicable per se. Seems lite something that was missed when the lite image was established. For now I disabled the service to avoid the error, which as far as I can see, just informational if you are not using the GL driver in the first place.
The actuall script executed...
`#!/bin/bash
if grep -s -q okay /proc/device-tree/soc/v3d@7ec00000/status \
/proc/device-tree/soc/firmwarekms@7e600000/status \
/proc/device-tree/v3dbus/v3d@7ec04000/status ||
[ "$(arch)" == aarch64 ]; then
if [ -e /usr/share/X11/xorg.conf.d/99-fbturbo.conf ] ; then
rm /usr/share/X11/xorg.conf.d/99-fbturbo.conf
fi
else
if ! [ -e /usr/share/X11/xorg.conf.d/99-fbturbo.conf ] ; then
cat > /usr/share/X11/xorg.conf.d/99-fbturbo.conf << EOF
# This is a minimal sample config file, which can be copied to
# /etc/X11/xorg.conf in order to make the Xorg server pick up
# and load xf86-video-fbturbo driver installed in the system.
#
# When troubleshooting, check /var/log/Xorg.0.log for the debugging
# output and error messages.
#
# Run "man fbturbo" to get additional information about the extra
# configuration options for tuning the driver.
Section "Device"
Identifier "Allwinner A10/A13 FBDEV"
Driver "fbturbo"
Option "fbdev" "/dev/fb0"
Option "SwapbuffersWait" "true"
EndSection
EOF
fi
fi
`
Running the script it fails with...
'/usr/lib/systemd/scripts/gldriver_test.sh: line 12: /usr/share/X11/xorg.conf.d/99-fbturbo.conf: No such file or directory'
For Pi OS lite image. |
NizarSemlali/INF3172-TP2 | 220358444 | Title: Questions à poser Lundi :
Question:
username_0: - Si l'exécutable tsh doit être dans le bin ou simplement dans le répertoire TP2 .
- Valider l'arborescence du programme .
-Si le fichier à afficher avec fin est vide, qu'est ce qu'on affiche ?
-.DS_Store , es ce que le fichier doit être considéré dans nos calculs ou pas ?<issue_closed>
Status: Issue closed |
rclone/rclone | 606600006 | Title: (req) search box
Question:
username_0: 
Answers:
username_1: You can use the forums for any site feedback rather than raising an issue.
https://forums.rclone.org
Thanks!
Status: Issue closed
|
ripperhe/Bob | 737383063 | Title: 百度翻译使用自动检测可能出现无法准确翻译问题
Question:
username_0: <!--
⚠️ 反馈前请确保已阅读
⚠️ 反馈前请确保已阅读
⚠️ 反馈前请确保已阅读
1. 请确保你已经认真阅读了 README 文件,可能你的问题不是「问题」。
2. 请在 issues 页面搜索你的问题,很可能已被解决。
3. 如果仍旧有问题,请填写模板描述问题,以便大家理解、定位和解决问题。
-->
<!-- 这是隐藏的信息 -->
<!-- 👆👆👆这样括起来的信息将被隐藏,填写时注意不要写在里面。 -->
<!-- 点击编辑器上方的 preview 可预览你填写的效果 -->
## 运行环境
```
❯ neofetch
'c. <EMAIL>
,xNMM. ----------------------
.OMMMMo OS: macOS Catalina 10.15.7 19H2 x86_64
OMMM0, Host: MacBookPro11,5
.;loddo:' loolloddol;. Kernel: 19.6.0
cKMMMMMMMMMMNWMMMMMMMMMM0: Uptime: 2 days, 13 hours, 55 mins
.KMMMMMMMMMMMMMMMMMMMMMMMWd. Packages: 156 (brew)
XMMMMMMMMMMMMMMMMMMMMMMMX. Shell: zsh 5.7.1
;MMMMMMMMMMMMMMMMMMMMMMMM: Resolution: 1920x1080
:MMMMMMMMMMMMMMMMMMMMMMMM: DE: Aqua
.MMMMMMMMMMMMMMMMMMMMMMMMX. WM: Quartz Compositor
kMMMMMMMMMMMMMMMMMMMMMMMMWd. WM Theme: Red (Dark)
.XMMMMMMMMMMMMMMMMMMMMMMMMMMk Terminal: iTerm2
.XMMMMMMMMMMMMMMMMMMMMMMMMK. Terminal Font: MesloLGS-NF-Regular 13
kMMMMMMMMMMMMMMMMMMMMMMd CPU: Intel i7-4870HQ (8) @ 2.50GHz
;KMMMMMMMWXXWMMMMMMMk. GPU: AMD Radeon R9 M370X, Intel Iris Pro
.cooc,. .,coo:. Memory: 9661MiB / 16384MiB
```
## 描述问题
在百度私有api中测试
使用自动检测

指定语言

Answers:
username_1: 了解了,感谢反馈 🤝
username_2: 更新了以后用上百度的 识别经常失败 不如上一版有道
Status: Issue closed
username_1: 新版加了语种检查按钮,如果不幸检查错误,可以手动点击这个按钮切换下语言
 |
ssborbis/ContextSearch-web-ext | 285073295 | Title: possible quick menu style change?
Question:
username_0: Quick Menu Style change?
select text and mouse left-click( not hold)
Quick Context Search
https://addons.mozilla.org/ko/firefox/addon/quickcontextsearch/
Answers:
username_1: I'll see about adding that. In the meantime, you might like the latest betas where I added an option for opening the quick menu automatically after selecting text.
https://addons.mozilla.org/en-US/firefox/addon/contextsearch-web-ext/versions/beta
username_1: If you're wanting the grid in the context menu like in the add-on you linked to, it's not possible with the new web extensions api
username_0: thanks.
Status: Issue closed
|
everpeace/banditsbook-scala | 208867080 | Title: Hedge Algorithm run results in a NotANumberException
Question:
username_0: [info] Running com.github.everpeace.banditsbook.algorithm.hedge.TestHedge
-------------------------------
Hedge Algorithm
-------------------------------
arms = (μ=0.1), (μ=0.9), (μ=0.1), (μ=0.1), (μ=0.1) (Best Arm = 1)
horizon = 500
nSims = 5000
η = (0.2,0.4,0.8,0.9,0.95,1.0)
starts simulation on η=0.2.
`results are written to output/test-hedge-results.csv`
`[error] (run-main-0) org.apache.commons.math3.exception.NotANumberException: NaN is not allowed
org.apache.commons.math3.exception.NotANumberException: NaN is not allowed
at org.apache.commons.math3.distribution.EnumeratedDistribution.<init>(EnumeratedDistribution.java:121)
at org.apache.commons.math3.distribution.EnumeratedIntegerDistribution.<init>(EnumeratedIntegerDistribution.java:100)
at org.apache.commons.math3.distribution.EnumeratedIntegerDistribution.<init>(EnumeratedIntegerDistribution.java:68)
at com.github.everpeace.banditsbook.algorithm.package$$anon$1.<init>(package.scala:34)
at com.github.everpeace.banditsbook.algorithm.package$.CategoricalDistribution(package.scala:32)
at com.github.everpeace.banditsbook.algorithm.hedge.Hedge$$anon$1.selectArm(Hedge.scala:55)
at com.github.everpeace.banditsbook.algorithm.hedge.Hedge$$anon$1.selectArm(Hedge.scala:45)
at com.github.everpeace.banditsbook.algorithm.Algorithm$$anonfun$selectArm$1.apply(Algorithm.scala:42)
at com.github.everpeace.banditsbook.algorithm.Algorithm$$anonfun$selectArm$1.apply(Algorithm.scala:40)
at cats.data.StateFunctions$$anonfun$inspect$2.apply(StateT.scala:184)
at cats.data.StateFunctions$$anonfun$inspect$2.apply(StateT.scala:184)
at cats.data.StateFunctions$$anonfun$apply$10.apply(StateT.scala:169)
at cats.data.StateFunctions$$anonfun$apply$10.apply(StateT.scala:169)
at cats.data.StateT$$anonfun$run$1.apply(StateT.scala:36)
at cats.data.StateT$$anonfun$run$1.apply(StateT.scala:36)
at cats.Eval$$anon$5$$anonfun$1$$anon$6$$anonfun$2.apply(Eval.scala:81)
at cats.Eval$$anon$5$$anonfun$1$$anon$6$$anonfun$2.apply(Eval.scala:81)
at cats.Eval$Compute.loop$1(Eval.scala:276)
at cats.Eval$Compute.value(Eval.scala:290)
at com.github.everpeace.banditsbook.algorithm.TracedAlgorithmDriver$$anonfun$2.apply(TracedAlgorithmDriver.scala:60)
at com.github.everpeace.banditsbook.algorithm.TracedAlgorithmDriver$$anonfun$2.apply(TracedAlgorithmDriver.scala:59)
at cats.data.StateT$$anonfun$map$1.apply(StateT.scala:30)
at cats.data.StateT$$anonfun$map$1.apply(StateT.scala:30)
at cats.data.StateT$$anonfun$transform$1$$anonfun$apply$6.apply(StateT.scala:72)
at cats.data.StateT$$anonfun$transform$1$$anonfun$apply$6.apply(StateT.scala:72)
at cats.Eval$$anonfun$map$1.apply(Eval.scala:58)
at cats.Eval$$anonfun$map$1.apply(Eval.scala:58)
at cats.Eval$Compute.loop$1(Eval.scala:286)
at cats.Eval$Compute.value(Eval.scala:290)
at com.github.everpeace.banditsbook.algorithm.TracedAlgorithmDriver.runFrom(TracedAlgorithmDriver.scala:95)
at com.github.everpeace.banditsbook.algorithm.TracedAlgorithmDriver.run(TracedAlgorithmDriver.scala:81)
at com.github.everpeace.banditsbook.algorithm.TracedAlgorithmDriver.run(TracedAlgorithmDriver.scala:88)
at com.github.everpeace.banditsbook.testing_framework.TestRunner$$anonfun$run$1.apply$mcVI$sp(TestRunner.scala:84)
at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160)
at com.github.everpeace.banditsbook.testing_framework.TestRunner$.run(TestRunner.scala:79)
at com.github.everpeace.banditsbook.algorithm.hedge._TestHedge$$anonfun$1.apply$mcVD$sp(TestHedge.scala:69)
at com.github.everpeace.banditsbook.algorithm.hedge._TestHedge$$anonfun$1.apply(TestHedge.scala:64)
at com.github.everpeace.banditsbook.algorithm.hedge._TestHedge$$anonfun$1.apply(TestHedge.scala:64)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
at scala.collection.AbstractTraversable.map(Traversable.scala:104)
at com.github.everpeace.banditsbook.algorithm.hedge._TestHedge$class.run(TestHedge.scala:64)
at com.github.everpeace.banditsbook.algorithm.hedge.TestHedge$.run(TestHedge.scala:35)
[Truncated]
at com.github.everpeace.banditsbook.algorithm.hedge.TestHedge$delayedInit$body.apply(TestHedge.scala:35)
at scala.Function0$class.apply$mcV$sp(Function0.scala:34)
at scala.runtime.AbstractFunction0.apply$mcV$sp(AbstractFunction0.scala:12)
at scala.App$$anonfun$main$1.apply(App.scala:76)
at scala.App$$anonfun$main$1.apply(App.scala:76)
at scala.collection.immutable.List.foreach(List.scala:381)
at scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:35)
at scala.App$class.main(App.scala:76)
at com.github.everpeace.banditsbook.algorithm.hedge.TestHedge$.main(TestHedge.scala:35)
at com.github.everpeace.banditsbook.algorithm.hedge.TestHedge.main(TestHedge.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
[trace] Stack trace suppressed: run last compile:runMain for the full output.
java.lang.RuntimeException: Nonzero exit code: 1
at scala.sys.package$.error(package.scala:27)
[trace] Stack trace suppressed: run last compile:runMain for the full output.
[error] (compile:runMain) Nonzero exit code: 1
[error] Total time: 1 s, completed Feb 20, 2017 1:31:39 PM` |
Purii/react-native-tableview-simple | 200316689 | Title: CustomCell doesn't render >
Question:
username_0: Hi there,
Is there a reacon I don't see ">" when I use CustomCell?
Also, how can I use the ">" in my code? How is it rendered?
Thanks.
Answers:
username_0: Found it: it's this:
<View style={styles.accessory_disclosureIndicator} />
Status: Issue closed
|
enzoampil/fastquant | 678490010 | Title: [FEATURE] How to control the size plots when backtrader plots are embedded into notebook output
Question:
username_0: Currently, it's common to have workflows where backtrader plots are rendered straight from a jupyter or colab notebook. The issue is that the size of the plot is quite small by default, which makes it difficult to see through the busy contents of the backtesting results.
It would be great if `backtest` could have a `figsize` argument similar to `matplotlib.pyplot`, where the default is much bigger than the current setup. E.g. `figsize=(12, 7)` could be a good default.
Answers:
username_0: Sample default size:

username_1: I looked into this problem and found a solution below.
We just need to add these lines on top of the jupyter notebook.
```python
import matplotlib.pyplot as pl
pl.style.use("default") #ggplot is also fine
pl.rcParams["figure.figsize"] = (15,12)
#all styles: print(pl.style.available)
```
Documentation: https://matplotlib.org/3.3.0/tutorials/introductory/customizing.html
We may be able to set this global config automatically if code is being run in jupyter but there may be [caveats](https://stackoverflow.com/questions/15411967/how-can-i-check-if-code-is-executed-in-the-ipython-notebook).
username_0: Understand @username_1 . Thanks will use this as quick fix. I think this should be automated as much as possible straight from `backtest`.
Yeah, think it makes sense to start chipping away at these plotting issues. Really affects the UX of the tool.
username_0: Also, great to refer these two relevant threads on backtrader docs: [link1](https://community.backtrader.com/topic/291/ipython-notebook-plot-sizing/5), [link2](https://community.backtrader.com/topic/22/plot-maximize-windows/2)
Status: Issue closed
username_2: I added support for `figsize()` parameter in the `backtest` function in https://github.com/username_0/fastquant/pull/286/files#diff-9081e26fa869af389e67c6f9e0872cb6eb190052979673832a9b2f4a6215343eR67
username_0: Thanks @username_2 ! :smile: |
jenkins-x/enhancements | 584994339 | Title: Enhance JX UI
Question:
username_0: JX UI looks great and it gives a high-level view of the current status of the builds, environments, projects and so on.
As user, it would be really great to be able to:
- Configure namespaces (at the moment 'jx' is hardcoded in some places and it is not possible to configure it)
- Support GHE or a configurable wat for Git Providers/Services (e.g. links/images seem to point towards Github only - so they are broken)
- Have covered edge-cases in builds' completion/status (collected builds, old ones, stuck, etc..)
- Check the Codebase for potential contributions
Answers:
username_1: we've an OSS plugin for octant now: https://github.com/jenkins-x/octant-jx which hopefully has the features you need; if not we could create the necessary issues there & hack on the code to improve it?
username_0: thank you very much @username_1! I have just tried it and it looks great (and those features seem implemented there)!
what about users not able to access the cluster resources directly and not able to use octant? (e.g. a developer who is taking care of a project managed by jx - or non-technical users such as Scrum Master or Product Owner) |
PAhelper/PhoenixAdult.bundle | 993763209 | Title: HotCrazyMess [Nubiles] Error Loading Actor
Question:
username_0: **Describe the bug**
When matching or refreshing metadata for any scene from HotCrazyMess the actor fails to load with an error in the logs.
**Include**
1. All Scenes from https://hotcrazymess.com/video/
2. HotCrazyMess - Scene Title
```
2021-09-11 02:09:15,492 (15bc) : INFO (logkit:16) - ******UPDATE CALLED*******
2021-09-11 02:09:15,492 (15bc) : INFO (logkit:16) - SiteNum: 543
2021-09-11 02:09:15,492 (15bc) : INFO (logkit:16) - Provider: networkNubiles
2021-09-11 02:09:15,493 (15bc) : DEBUG (preferences:258) - Loaded preferences from DefaultPrefs.json
2021-09-11 02:09:15,494 (15bc) : DEBUG (preferences:178) - Loaded the user preferences for com.plexapp.agents.phoenixadult
2021-09-11 02:09:15,496 (15bc) : INFO (logkit:16) - Requesting GET "https://hotcrazymess.com/video/watch/72198"
2021-09-11 02:09:15,844 (15bc) : INFO (logkit:16) - Requesting GET "https://hotcrazymess.com/model/profile/10220/chloe-temple"
2021-09-11 02:09:16,118 (15bc) : CRITICAL (core:574) - Exception in the update function of agent named 'PhoenixAdult', called with guid 'com.plexapp.agents.phoenixadult://72198|543?lang=en' (most recent call last):
File "C:\Program Files (x86)\Plex\Plex Media Server\Resources\Plug-ins-2b1b51db9\Framework.bundle\Contents\Resources\Versions\2\Python\Framework\api\agentkit.py", line 1093, in _update
agent.update(obj, media, lang, **kwargs)
File "C:\Plex\Plex Media Server\Plug-ins\PhoenixAdult.bundle\Contents\Code\__init__.py", line 136, in update
provider.update(metadata, lang, siteNum, movieGenres, movieActors)
File "C:\Plex\Plex Media Server\Plug-ins\PhoenixAdult.bundle\Contents\Code\networkNubiles.py", line 138, in update
galleryURL = PAsearchSites.getSearchBaseURL(siteNum) + detailsPageElements.xpath('//div[contains(@class, "content-pane-related-links")]/a[contains(., "Pic")]/@href')[0]
File "C:\Program Files (x86)\Plex\Plex Media Server\Resources\Plug-ins-2b1b51db9\Framework.bundle\Contents\Resources\Versions\2\Python\Framework\code\sandbox.py", line 108, in <lambda>
_getitem_ = lambda x, y: x.__getitem__(y),
IndexError: list index out of range
```<issue_closed>
Status: Issue closed |
kapi2289/fb-anti-unsend | 894236183 | Title: No license
Question:
username_0: EN: This software isn't licensed. This means that nobody can use this software without permission. Please add license file. I recommend MIT license for this software. You can find another license at https://choosealicense.com/
If you choose GNU GPL as your license, read their guidelines for using the GPL at https://www.gnu.org/licenses/gpl-howto.html.en or read guidelines at the end of gpl text. If you choose GPL, please add a note about using newer versions of GNU GPL.
PL: Te oprogramowanie nie jest udostępnione z licencją. To oznacza, że nikt nie może go używać bez zgody. Proszę, dodaj licencję. Polecam licencję MIT dla tego programu. Jeśli wolisz, możesz znaleźć inną licencję na https://choosealicense.com/ .
Jeżeli wybierzesz GNU GPL jako licencję, przeczytaj ich wskazówki używania GPL na https://www.gnu.org/licenses/gpl-howto.html.en lub na końcu tekstu GPL. Dodatkowo: jeśli wybierzesz GPL, dodaj informacje o używaniu nowszych wersji GNU GPL.
Answers:
username_1: Yeah, you're right. It's old, small and not used anymore because of problems with Facebook project so I haven't thought of any license yet. I'll add it, thanks!
Status: Issue closed
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.