repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
neo4j-contrib/neo4j-apoc-procedures
693122259
Title: Espace AsciiDoc attribute references to suppress warnings when building the documentation with Antora Question: username_0: Currently, when we build the documentation with Antora from https://github.com/neo4j-documentation/docs-refresh, Asciidoctor prints almost 300 warnings because the documentation contains unresolved attribute references. For reference, the syntax of an attribute reference in AsciiDoc is: `{myattribute}`. The main issue is that the description of the `apoc.*` procedures frequently contain this syntax. For instance: ``` apoc.cypher.runTimeboxed('cypherStatement',{params}, timeout) - abort kernelTransaction after timeout ms if not finished ``` As you can see in this example, AsciiDoc will try to resolve the attribute reference `{params}`. One way to solve this issue is to escape the attribute reference with `\`: ``` apoc.cypher.runTimeboxed('cypherStatement',\{params}, timeout) - abort kernelTransaction after timeout ms if not finished ``` We could also prevent substitution using a single plus symbol. While doing so we could also use a monospaced font formatting: ``` `+apoc.cypher.runTimeboxed('cypherStatement',{params}, timeout)+` - abort kernelTransaction after timeout ms if not finished ``` In my opinion, the result will be more readable: **Current** apoc.cypher.runTimeboxed('cypherStatement',{params}, timeout) - abort kernelTransaction after timeout ms if not finished **After** `apoc.cypher.runTimeboxed('cypherStatement',{params}, timeout)` - abort kernelTransaction after timeout ms if not finished However, this solution is more complex to implement because we need to accurately find the start and the end of the procedure (and if we guess wrong it will produce unexpected results). Status: Issue closed Answers: username_0: Thanks @username_1 I think we should also backport this change to 4.0. Should I open a new pull request or do you want to cherry-pick this commit to the 4.0 branch?
sheredom/utf8.h
172902516
Title: Crash waiting to happen? Question: username_0: So I stumbled upon this: https://github.com/username_1/utf8.h/blob/master/utf8.h#L697 Is this a crash waiting to happen, or am I reading the logic wrong? If both `h` and `n` are equal strings, this will read past the memory of those strings. Answers: username_1: Hmm you are probably right! I'll add a test for that and we'll make sure it fails, then probably have to add on a check for null terminator there too. Good find thanks! :smile: Status: Issue closed
tinygo-org/tinygo-site
942185137
Title: installation instructions still show v0.18.0 Question: username_0: Installation instruction still show old 0.18.0 version i.e. Linux: ``` wget https://github.com/tinygo-org/tinygo/releases/download/v0.18.0/tinygo_0.18.0_amd64.deb sudo dpkg -i tinygo_0.18.0_amd64.deb ```
Stremio/stremio-translations
287498419
Title: Same string keys multiple times Question: username_0: I see that many string keys string keys are multiple times in `Caps`, `nocaps` and `ALLCAPS` forms when string translation is the same. Is that necessary? Because maintenance is more easy if the same translation string is only in one string key, use only one version could be a good option. Answers: username_1: where is that? could be by mistake username_0: For example in `en-US.json`. Some repeated translation strings are: ````json "INTRO_TASTE_COMEDY": "Comedy", "Comedy": "Comedy", "COMEDY": "Comedy", "INTRO_TASTE_MUSIC": "Music", "Music": "Music", "MUSIC" : "Music", "ADDONS": "Add-ons", "WEBSITE_ADD_ONS": "Add-ons", "INTRO_TASTE_HELP_US": "Help us learn more about you.", "HELP_LEARN_MORE": "Help us learn more about you.", "WEBSITE_CREATE_AN_ADDON": "Create an add-on", "WEBSITE_ADDONS_CREATE": "Create an add-on", ```` And because this is the template for all you can see in more files. username_0: Hi @username_1 , also FAQ13 and FAQ18 are the same. username_0: "WEBSITE_LINK_ADDON_SDK" is a duplicate object key.
inspursoft/board
769697987
Title: change node group label failed Question: username_0: In GitLab by @yhua123 on May 31, 2018, 17:48 Can't saw deployment and service yaml file. Answers: username_0: In GitLab by @tmon546596046 on Nov 8, 2019, 07:51 can't reproduce, so close it username_0: In GitLab by @sokril on Nov 8, 2019, 08:21 closed Status: Issue closed
godotengine/godot
1159077436
Title: Project can be opened multiple times from Project Manager Question: username_0: ### Godot version 3f69ea4 ### System information W10 ### Issue description When you double click a project on project list, it opens the project. But if you keep clicking it until project manager closes, it will spawn multiple instances of the editor. If the project is starting, further clicks should have no effect. ### Steps to reproduce 1. Open Project Manager 2. Click a project on the list 3. Click it 4. Click again 5. Click 6. Click 7. Keep clicking ### Minimal reproduction project _No response_ Answers: username_1: Not bug but feature :D username_2: Running multiple Godot editor instances *on the same project* isn't an officially supported use case. Either way, if you really need it, you can start multiple project manager instances (which can then start multiple editor instances on the same project). Status: Issue closed
CDAT/cdat
467352839
Title: Build with hdf5=1.0.5 Question: username_0: The latest packages on conda-forge are build with hdf5=1.0.5, but cdat v8.1 and its dependencies are built with hdf5=1.0.4. Could a new build be made with hdv5=1.0.5 for compatibility with conda-forge? Answers: username_1: @username_2 can we do this? username_2: The issue I had was netcdf 4.7 and other libraries (libcf, libcdms,...) I had to pin netcdf=4.6.2 Last time I tried to upgrade libnetcf to 4.7 some conda-forge libraries were still at 4.6.2, I can try again and see if I can make it work. Issues often arise on OSX, but not Linux due to the some version metadata. username_0: @username_2, I believe this is the official place to see what the current conda-forge pinnings are: https://github.com/conda-forge/conda-forge-pinning-feedstock/blob/master/recipe/conda_build_config.yaml If you look there, they're still using libnetcdf 4.6.2: https://github.com/conda-forge/conda-forge-pinning-feedstock/blob/master/recipe/conda_build_config.yaml#L455 But hdf5=1.10.5: https://github.com/conda-forge/conda-forge-pinning-feedstock/blob/master/recipe/conda_build_config.yaml#L416 So I don't believe you need to update to libnetcdf 4.7 in order to use hdf5 1.10.5. conda-forge is not. username_0: I would imagine it would make all of our lives easier if more cdat packages made their way to conda-forge. If you're interested in doing that and need another maintainer to help with the process, let me know. I'm vaguely aware that some packages are on conda-forge and others are not but I'm not quite clear on which and why. username_1: @username_0 we have recipe for all but vtk-cdat pkg takes too long to build and is killed by conda-forge... Hence we could never push it to conda-forge. And therefore all of our graphics are NOT on conda-forge. But this is not the issue here as cdms2 is on conda-forge. username_2: @username_0 I had to recompile CDMS2 2 weeks ago since conda-forge changed the `gcc` compiler to `clang` on OSX. The `cdat/label/nightly` channel has 1.10.5 for hdf5. I could try to make a point release on conda-forge. @username_1 is this ok with you? username_0: @username_1, ah, that makes sense. I can see why vtk-cdat might be a problem beyond building -- it vendors a lot of other stuff from what I recall and conda-forge doesn't allow that. Thanks for the clarification. username_0: cdms2 from conda-forge is working fine with hdf5=1.10.5. I'll try to figure out which package wasn't happy with that hdf5 version. I assume it was something in `cdat/label/v81` but I didn't figure out for sure which. username_0: I'm not sure if either of you did anything, but I am now able to build the environment I need with `hdf5=1.10.5`. Not sure what changed or what wasn't working before. But thank you both either way! Status: Issue closed
grails/grails-core
115082757
Title: forward does not work in Grails 3.0.9 Question: username_0: forward command does not work in Grails 3.0.9 To reproduce the issue: *1. create a new application with grails 3.0.9* *2. create a controller like following:* class ForwardController { def index() { } def toAim() { forward action: 'aim' } def aim() { } } *3. create two views like following* **index.gsp:** <html> <head> <title>INDEX</title> </head> <body> <p>Hello INDEX</p> <g:link action="toAim">toAim</g:link> </body> </html> **aim.jsp:** <html> <head> <title>AIM</title> </head> <body> <p>Hello AIM</p> </body> </html>``` Answers: username_1: I'm confused you're forwarding to an action and expecting it to go to a JSP view? username_0: Oh, sorry. It's a .gsp view. Corresponding to the action in the controller. I change it in the issue username_1: This works fine and we have a functional test covering this behaviour https://github.com/grails/grails3-functional-tests/blob/master/app1/src/integration-test/groovy/functionaltests/ForwardingSpec.groovy#L46 If you can come up with an example that reproduces the problem we can take another look Status: Issue closed username_0: @username_1 https://www.dropbox.com/s/7urqdklrrqdcf6i/testForward.zip?dl=0 check out this example...
junegunn/fzf
1112696083
Title: toggle whether to show hidden files on the fly Question: username_0: <!-- ISSUES NOT FOLLOWING THIS TEMPLATE WILL BE CLOSED AND DELETED --> <!-- Check all that apply [x] --> - [x] I have read through the manual page (`man fzf`) - [x] I have the latest version of fzf - [x] I have searched through the existing issues ## Info - OS - [x] Linux - [ ] Mac OS X - [ ] Windows - [ ] Etc. - Shell - [ ] bash - [x] zsh - [ ] fish ## Problem / Steps to reproduce How can I toggle between show hidden files and not show hidden files on the fly? I know I can set whether to show them in FZF_DEFAULT_COMMAND using `--hidden` flag (for `fd`) But I think to toggle this on the fly is more useful. Answers: username_1: You can set up `reload` bindings to update the input list dynamically. See https://github.com/username_1/fzf/blob/master/ADVANCED.md#dynamic-reloading-of-the-list Status: Issue closed
prolike/prolike.io
590988251
Title: Write the Gigs Question: username_0: ### Problem Right now we don't have any stories of our customers, we don't have anything to show for it. ### Solution We write down in the gigs all the good stories we have of our customers, the relationships, our workflow with them and what we have accomplished. ### Value We got an archive of good and successful stories of our customers, that we can post, share and reference/show to potential customers who is possibly looking for the same solution, whether it is IT support, a website or software. #### Witch customers should we use? - [ ] Copenhagen Coaching Center - [ ] Cykling Uden Alder - [ ] Erhvervshusene (Hvad hedder de??) - [ ] IT System Design - [ ] Omnihorse - [ ] Champagnekælderen - [ ] Oticon - [ ] ..more? Answers: username_0: @username_1 Do we need permission from our customers to write these stories? username_1: Make it whole! As in #626 Whats that plan? Why do we do it? what are we hoping to achieve? think "Landing page" and SoMe ...and alla that jazz! username_1: Quote template: https://docs.google.com/document/d/1-kX5RCPnM2ctLaPDpJlVYgXK_Z1Eb5pQYPP4bkE8lfY username_1: Maybe develop a format for what a gig story is? username_1: More Issues? * SoMe campaign? * Update the gig layout? * Improve the call-to-action?
patternfly/patternfly-org
469251868
Title: Fix a11y issues in Documentation/React/Wizard Question: username_0: https://www.patternfly.org/v4/documentation/react/components/wizard/ There is 1 type of a11y issue in this page: 1. The skip-link target should exist and be focusable <img width="486" alt="Screen Shot 2019-07-17 at 10 32 25 AM" src="https://user-images.githubusercontent.com/20118816/61384229-303a8d00-a87e-11e9-97d0-b4757c726353.png"> Steps to reproduce the a11y errors: 1. Right click on the page and choose inspect 2. In the developer tools, choose "axe" from the top menu 3. Run the test<issue_closed> Status: Issue closed
dominique-mueller/angular-package-builder
288481736
Title: fix(compile): build error when not in debug mode Question: username_0: Hi, First of all, thanks for this awesome package builder which is very useful 😍 But, I think I've found a bug / regression in the lib. All the code I present here is extracted from a demo repo I've create for this issue (https://github.com/username_0/demo-angular-package-builder) When I use dependencies coming from Angular (or maybe from something else, I haven't test...), the build doesn't work with the following result: ```shell iMac-de-Kevin:~/W/g/d/demo-angular-package-builder kevin $ yarn build yarn run v1.3.2 $ angular-package-builder Angular Package Builder ➜ Configuration ➜ Prepare (line breaks, resources) ➜ Compile TypeScript into JavaScript (ES2015, ES5) ERROR: An error occured while trying to compile the TypeScript sources using the Angular Compiler. [TypeScript] : Error: ENOENT: no such file or directory, realpath '/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/@angular/core/core.d.ts' [TypeScript] at ImportDeclaration in /Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/dist-angular-package-builder/lib/index.ts:1:1 [TypeScript] at SourceFile in /Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/dist-angular-package-builder/lib/index.ts:1:1 [TypeScript] at createError (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:105:17) [TypeScript] at throwError (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:114:11) [TypeScript] at Volume.realpathBase (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:1098:13) [TypeScript] at Volume.realpathSync (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:1106:21) [TypeScript] at Object.realpath (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:3989:32) [TypeScript] at Object.realpath (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:69063:76) [TypeScript] at realpath (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26373:42) [TypeScript] at tryResolve (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26360:87) [TypeScript] at nodeModuleNameResolverWorker (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26342:68) [TypeScript] at nodeModuleNameResolver (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26321:16) error Command failed with exit code 1. info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command. ``` But, if I run the process in debug mode, without any modification, no error appears and the lib are correctly built. ``` iMac-de-Kevin:~/W/g/d/demo-angular-package-builder kevin $ yarn build --debug yarn run v1.3.2 $ angular-package-builder --debug Angular Package Builder ➜ Configuration ➜ Prepare (line breaks, resources) ➜ Compile TypeScript into JavaScript (ES2015, ES5) ➜ Generate JavaScript bundles (ES2015, ES5, UMD) ➜ Compose package Success! [4.15 seconds] ✨ Done in 4.35s. ``` This is the full example where I run the command one after the other ![image](https://user-images.githubusercontent.com/1970922/34927799-3820287e-f9b9-11e7-85d1-fd3acfe3fa44.png) The `index.ts` of my app is very simple : ```ts import {InjectionToken} from "@angular/core"; export const foo = new InjectionToken('Foo'); ``` And, of course, the error shown during the standard build process isn't coherent, because I have the file locally on my file system. ![image](https://user-images.githubusercontent.com/1970922/34927840-8079f10e-f9b9-11e7-8ff8-3b5b03b4aa44.png) If you need more information, let me know 😉 /cc @neonox31 @Yann29 @yakoust and @aymeric-duchein Answers: username_1: First, thanks for making us aware of your issue, appreciate it very much :) I actually have a few questions: Which versions of Angular and TypeScript are you using? What NodeJS version do you have? And finally, did it ever work properly or did the bug came in later on, e.g. after updating dependencies? username_1: I also see that you're using Yarn and MacOS, whereas I always developed with npm on Windows. Yarn shouldn't be a problem, MacOS might be though ... username_0: Angular Package Builder ➜ Configuration ➜ Prepare (line breaks, resources) ➜ Compile TypeScript into JavaScript (ES2015, ES5) ERROR: An error occured while trying to compile the TypeScript sources using the Angular Compiler. [TypeScript] : Error: ENOENT: no such file or directory, realpath '/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/@angular/core/core.d.ts' [TypeScript] at ImportDeclaration in /Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/dist-angular-package-builder/lib/index.ts:1:1 [TypeScript] at SourceFile in /Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/dist-angular-package-builder/lib/index.ts:1:1 [TypeScript] at createError (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:105:17) [TypeScript] at throwError (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:114:11) [TypeScript] at Volume.realpathBase (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:1098:13) [TypeScript] at Volume.realpathSync (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:1106:21) [TypeScript] at Object.realpath (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:3989:32) [TypeScript] at Object.realpath (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:69063:76) [TypeScript] at realpath (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26373:42) [TypeScript] at tryResolve (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26360:87) [TypeScript] at nodeModuleNameResolverWorker (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26342:68) [TypeScript] at nodeModuleNameResolver (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26321:16) npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! [email protected] build: `angular-package-builder` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the [email protected] build script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /Users/kevin/.npm/_logs/2018-01-15T07_22_54_835Z-debug.log MacBook-Pro-de-Kevin:~/w/g/d/demo-angular-package-builder kevin | master $ ``` And, I don't right now have any Windows machine to test it... maybe @yakoust or @neonox31 can test it for me 😀 username_2: Angular Package Builder → Configuration → Prepare (line breaks, resources) → Compile TypeScript into JavaScript (ES2015, ES5) ERROR: An error occured while trying to compile the TypeScript sources using the Angular Compiler. [TypeScript] : Error: ENOENT: no such file or directory, realpath 'C:/dev /github/demo-angular-package-builder/node_modules/@angular/core/core.d.ts' [TypeScript] at ImportDeclaration in C:/dev/github/demo-angular-package- builder/dist-angular-package-builder/lib/index.ts:1:1 [TypeScript] at SourceFile in C:/dev/github/demo-angular-package-builder /dist-angular-package-builder/lib/index.ts:1:1 [TypeScript] at createError (C:\dev\github\demo-angular-package-build er\node_modules\memfs\lib\volume.js:105:17) [TypeScript] at throwError (C:\dev\github\demo-angular-package-builde r\node_modules\memfs\lib\volume.js:114:11) [TypeScript] at Volume.realpathBase (C:\dev\github\demo-angular-packa ge-builder\node_modules\memfs\lib\volume.js:1098:13) [TypeScript] at Volume.realpathSync (C:\dev\github\demo-angular-packa ge-builder\node_modules\memfs\lib\volume.js:1106:21) [TypeScript] at Object.realpath (C:\dev\github\demo-angular-package-b uilder\node_modules\typescript\lib\typescript.js:3989:32) [TypeScript] at Object.realpath (C:\dev\github\demo-angular-package-b uilder\node_modules\typescript\lib\typescript.js:69063:76) [TypeScript] at realpath (C:\dev\github\demo-angular-package-builder\ node_modules\typescript\lib\typescript.js:26373:42) [TypeScript] at tryResolve (C:\dev\github\demo-angular-package-builde r\node_modules\typescript\lib\typescript.js:26360:87) [TypeScript] at nodeModuleNameResolverWorker (C:\dev\github\demo-angu lar-package-builder\node_modules\typescript\lib\typescript.js:26342:68) [TypeScript] at nodeModuleNameResolver (C:\dev\github\demo-angular-pa ckage-builder\node_modules\typescript\lib\typescript.js:26321:16) npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! [email protected] build: `angular-package-builder` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the [email protected] build script. npm ERR! This is probably not a problem with npm. There is likely additional log ging output above. npm ERR! A complete log of this run can be found in: npm ERR! C:\Users\Admin_WebFactory\AppData\Roaming\npm-cache\_logs\2018-01-1 5T07_30_59_890Z-debug.log ``` username_1: Thanks for the info! So, it turns out that Angular 5.2 breaks but Angular 5.1 works as intended. Further analysis showed that it's probably an issue with `tsickle` - Angular 5.1 uses tsickle 0.25.6 and Angular 5.2 uses 0.26.0. Either downgrading to Angular 5.1 or setting `annotateForClosureCompiler` to `false` fix this issue, at least temporarily (or using debug mode, of course). I will try to find a fix over the course of the next few days. username_1: Changes: https://github.com/angular/tsickle/compare/7d20a06330e342eb72ccb729da7744171082e8ea...82992d1b5afac4948cc3b56344251d4ccbc0ded5 username_0: The fact that the `--debug` flag allows to do the compilation is still strange. Do you think the lib is well built after this or some errors are still present in the package ? username_1: The `--debug` flag doesn't change the build chain itself, it only decides whether interim build outputs should be emitted to the disk (if enabled) - or kept in memory (if disabled). To make this work, however, I had to use some not-so-fancy techniques of which one seems to break with the latest 'tsickle' ... So no worries, the angular-package-builder still does it's job, even with the `debug` flag enabled :) Status: Issue closed username_1: Fixed with https://github.com/username_1/angular-package-builder/pull/41, will also make a documentation update regarding Angular & TypeScript compatibility - and hopefully publish a new release tomorrow. Also, the CI now covers all Angular & TypeScript combinations. ![travisci](https://user-images.githubusercontent.com/7271961/35070936-5b72b2a6-fbdf-11e7-8021-c56bd3d03a03.PNG) username_1: Here we go, version `1.0.1` is published! Have fun! Changelog: https://github.com/username_1/angular-package-builder/blob/master/CHANGELOG.md
makewebau/shipper
270116890
Title: Does not work with themes Question: username_0: ![image](https://user-images.githubusercontent.com/13425337/32248656-755b657e-beda-11e7-96cf-34e26ae64b40.png) No version number (to be expected) Not stripping out .git (weird because it takes out .shipignore) Status: Issue closed Answers: username_0: @unetics Could not replicate the failure to delete `.git` folder, but the issues with the version should now be resolved.
frioux/DBIx-Class-DeploymentHandler
88491250
Title: (minor) Readme mis-generated Question: username_0: https://metacpan.org/source/FREW/DBIx-Class-DeploymentHandler-0.002216/README#L161 Answers: username_1: As of https://metacpan.org/source/ETJ/DBIx-Class-DeploymentHandler-0.002223/README this is no longer the case! (Not my doing though) Thanks for the report. Status: Issue closed
github/docs
928083359
Title: Some filenames are too long for Git for Windows Question: username_0: ### What article on docs.github.com is affected? `translations/<lang>/content/github/setting-up-and-managing-your-enterprise-account/setting-policies-for-organizations-in-your-enterprise-account/configuring-the-retention-period-for-github-actions-artifacts-and-logs-in-your-enterprise-account.md` (German, Japanese, Brazilian Portuguese, and Chinese translation) ### What part(s) of the article would you like to see updated? Git for Windows may complain about the file paths being too long, e.g. if you attempt to checkout the latest main branch of this repo. The length of just the relative path is 244 characters, with the limit being 260, but the absolute path is what counts (so you have to include `C:\Users\<name>\Documents\...` or wherever your working copy is located at. - Cloning the repo directly to `C:\` would shorten the path, but it's just a workaround. - Shortening the folder names and the file name of the actual content would be possible, but alter doc URLs and would make the URL slug less self-descriptive and SEO-friendly. - [Git supports long paths](https://stackoverflow.com/a/22575737/2044940) (apparently at the expense of compatibility), but it needs to be enabled manually, e.g. with `git config core.longpaths true` for the current repo. This worked for me. I suggest to add a remark about this to the `CONTRIBUTING.md`. ### Additional information ``` error: cannot stat 'translations/de-DE/content/github/setting-up-and-managing-your-enterprise-account/setting-policies-for-organizations-in-your-enterprise-account/configuring-the-retention-period-for-github-actions-artifacts-and-logs-in-your-enterprise-account.md': Filename too long error: cannot stat 'translations/ja-JP/content/github/setting-up-and-managing-your-enterprise-account/setting-policies-for-organizations-in-your-enterprise-account/configuring-the-retention-period-for-github-actions-artifacts-and-logs-in-your-enterprise-account.md': Filename too long error: cannot stat 'translations/pt-BR/content/github/setting-up-and-managing-your-enterprise-account/setting-policies-for-organizations-in-your-enterprise-account/configuring-the-retention-period-for-github-actions-artifacts-and-logs-in-your-enterprise-account.md': Filename too long error: cannot stat 'translations/zh-CN/content/github/setting-up-and-managing-your-enterprise-account/setting-policies-for-organizations-in-your-enterprise-account/configuring-the-retention-period-for-github-actions-artifacts-and-logs-in-your-enterprise-account.md': Filename too long ``` Answers: username_1: Thanks for the issue @username_0! I'll triage this for the team to take a look 👀 username_2: In the [Windows](https://github.com/github/docs/blob/main/CONTRIBUTING.md#windows) section possibly add a `curl` command to download the .zip file. username_3: curl username_4: @username_0 Thank you for reporting this and explaining some of the viable workarounds. We have added this issue to our backlog. 👍🏻 username_5: > Depending on your Windows Version >> Steps: to fix |-- > > Step A: > 1. Open: cmd.exe > 2. Run: regedit.exe > 3. Edit: HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\FileSystem > 4. Look: for LongPathsEnabled (If Value Missing) > |--> Do Step B: > 4a. Right click : "FileSystem" Folder > 4b. Choose: New Dword Value > 4c. Add: "LongPathsEnabled" text > Else: > 5. Change Value 0 To 1 > 6. Then : sound=Applause Author: username_5 <the_kat6<EMAIL>> AuthorDate: 2021-July-29 12:52:00 -0500 Commit: username_5 <<EMAIL>> CommitDate: 2021-July-29 12:52:00 -0500 username_1: We've added information about this to our contributing guide! https://github.com/github/docs-internal/blob/main/CONTRIBUTING.md#windows Status: Issue closed
facebook/flow
132446891
Title: Type narrowing on function args doesn't work as expected Question: username_0: ```js function getValue(): Promise<{prop: {id: string, name: string}}> { return Promise.resolve({prop: {id: 'whatever', name: 'whatever'}}); } function onValue(value: {prop: {id: string}}) { return value.prop.id; } getValue().then(onValue); ``` Gives me the error: ```js function getValue(): Promise<{prop: {id: string, name: string}}> { ^^^^^^^^^^^^^^^^^^^^^^^^^^ property `name`. Property not found in function onValue(value: {prop: {id: string}}) { ^^^^^^^^^^^^ object type ``` The type of both functions is correct, the `onValue` intentionally identifies that it is not interested in `value.prop.name`. Incidentally, this works for properties that are only one level deep. Answers: username_1: This seems to be related to how we type object properties. You can simplify this further to the following (which suffers from the same errors): ```javascript function getValue(): {prop: {id: string, name: string}} { return {prop: {id: 'whatever', name: 'whatever'}}; } function onValue(value: {prop: {id: string}}) { return value.prop.id; } let obj: {prop: {id: string}} = getValue(); ``` Status: Issue closed username_0: This can now be handled by simply marking `prop` as covariant: ```js function getValue(): Promise<{prop: {id: string, name: string}}> { return Promise.resolve({prop: {id: 'whatever', name: 'whatever'}}); } function onValue(value: {+prop: {id: string}}) { return value.prop.id; } getValue().then(onValue); ``` Or in your minimal example: ```js function getValue(): {prop: {id: string, name: string}} { return {prop: {id: 'whatever', name: 'whatever'}}; } let obj: {+prop: {id: string}} = getValue(); ```
wso2/carbon-dashboards
358734164
Title: Report Generation / Export Feature does not work Question: username_0: **Description:** Report Generation / Export Feature does not work **Suggested Labels:** Dashboard , PDF Generation ![report](https://user-images.githubusercontent.com/2908631/45315748-cad98980-b552-11e8-987b-896139ef52bb.png) **Suggested Assignees:** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assingnees, so this will help issue creators who are not a committer to suggest possible assignees--> **Affected Product Version:** **OS, DB, other environment details and versions:** **Steps to reproduce:** **Related Issues:** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --><issue_closed> Status: Issue closed
ActoKids/AD440_W19_CloudPracticum
407560754
Title: Configure event trigger Google Calendar Question: username_0: Triggers will need to be set for the google calendar crawler. Please use Cloud Watch to create a trigger based on time, and on URL. You have already done something similar in #44 but this time work with Ryan to create several triggers in case he needs to break up his Google Crawler. Answers: username_1: Update: Now have access to project AWS account but I do not currently have permissions for Cloudwatch in order to create the trigger. I need permission to call **CloudWatch Events API** so that I can create a trigger. username_1: I seem to have all the permissions needed to complete my task. I'm currently waiting on Ryan's crawler code to be approved so that I can create the event trigger. username_1: Successfully created trigger for GoogleCrawler on AWS. ![eventtriggerconfirm](https://user-images.githubusercontent.com/34496119/53132699-57866b80-3525-11e9-9065-f2817143ac68.JPG) username_1: At this point the Trigger is created and will trigger the lambda function every 24 hours. @rberry206 is working on get his code into what is currently an empty lambda function that was just created a few minutes ago by the DevOps team. This sprint assignment may need to be extended one week in order to finish testing. Will provide updates on this thread. username_1: ![invocations](https://user-images.githubusercontent.com/34496119/53139147-90cad580-353d-11e9-94c5-11b7a1fa62ea.JPG)
actor-framework/actor-framework
552375024
Title: CAF produces unexpected log files Question: username_0: It seems like CAF produces a log file when hitting a `CAF_LOG_ERROR` even if CAF was compiled with log level set to quiet. Answers: username_0: Already fixed in `topic/0.17.4`. The branch `issue/1024-integration` contains the same fix for `master` that applies once #1007 got merged. Status: Issue closed
symengine/symengine.py
116175529
Title: symengine.py does not support current symengine Question: username_0: Building symengine.py requires a specific commit of symengine. This is undocumented. ``` git checkout 3a1152fdafc83f8d1f35004b5bcda08f355e57c3 ``` Answers: username_1: Thanks @username_0 for reporting it. Indeed, we need to document this better, currently this is only "documented" here: https://github.com/symengine/symengine.py/blob/83c1a70cd2516a597a7a8a12ef0c0ce21ed825de/.travis.yml#L111 username_2: The README does it have now https://github.com/symengine/symengine.py/blob/master/README.md#installation Also after this PR https://github.com/symengine/symengine.py/pull/29 there will be a `symengine_version.txt` file. username_3: I have problems understanding the issue because current symengine.py does not compile with that version mentioned, while it does with current symengine. So, is this still an issue? Status: Issue closed username_4: Closing the issue as this is now documented in README.md username_4: To clarify, `symengine.py` supports only the version mentioned in README.md, but current `symengine` master may work as well. `symengine.py` will be updated from time to time to work with the latest `symengine` master
Anaconda-Platform/anaconda-client
906688734
Title: API call results too large Question: username_0: This is a feature request for the anaconda.org API. I don't see any repository where I could make a PR, so I asked around, and @chenghlee recommended posting here so that I reach the right people. I noticed that the anaconda.org API sometimes returns huge results for packages with a lot of release files. For example, https://api.anaconda.org/package/conda-forge/dvc is 13 MB. (This is breaking some conda-forge shields/badges due to a 10 MB limit, see conda-forge/dvc-feedstock#215 ). Currently, to compute the total number of downloads, the whole API result is downloaded and a sum is computed over all files. In order to address the problem, I think it would be helpful to have more endpoints like * `/packages/{owner_login}/{package_name}/latest-version` * `/packages/{owner_login}/{package_name}/total-downloads` * `/packages/{owner_login}/{package_name}/platforms` * `/packages/{owner_login}/{package_name}/license` These endpoints would match the functionality of the badges, and would be very useful to have in general.
huggingface/transformers
925628111
Title: [performance] module init w/ `from_pretrained` skip storage allocation Question: username_0: # 🚀 Feature request pt-1.9.0 added `torch.nn.utils.skip_init()` which (1) skips the module init (2) doesn't allocate any memory https://pytorch.org/tutorials/prototype/skip_param_init.html note: `torch.nn.utils.skip_init()` itself will be in 1.9.1, but the rest of the code should be in 1.9.0 We already implemented part 1 (skipping the custom init) in https://github.com/huggingface/transformers/pull/11471. We could further speed up the start up time and reduce CPU memory usage, by not allocating any storage for module init since `load_state_dict` will already have allocated `state_dict` from the pretrained weights (and some sub-modules that don't have pre-trained weights - will have to go through normal init). See https://pytorch.org/tutorials/prototype/skip_param_init.html#implementation-details another note: currently deepspeed needs to have the module storage pre-allocated for its `zero.Init` gather/scatter, but if the initial model's weights aren't allocated, then we can probably get rid of `zero.Init` altogether https://github.com/huggingface/transformers/issues/12273
pulibrary/figgy
276262825
Title: Titles in show view should show up on different lines instead of joined by `and` Question: username_0: Figgy: <img width="263" alt="screen shot 2017-11-22 at 8 16 35 pm" src="https://user-images.githubusercontent.com/2806645/33158819-1b5ec072-cfc2-11e7-9831-1a5f293d0e90.png"> Plum: <img width="170" alt="screen shot 2017-11-22 at 8 16 38 pm" src="https://user-images.githubusercontent.com/2806645/33158821-21099f10-cfc2-11e7-9e76-a1b0d000e517.png"> Answers: username_1: In the breadcrumb area, we probably want to use the first title. In the header above, we want all the titles to display, each on its own line (to gracefully handle RTL text) username_2: Probably a different issue, but it's weird to have a breadcrumb with no links in the case where there's not a parent. username_2: Other places that use header: * member lists (substitute semicolon-separated list? so translated values will be shown?) (e.g. https://github.com/pulibrary/figgy/blob/master/app/views/catalog/_members_multi_volume_work.html.erb#L15) * https://github.com/pulibrary/figgy/blob/master/app/models/concerns/linked_data/linked_ephemera_folder.rb#L84 ? use one-line list? * https://github.com/pulibrary/figgy/blob/master/app/services/manifest_builder.rb#L50 ? one-line list? username_2: uses of `heading` (currently implemented as first title): * https://github.com/pulibrary/figgy/blob/master/app/document_presenters/valkyrie_show_presenter.rb#L4 used by blacklight (catalog_helper) to set heading on record page that's really the only one I guess; I'll change this directly to insert the newlines username_2: Not sure how that manifest `to_s` is used; that method isn't in plum. The manifests themselves look different in the label: https://plum.princeton.edu/concern/multi_volume_works/pb5647p872/manifest vs https://figgy.princeton.edu/concern/scanned_resources/9c145976-1525-4047-80dd-49a4b27d2457/manifest will look into how that's done. username_2: remaining instances of the sentence format that I've noticed: - [ ] manifest label - [ ] search results show view (not sure where this comes from; it may be that blacklight uses 'header' and I'll just need to finally change how that's formatted (in plum it's separated with a comma) username_2: I sort of think plum was doing the manifest label wrong. it seems like it should be a string. http://iiif.io/api/presentation/2.1/#label Status: Issue closed
open-telemetry/opentelemetry-java
685605587
Title: BatchSpanProcessor performance goals Question: username_0: During recent optimisations of `BatchSpanProcessor` several questions have arisen regarding further performance optimisations. But we don't currently have proper benchmarks that could clearly demonstrate the benefit of any performance gains as opposed to induced code complexity. E.g. we cannot it automatic way compare the percent of dropped spans. We have to come up with such benchmarking setup and agree on acceptable performance characteristics of `BatchSpanProcessor`. Answers: username_1: Probably related to #1438 . username_1: There is also an issue to write micro-benchmarks for the BSP: #790 Status: Issue closed username_1: We have multiple benchmarks now for the BSP, so I'm going to close this.
nutonomy/nuscenes-devkit
503468487
Title: get the yaw in global coordinates from the yaw in local coordinate Question: username_0: Hi, I find strange that the cameras orientation have large pitches and rolls. How is that possible? I am trying to get the global yaw from only the yaw in camera coordinate (and not the pitch and roll). Do you have any idea how to do that? Thanks for your inputs Answers: username_1: How much is the pitch and roll? If you look at the videos I'd say they are imperceptible and therefore pretty small: https://www.youtube.com/watch?time_continue=4&v=Sipud0IonvQ username_0: This code shows that boxes in camera frame have a large pitch and roll. This is caused by the camera (calibrated sensor) rotation which has a non zero pitch and roll. Am I missing something? ` nusc = NuScenes(version='v1.0-trainval', dataroot="./nuScenes/v1.0-trainval", verbose=True) my_scene = nusc.scene[0] first_sample_token = my_scene['first_sample_token'] sensor = 'CAM_FRONT' my_sample = nusc.get('sample', first_sample_token) cam_front_data = nusc.get('sample_data', my_sample['data'][sensor]) my_annotation_token = my_sample['anns'][0] box_global = nusc.get('sample_annotation', my_annotation_token) nusc.get('sample_annotation', my_annotation_token) data_path, boxes, camera_intrinsic = nusc.get_sample_data(my_sample['data'][sensor]) print('box orientation in global frame: ', box_global['rotation']) print('box orientation in camera frame: ', boxes[0].orientation.elements) sd_record = nusc.get('sample_data', my_sample['data'][sensor]) cs_record = nusc.get('calibrated_sensor', sd_record['calibrated_sensor_token']) print('camera rotation:',cs_record['rotation'])` username_1: Hi, I refactored your code a bit, used the mini split and converted from quaternions to euler angles: ``` from pyquaternion import Quaternion import numpy as np from nuscenes import NuScenes nusc = NuScenes(version='v1.0-mini', dataroot="/data/sets/nuscenes-mini", verbose=True) my_scene = nusc.scene[0] first_sample_token = my_scene['first_sample_token'] sensor = 'CAM_FRONT' my_sample = nusc.get('sample', first_sample_token) cam_front_data = nusc.get('sample_data', my_sample['data'][sensor]) my_annotation_token = my_sample['anns'][0] box_global = nusc.get('sample_annotation', my_annotation_token) data_path, boxes, camera_intrinsic = nusc.get_sample_data(my_sample['data'][sensor]) sd_record = nusc.get('sample_data', my_sample['data'][sensor]) cs_record = nusc.get('calibrated_sensor', sd_record['calibrated_sensor_token']) angles_box_global = np.array(Quaternion(box_global['rotation']).yaw_pitch_roll) / np.pi * 180 angles_box_camera = np.array(Quaternion(boxes[0].orientation.elements).yaw_pitch_roll) / np.pi * 180 angles_camera = np.array(Quaternion(cs_record['rotation']).yaw_pitch_roll) / np.pi * 180 print('box orientation in global frame: ', angles_box_global) print('box orientation in camera frame: ', angles_box_camera) print('camera rotation:', angles_camera) ``` Results are: ```` box orientation in global frame: [-21.091 0. 0. ] box orientation in camera frame: [178.79077403 1.27037324 90.88688192] camera extrinsics: [ 44.75652071 89.54131123 -134.80356665] ```` So the annotation lives in the global frame with no pitch or roll. When transformed to camera frame we can see that the orientations are less than 2 degrees off. The camera extrinsics are not easy to read as ego vehicle (IMU) and camera use different coordinate frames: https://www.nuscenes.org/public/images/data.png Status: Issue closed username_0: Thanks a lot, I did not thought the axis orientation were different between vehicle and sensor frame.
finos/perspective
770185325
Title: Defining calculated column using 'final aggregate column values' Question: username_0: ## Support Question Is there a way to define a calculated column that will be calculated using aggregated columns values rather than an aggregation over itself? Example: I'm looking at a portfolio using underlying positions data. The portfolio level values are generated using a 'group by' clause. However, something like fill rate won't aggregate properly. fill rate = sum(executed_qty) / sum(qty) I would need fill rate to be calculated using aggregate values of both sum(executed_qty) and sum(qty) rather than the individual executed_qty/qty thx. Answers: username_1: This can only be done explicitly in the API by _flattening_ ([Docs here](https://perspective.finos.org/docs/md/js.html#flattening-a-view-into-a-table)) an aggregated `View` into a new `Table`, then adding a computed column. For your case, I'm assuming you are using a Perspective Web Component, so you just need to create the flat `Table` and `load()` it, something like this: ```javascript const agg_view = table.view({ row_pivots: ["portfolio"], columns: ["executed_qty", "qty"] }); workspace.load(worker.table(agg_view)); ``` This is not currently possible through the GUI alone, starting with just `table` - you must specify the intermediate aggregate table ahead of time. The plan is to add this as the `Flatten` command in the `<perspective-workspace>` right-click context menu. Status: Issue closed
DJ73/ShellCrypt
733766220
Title: Add more pseudo random number generators Question: username_0: Check `middle_square_method.py` for format Answers: username_1: Assign this to me username_1: Please explain what I need to do. Your readme file is not detailed username_0: Need to create new files like `middle_square_method.py`. The file should have one generator function that returns a value in [0, 25] and accepts a seed to generate said values username_1: Okay, I will see to it
yegor256/syncem
433096298
Title: confusion with method send() Question: username_0: If the method's name is `send`, we get confused. Answers: username_0: @username_1 release, tag is `0.1.2` username_1: @username_0 OK, I will release it now. Please check the progress [here](http://www.username_1.com/t/19302-504747552) username_1: @username_0 Done! FYI, the full log is [here](http://www.username_1.com/t/19302-504747552) (took me 2min) Status: Issue closed
hurlbertlab/dietdatabase
771139823
Title: complete search for Parulidae diet data Question: username_0: The following species in Parulidae are missing from the `NA_avian_diet_refs.txt` file and therefore may never have been intentionally searched: Golden-winged warbler Black-throated gray warbler Yellow-throated warbler Northern Parula* Tropical Parula Ovenbird* Louisiana waterthrush* Connecticut warbler* Mourning warbler* Canada warbler* For each of the above species, 1) go to https://birdsoftheworld.org/bow/home (subscription only, so you'll need to be connected to UNC's VPN), and visit the species' *Diet and Foraging* page. 2) Look up each of the references cited on that page, and add them to the `NA_avian_diet_refs.txt` and indicate whether they do or do not have any quantitative diet data. Pay attention to whether any studies are cited within those papers that might potentially contain such data. 3) On Google Scholar, put the following search terms in quotes: "[common name] AND (diet OR foraging OR food)", for example "golden-winged warbler AND (diet OR foraging OR food)". Follow up with any papers that have titles indicating they might contain quantitative data. Add all papers examined to the `NA_avian_diet_refs.txt` file regardless of whether they have data. 4) For species above with an *, some diet data exists currently in our Avian Diet Database. Be sure to check that any papers you find in your literature search are not already entered. Answers: username_0: Note that "golden-winged warbler" has actually been listed as its own issue, so ignore that species here. Update in this issues thread as you finish searching a species on the list above. username_1: Finished with Black-throated Gray Warbler Status: Issue closed
Radnen/spherestudio
421686467
Title: Null reference exception when creating script in blank project Question: username_0: Just opened the editor for the first time, set plugins for everything except the option for the default editor to use with unknown file types. Then created a script, and got a null reference exception. Log: ```************** Exception Text ************** System.NullReferenceException: Object reference not set to an instance of an object. at SphereStudio.Ide.DocumentTab..ctor(IdeWindow ide, DocumentView view, String fileName, Boolean restoreView) at SphereStudio.Ide.IdeWindow.AddDocument(DocumentView view, String filepath, Boolean restoreView) at SphereStudio.Ide.IdeWindow.<>c__DisplayClass52_0.<menuNew_DropDownOpening>b__3(Object s, EventArgs ea) at System.Windows.Forms.ToolStripItem.RaiseEvent(Object key, EventArgs e) at System.Windows.Forms.ToolStripMenuItem.OnClick(EventArgs e) at System.Windows.Forms.ToolStripItem.HandleClick(EventArgs e) at System.Windows.Forms.ToolStripItem.HandleMouseUp(MouseEventArgs e) at System.Windows.Forms.ToolStrip.OnMouseUp(MouseEventArgs mea) at System.Windows.Forms.ToolStripDropDown.OnMouseUp(MouseEventArgs mea) at System.Windows.Forms.Control.WmMouseUp(Message& m, MouseButtons button, Int32 clicks) at System.Windows.Forms.Control.WndProc(Message& m) at System.Windows.Forms.ToolStrip.WndProc(Message& m) at System.Windows.Forms.ToolStripDropDown.WndProc(Message& m) at System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam) ************** Loaded Assemblies ************** mscorlib Assembly Version: 4.0.0.0 Win32 Version: 4.7.3362.0 built by: NET472REL1LAST_C CodeBase: file:///C:/Windows/Microsoft.NET/Framework64/v4.0.30319/mscorlib.dll ---------------------------------------- Sphere Studio Assembly Version: 0.0.0.0 Win32 Version: 0.0.0.0 CodeBase: file:///C:/Program%20Files/miniSphere/ide/Sphere%20Studio.exe ---------------------------------------- System.Windows.Forms Assembly Version: 4.0.0.0 Win32 Version: 4.7.3324.0 built by: NET472REL1LAST_C CodeBase: file:///C:/WINDOWS/Microsoft.Net/assembly/GAC_MSIL/System.Windows.Forms/v4.0_4.0.0.0__b77a5c561934e089/System.Windows.Forms.dll ---------------------------------------- System Assembly Version: 4.0.0.0 Win32 Version: 4.7.3362.0 built by: NET472REL1LAST_C CodeBase: file:///C:/WINDOWS/Microsoft.Net/assembly/GAC_MSIL/System/v4.0_4.0.0.0__b77a5c561934e089/System.dll ---------------------------------------- System.Drawing Assembly Version: 4.0.0.0 Win32 Version: 4.7.3056.0 built by: NET472REL1 CodeBase: file:///C:/WINDOWS/Microsoft.Net/assembly/GAC_MSIL/System.Drawing/v4.0_4.0.0.0__b03f5f7f11d50a3a/System.Drawing.dll ---------------------------------------- SphereStudio.Base Assembly Version: 0.0.0.0 Win32 Version: 0.0.0.0 CodeBase: file:///C:/Program%20Files/miniSphere/ide/SphereStudio.Base.DLL ---------------------------------------- System.Core Assembly Version: 4.0.0.0 Win32 Version: 4.7.3362.0 built by: NET472REL1LAST_C CodeBase: file:///C:/WINDOWS/Microsoft.Net/assembly/GAC_MSIL/System.Core/v4.0_4.0.0.0__b77a5c561934e089/System.Core.dll ---------------------------------------- System.Configuration Assembly Version: 4.0.0.0 [Truncated] CodeBase: file:///C:/Program%20Files/miniSphere/ide/ScintillaNET.DLL ---------------------------------------- ************** JIT Debugging ************** To enable just-in-time (JIT) debugging, the .config file for this application or computer (machine.config) must have the jitDebugging value set in the system.windows.forms section. The application must also be compiled with debugging enabled. For example: <configuration> <system.windows.forms jitDebugging="true" /> </configuration> When JIT debugging is enabled, any unhandled exception will be sent to the JIT debugger registered on the computer rather than be handled by this dialog box. ```
tdurieux/leboncoin-api
372082926
Title: TypeError: Cannot read property 'value' of undefined Question: username_0: Hello When I try the example provided in main page (https://github.com/username_1/leboncoin-api), I have the following error : ``` C:\lbc\node_modules\leboncoin-api\lib\search.js:235 searchLabel = searchId[0].value; ^ TypeError: Cannot read property 'value' of undefined at search.map.searchLabel (C:\lbc\node_modules\leboncoin-api\lib\search.js:235:39) at Array.map (<anonymous>) at getIdsForLabels (C:\lbc\node_modules\leboncoin-api\lib\search.js:232:21) at _.forEach (C:\lbc\node_modules\leboncoin-api\lib\search.js:277:30) at C:\lbc\node_modules\lodash\lodash.js:4911:15 at baseForOwn (C:\lbc\node_modules\lodash\lodash.js:2996:24) at C:\lbc\node_modules\lodash\lodash.js:4880:18 at Function.forEach (C:\lbc\node_modules\lodash\lodash.js:9344:14) at LodashWrapper.object.(anonymous function) [as forEach] (C:\lbc\node_modules\lodash\lodash.js:15716:25) at getSubCategoriesExtras (C:\lbc\node_modules\leboncoin-api\lib\search.js:269:21) at Search.getBodyParams (C:\lbc\node_modules\leboncoin-api\lib\search.js:324:22) at Search.run (C:\lbc\node_modules\leboncoin-api\lib\search.js:369:51) at Object.<anonymous> (C:\lbc\lbc.js:18:8) at Module._compile (internal/modules/cjs/loader.js:688:30) at Object.Module._extensions..js (internal/modules/cjs/loader.js:699:10) at Module.load (internal/modules/cjs/loader.js:598:32) ``` Any idea why ? Thanks for advance. Sybaris Answers: username_1: That is strange, I never faced this error. And the example is working right now :/ I will try to fix this. Thanks for the report. username_2: Hello, I have still the issue… Any idea ? Sybaris username_0: Hello, I have still the issue… Any idea ? Sybaris Status: Issue closed
Captain-K-101/Ssrf-labs
779390170
Title: Could you possibly share writeup? Question: username_0: Hi Captain! I am creating lab for SSRF. COuld you possibly share writeup for your tasks? Thank you Answers: username_1: Hi There yea sure , i'll be happy to but i am currently a bit busy, i'll send it in the upcoming 2-3 days *Regards * <NAME> Third Year B.tech ,Computer Science *Twitter <https://twitter.com/@Captainkay11> | Github <https://github.com/username_1>* Status: Issue closed username_1: Hi Captain! I am creating lab for SSRF. COuld you possibly share writeup for your tasks? Thank you username_2: Hi Captain,I think you forgot. I take the liberty to ask you for a writeup. Thanks! username_3: Can we fetch forbidden file using fopen() in php username_1: Hey @username_2, sorry for the late response, but for writeup, i am currently swarmed with work so might not be able to make a writeup for this in the upcoming future but once free will definitely get on this . The attacks are pretty simple to exploit basic understanding of SSRF should be enough for exploitation for any reference materials or stuff u can ping me on twitter @Captainkay11 username_1: @username_3 if u do have control over the fopen name parameter being passed it's possible to read files using basic lfi, but again there would be certain criteria too ,ie adequate permissions be there to read etc... username_3: Bro i ask it that can we fetch forbidden file (flag.php ) using ssrf Vulernbility because when I try to fetch it it's only give me a forbidden error :(
nuxt/modules
821197505
Title: Add Agility CMS Module Question: username_0: Hey there, Joshua from Agility CMS here 👋 Agility is a cloud-native, API-first CMS with built-in **Page Management** & **Sitemap Control** - learn more [here](https://help.agilitycms.com/hc/en-us/articles/360002996392)! One of our folks recently wrote a module to help integrate Agility CMS with Nuxt.js and we're interested in getting it listed into the list of modules: https://github.com/agility/agilitycms-nuxt-module This module pairs with our [Agility CMS & Nuxt.js Blog Starter](https://github.com/agility/agility-nuxt-blog). Let me know how we can go about getting it added! Thank you 😁<issue_closed> Status: Issue closed
baidu/amis
631434624
Title: 按钮添加disabledOn属性后,各行的操作按钮无法对齐 Question: username_0: ![image](https://user-images.githubusercontent.com/64950961/83858033-70cd1e80-a74e-11ea-8e0a-b7425bbc7ec4.png) Answers: username_1: 我记得我好想修复了,你换个新版试试? username_0: 刚改了下依赖,然后本地重启了项目,还是不行 ![image](https://user-images.githubusercontent.com/64950961/83860368-c2c37380-a751-11ea-8822-d40381927ee0.png) username_0: 现在好了 Status: Issue closed
python-cmd2/cmd2
813779919
Title: Cmd2 Application wide settings file via cmd2 extension. Question: username_0: I've been writing a few application-specific configuration file modules in different ways and haven't found something quite elegant enough. Ideally, I'd like to have a cmd2 add in module that took care of adding these configuration issues. It would work like this. When a cmd2 application starts up you would attempt to load a configuration file (user specified json, toml, ini.. etc) if successful their would be a Cmd.Configuration object available to all of cmd and subcommands and CommandSets via self._cmd.Config. My reasoning is that I would like to have CommandSets that are API clients for specific endpoints but each of those endpoints shouldn't be hard coded. But I cant imagine using Settables to hold these configs unless I could read them from a config file at start up in the event a url changed. Answers: username_1: Settables are really designed for settings you want users to be able to change at runtime. So far we have left how to implement at-startup configuration to developers and haven't attempted to provide anything to make this easier. Typically I have seen developers use either command-line arguments via `argparse`, some sort of configuration file (JSON, YAML, ini, etc.), or a combination thereof. But I'm open to the concept of providing something to make this easier, particularly if it integrates well with CommandSets. Any solution needs to be completely cross-platform and work well on Windows, Mac, Linux, Python 3.6+, x86, ARM, etc. username_2: @username_0 Since `CommandSets` already have access to their parent `cmd2` object, what is blocking your application from having a configuration structure at the top level which all `CommandSets` have access to? I'm worried a feature like this isn't generic enough to add to the cmd2 framework since applications vary greatly on how they load config settings. username_0: ![cmd2](https://user-images.githubusercontent.com/29623/110249124-8fad4b00-7f42-11eb-8f32-3a04e68ecd4e.jpg) Here's my external plugin implementation. Because I am creating the CommandSet instances and passing them into the cmd2 app constructor I lose the ability for each Commandset to simply call self.config. I fully admit that I could be thinking of this incorrectly or misunderstanding how inheritance works.. Am I doing this correctly? username_0: I could also be easily convinced that Configuration storage is NOT a framework issue but an implementation one. Perhaps I should just look more into how other applications are implementing persistent configurations. When I wrote CLI apps pre cmd2 I would first look to a default ini file to load variables, If they were not present or available I made the arguments mandatory. like.. if the ini had a 'port' variable that was set to 8080 then each command would automatically assume this is true. But I could override this variable with the cli argument --port. I'm trying to implement something similar in my cmd2 apps. I want to load a file if present at app startup that then is available to all my other commandsets. I'll write up a quick POC and share it. username_0: https://github.com/username_0/cmd2-app-poc-plugin Here is the POC.. https://github.com/username_0/cmd2-app-poc-plugin/blob/main/sample_plugin/src/sample_plugin/fruits.py This function does indeed show that a global config object is accessible via self._cmd def do_really_like_apples(self, _:Statement): apple_setting = self._cmd.config['apple'].getboolean('like') but this is NOT accessible via the __init__. Is this by design? username_2: @username_0 Yes, it is by design. `CommandSets` have a lifecycle and various events occur within this lifecycle. Refer to the [event handler](https://cmd2.readthedocs.io/en/stable/features/modular_commands.html#event-handlers) documentation for a description of each event. The `cmd2` object becomes available to the `CommandSet` upon registration (`on_register()`). Here is a link to the `CommandSet` [code](https://github.com/python-cmd2/cmd2/blob/master/cmd2/command_definition.py) as well so you can see what each event handler does. The class is at the bottom of that file. username_2: The main reason I advocate for a common config object being implementation-based is because `cmd2` itself has no common configuration settings that aren't already individual class members. Therefore `cmd2` would have this empty `self.config` member. Another question that arises is, what type of object is `self.config`? Your example makes it a dictionary. Other developers may prefer a `namedtuple` or `dataclass`. I would not want to enforce such a standard from the framework level. username_0: I can agree.. Perhaps it's just better to submit an advanced example. Enforcing a standard from the framework is indeed a "bad" thing but looking at all of the different cmd2 implementations there is a lot of rope available to hang one's self ;) Status: Issue closed
imixs/imixs-docker
481060865
Title: Can I use another postgreSQL version? Question: username_0: I am using postgres:11-alpine docker image in my docker-compose.yml file. But your Dockerfile is using postgres:9.6.1. When I run docker-compose up, it shows that: pg_dump: server version: 11.2; pg_dump version: 9.6.11 pg_dump: aborting because of server version mismatch So I must use an older version? Answers: username_1: no - you can run also newer versions of Postgress - Postgres:11 is fine username_1: note: the image just imports the JDBC Postgres Driver. This driver is based on version 9.x but it will also work fine with newer Postgress Server versions like 11. But of course you can upgrade the JDBC driver also by overwriting the image, or we can do this in this project. We just need to replace the jar file with a newer version. Status: Issue closed
yoheimuta/hubot-aws
76321702
Title: typo: hubot ec2 run --image-id=[ami-id] Question: username_0: `hubot ec2 run --image_id=[ami-id]` is right. - https://github.com/username_0/hubot-aws/blob/master/scripts/ec2/run.coffee#L17 Answers: username_0: Fixed in [Release v0.0.5 · username_0/hubot-aws](https://github.com/username_0/hubot-aws/releases/tag/v0.0.5) Status: Issue closed
telegraphic/hickle
745808539
Title: Dumping to io.BufferedReader Fails Question: username_0: I am trying to dump to in-memory bytes so that I can then compress these bytes with zlib before writing to disk. ```python raw = io.BytesIO() writer = io.BufferedWriter(raw) hickle.dump(obj, writer) ``` The last line raises `AttributeError: '_io.BytesIO' object has no attribute 'name'`. Answers: username_1: I think the error comes from these lines in file_opener function ``` if isinstance(f, (io.TextIOWrapper, io.BufferedWriter)): filename, mode = f.name, f.mode f.close() mode = mode.replace('b', '') h5f = h5.File(filename, mode) ``` instead of passing the file like object to h5py as is or at least specifiying driver as 'fileobj' ``` h5f = h5.File(f,mode,driver = 'fileobj') ``` it is tried to get the file name form the passed in TextIOWrapper or BufferedWrite (any other filelike objects are ignored) and its mode and than the file is closed and a new hdf5 file with specified file name and node is opened. Alternatively only an already opened h5py.File object or a plain file path string are accepted any thing else causes an exception. Possibly on finalization of next minor or major release somebody shall have a look to it h5py from 2.10 on definitely is capable of handling filelike objects as long as they provide read, seek, tell and write method. (see [h5py.File](https://docs.h5py.org/en/stable/high/file.html)) username_2: Oh, I guess I missed this one. I am not entirely sure why that does not work, as we do test for it, but I will have a look. However, as h5py was recently updated to 3.0.0, which brought a ton of changes (that are also incompatible with hickle), I am more planning on doing a pass over the entire package to account for that. I will add this issue to that list, but it may take a while before it will be fixed. username_1: @username_2 @username_3 just in case it might be of any interest to you or even any help at all i wanted to let you know that: Beeing a bit boored while waiting for @username_3 to decide upon pull request #138 tried to do some proof of concept for handling file and file like objects as supported by h5py. The results of this trial and error can also be found in the detached concept_memp_compact_expand branch of my hickle fork. username_1: The same here, would be included in my finalize and cleanup pullrequest after #138, and upcomming for #139 and #145. username_2: After looking into it, I realize that this is not an error. `hickle` can solely be used to dump to HDF5-files. A `BufferedWriter` is not an HDF5-file, so `hickle` cannot dump to it. Status: Issue closed username_0: Are there plans to support writing to in-memory bytes rather than files? username_2: Not at the moment, no. username_3: @username_0 just to chime in here: `hickle` is indeed designed specifically for dumping to HDF5 files, and uses `h5py` as its API -- which doesn't support `BufferedWriter`. If you really wanted a HDF5 file in memory, you could try setting up a ramdisk? However I think there are probably better solutions out there for in-memory data storage... username_1: @username_3 @username_2 not so true, according to documentation for h5py 2.10 and onward they support any file like object which is capable of reading and writing binary data and which is seekable and `io.BytesIO` exactly full fills that, one can find that example in h5py [manual](https://docs.h5py.org/en/2.9.0/high/file.html#python-file-like-objects) . Thus the questions is rather is it worth the efforts to add all the required checks if passed in file-like object conforms to requirements of h5py or not and in worst remove support for file-like objects and Python file handles from `hickle` at all. Only supporting native h5py files. @username_0 why would you need the `io,BufferedWriter`. `io.BytesIO` already a `io,BufferedIOBase` object like `io.BufferedWriter` and `io.BufferedReader` are and thus is already buffered. So replace `io.BufferedWriter` simply by `h5py.File` to make your example work. ``` raw = io.BytesIO writer = h5py.File(raw) hickle.dump(obj,writer,mode='w') ``` and on read ``` reader = h5py.File(raw) hickle.load(obj,reader,mode='r') ``` So you see no need for `io,BufferedWriter` at all or in other words `h5py.File` acts as wrapping writer and reader. username_2: @username_1 It is still true what we are saying. It does not matter if h5py allows writing to other filetypes, hickle does not support it. username_1: Yes you are right, bad wording from my side. What i wanted to say, is if it does not support it it should not allow to pass file objects and file-like objects at all. As how it is done now is broken and against expectations when passing file objects, with the consequence that this will not stay the only issue related to strange or broken support of file and file like objects.### ### An example: ``` fid = open('/tmp/somefile.h5','w+b') hickle.dump(obj,fid) somesocket.write(fid.read()) ``` But that does not work as `hickle` will just takes the filename and closes the original file or file-like object and replaces the underlying file on disk with a completely new file with the same name and the default access rights files owned by the process running `hickle` not the ones of the original file and also not necessarily with the same rights of the original file. This does not make sense at all to me . Why should i first open a file which is never used or even worse when reading for an already written hickle file the file is deleted. When i open a file beforehand i want `hickle` to place the hdf5 file content exactly in that file and nothing else and the wrapping inside 'io.BufferedWriter' or 'io.TextIOWrapper' should not make any difference here. So either and that is what i meant take the decision to properly support file and file like-objects eg from hickle >= 5.0 on and take the efforts to fix it until then or decide not to support file and file-like objects at all beyond indirect support through passed in h5py.File objects. Than remove support completely only allowing filename strings and h5py.File objects to be passed.
mampfes/hacs_waste_collection_schedule
1177695454
Title: Not Able to Pull data for Auckland Question: username_0: Hi, What a wonderful idea to automatically update it. Great concept. I think Documentation is lacking for new users. Probably would be ideal to have more step by step guides. I have got the following config in YAML, took ages to generate it without errors. My sensors show up but they are blank, no information is getting pulled. Any help will be much appreciated. waste_collection_schedule: sources: - name: aucklandcouncil_govt_nz args: area_number: 12342560766 customize: - type: brown bin show: no - type: Rubbish - type: Recycling sensor: # next collection - platform: waste_collection_schedule name: next_collection # next yard waste collection - platform: waste_collection_schedule name: next_yard_collection types: - Rubbish # next garbage collection - platform: waste_collection_schedule name: next_garbage_collection types: - Recycling Answers: username_1: Try (note that yaml is case sensitive and that indentation counts!): ```yaml waste_collection_schedule: sources: - name: aucklandcouncil_govt_nz args: area_number: 12342560766 customize: - type: brown bin show: no sensor: - platform: waste_collection_schedule name: next_collection - platform: waste_collection_schedule name: next_yard_collection types: - rubbish - platform: waste_collection_schedule name: next_garbage_collection types: - recycle ``` Could you explain the problems you had? This would be helpful to improve to documentation. Or even better: Provide a suggestion :-) username_0: Sorry my code was right the formatting got stuffed when i pasted here. So its exactly same as you have mentioned above, configurations check out. But when i insert the sensor in a button , it shows as Next Garbage collection (Unknown) it means its not picking up the values from the website, how do i test or confirm that. I will create a video for step by step to help you out. instead of suggestions. username_1: I had to adjust `rubbish `and `recycle` to get it working! Please check if you use exactly the same values. username_0: Hi, Thanks for the prompt reply, I used it exactly as mentioned above but when i put the sensor to home screen it has unknown value. How to Troubleshoot ? how do i confirm if its picked up the value from that site ? and what variables or names its using. Below is a copy paste from my Configuration.Yaml and attached is my button ```` waste_collection_schedule: sources: - name: aucklandcouncil_govt_nz args: area_number: 12342560766 customize: - type: brown bin show: no sensor: - platform: waste_collection_schedule name: next_collection - platform: waste_collection_schedule name: next_yard_collection types: - rubbish - platform: waste_collection_schedule name: next_garbage_collection types: - recycle ```` ![image](https://user-images.githubusercontent.com/74592813/159843979-c84e25bf-c31a-4a7f-8192-352de67acd06.png) username_1: Ok, strange. Here is a small step-by-step instruction: 1. Check that there are no errors in the HA logs. 2. Check that the dates appear in the HA calendar. It should look like this with your configuration: ![image](https://user-images.githubusercontent.com/375950/159855180-1c036c52-b720-49de-b146-f25520067fdd.png) 3. Open the HA Developer Tools and check the state of the 3 sensors. They should look like this: ![image](https://user-images.githubusercontent.com/375950/159855345-2fab0364-b977-426b-9ce9-a0f1a7f2b6ed.png) If I use the configuration above, the entity cards look like this: ![image](https://user-images.githubusercontent.com/375950/159855691-157bc627-37a8-4f7d-9d15-63d473d3e3bf.png) username_0: Excellent Help, At least now i know where to go and find the info. so I have it half working. Its only pulling the rubbish and rubbish, recycle both under same sensor. how do i fix it or do i just leave that as is. ![image](https://user-images.githubusercontent.com/74592813/159857591-2a3b8eff-a31c-473a-b2ed-483a167cfc19.png) username_0: All working now, guess took a while to pull that through. We can close this now please. Thanks Status: Issue closed username_2: I wish I'd seen this sooner @username_0 as I could have helped too. Here is what I am using myself: In configuration.yaml: ``` waste_collection_schedule: sources: - name: aucklandcouncil_govt_nz args: area_number: 12343981102 # from https://www.aucklandcouncil.govt.nz/rubbish-recycling/rubbish-recycling-collections/Pages/rubbish-recycling-collection-days.aspx customize: - type: rubbish alias: Rubbish icon: mdi:trash-can - type: recycle alias: Recycling icon: mdi:recycle calendar_title: Auckland Council Rubbish & Recycling Collections day_switch_time: "17:00" ``` In my sensors file (I have sensors in a separate .yaml file, but you could have these in configuration.yaml under `sensors:`) They are all slight tweaks on the same data so I can display it in different formats for different purposes - the "this week" set is what I use the most. ``` - platform: waste_collection_schedule name: Next Rubbish Collection Date types: - Rubbish value_template: '{{value.date.strftime("%d %b")}}' - platform: waste_collection_schedule name: Next Recycling Collection Date types: - Recycling value_template: '{{value.date.strftime("%d %b")}}' - platform: waste_collection_schedule name: Next Rubbish Collection Days types: - Rubbish value_template: '{% if value.daysTo == 0 %}Today{% elif value.daysTo == 1 %}Tomorrow{% else %}in {{value.daysTo}} days{% endif %}' - platform: waste_collection_schedule name: Next Recycling Collection Days types: - Recycling value_template: '{% if value.daysTo == 0 %}Today{% elif value.daysTo == 1 %}Tomorrow{% else %}in {{value.daysTo}} days{% endif %}' - platform: waste_collection_schedule name: Next Rubbish Collection This Week types: - Rubbish value_template: '{% if value.daysTo == 0 %}Today{% elif value.daysTo == 1 %}Tomorrow{% elif value.daysTo <= 6 %}This {{value.date.strftime("%A")}}{% else %}in {{value.daysTo}} days{% endif %}' - platform: waste_collection_schedule name: Next Recycling Collection This Week types: - Recycling value_template: '{% if value.daysTo == 0 %}Today{% elif value.daysTo == 1 %}Tomorrow{% elif value.daysTo <= 6 %}This {{value.date.strftime("%A")}}{% elif value.daysTo < 13 %}Next {{value.date.strftime("%A")}}{% else %}in {{value.daysTo}} days{% endif %}' ``` Glad you have it working now!
Apicurio/apicurito
984949028
Title: Operator projects using the removed APIs in k8s 1.22 requires changes. Question: username_0: For further information and tips see the [comment](https://github.com/redhat-openshift-ecosystem/community-operators-prod/discussions/138#discussioncomment-1245759). Answers: username_1: The code for the apicurito operator has moved to Apicurio/apicurio-operators and has been updated to remove all deprecated apis. Please see [here](https://github.com/Apicurio/apicurio-operators/commit/8ffd67465a1fb1c20c249a705f9131207bffa2cf). username_0: So, can you please publish it on the community repo https://github.com/redhat-openshift-ecosystem/community-operators-prod/tree/main/operators/apicurito ? username_1: New version of apicurito underway: https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/216 Status: Issue closed
power-assert-js/espower-typescript
435535316
Title: "directories" property in package.json conflicting with other npm package Question: username_0: Currently, "directories" property is used to specify the arbitrary test directory. [electron-builder](https://github.com/electron-userland/electron-builder) used the same property in the past and it is deprecated now. When I try to use electron-builder with"directories" property specified in package.json, electron-builder asserts that "directories" property is deprecated and aborts. The problem is that "directories" is overly general. If it is specified like this, it will never conflict with other packages. ``` { "name": "your-module", "description": "Your module", "version": "0.0.1", "espower-typescript": { "directories": { "test": "spec/" }, } ... } ``` Is it possible?
hankcs/HanLP
712912638
Title: Which word segmentation is faster in python or java Question: username_0: <!-- Thank you for suggesting an idea to make HanLP better. Please fill in the template below to bypass our spam filter. 以下必填,否则直接关闭。 --> **Describe the feature and the current behavior/state.** I want to try HanLP in spark in scala(in scala, i can directly use HanLP Java version 1.x) or pyspark(HanLP 2.x), the dataset is billion level, so i want to know which one is better **Will this change the current api? How?** No **Who will benefit with this feature?** No **Are you willing to contribute it (Yes/No):** No **System information** - OS Platform and Distribution : Linux Ubuntu 16.04 - Python version: 3.7 - HanLP version: 2.x in Python, 1.x in Java - Scala version: 2.11.x - Java version: 1.8 **Any other info** thank you for your support, i tried to find the post button in its bbs but couldn't find it. BTW, hava a happy national day! * [ ] I've carefully completed this form. Answers: username_1: [auto-reply] Thanks for your comment. However, the essential information is required. Please carefully fill out the form when open a new issue. Status: Issue closed
nicothin/NTH-start-project
437913288
Title: Что уже должно быть установлено? Question: username_0: Привет. Прошу помочь разобраться. Я пытался установить зависимости на своем ноуте. Выдает ошибки что не хватает компонентов, не достаточно прав. Система Убунту, устанавливал через судо. Я новичок ) Answers: username_1: Зависимости ставятся без sudo. Должен быть установлен nodejs. Без скрина ошибки точно ничего нельзя сказать об ошибке. username_0: Приложил скриншот. Пытался сделать как пишут здесь - https://docs.npmjs.com/resolving-eacces-permissions-errors-when-installing-packages-globally Не помогло ( username_1: нет. username_0: С празником Пасхи! Пробую еще раз, спасибо за терпение ) username_0: Проблема решена установкой nvm. Спасибо. Чайники они такие чайники ) вс, 28 апр. 2019 г. в 07:12, <NAME> <<EMAIL>>: > С празником Пасхи! Пробую еще раз, спасибо за терпение ) > Status: Issue closed
opencv/opencv
201097103
Title: CascadeClassifier::detectMultiScale levelWeights & rejectLevels Question: username_0: ##### System information (version) - OpenCV => 3.2.0 - Operating System / Platform => Ubuntu 14.04 - Compiler => gcc 4.8.4 ##### Detailed description I'm having a problem trying to implement a code that determines if an image is a face. I created a minimal [repo](https://github.com/username_0/cascade-weight-test/tree/master) to test this out. There are six face, from a good relatively fontal face to a decentered face. The output I get is strange: ``` Detection ./cropped_face1.png [109 x 109 from (2, 14)] with weight 108.512 and level: 22 and size: 1 Detection ./cropped_face2.png [111 x 111 from (9, 9)] with weight 108.411 and level: 22 and size: 1 Detection ./cropped_face3.png [109 x 109 from (2, 14)] with weight 108.512 and level: 22 and size: 1 ``` The last 3 faces do not even get a single output, even if face#4 is reasonably nice: Face#3:![cropped_face3](https://cloud.githubusercontent.com/assets/17952091/21995309/2e7e0d12-dbf2-11e6-8768-2b7305d83a2f.png) Face#4:![cropped_face4](https://cloud.githubusercontent.com/assets/17952091/21995009/9b530d22-dbf0-11e6-88f4-591d9b1de1f6.png) The level always seems to be 22 (on my very little sample, sorry). Stage 22 seems to be the final stage of the classifier based on the xml file of the classifier. I am also unclear on to what the weights actually represent with such strange number. I didn't find anything in the documentation to help me. I started going through the source code, but I didn't find anything to help me yet. I do not quite understand why face#3 gets to the final stage of the classifier, with ~108 of weight, and the face#4 has no output, it didn't even get to stage 1? Am I doing something wrong? Is something wrong with the function? Can there be an explanation of the usage and what the weight is? ##### Steps to reproduce See [repo](https://github.com/username_0/cascade-weight-test/tree/master) Answers: username_1: You should get yourself a copy of OpenCV 3 Blueprints, Chapter 5, which discusses all these parameters in the finest details. That being said, the weights are the decision level of your final stage classifier. IN GENERAL It seems that you are using a detector for classification, which it is not designed for. My guess, your faces are to close to the border, which makes the detector fail. Again this is designed to locate face areas in an image with background information. username_0: So, basically this overloaded function doesn't have much use now? Ok, it makes sense that the faces are too close to the border and that's why they aren't classified or detected. Do you have a recommendation on how I can achieve a classification score from a detector or OpenCV? username_2: This is what exactly the docs in [opencv](https://github.com/opencv/opencv). even the face recognizers use this method :smiley: :smile_cat: :small_blue_diamond: username_3: `cv::CascadeClassifier` is almost deprecated. i think this issue can be closed down. username_1: I am sorry @username_3 but I do not agree that cascade classifiers are deprecated. Yes DL detectors work as smooth and fast and are more acurate, but these still have their applications in embedded setups :) username_4: Citation needed. Status: Issue closed
int-brain-lab/ibllib
411410449
Title: Ambient sensor data access Question: username_0: They are on the Flatiron: ``` sessions, details = one.search(dataset_types='_iblrig_ambientSensorData.raw', lab='churchlandlab') asd = one.load(sessions[0], dataset_types='_iblrig_ambientSensorData.raw') ``` However @username_2 , is there a point in storing all those values without a timestamp (albeit imprecise) ? Status: Issue closed Answers: username_1: The commands above return `asd = None` even though the jsonable loads onto disk fine username_0: They are on the Flatiron: ``` sessions, details = one.search(dataset_types='_iblrig_ambientSensorData.raw', lab='churchlandlab') asd = one.load(sessions[0], dataset_types='_iblrig_ambientSensorData.raw') ``` There is one temperature/pressure value per trial. username_0: Hmm this is not reproducible here. Possible explanations: - this happens to most recent datasets: the datasets are registered, but do not exist yet on flatIron and `none` is the expected behaviour - it happens to all datasets: fb757883278f770b9145fbaa4b3facc20a3d7974 was implemented 19 days ago on master. You should be up to date but still worth checking Otherwise, please send a specific session that has this behaviour, and I"ll see if I can reproduce. Which OS ? username_2: the code above is missing a `..., details=True)` username_1: Updating to the latest develop branch fixed the `None` problem. Could importing `one.ibllib` throw a warning when the user is on an old version? Also, the output returned is a cumbersome list of dicts of lists (I think?): ``` 'RelativeHumidity': [32.5419921875]}, {'Temperature_C': [26.780000686645508], 'AirPressure_mb': [1011.332763671875], 'RelativeHumidity': [32.552734375]}, {'Temperature_C': [26.770000457763672], 'AirPressure_mb': [1011.30322265625], 'RelativeHumidity': [32.5634765625]}, {'Temperature_C': [26.790000915527344], 'AirPressure_mb': [1011.3099365234375], 'RelativeHumidity': [32.564453125]}, {'Temperature_C': [26.780000686645508], 'AirPressure_mb': [1011.3465576171875], 'RelativeHumidity': [32.564453125]}]] ``` So that `ambient = pd.DataFrame(asd[0])` returns ``` AirPressure_mb RelativeHumidity Temperature_C 0 [1010.5851440429688] [33.04296875] [26.219999313354492] 1 [1010.7210083007812] [33.3408203125] [25.989999771118164] 2 [1010.6853637695312] [33.3408203125] [25.979999542236328] ``` Could the dict values be numeric, rather than wrapped in a list? And what datasetType should I grab to get the timing for each of these (`stimulus_onTime`?) Status: Issue closed
pyrocms/pyrocms
242510973
Title: [files-module] Issue with Installer when installing PyroCMS Question: username_0: This is a very specific issue. The files-module throws an exception when trying to run the "install" command from a controller or a queue job to generate a new Application and the Application is configured to use a different database that the one used by the main site. Answers: username_0: https://github.com/anomalylabs/files-module/pull/47
poooi/poi
438139989
Title: 接7个任务后下方出现未解锁 Question: username_0: 接6个任务的时候没有,接了第7个就出现这个未解锁了 ![001](https://user-images.githubusercontent.com/21211204/56875507-130ed300-6a74-11e9-9929-ddf834d475f1.jpg) <!-- Love poi? Please consider supporting our collective: 👉 https://opencollective.com/poi/donate --> ![002](https://user-images.githubusercontent.com/21211204/56875473-d216be80-6a73-11e9-9cbb-ce4d5f318104.jpg) Answers: username_1: 已修复等发布 Status: Issue closed username_2: Poi 10.2.4, Win10 x64 17763仍存在该问题。
smorgo/3DDesigns
417891523
Title: Bayonet_Locks library generates locking bump incorrectly Question: username_0: Having changed the way in which the locking slots were generated from Alexandre's original, I haven't yet worked out the best way to generate the small 'bump' (for want of a better word) that holds the locking pin in position.
nus-cs2030/2021-s2
860725261
Title: Multiple classes/interfaces/enums found in a class Question: username_0: Hello! I only declared one enum in the Event class and did not declare any other interfaces/classes in my Event.java file, so does this mean I have to declare my enum separately? Did anyone face this issue before, and how did y'all go about solving it? Thanks! <img width="585" alt="Screenshot 2021-04-19 at 1 57 01 AM" src="https://user-images.githubusercontent.com/65957285/115155433-8a4f2e80-a0b2-11eb-8ab4-f8aba08204e5.png"><issue_closed> Status: Issue closed
aws-amplify/amplify-js
744539966
Title: unauthorized users can write to local datastore Question: username_0: **Describe the bug** unauthorized users in the schema.graphql file can still write data to the local datastore. A warning message is displayed in the console, but data still resides in the local datastore. Essentially, data is written to the local datastore, but then fails to sync with DynamoDB. This leads to data inconsistencies between the front and back ends. Obviously I could write logic in the application to prevent this occurring, but the graphql auth permissions should prevent this on both front and backend. Warning message received: ``` {localModel: Model, message: "Not Authorized to access createAsset on type Asset", operation: "Create", errorType: "Unauthorized", errorInfo: null, …} errorInfo: null errorType: "Unauthorized" localModel: Model {ModelData...} message: "Not Authorized to access createAsset on type Asset" operation: "Create" remoteModel: null } ``` **To Reproduce** Ensure cloud sync is configured as per documentation: https://docs.amplify.aws/lib/datastore/sync/q/platform/js **schema.graphql file :** ``` type Asset @model @auth( rules: [ { allow: groups, groups: ["Admins"], queries: [get, list], mutations: [create, update, delete] } { allow: groups, groups: ["Users"], queries: [get, list], mutations: null } ] ) { id: ID! name: String! } ``` **Perform a write to the datastore:** ``` DataStore.save( new Asset({ name: 'assetName', }) ); ``` **Expected behavior** If data is not permitted to be written to dynamoDB, it shouldn't be allowed to be written to the local datastore. Amplify version: aws-amplify: ^3.3.8 => 3.3.8 aws-amplify-angular: ^5.0.37 => 5.0.37 Note: Data sync works fine when users are permitted via the schema.graphql file. Answers: username_1: Hey @username_0, currently, DataStore does not enforce auth rules clientside. This is by design, as we want the AppSync back end to serve as the only source of truth for authorization rules. We recommend that customers add their own business logic around this in their applications. Also, just want to point out that the `queries` and `mutations` arguments [are deprecated](https://docs.amplify.aws/cli/graphql-transformer/auth#definition) and are not supported with DataStore. When establishing subscriptions for the Sync Engine, DataStore will ignore those arguments. I suggest using `operations` instead: ```graphql type Asset @model @auth( rules: [ { allow: groups, groups: ["Admins"] operations: [create, read, update, delete] } { allow: groups, groups: ["Users"], operations: [read] } ] ) { id: ID! name: String! } ``` username_0: I thought that might be the case, thanks @username_1 Status: Issue closed
glidejs/glide
946201715
Title: Automatic value "Perview"? Question: username_0: I am using multiple sliders on one page. Each slider has a different number of slides. How do I set the "perview" value to auto? Or how to set a separate "Perview" for each slider? Answers: username_1: You need to initialize each instance of glide with different settings. There are no `auto` option. Status: Issue closed
colin-dumitru/F.js
53129037
Title: Add support for event streams. Question: username_0: Implemented with the 0.3 release. The event stream now follows the Promise api. So when a value is pushed into a stream, a consumer is notified. This way, we can implement a pull-push mechanism for iterable chains. Status: Issue closed Answers: username_0: Implemented with the 0.3 release. The event stream now follows the Promise api. So when a value is pushed into a stream, a consumer is notified. This way, we can implement a pull-push mechanism for iterable chains. Status: Issue closed
wesselb/stheno
1127458489
Title: Documentation about Multi-Output Regression Question: username_0: Hi @username_1, I am trying to use your example of Multi-Output Regression with some data I have. I don't understand how to correctly give them to the VGP and them make a prediction. My data as input x_obs are not the same, so it's not exactly as the example. I have nine x observation as [x1,x2,x3,x4,x5,x6,x7,x8,x9] with their y observation as [y1,y2,y3,y4,y5,y6,y7,y8,y9]. Also, with your example provided, is it possible to optimize some hyperparameters if we had some in the VGP ? Here are my code I was trying to use, with 3 different outputs to simulate data. Thank you in advance for your help. ``` import matplotlib.pyplot as plt from wbml.plot import tweak from stheno import B, Measure, GP, EQ, Delta, Matern52 class VGP: """A vector-valued GP.""" def __init__(self, ps): self.ps = ps def __add__(self, other): return VGP([f + g for f, g in zip(self.ps, other.ps)]) def lmatmul(self, A): m, n = A.shape ps = [0 for _ in range(m)] for i in range(m): for j in range(n): ps[i] += A[i, j] * self.ps[j] return VGP(ps) # Define points to predict at. x = B.linspace(0, 10, 5) # Create some sample data. x1 = np.atleast_2d(np.linspace(0, 10, 5)).T x2 = np.atleast_2d(np.linspace(0, 9, 5)).T x3 = np.atleast_2d(np.linspace(0, 7, 5)).T y1 = np.atleast_2d(np.linspace(0, 10, 5)).T y2 = np.atleast_2d(np.linspace(0, 10, 5)).T y3 = np.atleast_2d(np.linspace(0, 10, 5)).T x_obs = [x1,x2,x3] y_obs = [y1,y2,y3] # Model parameters: m = 3 p = 3 H = B.randn(p, m) with Measure() as prior: # Construct latent functions. us = VGP([GP(Matern52()) for _ in range(m)]) # Construct multi-output prior. fs = us.lmatmul(H) # Construct noise. e = VGP([GP(0 * Delta()) for _ in range(p)]) # Construct observation model. ys = e + fs # Sample a true, underlying function and observations. [Truncated] post = prior.condition(*((p(x_obs), y_obs) for p, y_obs, x_obs in zip(ys.ps, ys_obs, x_obs))) preds = [post(p(x)) for p in fs.ps] # Plot results. def plot_prediction(x, f, pred, x_obs=None, y_obs=None): plt.plot(x, f, label="True", style="test") if x_obs is not None: plt.scatter(x_obs, y_obs, label="Observations", style="train", s=20) mean, lower, upper = pred.marginal_credible_bounds() plt.plot(x, mean, label="Prediction", style="pred") plt.fill_between(x, lower, upper, style="pred") tweak() plt.figure(figsize=(10, 6)) for i in range(3): plt.subplot(3, 1, i + 1) plt.title(f"Output {i + 1}") plot_prediction(x, fs_true[i], preds[i], x_obs, ys_obs[i]) plt.show() ``` Answers: username_1: Hi @username_0! I've put together a simple MOGP model (not using the example) which might better suit your use case. The script uses JAX to learn hyperparameters. (You can also use another AD framework if you like.) ```python from stheno.jax import GP, Matern52, Measure from varz.jax import Vars, minimise_l_bfgs_b from wbml.plot import tweak import matplotlib.pyplot as plt import jax.numpy as jnp import numpy as np x1 = np.linspace(0, 10, 30) x2 = np.linspace(0, 9, 40) x3 = np.linspace(0, 7, 50) # Generate some test data. f = GP(Matern52()) y1 = f(x1, 0.2).sample().flatten() y2 = f(x2, 0.2).sample().flatten() y3 = f(x3, 0.2).sample().flatten() p = 3 # Number of outputs m = 3 # Number of latent processes def model(vs): ps = vs.struct with Measure() as prior: # Create independent processes with learnable length scales initialised to `1`. us = [ GP(Matern52().stretch(ps_u.scale.positive(1))) for ps_u, _ in zip(ps.us, range(p)) ] # Mix processes together to induce correlations between the outputs. H = ps.mixing_matrix.unbounded(shape=(p, m)) fs = [0 for _ in range(p)] for i in range(p): for j in range(m): fs[i] = fs[i] + H[i, j] * us[j] # Create learnable observation noises initialised to `0.1` noises = ps.noises.positive(0.1, shape=(p,)) return prior, fs, noises def objective(vs): prior, fs, noises = model(vs) return -prior.logpdf( (fs[0](x1, noises[0]), y1), (fs[1](x2, noises[1]), y2), (fs[2](x3, noises[2]), y3), ) # Perform learning. [Truncated] # Plot results. plt.figure(figsize=(10, 6)) x_to_plot = np.linspace(0, 10, 200) plt.subplot(3, 1, 1) plt.title("Output 1") plot_posterior(x_to_plot, f1_post, x1, y1) plt.subplot(3, 1, 2) plt.title("Output 2") plot_posterior(x_to_plot, f2_post, x2, y2) plt.subplot(3, 1, 3) plt.title("Output 3") plot_posterior(x_to_plot, f3_post, x3, y3) plt.show() ``` The script produces the following plot: ![Output](https://i.ibb.co/MNQpD8p/figure.png) Let me know if this suits your needs. :) username_0: Thank you very much for your code example. It is running on my side too, and I have the same results by using my CPU. Because the computational time is high for nine inputs by using a CPU, I wanted to use my GPU to see if it will be faster. I followed the steps to use CUDA with the Jax library and was able to link both of them. However, by using the same code as you give me, I obtained this time an error: ``` --------------------------------------------------------------------------- RuntimeError Traceback (most recent call last) <ipython-input-4-45889ae4f67a> in <module> 55 # Perform learning. 56 vs = Vars(jnp.float64) ---> 57 minimise_l_bfgs_b(objective, vs, trace=True, jit=True) 58 vs.print() # Display learned parameters. 59 ~/python-env/lib/python3.6/site-packages/varz/minimise.py in minimise_l_bfgs_b(f, vs, f_calls, iters, trace, names, jit) 77 trace=trace, 78 names=names, ---> 79 jit=jit, 80 ) 81 ~/python-env/lib/python3.6/site-packages/varz/minimise.py in _minimise_l_bfgs_b(f, vs, f_calls, iters, trace, names, jit) 154 # Run function once to ensure that all variables are initialised and 155 # available. --> 156 res = convert(f(vs, *args), tuple) 157 val_init, args = res[0], res[1:] 158 <ipython-input-4-45889ae4f67a> in objective(vs) 49 (fs[0](x1, noises[0]), y1), 50 (fs[1](x2, noises[1]), y2), ---> 51 (fs[2](x3, noises[2]), y3), 52 ) 53 ~/python-env/lib/python3.6/site-packages/plum/function.cpython-36m-x86_64-linux-gnu.so in plum.function._BoundFunction.__call__() ~/python-env/lib/python3.6/site-packages/plum/function.cpython-36m-x86_64-linux-gnu.so in plum.function.Function.__call__() ~/python-env/lib/python3.6/site-packages/stheno/model/measure.py in logpdf(self, *pairs) 461 """ 462 fdd, y = combine(*pairs) --> 463 return self(fdd).logpdf(y) 464 465 @_dispatch ~/python-env/lib/python3.6/site-packages/stheno/random.py in logpdf(self, x) 210 B.logdet(self.var)[..., None] # Correctly line up with `iqf_diag`. 211 + B.cast(self.dtype, self.dim) * B.cast(self.dtype, B.log_2_pi) --> 212 + B.iqf_diag(self.var, B.subtract(x, self.mean)) 213 ) 214 / 2 ~/python-env/lib/python3.6/site-packages/plum/function.cpython-36m-x86_64-linux-gnu.so in plum.function.Function.__call__() ~/python-env/lib/python3.6/site-packages/matrix/ops/iqf_diag.py in iqf_diag(a, b) 33 @B.dispatch 34 def iqf_diag(a, b): ---> 35 return iqf_diag(a, b, b) 36 [Truncated] 608 ~/python-env/lib/python3.6/site-packages/jax/interpreters/xla.py in _xla_call_impl(***failed resolving arguments***) 593 *unsafe_map(arg_spec, args)) 594 try: --> 595 return compiled_fun(*args) 596 except FloatingPointError: 597 assert config.jax_debug_nans or config.jax_debug_infs # compiled_fun can only raise in this case ~/python-env/lib/python3.6/site-packages/jax/interpreters/xla.py in _execute_compiled(compiled, avals, handlers, kept_var_idx, *args) 891 for i, x in enumerate(args) 892 if x is not token and i in kept_var_idx)) --> 893 out_bufs = compiled.execute(input_bufs) 894 check_special(xla_call_p.name, out_bufs) 895 return [handler(*bs) for handler, bs in zip(handlers, _partition_outputs(avals, out_bufs))] RuntimeError: Internal: Unable to launch triangular solve for thunk 0x2c46c570 ``` Do I need to add something to the code to make it work with a GPU ? username_1: Ouch! That doesn't look good. Could you confirm that running other JAX code on the GPU works fine? If that's the case, I can look into this more closely to see what's going on. username_0: I tried some examples of JAX code with my GPU (like these one: https://jax.readthedocs.io/en/latest/notebooks/quickstart.html) and it was working. I think the issue come from the library. I have jax-0.2.17 and jaxlib-0.1.65+cuda110 install on my computer
rebekahsong/bb_swe
855468677
Title: Add Ratings to Songs Question: username_0: Edit Django Model and Views to include ratings as an attribute of songs. Could potentially write some backend code that calculates the rating as an average of all ratings on a song. Edit React components to include ratings in the request and the display of the responses for songs.
exastro-suite/it-automation
1044465971
Title: 【エクスポート/インポート】メニューインポートのREST APIのXcommandがUPLOADとEXECUTEの時にメニューIDの形式が異なっている Question: username_0: メニューインポートのREST APIのXcommandがUPLOADの場合はメニューIDが「0000000001」と左0埋めで返却されるが、 XcommandがEXECUTEの場合は「1」という形で左0埋めを消して実行しないと以下のエラーになる。 メニューIDの設定値が不正です。(正規表現(/^0$|^-?[1-9][0-9]*$/s)に一致するデータを入力してください 。 EXECUTE時は「0000000001」でも「1」でもどちらの指定方法でも正常となるように修正する。<issue_closed> Status: Issue closed
aplpy/aplpy
512550
Title: Offset tick position option Question: username_0: Add the option to have the axes labeled in offset coordinates relative to a chosen point in the field of view. E.g. with ticks at [ -1, -0.5, 0, 0.5, 1] arcseconds for the 2.4" FOV. Example code: f = APYpy.FITSFigure('some_tiny_AO_image.fits') f.show_greyscale(stretch='log') f.set_tick_relative(True) f.set_tick_relative_to( (200,150) ) <NAME> suggested the idea. Thanks! Answers: username_1: A very late bump - @jpinedaf asked for this, especially for interferometer images. username_0: What we need to do is implement this in WCSAxes and then we'll get it for free once I merge in the changes to use WCSAxes internally. username_1: Is there an open issue in wcsaxes for this that should be linked? username_0: Here it is: https://github.com/username_0/wcsaxes/issues/8 username_2: Is there any chance of implementing this soon? I would really like to use it to simplify a multi-panel figure in an ApJL that I am about to submit. It would be great to have it in time for the second submission. username_2: For others who arrive here, below is some code to do it in a kludged way: def ra2deg(s): token = s.split(':') deg = 15*(float(token[0])+float(token[1])/60.+float(token[2])/3600.) return(deg) def dec2deg(s): token = s.split(':') deg = (abs(float(token[0]))+float(token[1])/60.+float(token[2])/3600.) if (token[0].find('-') >= 0): deg = -deg return(deg) target_ra=ra2deg('17:20:53.32') target_dec=dec2deg('-35:47:00') image_radius_arcsec=4.5 image_radius=image_radius_arcsec/3600. if True: # Try to draw relative tick marks length = 0.20/3600. cosdec = cos(radians(target_dec)) xleft = target_ra + image_radius/cosdec xright = target_ra - image_radius/cosdec ytop = target_dec + image_radius ybottom = target_dec - image_radius # x-axis ticks for x in np.arange(-np.ceil(image_radius_arcsec),image_radius_arcsec): xt = target_ra + x/3600./cosdec if int(abs(np.round(x))) % 4 == 0: lengthFactor = 1.6 else: lengthFactor = 1.0 iline = np.array([[xt,xt],[ybottom,ybottom+length*lengthFactor]]) f.show_lines([iline], color='w') iline = np.array([[xt,xt],[ytop,ytop-length*lengthFactor]]) f.show_lines([iline], color='w') # y-axis ticks for y in np.arange(-np.ceil(image_radius_arcsec),image_radius_arcsec): if int(abs(np.round(y))) % 4 == 0: lengthFactor = 1.6 else: lengthFactor = 1.0 yt = target_dec + y/3600. iline = np.array([[xleft,xleft-length*lengthFactor], [yt,yt]]) f.show_lines([iline], color='w') iline = np.array([[xright,xright+length*lengthFactor], [yt,yt]]) f.show_lines([iline], color='w') username_0: @username_2 - sorry for the delay, in the mean time the easiest way to do this with minimal hacking is to use this function: ```python from astropy.wcs import WCS from astropy.wcs.utils import skycoord_to_pixel, proj_plane_pixel_scales def linear_offset_coords(wcs, center): """ Returns a locally linear offset coordinate system. Given a 2-d celestial WCS object and a central coordinate, return a WCS that describes an 'offset' coordinate system, assuming that the coordinates are locally linear (that is, the grid lines of this offset coordinate system are always aligned with the pixel coordinates, and distortions from spherical projections and distortion terms are not taken into account) Parameters ---------- wcs : `~astropy.wcs.WCS` The original WCS, which should be a 2-d celestial WCS center : `~astropy.coordinates.SkyCoord` The coordinates on which the offset coordinate system should be centered. """ # Convert center to pixel coordinates xp, yp = skycoord_to_pixel(center, wcs) # Set up new WCS new_wcs = WCS(naxis=2) new_wcs.wcs.crpix = xp + 1, yp + 1 new_wcs.wcs.crval = 0., 0. new_wcs.wcs.cdelt = proj_plane_pixel_scales(wcs)*3600. new_wcs.wcs.ctype = 'XOFFSET', 'YOFFSET' new_wcs.wcs.cunit = 'arcsec', 'arcsec' return new_wcs ``` then you can do something like: ```python import aplpy from astropy.io import fits from astropy.wcs import WCS from offset import linear_offset_coords from astropy import units as u from astropy.coordinates import SkyCoord galcen = SkyCoord(0 * u.deg, 0 * u.deg, frame='galactic') hdu = fits.open('2MASS_k.fits')[0] wcs = WCS(hdu.header) wcs_offset = linear_offset_coords(wcs, galcen) hdu_linear = hdu.copy() hdu_linear.header = wcs_offset.to_header() fig = aplpy.FITSFigure(hdu_linear) fig.show_colorscale() ``` username_3: So I've been using this "linear_offset_coords" function with some success to make plots, but have run into the issue that using this method to change the axes labels makes it difficult over-plot contours from data with different wcs information. I've been using reproject.reproject_interp to put the contour data into the same pixels as the image data but this involves seriously down-sampling the contours which I'd like to avoid. Any pointers on ways to do this? I'm using wcsaxes right now. username_4: Just posting to receive updates on this. I am attempting to use the linear_offset_coords function. It (mostly) seems to work well for simple image display, but unfortunately when I try to overlay the beam via add_beam() (they're interferometric images) or even show_ellipses(), I get the error WCS is not celestial, cannot determine celestial pixel scale. @username_3, can you describe how you are trying to do your overlay? I wonder if I can adapt your method for what I'm trying to do. username_3: @username_4 For the overlay I've been using reproject.reproject_interp (http://reproject.readthedocs.io/en/stable/api/reproject.reproject_interp.html) which pretty much just rotates and rebins my contour data to the same pixel scale as the image then I can just plot them over each-other without using the wcs transformation. I've been mostly using wcsaxes not aplpy though, not 100% certain how that changes things. Using wcsaxes at least, if you can get the dimensions of the beam in pixel units you should just be able to over-plot it using the Ellipse patch in matplotlib (http://matplotlib.org/examples/pylab_examples/ellipse_demo.html). username_5: I would LOVE to see this feature as soon as possible! Tried the "minimal hacking" approach suggested by @username_0 , but I get the error "ValueError: 'cdelt' array is the wrong shape, must be 2" I'm really not a pro, but guess this is because I'm working with interferometric data and there is a spectra axis as well? What I'm trying to make is a velocity map plot, so I need my spectral axis but I still would like to show relative coordinates instead of world coordinates to show how big the structure in the images are. username_6: would be awesome to see this implemented!
JuliaLang/julia
530245383
Title: Extend at-specialize to force specialization of arguments Question: username_0: A trick currently used in `Base` (and probably elsewhere) to force specialization on functions is to artificially declare a type parameter, e.g. `foo(f::F) where F = f(1)`. I would bet that the majority of Julia users don't know about this and would look at this piece of code without any idea about the role of `F`, and would even fix what appears as a naive use of type parameters by simplifying it to `foo(f) = f(1)`. What about adding some explicit means to this end, for example reusing `@specialize`: `foo(@specialize(f)) = f(1)` ? xref: https://github.com/JuliaLang/julia/pull/33917#discussion_r349864113 Answers: username_1: Since in `foo(f::F) where F = f(1)` the function `f` is called, it would be equivalent to `foo(f) = f(1)`, right? username_0: I'm not the right person to answer, but this kind of prove the point that the exact rules are not obvious to everyone (talking of me!) username_1: It is worth mentioning that #32817 added documentation on when a `Function` argument is specialized. After thinking more about this, my opinion shifted somewhat: Having `@specialize` act in this way would not really clear up the confusion. The current syntax `foo(f::F) where F` is hacky, but so is overriding compiler heuristics. :) username_2: But I do think the intent of `@specialize(f)` is much more transparent than `f::F ... where F`. For a person with just enough knowledge of Julia to be dangerous :smiling_imp:, the latter might induce him/her to "clean up" and remove that little annotation. username_0: Nice, I was wondering whether it was documented somewhere, but got too lazy to check. Otherwise I agree with Tim, the current way to force specialization is too easy to miss out, and is brittle. Overrriding comiler heuristics might be hacky, but we all love to do it :D And `@inline` gets its own explicit name. If you think `@specialize` is not the right word to clear up the confusion, would another name do it? `@force_specialize` ? username_3: I agree we need a way to do this. In the last year I've noticed people start to use the static parameter trick more often, and each time I have to ask the question "Is `foo(x::T) where {T}` an attempt to force specialization, or an ugly way to get at `typeof(x)`?" Furthermore, we're then forced to leave a comment that this is indeed to force specialization, lest someone clean up the code later. To make matters worse, the specialization heuristics are somewhat subtle (and presumably subject to change) so I've seen people start to throw this in at the first sign of trouble in the hope that it will help. username_3: So looking at the code which processes `@specialize`, it seems that the per-argument form is currently ignored, ie the following doesn't actually do anything: ```julia function foo(@specialize(x)) body end ``` See https://github.com/JuliaLang/julia/blob/52c55d7934f71c5b2d9f6e6fa98cb48817def57c/src/method.c#L554-L558 So in principle we could steal `@specialize(x)` to mean "force specialize". However this form would be inconsistent with ```julia function foo(x) @specialize body end ```
ghusta/android-docker-search
808383667
Title: Refactoring : Use ConstraintLayout Question: username_0: Replace Android's `RelativeLayout` by `ConstraintLayout` See : - https://developer.android.com/reference/androidx/constraintlayout/widget/ConstraintLayout - https://developer.android.com/training/constraint-layout<issue_closed> Status: Issue closed
yii2tech/config
213524624
Title: Dynamic modules connection Question: username_0: Can I add new modules dynamicly, but not only parametrs of nested modules? For example:         ``` 'NewModules' => [                      'Path' => 'modules',                      'Label' => new Modules',                  ], ``` and db row id: `NewModules` 'value: `['user'=>'path/to/Module']` Status: Issue closed Answers: username_1: It is impossible to manipulate entities, which do not exist in current application. Dynamic componenst or modules can not be added.
platformio/platformio-vscode-ide
435018356
Title: Start PIO Home Server Question: username_0: # Description of problem Leave a comment... BEFORE SUBMITTING, PLEASE SEARCH FOR DUPLICATES IN - https://github.com/platformio/platformio-vscode-ide/issues # Configuration VSCode: 1.33.1 PIO IDE: v1.7.0 System: Windows_NT, 10.0.17134, x64 # Exception ``` Error: Error: Detected unknown package 'contrib-pysite' at P.Promise (C:\Users\Aj\.vscode\extensions\platformio.platformio-ide-1.7.0\node_modules\platformio-node-helpers\dist\index.js:1:14992) at ChildProcess.h (C:\Users\Aj\.vscode\extensions\platformio.platformio-ide-1.7.0\node_modules\platformio-node-helpers\dist\index.js:1:5916) at ChildProcess.emit (events.js:182:13) at ChildProcess.cp.emit (C:\Users\Aj\.vscode\extensions\platformio.platformio-ide-1.7.0\node_modules\cross-spawn\lib\enoent.js:34:29) at maybeClose (internal/child_process.js:961:16) at Process.ChildProcess._handle.onexit (internal/child_process.js:248:5) ``` Status: Issue closed Answers: username_1: Duplicate of https://github.com/platformio/platformio-core/issues/2341
EarthSystemCoG/COG
51328579
Title: Add new metrics to the system users page Question: username_0: WHO: Sylvia For security purposes, the following items would be useful to know about folks who have registered for cog. I propose: 1) adding these 3 metrics to the system users page: a) Number of projects to which they are members. Looking for 0 b) Number of web URLs they have. 0 projects + a web url may be a spam bot c) Date of last login. Eventually we may end up culling folks who never used the site. 2) changing the template for this page so that it has neither the left nor right widgets and can span the entire page. Answers: username_0: The number of projects is showing up as zero for all users. Will have to investigate. The number of projects is showing up correctly on individual profiles. username_0: Ok. I determined that {{projects|length}} will not work on this page because the function that creates projects is not associated with this template now. I created a new function that created a list of the projects|length for each user. Since it can not be associated with the user object, I had to fudge the table by breaking it it in two, drawing a separate table with the project info and then making them look seamless. username_0: Luca pointed out a better way to do this. In testing, the search is not working on it, so we will postpone to 3.0 since 2.9 is coming out in a rush. username_0: We realized that gathering the number of projects does not work across nodes. the information is irrelevant. Redesign this page as necessary. username_0: Removed the number of projects value. It is not federated and is bogus. Turned the page back on for system administrators. Need to check the links to profiles work on dev-hydra, can't test locally because created accounts have no openid and clicking on the name fails. username_0: verified on dev-hydra. Closing. Status: Issue closed
suyashkumar/dicom
1146132301
Title: De-identify DICOM tags Question: username_0: I'm trying to update/de-identify some of the metadata tags (such as Patient ID). Is there a way to update the file directly without reading the file and writing back to the file? Answers: username_0: @suyashkumar @jstutters
KIST-Iceberg/Iceberg
288070574
Title: Image Crop Question: username_0: ![figure_1](https://user-images.githubusercontent.com/6357456/34870390-36489166-f78a-11e7-91bb-8a043e57fcec.png) ![figure_1-1](https://user-images.githubusercontent.com/6357456/34870391-3664c278-f78a-11e7-9f41-56a1bb5e741b.png) ![figure_1-2](https://user-images.githubusercontent.com/6357456/34870392-36814d76-f78a-11e7-8794-e0adeae2b173.png) ![figure_1-3](https://user-images.githubusercontent.com/6357456/34870393-36a00f18-f78a-11e7-8ccf-3ee55fe5d544.png) ![figure_1-4](https://user-images.githubusercontent.com/6357456/34870394-36bc75ae-f78a-11e7-937b-7e460cc1175b.png) Status: Issue closed Answers: username_0: not need to crop https://www.kaggle.com/brassmonkey381/viewing-leak-and-machine-images
moby/moby
341268075
Title: .Net ASP.Net Webapp in Container loose Primary Domain Trust randomly after some days runtime Question: username_0: **Reference:** https://social.msdn.microsoft.com/Forums/en-US/542286aa-1095-4af0-91f2-43ff0c0f9469/docker-container-with-gsma-loose-trust-relationship-after-a-few-days?forum=windowscontainers Answers: username_0: I rebuild all my images to Server 2019 LTSC Container on new Server 2019 Docker Host VMs like descibe above. Problem is better on a factor of 1000 :-) but still sometimes i loose trust on container starts like username_1: Do you use a time synchronization tool to have both the container host and the DC sync their clocks? A difference in clocks can account for the behavior that you describe. username_2: We are also experiencing this issue. @username_0 did you get it to work as expected? username_0: No we still have this error but its not as often if we use win2019 😟 You have any success on this? username_0: @username_2 you also reverseproxy with haproxy infront? username_2: @username_0 Nope. We have interlock in front and it fully supports the Kerberos according to their documentation. username_0: @memonfaheem You find anything after a near ayear? username_3: @username_0 have you found any more information about this? username_0: no sorry still investigating you got same Problem? and als no intel from ms or docker username_3: Hey @username_0 , do your containers all use the GMSA account name they are running as as their hostname? This was previously a requirement, but as of Windows Server 2019 it is no longer needed. In fact, setting the container hostnames to all be the same may be what causes this issue. If you are able, try allowing the hostnames to be automatically set by docker to the container ID instead of set to the GMSA account. username_0: Thx for your Input @username_3 We upgraded to 2019 because of the issue you mention. All Containers got a different Containername defined by an Ansible Playbook, i need to have so the Playbook can identify the instances. All Containers share the same GMSA Account Credentials ans running on 2 VMs Maybe thats the Problem? For each Application Containers we run them on both Docker hosts(VM). Both Dockerhosts have permission for the GMSA and the GMSA Account have the required SPNs for the Service, and we use a dedicated Container IP Adresses for each Container. If the Problem are the IPs because of Kerberos Auth cannot validarte SPN Hostname or something they will not work after Start right? But they work and after an unkown time they loose the trust. Will try Random Containernames in an Testenvironment Any other Ideas? username_0: And each Container have different Name on both host like App1-1,App1-2 or something, so on both dockerhosts the containernames are unique
deviant-dev/ds-issues
559307852
Title: System can't find Focused Char, can't perform 'Looming' Question: username_0: **Describe the bug** Playing through the demo, console shows an error that the system can't find a focused character to loom over. **To Reproduce** Playing through the demo with the console open, occurs near the end of goddess's chambers **Expected behavior** No errors to be displayed on the console **Screenshots** ![looming](https://user-images.githubusercontent.com/12089633/73687215-c9988780-4686-11ea-8475-44f8334da780.png) **System (please complete the following information):** - VR Device: KB+M - Win 10 64bit - Version: Demo-1 **Additional context** Screenshot taken after leaving goddess's chambers Answers: username_0: I'm not going to create a bug report around the 'Perform angry 50' as I'm guessing that fixing this focused character issue will likely fix any other issues around focused characters. Status: Issue closed
spatie/laravel-backup
504078022
Title: Problems with 32bit php Question: username_0: Hello, I run into some problems with a 32 bit php. I can't look up my backup list nor can I look it up in your nova tool. The main problem is described in #567 .The Fork from CharlieEtienne isn't updated so I don't want to use it. Maybe we can fix it in the real package? Thanks for the great package and keep up the good work! If you need more information just reply. Answers: username_1: I'm not running into the problem myself and don't have time to investigate. I'd accept a PR to fix it in this package. username_0: Did a pull request for the issue. It fixed the problem for me but it needs a validation for a 64bit php system. #976 Status: Issue closed username_1: We'll continue the conversation in #976
matplotlib/matplotlib
143285472
Title: Example embedding_in_tk.py freezes in Python3.5.1 Question: username_0: When I try and run the [example code on matplotlib's website](http://matplotlib.org/examples/user_interfaces/embedding_in_tk.html) for embedding matplotlib in tk, the GUI never appears and the kernel crashes. I'm using Python 3.5.1, matplotlib 1.5.1, and tk 8.5.18 vc14. Everything was installed using Anaconda. I'm pretty sure this is supposed to work, as I've seen other tutorials run it using Python 3. Interestingly, when I create a Python 2.7 environment, I'm able to run the code successfully. Answers: username_1: I can not reproduce this with a linux box. Do you get any error messages from the crash? username_2: When you say "the kernel", are you talking about the ipython kernel, or the linux kernel? If it is the linux kernel, I doubt this would be specific to matplotlib and wonder if it is a tk packaging issue in general? username_0: I'm a grad student in MechE, so forgive me if I get some of the terminology wrong. I'm talking about the instance of python. username_2: Ah, so you are on Windows? Which version? username_0: Windows 7 Professional, SP1 username_3: I suspect this is a problem specific to the Anaconda build and/or your particular system (e.g., are there Tk versions installed in addition to the one from Anaconda?) rather then being an MPL bug. You could check with a message to the Anaconda mailing list to see whether anyone there can reproduce it, or whether there is someone with a similar configuration who *cannot* reproduce it. username_0: I've done a fresh installed on a smiliar windows 7 machine and was able to reproduce the same issue, so I'm thinking its either an issue with anaconda or matplotlib. Is anyone else running a non-anaconda python interpreter on windows 7, and is able to reproduce the issue? I've started a [topic on the anaconda mailing](https://groups.google.com/a/continuum.io/forum/#!topic/anaconda/f-UwlCovN_I) list to see if someone else can reproduce it. I'm moving forward using Python 2.7, but hopefully we can get this resolved. username_0: Thanks for the help everyone! Status: Issue closed
stegos/stegos
447051502
Title: Panic during consensus. Question: username_0: ``` STEGOS_LOG 2019-05-22T08:37:53.712194541+00:00 [stegos_consensus::state] DEBUG - Precommit(11450:9): collected a pre-commit: from=SecurePKey(e54df9c0de923c8972b3e7984eab8dd3261d2530fd776d2dbb29e98e7ff3 b28002954ad4ce7001f379e61323c2e8d7a2cacd45906e0677784c544c88f5c6276301) STEGOS_LOG 2019-05-22T08:37:53.922420499+00:00 [stegos_consensus::state] DEBUG - Precommit(11450:9): collected a pre-commit: from=SecurePKey(<KEY> 240f8e7785d24b49efd49e738609aa508b15a1ea88b6c29311e60e6d3510ea13411501) STEGOS_LOG 2019-05-22T08:37:54.123815742+00:00 [stegos_consensus::state] DEBUG - Precommit(11450:9): collected a pre-commit: from=SecurePKey(4f6adf8ff3c224c7b181a12ce48c520b0173d20fee742ad92d70f6ee3868 90cc0fc413aaa2a635401e5007538695eb01ff0ad8a697de502c0c70ab2f001b8ae600) STEGOS_LOG 2019-05-22T08:37:54.123882127+00:00 [stegos_consensus::state] DEBUG - Precommit(11450:9) => Commit(11450:9) STEGOS_LOG 2019-05-22T08:37:54.124543151+00:00 [stegos_node] INFO - Sent sealed block to the network: height=11450, block=80ec0bed1a1c STEGOS_LOG 2019-05-22T08:37:54.124589928+00:00 [stegos_blockchain::validation] DEBUG - Validating a macro block: height=11450, block=80ec0bed1a1c STEGOS_LOG 2019-05-22T08:37:54.124715373+00:00 [stegos_network::pubsub] DEBUG - Sending broadcast message: topic=block, size=1414 thread 'tokio-runtime-worker-1' panicked at 'block is validated before: BlockError(NoLeaderSignatureFound(11450, H(80ec0bed1a1c26c3cbe2716a6b4207c605ff31fb3a304e1f428f1b8416070416))) ``` looks like create propose without leader signature, this could be caused because of "round lock". Answers: username_1: This was a transitional problem - it has been fixed already. Status: Issue closed
jlippold/tweakCompatible
417004354
Title: `SwipeForMore` working on iOS 12.1.1 Question: username_0: ``` { "packageId": "org.thebigboss.swipeformore", "action": "working", "userInfo": { "arch32": false, "packageId": "org.thebigboss.swipeformore", "deviceId": "iPhone6,1", "url": "http://cydia.saurik.com/package/org.thebigboss.swipeformore/", "iOSVersion": "12.1.1", "packageVersionIndexed": true, "packageName": "SwipeForMore", "category": "Tweaks", "repository": "BigBoss", "name": "SwipeForMore", "installed": "1.1.7", "packageIndexed": true, "packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 100% with 13 working reports.", "id": "org.thebigboss.swipeformore", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.1.2", "shortDescription": "Manage packages via swipe.", "latest": "1.1.7", "author": "PoomSmart", "packageStatus": "Working" }, "base64": "<KEY>", "chosenStatus": "working", "notes": "" } ```<issue_closed> Status: Issue closed
tminglei/slick-pg
92736392
Title: Slick 3.0 tsql type mapping Question: username_0: I'm trying to use the new typed sql statements in Slick 3.0 and I'm discovering that the type mapping that I have found in slick-pg doesn't seem to be working. Is this something to do with how I extend the postgres driver? Or is this something that needs direct support in slick-pg? Answers: username_1: Hi @username_0, do you mean those extension methods not working? _Those extension methods are based on `JdbcType`, but slick plain sql isn't based on it. So the types supported by slick-pg plain sql hasn't those extension methods._ username_0: I'm not sure I understand your response, so I'll give a little more information. In the normal slick methodology, I could map a case class to a table like this: import org.joda.time._ case class Foo(name: String, date: DateTime) class Foos(tag: Tag) extends Table[Foo](tag, "foos") { def name = column[String]("name") def date = column[DateTime]("date") def * = (name, date) <> (Foo.tupled, Foo.unapply) } val foos = TableQuery[Foos] With the new slick 3.0 `tsql` typed sql statements, I would like to be able to write the following: def selectAllFoos: DBIO[Seq[(String,DateTime)]] = tsql"select * from foos" but in this case the type mapping that I have come to expect from slick-pg for the joda DateTime class doesn't work. Are you saying that the underlying JDBC functions are different between Slick's normal usage and its `tsql` statements? The methods used for marshalling case classes [as explained here](http://slick.typesafe.com/doc/3.0.0/sql.html#result-sets) look like they are using normal JDBC column data types. username_1: Well, I got your means. I just check the related codes of `slick`, and seems it requires some special customization on `slick-pg`. I'll implement it ASAP when I have time. _Thanks for pointing it out! :-)_ username_1: Hi @username_0, to some degree, this is a problem same as #171, which requires `slick-pg` to auto bind a pg type to a scala type. This is not so easy, as you know, since a pg type can be bound to multiple scala types. I need more time to construct a mechanism to help do it easily when your customized postgres driver mixing some pg type supports. So it won't be ready soon. username_1: Hi @username_0, the implement is almost done, but we need some enhancement in `slick`, related pull request is slick/slick#1226 The working branch in `slick-pg` is tsql, you can check its codes [here](https://github.com/username_1/slick-pg/tree/tsql). _BTW, currently I can only support bind a pg type to a scala type without type parameter. For scala types with type parameter, I can't implement it, since it's not supported by `slick` by now._ username_1: Hi @username_0, the changes already merged into master and released in slick-pg v0.10.0-RC1. You can help check it. Thanks! ^^ Status: Issue closed username_0: Thanks, I'll test it out tonight. I appreciate the fast turnaround on this request!
ThomasWest/blog20160423
150682391
Title: latest endeavor 20160423 Question: username_0: I have now complete through chapter 6 in this particular project. I have had a number of irreconcilable bugs (actually some greater success in this project) and started over, essentially, a couple of times. As always, I have many questions (re: switching to postgres, schema, data, etc and purpose and action of some "requirements in README.md), but a miniscule bit more understanding, I hope. I would hope for a face-to-face for explanations and "demonstrations" at some point.
factor1/phoenix-design-week-2016
165963531
Title: About: Revise styling on learn about MM and check out events box headlines Question: username_0: These headlines need to be in Proxima Nova as well—same styling as previous specs given for H2s and H3s. Basically every H-tag needs to be in Proxima Nova. Only buttons use the Bebas, and a couple subheads in the hero areas. ![screenshot 2016-07-16 23 34 51](https://cloud.githubusercontent.com/assets/4261229/16899158/57393fd0-4bae-11e6-9c39-b8902bcb921c.png) Status: Issue closed Answers: username_1: fixed with issue #41 username_0: Again, I failed to provide all of the details here: They also need to be regular weight and sentence case (I think I noted this in Issue #41 as well)
tensorflow/tensorflow
396732197
Title: Support first_n in tf.print Question: username_0: **System information** - TensorFlow version (you are using): 1.12 - Are you willing to contribute it (Yes/No): no **Describe the feature and the current behavior/state.** `tf.Print` is deprecated and `tf.print` doesn't support `first_n`. Would be good to support `first_n` in `tf.print` as well. **Will this change the current api? How?** Add a `first_n` parameter to tf.print. **Who will benefit with this feature?** Anyone who has used `first_n` in tf.Print. **Any Other info.** See Goomics #50. Answers: username_0: It seems there's already a solution to the "where to store first_n state": store it in the op just like `tf.Print` does today. I have no experience with eager mode or autograph in TF, but it seems that the code you provided doesn't actually reproduce the behavior of `tf.Print`. First, there's just a single counter Variable - does that imply the state is shared across all calls to print_first_n? Second, it's my impression that *all* Variables are stored in checkpoints. If we had a per-print Variable, we'd have trouble restoring if someone decided to add / remove a print statement, no? username_1: In tf2 variables are only stored in checkpoints if you add them to checkpoints. Re how many counters you have, this is up to you; as you can see it's easy to control with this workaround. And about where to put the state, the solution in tf.Print is nondeterministic in many ways, something we do not want to reproduce in tf.print. username_0: Gotcha, so these are all fair points, and looks like things will be cleaned up in tf2. In the meantime, `tf.Print` is deprecated in tf1 so is there another solution I can use before we switch over to tf2 or will we just have to accept and ignore the deprecation warning? username_2: If you're willing to get a little dirty, you can look into the silence context manager inside deprecation.py. It's not part of the official API (and it won't be), but you can use it to hackily suppress deprecation warnings.
lvgamedev/simairport-modding
646736693
Title: At Baggage Scanner the trays not dissapier at the scan area Question: username_0: Also with right marker setting , it not works , trays are to see at the whole mod. (i post that for marineanukotilah , because his account here not works anymore) [RemoteBagScannerSmall.zip](https://github.com/lvgamedev/simairport-modding/files/4841078/RemoteBagScannerSmall.zip) ![image](https://user-images.githubusercontent.com/56348209/85929757-a5e51080-b8b7-11ea-9a29-7acf237b87d1.png) Answers: username_1: Added an extra field to the bag scanner utility named "maskSprite" - set it to the name of the sprite that should be used for masking, facings and large will automatically be added. For masking name "myMask" the large bag scanner facing to the left will use the texture "myMask_Large_Left". Status: Issue closed
flutter/flutter
705823097
Title: Local network permission issue on iPad, crash on launch Question: username_0: @username_2, following up from https://github.com/flutter/flutter/issues/57225 Steps to reproduce: 1. `flutter create testproject3` 2. `cd testproject 3` 3. `flutter run -v > output_1.txt` The app runs on the iPad, and presents the "testproject3 would like to find and connect to devices on your local network" dialog. The app then crashes (disappears from screen), leaving the dialog still there. stderr output: ```text testproject3$ flutter run -v > output_1.txt [ ] Failed to attach to the observatory. [ +2 ms] Try accepting the local network permissions popup, or enable "Settings > Privacy > Local Network" for your app. [ ] For more information, see https://flutter.dev/docs/development/ios-14#debugging-flutter [ ] Error launching application on Stephen’s iPad. [ +5 ms] #0 throwToolExit (package:flutter_tools/src/base/common.dart:14:3) #1 RunCommand.runCommand (package:flutter_tools/src/commands/run.dart:606:7) <asynchronous suspension> #2 FlutterCommand.verifyThenRunCommand (package:flutter_tools/src/runner/flutter_command.dart:985:18) #3 _rootRunUnary (dart:async/zone.dart:1198:47) #4 _CustomZone.runUnary (dart:async/zone.dart:1100:19) #5 _FutureListener.handleValue (dart:async/future_impl.dart:143:18) #6 Future._propagateToListeners.handleValueCallback (dart:async/future_impl.dart:696:45) #7 Future._propagateToListeners (dart:async/future_impl.dart:725:32) #8 Future._completeWithValue (dart:async/future_impl.dart:529:5) #9 _AsyncAwaitCompleter.complete (dart:async-patch/async_patch.dart:40:15) #10 _completeOnAsyncReturn (dart:async-patch/async_patch.dart:311:13) #11 RunCommand.usageValues (package:flutter_tools/src/commands/run.dart) #12 _rootRunUnary (dart:async/zone.dart:1198:47) #13 _CustomZone.runUnary (dart:async/zone.dart:1100:19) #14 _FutureListener.handleValue (dart:async/future_impl.dart:143:18) #15 Future._propagateToListeners.handleValueCallback (dart:async/future_impl.dart:696:45) #16 Future._propagateToListeners (dart:async/future_impl.dart:725:32) #17 Future._completeWithValue (dart:async/future_impl.dart:529:5) #18 Future._asyncCompleteWithValue.<anonymous closure> (dart:async/future_impl.dart:567:7) #19 _rootRun (dart:async/zone.dart:1190:13) #20 _CustomZone.run (dart:async/zone.dart:1093:19) #21 _CustomZone.runGuarded (dart:async/zone.dart:997:7) #22 _CustomZone.bindCallbackGuarded.<anonymous closure> (dart:async/zone.dart:1037:23) #23 _microtaskLoop (dart:async/schedule_microtask.dart:41:21) #24 _startMicrotaskLoop (dart:async/schedule_microtask.dart:50:5) #25 _runPendingImmediateCallback (dart:isolate-patch/isolate_patch.dart:118:13) #26 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:169:5) ``` stdout output attached as `output_1.txt` [output_1.txt](https://github.com/flutter/flutter/files/5257450/output_1.txt) Other things happen if I run again, or if I erase the app from the iPad, and run again. But maybe this is enough for now? Please let me know if I can provide anything else useful. ```text $ flutter doctor -v [Truncated] • Xcode 12.0, Build version 12A7209 • CocoaPods version 1.9.3 [✓] Android Studio (version 4.0) • Android Studio at /Applications/Android Studio.app/Contents • Flutter plugin version 47.1.2 • Dart plugin version 193.7361 • Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593) [✓] VS Code (version 1.49.0) • VS Code at /Applications/Visual Studio Code.app/Contents • Flutter extension version 3.14.1 [✓] Connected device (1 available) • Stephen’s iPad (mobile) • b0975f3aa3ebe98c71462a866b0356a629fc6aa3 • ios • iOS 14.0 ! Error: Stephen’s iPad is busy: Copying cache files from device. Xcode will continue when Stephen’s iPad is finished. (code -10) • No issues found! ``` Answers: username_1: Have you tried this at all? username_2: Oh the app is crashing, that's why the observatory isn't found. The `Try accepting the local network permissions popup` comment is just confusing in this spot. Can you attach the crash log from the device? username_0: It doesn't look like a crash log was produced. I'm looking in Xcode under "view device logs" in "devices and simulators". There are other crash logs here (e.g. one from Skype, looks like it was terminated for using too much CPU when backgrounded). I tried reproducing the behaviour while viewing the console to see what's going on there. This time (starting by creating a new test project with a different name) the app launched. However, the observatory never attached. I tried a couple more times, and I have not been able to reproduce the crash on launch. username_2: @username_0 Now that the app doesn't crash on launch (weird there's no log), do you see `Settings > Privacy > Local Network` now? Is it checked? username_0: @username_2 I appreciate you following up on this. Here's what I see with the latest master (1.22.0-10.0.pre.318). Summary: default app crashes, permission setting is available but doesn't seem to do anything useful. First, I create a new project. I'm using new names for each project to ensure there is no interference from previous state on my iPad. ```text $ flutter create testproject8 ... All done! [✓] Flutter: is fully installed. (Channel master, 1.22.0-10.0.pre.318, on Mac OS X 10.15.6 19G2021 x86_64, locale en-GB) [✓] Android toolchain - develop for Android devices: is fully installed. (Android SDK version 28.0.3) [✓] Xcode - develop for iOS and macOS: is fully installed. (Xcode 12.0) [✓] Android Studio: is fully installed. (version 4.0) [✓] VS Code: is fully installed. (version 1.49.1) [✓] Connected device: is fully installed. (1 available) ``` `$ cd testproject8/` `flutter run -v > >(tee -a stdout.log) 2> >(tee -a stderr.log >&2)` The usual stuff happens. The app appears on the iPad homescreen. The app launches. The app displays the "In iOS14+, ..." message. Meanwhile the `flutter run` command has exited. [stdout.log](https://github.com/flutter/flutter/files/5268463/stdout.log) [stderr.log](https://github.com/flutter/flutter/files/5268465/stderr.log) I open the Settings app on the iPad and look under Privacy, Local Network. No app is visible for me to allow local network access. There is no crash log for this app available via Xcode. ```text $ flutter doctor -v [✓] Flutter (Channel master, 1.22.0-10.0.pre.318, on Mac OS X 10.15.6 19G2021 x86_64, locale en-GB) • Flutter version 1.22.0-10.0.pre.318 at /Users/steve/code/flutter • Framework revision 6b39acdc53 (62 minutes ago), 2020-09-23 08:27:04 -0400 • Engine revision 3e7ca5b869 • Dart version 2.10.0 (build 2.10.0-156.0.dev) [✓] Android toolchain - develop for Android devices (Android SDK version 28.0.3) • Android SDK at /Users/steve/Library/Android/sdk • Platform android-29, build-tools 28.0.3 • Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java • Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593) • All Android licenses accepted. [✓] Xcode - develop for iOS and macOS (Xcode 12.0) • Xcode at /Applications/Xcode.app/Contents/Developer • Xcode 12.0, Build version 12A7209 • CocoaPods version 1.9.3 [✓] Android Studio (version 4.0) • Android Studio at /Applications/Android Studio.app/Contents • Flutter plugin version 47.1.2 • Dart plugin version 193.7361 • Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593) [✓] VS Code (version 1.49.1) • VS Code at /Applications/Visual Studio Code.app/Contents • Flutter extension version 3.14.1 [✓] Connected device (1 available) • Stephen’s iPad (mobile) • b0975f3aa3ebe98c71462a866b0356a629fc6aa3 • ios • iOS 14.0 ! Error: Stephen’s iPad is busy: Copying cache files from device. Xcode will continue when Stephen’s iPad is finished. (code -10) • No issues found! ``` username_3: Looks like it's just not iPad, happens on iPhone as well https://github.com/flutter/flutter/issues/66408 username_4: This error message may be the cause. `Error: Stephen’s iPad is busy: Copying cache files from device. Xcode will continue when Stephen’s iPad is finished. (code -10)` username_2: So it's not crashing anymore, it just sits on the error message? I wouldn't expect the permissions pop up to happen when it's sitting on that message. What happens when you run `ios/Runner.xcworkspace` from Xcode? username_0: Running from Xcode, the app launches with a blank white screen, and the iOS dialog "testproject9 would like to find and connect to devices on your local network" I press "OK" and then the app fully launches. In the output window of Xcode is the following: ```text 2020-09-24 12:20:41.808173+0200 Runner[4478:1907931] Metal API Validation Enabled 2020-09-24 12:20:42.338640+0200 Runner[4478:1908359] flutter: Observatory listening on http://127.0.0.1:49305/u3r3W6zN1P8=/ 2020-09-24 12:21:33.790822+0200 Runner[4478:1908352] fopen failed for data file: errno = 2 (No such file or directory) 2020-09-24 12:21:33.790937+0200 Runner[4478:1908352] Errors found! Invalidating cache... 2020-09-24 12:21:33.918937+0200 Runner[4478:1908352] fopen failed for data file: errno = 2 (No such file or directory) 2020-09-24 12:21:33.919056+0200 Runner[4478:1908352] Errors found! Invalidating cache... ``` username_0: Immediately after this, I stopped the app from Xcode, closed Xcode, and tried again with `flutter run` I got the following terminal output, and the app launches with the "In iOS 14+, Flutter application in debug mode..." message. ```text $ flutter run Launching lib/main.dart on Stephen’s iPad in debug mode... Automatically signing iOS for device deployment using specified development team in Xcode project: RPX9H7TNPD Running Xcode build... └─Compiling, linking and signing... 9.2s Xcode build done. 14.7s Failed to attach to the observatory. Try accepting the local network permissions popup, or enable "Settings > Privacy > Local Network" for your app. For more information, see https://flutter.dev/docs/development/ios-14#debugging-flutter Installing and launching... 28.3s Error launching application on Stephen’s iPad. ``` username_2: That's really odd. ``` [ +23 ms] (lldb) run [ +85 ms] success [ ] Application launched on the device. Waiting for observatory port. ``` `flutter` attached the debugger, so it should never show that "In iOS 14+" error message. It's not surprising the observatory isn't connecting at that point though. Would you mind opening the Console app, selecting your app, and reproducing the problem, and attaching any `Runner` logs? Not sure why the engine thinks it can't run ptrace. <img width="945" alt="Screen Shot 2020-09-24 at 7 08 08 PM" src="https://user-images.githubusercontent.com/682784/94218407-4936c900-fe99-11ea-9cc5-2f5939c41f93.png"> username_5: Hi All. I am having this problem too. FYI, my iPad has been busy copying files for 7 days now.... I've tried flutter clean multiple times. Anyone have any solutions or workarounds? flutter doctor --verbose [✓] Flutter (Channel master, 1.22.0-10.0.pre.380, on Mac OS X 10.15.6 19G2021 x86_64, locale en-US) • Flutter version 1.22.0-10.0.pre.380 at /Users/rob/dev/flutter • Framework revision ec40df9576 (17 hours ago), 2020-09-25 21:27:22 -0700 • Engine revision 3a73d073c8 • Dart version 2.11.0 (build 2.11.0-161.0.dev) [✓] Android toolchain - develop for Android devices (Android SDK version 29.0.2) • Android SDK at /Users/rob/Library/Android/sdk • Platform android-29, build-tools 29.0.2 • Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java • Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593) • All Android licenses accepted. [✓] Xcode - develop for iOS and macOS (Xcode 12.0.1) • Xcode at /Applications/Xcode.app/Contents/Developer • Xcode 12.0.1, Build version 12A7300 • CocoaPods version 1.9.3 [✓] Chrome - develop for the web • Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome [✓] Android Studio (version 4.0) • Android Studio at /Applications/Android Studio.app/Contents • Flutter plugin version 45.1.1 • Dart plugin version 192.7761 • Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593) [✓] IntelliJ IDEA Ultimate Edition (version 2019.2.4) • IntelliJ at /Applications/IntelliJ IDEA.app • Flutter plugin version 41.1.4 • Dart plugin version 192.7402 [✓] VS Code (version 1.49.2) • VS Code at /Applications/Visual Studio Code.app/Contents • Flutter extension version 3.13.2 [✓] Connected device (2 available) • Web Server (web) • web-server • web-javascript • Flutter Tools • Chrome (web) • chrome • web-javascript • Google Chrome 85.0.4183.121 ! Error: CM2 is not connected. Xcode will continue when CM2 is connected. (code -13) username_2: `Try accepting the local network permissions popup` was just suggesting that it could possibly be a possible reason for a failure to connect to the observatory, but people took it literally. I removed it here https://github.com/flutter/flutter/pull/66621. username_0: I can no longer reproduce this problem with the latest Master, using a fresh project created with `flutter create`. I first experienced problems when I was using packages that had native parts, using cocoapods. I'll try that again, and check. ```text $ flutter doctor -v [✓] Flutter (Channel master, 1.22.0-10.0.pre.428, on Mac OS X 10.15.6 19G2021 x86_64, locale en-GB) • Flutter version 1.22.0-10.0.pre.428 at /Users/steve/code/flutter • Framework revision a06f385145 (12 hours ago), 2020-09-29 21:37:04 -0400 • Engine revision d18c3ea75c • Dart version 2.11.0 (build 2.11.0-178.0.dev) [✓] Android toolchain - develop for Android devices (Android SDK version 28.0.3) • Android SDK at /Users/steve/Library/Android/sdk • Platform android-29, build-tools 28.0.3 • Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java • Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593) • All Android licenses accepted. [✓] Xcode - develop for iOS and macOS (Xcode 12.0) • Xcode at /Applications/Xcode.app/Contents/Developer • Xcode 12.0, Build version 12A7209 • CocoaPods version 1.9.3 [✓] Android Studio (version 4.0) • Android Studio at /Applications/Android Studio.app/Contents • Flutter plugin version 47.1.2 • Dart plugin version 193.7361 • Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593) [✓] VS Code (version 1.49.1) • VS Code at /Applications/Visual Studio Code.app/Contents • Flutter extension version 3.14.1 [✓] Connected device (1 available) • Stephen’s iPad (mobile) • b0975f3aa3ebe98c71462a866b0356a629fc6aa3 • ios • iOS 14.0.1 • No issues found! ``` username_6: Hello, are there any updates for this? :)
sultan-of-rum/morrowind-scripting-for-dummies
982701702
Title: Add discussion on OpenMW-CS Question: username_0: This is a bit of a stretch goal. With cc9cii's and unelsson's work on [OpenMW-CS](https://openmw.readthedocs.io/en/latest/manuals/openmw-cs/index.html) (example [1](https://gitlab.com/OpenMW/openmw/-/merge_requests/1052) and [2](https://gitlab.com/OpenMW/openmw/-/merge_requests/1176)), the program is rapidly approaching the point where it may compete with TESCS in dialogue and quest development workflows. Hence, it might pay to include mention of it in the book and to bring out the respective caveats for both programs. An important thing to mention here is that OpenMW-CS does not compile script bytecode -- hence, to make the generated plugin files compatible with vanilla CS, the plugin file needs to be resaved in TESCS. Answers: username_1: I will indeed be mentioning the OpenMW-CS because I will mostly be using their default syntax highlighting.
apache/pulsar
590669167
Title: Broker and Proxy support multiple CA certs Question: username_0: I'm requesting the Broker and Proxy be updated to support multiple CA certs. I have a requirement that internal communication uses an internal CA cert and we allow customers to define their own CA cert alongside. Answers: username_1: If you're talking about your brokers accepting multiple CAs in tlsTrustCertsFilePath I _think_ that is already supported. You should be able to cat all the ca pem files into one big file and use that as your trust cert store. I have not done it personally because we have no such requirement but I think it will work. You will get a pem file like this: `-----BEGIN CERTIFICATE-----` `<your CA>` `-----END CERTIFICATE-----` `-----BEGIN CERTIFICATE-----` `<their CA>` `-----END CERTIFICATE-----` Status: Issue closed
w3c/csswg-drafts
280149848
Title: [cssom] using scroll wheel before scrollIntoView() finishes Question: username_0: I'm looking at [the spec](https://drafts.csswg.org/cssom-view/#dom-element-scrollintoview) and maybe I am missing it, but it doesn't seem to explain what happens when a user takes some contradictory action before the scrollIntoView() finishes, like use the scroll wheel (on their mouse), press a scroll key on keyboard (page up, page down), or click another element that triggers a different scrollIntoView(). Should the initial scrollIntoView() abort? Or should it finish the entire scroll sequence and delay the contradictory action? Or just cancel out the contradictory action?
mrkkrp/megaparsec
112199261
Title: Make expression parser parse repeated prefix/postfix operator Question: username_0: For more information see this SO question: https://stackoverflow.com/questions/10475337/parsec-expr-repeated-prefix-postfix-operator-not-supported Note to self: we cannot remove chaining combinators until this is implemented. Answers: username_0: I think it should be implemented in 4.2.0, not in 4.3.0. Obsolete combinators of `chain` family fill be removed in 4.2.0 too. This version will be released in December probably, don't want to release it too soon because 4.1.1 should be fine for most users. Status: Issue closed username_1: Not sure if this is the appropriate place to ask, but how do can I parse multiple "lookups" with `Postfix`? For example, for the array lookup syntax you can find in several languages: [1,[2,3]][1][0] username_0: @username_1, Unfortunately I have no time right now to write working example, but see in documentation how to parse several unary operators in a row: https://hackage.haskell.org/package/megaparsec-4.4.0/docs/Text-Megaparsec-Expr.html#v:makeExprParser Basically, you first write a parser for `[x]` things that should actually allow nested indices and then you: ```haskell parseManyPostfixIndices = foldr1 (.) <$> some parseOneIndex ``` This should do the trick. username_1: I should have read the docs! Thanks a lot, and sorry for the noise. username_2: Sorry to rehash this, but I can't for the life of me figure out goes in the `parseOneIndex` space. The type, according to ghci, is `Functor f => f (a -> a)` but it's very unclear to me if that is the expression parser, the term parser, the operator, or something else. username_0: @username_2, `parseOneInex` should parse an index, such as `[1]` for example, and return a function that would change some `a`, which is probably what you want to return from the whole parser, i.e. a term. For example: ``` my_array[0][1] ``` * `my_array` could on its own return something like `Var "my_array"`. * The indices could add something like this: `Indexed (Indexed (Var "my_array") 0) 1`. So the function that `parseOneIndex` would return could be: `\x -> Indexed x n`, where `n` comes from whatever you parse in that parser, of course. username_2: Ah, I see. So, to parse multiple prefix operators, as in the example in the haddocks, one would change `prefix name f = Prefix (f <$ symbol name)` to `prefix name f = Prefix $ foldr1 (.) <$> some (f <$ symbol name)`. Thank you very much for your help, and for writing a great library! username_3: ``` opSubscript = Postfix $ foldr1 (.) <$> some singleIndex singleIndex = do index < brackets expr return $ \l -> ArrayIndex l index ``` array[1][2] is expected to be parsed as `ArrayIndex (ArrayIndex (Var "array") 1)) 2` However, it was `ArrayIndex (ArrayIndex (Var "array") 2)) 1` What is going wrong here? username_3: [Fixed](https://www.reddit.com/r/haskell/comments/lfnd3o/issue_when_parsing_array12_using_megaparsec/) with `foldr1 (flip (.))`
mpetroff/pannellum
1119959999
Title: Adsense Question: username_0: Is it possible to add adsense to my pano player, if people hotlink and use my popular images by throttling my resources it wouldn't be a problem if i could enable adsense, i use your amazing panorama player as a nasa nerd star guide, if one of my 20mb milkyway images becomes amazingly popular it'll slow my modest server resources. If i can enable adsense it would fund my modest resources instead of burdening my server. Answers: username_0: update: - oh yeah! forgot to mention with an option to hide the adsense (x) so people can gaze at my nerdy 20mb images without visible obstruction. username_1: You can add whatever you want to the page you're using it on. You can also overlay content on the viewer, e.g., as is done in [the custom controls example](https://pannellum.org/documentation/examples/custom-controls/). username_0: i can't find a file labelled Overlay content or Custom Controls, i'm not a coder. . I'll probably add an infolinks instead of my adsense cause infolinks has a neat close ability, i don't know which file to edit to add my javascript so it shows on my player. . Trying to add this ``` <script type="text/javascript"> var infolinks_pid = 3356228; var infolinks_wsid = 0; </script> <script type="text/javascript" src="//resources.infolinks.com/js/infolinks_main.js"></script> ``` username_1: There's nothing named "overlay." The custom controls in that example are overlaid on the viewer. You can overlay other content in the same manner that that was done. username_0: That tutorial is very limited it doesn't tell you what file to edit, i don't know which file to edit username_1: It's an example, not a tutorial. You can overlay content on the viewer the same way you'd overlay it on any other `<div>` element. I've pointed you in the direction of what you need, but I have neither the time nor the motivation to explain basic HTML / CSS, so the rest is up to you. Unless you formulate a concrete question specific to how such content interacts with Pannellum, there's not much more for me to do. username_0: Okay fine, i was just a little confused with the Pannellum 50 kb version VS the 12mb version. So i uploaded both onto my server and forgot which one i was linking too in my spaghetti code. So i messed around with the 50kb version first and it seemed to work fine. I need to ask this however, i'm freelancing for a large known big brand, would i be correct on telling them that they can place advertisements on your creation or is that illegal from your optics username_0: They're now asking for ad html placements i hope i'm not breaking your copyrights, i am sincere which is why i ask. username_1: Your client's attorneys are free to review the license, but it allows for commercial use: https://github.com/username_1/pannellum/blob/master/COPYING Status: Issue closed
google/gapid
297537391
Title: Frames shown are different than output on device screen (Unity3D) Question: username_0: When using GAPID to debug Unity3D applications built for Android using OpenGL ES 3, the frames shown in the debugger are extremely different from the what I see on the device's screen in certain cases. It probably has something to do with transparency since it's only when I introduce alpha mask textures in my shaders that it displays this behavior. Typically, in GAPID I see what I would actually like to see on my device while my device shows something different, hence the debugging. Is there some underlying reason that there is a difference? It might help in actually debugging the erroneous behavior on the device. Answers: username_1: Hi @username_0, GAPID does some translation to replay OpenGL ES on desktop GL. It is possible there's a bug in this code. Are you willing and able to share a trace file demonstrating this issue? It would be a great help in figuring out the problem. Cheers, Ben username_0: [gapid_trace.zip](https://github.com/google/gapid/files/1728862/gapid_trace.zip) This is the trace. That is what I was expecting to see, but on my device I see a solid grey with a slightly darker row on the right. Texture 31 is supposed to be my alpha mask from which I sample the red channel. In this debug build I simply used a shader that displayed this texture full screen to see what was going on. username_0: [gapid_trace2.zip](https://github.com/google/gapid/files/1759359/gapid_trace2.zip) @username_1 this is another trace with the same result: What I see in GAPID is what I would want to see on my device but instead on the device all I see is a dark grey fill. One additional detail since last time is that this only happens on Nougat devices while Oreo devices work fine. Also, I get a repeating "bindTextureImage: clearing GL error: 0x502" in the adb log when running on Nougat devices. username_1: Hi @username_0 , Thank you for the reports. Sorry for the delay in looking into this, I've been working on something that is particularly high priority (it's blocking us from doing any more releases). Fortunately, I'm very close to finishing this work, so I'll hopefully be taking a look into this later this week. Thank you for your patience. Cheers, Ben username_0: @username_1 Thank you for your interest. Just wanted to add another clue: by coincidence the Nougat device has a Mali G-71 (Galaxy Note 8) while the Oreo devices (Nexus 6P and OnePlus 3) both have an Adreno. Wondering if maybe the android version has nothing to do with it. username_1: Hi @username_0, I've finally found the time to look into this. There are two suspicious things with these traces which are highlighted in the report tab: From `gapit_trace.gfxtrace`: Invalid Operation * `30861`: Error `glBindTexture` * `30869`: Error `glBindTexture` Both of these are erroring because you've bound the texture as an external texture (`GL_TEXTURE_EXTERNAL_OES`), and then (where it errors) you attempt to bind it as a texture 2D (`GL_TEXTURE_2D`). This is illegal. The texture takes the kind on the first binding, so texture 31 will be an external texture. As for why it looks different - it's probable this is just caused by undefined behavior due to the errors described above. Does this resolve your original question? Many thanks, Ben username_0: Thanks for the response but unfortunately I don't think that was it as those two keep showing up in the report even now that everything works correctly. I think those two glBindTextures fail because something is still initializing at that point but they get called every frame so they should be working. username_1: I'm happy to take another look if you don't mind sending another trace. username_0: Is there a way to cut frames from a trace? Trace capturing is slowing the application down quite significantly which means that the trace has to run for quite some time before the app reaches the point where everything is set up and it is executing the aforementioned alphamasking operation. The trace files way up to 1gb. username_1: Currently, I'm afraid not. We are working on adding mid-execution capture support for OpenGL ES, which should fix this issue. Status: Issue closed
gusaul/go-dynamock
942255999
Title: TransactWriteItems TableName expectation check for difference string references Question: username_0: Hello, In my code the assignment of tableName is done when constructing the DynamoDB SDK payloads, by using a string value that is passed around the different packages of my application. I assume most clients would rather use the String value than the pointer, and use `aws.String(myTable)` when constructing the `dynamodb.TransactWriteItemsInput` items. The side-effect of this, is that aws.String() creates a new reference each time, and therefore the equality check fails on go-dynamock, because it compares the references and not the value. Please find the following PR with my naive approach to fix the problem : https://github.com/gusaul/go-dynamock/pull/41
mageplaza/magento-2-webhook
535695724
Title: Webhook for customer newsletter subscription Question: username_0: Hello, is there a webhook I can use when a customer subscribes to the newsletter? If not it would be nice to have to be able to integrate with some CRMs like Agile CRM for example. Thank you! Answers: username_1: Hi @username_0 I’m Shin from Mageplaza.com and I will support you on this issue. For your question, you can create a new hook with the 'subscriber' option. Please see this screenshot: - http://prntscr.com/q93g80 Thanks so much! username_0: Thank you very much! Status: Issue closed username_0: Hello, is there a webhook I can use to trigger when a customer subscribes to the newsletter? If not it would be nice to have in order to be able to integrate this action with some CRMs like Agile CRM for example. Thank you! username_0: Hello I tested it (Subscriber Webhook) on 2.3.3 (with PHP 7.2), does not work, too. There is no log. Any suggestion? Thanks username_0: Found the fix here: https://github.com/mageplaza/magento-2-webhook/issues/12#issuecomment-521190268 Status: Issue closed
TheImagingSource/tiscamera
398929327
Title: Raspberry PI 2 not able to access some camera properties, cannot set Trigger mode Question: username_0: When trying to access some properties of the camera on a Raspberry PI 2, they are not available. One example is the Trigger mode, which I can't access. When I run tcam-capture I can see the camera, the videofeed and set several options, like exposure and white balance. However on my laptop I can control many more properties, including partial readout modes and trigger modes. I compiled the code on the Pi from the newest source. This is a message I get in the console after starting tcam-capture. libEGL warning: DRI2: failed to authenticate qt5ct: using qt5ct plugin 14-01-2019:15:57:09,744 WARNING [CapsDesc.py:167] Could not interpret structure. Omitting. video/x-raw, format=(string){ RGBx, xRGB, BGRx, xBGR, RGBA, ARGB, BGRA, ABGR }, width=(int)640, height=(int)480, framerate=(fraction){ 60/1, 30/1, 15/1, 5/1, 1/1 }; 14-01-2019:15:57:09,746 WARNING [CapsDesc.py:167] Could not interpret structure. Omitting. video/x-raw, format=(string){ RGBx, xRGB, BGRx, xBGR, RGBA, ARGB, BGRA, ABGR }, width=(int)1920, height=(int)1080, framerate=(fraction){ 15/1, 5/1, 1/1 }; 14-01-2019:15:57:09,747 WARNING [CapsDesc.py:167] Could not interpret structure. Omitting. video/x-raw, format=(string){ RGBx, xRGB, BGRx, xBGR, RGBA, ARGB, BGRA, ABGR }, width=(int)2048, height=(int)2048, framerate=(fraction){ 5/1, 1/1 }; 14-01-2019:15:57:09,748 WARNING [CapsDesc.py:167] Could not interpret structure. Omitting. video/x-raw, format=(string){ RGBx, xRGB, BGRx, xBGR, RGBA, ARGB, BGRA, ABGR }, width=(int)3072, height=(int)2048, framerate=(fraction){ 5/1, 1/1 }; 14-01-2019:15:57:11,351 WARNING [TcamView.py:482] get_tcam_property failed for 'Trigger Mode' inotify_add_watch("/home/pi/.config/qt5ct") failed: "No such file or directory" My main goal is to be able to enable Trigger mode and trigger the camera externally. When trying to run the example Save an image on trigger in Python I get the following error: tcambin serial="48810687" name=source ! video/x-raw,format=BGRx,width=640,height=480,framerate=30/1 ! videoconvert ! appsink name=sink Traceback (most recent call last): File "Program.py", line 75, in <module> print("Gain Auto : %s " % Tis.Get_Property("Gain Auto").value) File "/home/pi/dev/Linux-tiscamera-Programming-Samples-bak/Save an image on trigger in Python/TIS.py", line 160, in Get_Property return CameraProperty(*self.source.get_tcam_property(PropertyName)) Other examples which use trigger work, but the Trigger mode True does not seem to apply, they just stream the frames continuously regardless of the trigger being sent. Answers: username_1: Hello usually the uvc extension xml file and is installed with the "sudo make install" process. I would like you to check, whether the xml is correcly installed. A documentation about this can be found at https://github.com/TheImagingSource/tiscamera/wiki/UVC-Extension-Units I would like you to send me the output of `tcam-ctrl -p 48810687` also `tcam-ctrl -c 48810687` would be nice. Stefan username_0: Hi, thanks for the quick reply. The ouput of tcam-ctrl -p 48810678: Found 7 propert(y/ies) Brightness (int) min=0 max=4095 step=1 default=-8193 value=240 category=Exposure Whitebalance Auto (bool) default=true value=true category=Color Whitebalance Red (int) min=0 max=255 step=1 default=64 value=64 category=Color Whitebalance Blue (int) min=0 max=255 step=1 default=64 value=64 category=Color Gain (int) min=0 max=480 step=1 default=57343 value=0 category=Exposure Exposure Auto (bool) default=false value=true category=Exposure Exposure (int) min=100 max=60000000 step=100 default=300 value=300 category=Exposure and of tcam-ctrl -c 48810687: Available gstreamer-1.0 caps: video/x-bayer, format=(string)rggb, width=(int)640, height=(int)480, framerate=(fraction){ 60/1, 30/1, 15/1, 5/1, 1/1 }; video/x-bayer, format=(string)rggb, width=(int)1920, height=(int)1080, framerate=(fraction){ 15/1, 5/1, 1/1 }; video/x-bayer, format=(string)rggb, width=(int)2048, height=(int)2048, framerate=(fraction){ 5/1, 1/1 }; video/x-bayer, format=(string)rggb, width=(int)3072, height=(int)2048, framerate=(fraction){ 5/1, 1/1 }; video/x-bayer, format=(string)rggb16, width=(int)640, height=(int)480, framerate=(fraction){ 30/1, 15/1, 5/1, 1/1 }; video/x-bayer, format=(string)rggb16, width=(int)1920, height=(int)1080, framerate=(fraction){ 7/1, 5/1, 1/1 }; video/x-bayer, format=(string)rggb16, width=(int)2048, height=(int)2048, framerate=(fraction){ 2/1, 1/1 }; video/x-bayer, format=(string)rggb16, width=(int)3072, height=(int)2048, framerate=(fraction){ 2/1, 1/1 } I will check the documentation about the xml file. Thanks, Cenek username_1: Hi Cenek thank you. You are right, many properties are missing. I wait for your result regarding the uvc extension. Stefan username_0: Hi Stefan, I tried the procedure described at https://github.com/TheImagingSource/tiscamera/wiki/UVC-Extension-Units without success. Unfortunately I cannot produce any meaningful results now, because both our cameras refuse to send any pictures due to broken connector. I need to wait for new cameras to arrive. username_1: I already answered to you by email. I do not think, the connector is broken, because the cameras worked fine, when they left the office. Plugin the plug into the camera, until there was faint click sound. You may need some power doing that. Stefan Status: Issue closed
OpenChemistry/tomviz
111692885
Title: C++11 Question: username_0: @username_1 I'm making this issue to record our discussions about starting to use C++11 features in tomviz since it's been a few weeks and I don't recall all the details anymore. Things we had talked about starting to use (at least in the short term): override std::nullptr Unlike VTK, we only care about newer compilers. So the question is do we need to make macros for these or can we just start using them directly and turn on the cmake switch that will fail to configure on compilers that do not support them? Answers: username_0: Notes: use check_cxx_compiler_flag for testing c++11 present see [this](http://review.source.kitware.com/#/c/16780/1/CMake/vtkCompilerExtras.cmake) windows: VS2013 is minimum supported version since we don't test anything else linux: looks like we need minimum of debian wheezy (7) to get a new enough gcc to build with c++11 Status: Issue closed
InFact-coop/create-refresh
391076478
Title: Support for different languages Question: username_0: - [ ] detect user language - [ ] Pull in content from external source in appropriate language Answers: username_0: @username_1 if this is something that is near the top of your priorities in the backlog then it would be a good idea to get the content ready, in case we find time to do this. If it's at the bottom of your list of priorities, then it's unlikely to get done so you don't need to worry 😄 username_1: Getting this covered early might be a bit difficult as we need to get final approval on copy (which we probably won't get until the new year) and then get that translated. What is the alternative if we didn't get this done? And what is the likelihood that you would have time for this? username_0: @username_1 the likelihood we'll have time depends on how near the top of your priority list it is. Did you move the items in the backlog around already? I suppose even if we don't get the final copy until we're nearly finished, we could build the external source that we pull the content in from, have all of them in English to begin with, and then you guys would have to update it with the other languages once you have translations. username_1: @username_0 Is the backlog an actual place on Github where I can prioritise things? Sorry - still getting used to this. I would say it would make a difference in getting sign ups if we had different language options so we should make it more a priority in the backlog - although not sure where to add this! username_0: @username_1 no worries! yes it is, if you go [here](https://github.com/InFact-coop/create-refresh/milestone/3) you can add priority labels to all of the issues, according to your feelings 😃 Hopefully this picture helps: ![screen shot 2018-12-19 at 12 46 43](https://user-images.githubusercontent.com/13594621/50221175-420d8d80-038c-11e9-8437-296e3cac9c2b.png)
badges/shields
55941451
Title: Node callback should be err, response Question: username_0: The accepted format is: ```js function (error, response) { } ``` If you don't use this format, then your API is not open to being promisified, e.g: ```js var Promise = require('bluebird'); var badge = Promise.promisify(require('gh-badges')); badge({..}) .then(function (svg) { // svg here }) .catch(function (error) { // error here }); ``` Basically how it currently is, you can't do the above. Answers: username_1: That's a fair point, although this change would obviously break backward-compatibility. I suppose we could have this for v2.0.0. Let's [make an issue][] for that. [make an issue]: https://github.com/badges/shields/issues/362 Status: Issue closed
very-good-science/data-hazards
1020891009
Title: Create Data Hazards lesson plan materials Question: username_0: - [ ] Create lesson plan page - [ ] Use explainer videos from asynchronous materials #77 - [ ] Make an explainer video for a project/projects or explain how to do it live (link to materials for workshop) - [ ] Add timeline for lesson - [ ] Add to project timeline on homepage Status: Issue closed Answers: username_1: These resources are now on the website and reusable :)
angular/angular
176270794
Title: Compiler should detect invalid property binding Question: username_0: **I'm submitting a ...** `[X ] feature request` If possible, Angular2 compiler should detect and flag the following as an error: Component A has an input property defined as follows: `@Input() rows: number;` The template of a parent component (hosting Component A) references the property `rows` as an attribute like this: `<comp-a ... rows="5">` instead of `<comp-a ... [rows]="5">` In Angular RC6/Typescript 2.0.2, this compiles, but the value of `this.rows` in Component A is the string `"5"` as opposed to the number `5`. **What is the motivation / use case for changing the behavior?** Developer productivity. **Please tell us about your environment:** Windows 8.1 x64, Visual Studio Code * **Angular version:** 2.0.0-rc.6 <!-- Check whether this is still an issue in the most recent Angular version --> * **Language:** TypeScript 2.0.2 Answers: username_1: This may help https://github.com/angular/angular/issues/6919 username_2: It is not an error. think `<p title="...">` it makes sense for directive to get the content of the title Status: Issue closed
brittyazel/Neuron
772008539
Title: Zone ability not showing sometimes Question: username_0: As you level zone abiltiy is empty from a start but sometimes it switches on later as you quest or you get second ability. Also yesterday i noticed that when i got ventyr ability to change execute into condem -it did change execute into condem but didnt change icon for it on bar. Answers: username_0: Also there is no CD showing for Condemn. username_0: Fixing bugs in Neuron lately? username_1: I'm very busy at the moment, and am fixing things when and where I can. Neuron is not my priority at the moment, no. Once I finish the immediate deadlines in my personal life, I'll likely ramp Neuron work back up. Until then, I'll continue to do drive by fixes when I have a spare hour or two. 80+ hour work weeks make it hard to devote time to side projects, sadly. username_0: Its not big problem . Its just hard to see if you have given up on addon or just too busy. I havent received comment on this bug from 21 dec 2020 and too many times i have seen authors give up on addon and never tell users of their addon that they dont have time/interest to maintain it any longer. And given history i had with you as authors - meaning you at least acknowledeged bugs preatty quick even i you didnt fix it that moment made me wonder whats happening with addon which i used for very long (back to Maul days ). username_0: Any chance you looked at Condemn problem? Now every time i log in for first time on warr and try to use Condemn i get spell not learned message and i have to take spell out of spellbook again and relog for Condemn to start working. username_1: Have you tried using the spellID? username_0: Is that even possiable for spells? For fun i tried it but nothing happens /cast 317349 username_1: Oh you're right! My mistake. I'll see what I can do to check this out. What class/spec is this? username_0: warr venthyr fury username_1: https://eu.forums.blizzard.com/en/wow/t/condemn-macro-doesnt-work/214663 username_1: Is this your issue? Seems like an issue with macros in general username_0: Not sure if its totally same since i dont get spell disapering from spellbook but in general yes. Plus i dont get 4sec cd timer on it .Just 1 sec global cooldown spin.
YingboMa/MaBLAS.jl
622853994
Title: The analytical model for blocking Question: username_0: ```julia julia> using Hwloc julia> function params(::Type{T}) where T micro_m = LoopVectorization.mᵣ * VectorizationBase.pick_vector_width(T) micro_n = LoopVectorization.nᵣ topology = Hwloc.topology_load() l3 = getdatacache(topology, :L3Cache) l2 = getdatacache(l3, :L2Cache) l1 = getdatacache(l2, :L1Cache) l1associativity = l1.attr.associativity l1associativity = l1associativity == 0 ? 8 : # assume 8 if unknown l1associativity == -1 ? l1.attr.size ÷ l1.attr.linesize : # fully associative l1associativity l1sets = l1.attr.size ÷ (l1.attr.linesize * l1associativity) ncachelines_per_panelA = floor(Int, (l1associativity - 1) / (1 + micro_n/micro_m)) cache_k = ncachelines_per_panelA * l1.attr.linesize * l1sets ÷ (micro_m * sizeof(T)) end params (generic function with 1 method) julia> getdatacache(topology, name) = for t in topology.children isdefined(t, :type_) || return nothing return t.type_ === name && t.attr.type_ in (:Unified, :Data) ? t : getdatacache(t, name) end getdatacache (generic function with 1 method) julia> params(Float64) 213 julia> params(Float32) 256 ``` someone may find this code useful. Answers: username_1: This is an, uhh, interesting choice ```julia julia> params(Float32) 76 julia> params(Float64) 76 ```
Azure/azure-functions-sql-extension
679365052
Title: Deserialization into POCO Question: username_0: Currently, whenever we have to deserialize the SQL rows from the user's table into a POCO (like [here](https://github.com/Azure/azure-functions-sql-extension/blob/dev/src/SqlBinding/SqlAsyncEnumerable.cs#L117) for the input binding and [here](https://github.com/Azure/azure-functions-sql-extension/blob/dev/src/SqlBinding/TriggerBinding/SqlTableWatcher.cs#L735) for the trigger binding), we first build up a dictionary mapping from column name to column value. We then serialize that dictionary into a JSON string, and then deserialize the dictionary into the POCO type. This is a pretty round-about and inefficient way to do it, so it might be worth looking into alternatives, such as building our own deserializer that uses reflection to map column names to property names.
udistrital/generator-oas
200680670
Title: Finalización del proceso del token. Question: username_0: El login se hace mediante AngularJS-OAuth2, pero tenia problemas para manipulacion del token. Hasta hora el token se almacena en localstorage usando usando ngStorage, a su vez es posible decriptarlo, para esto se uso https://jwt.io/ -> kjur-jsrsasign. una version funcional de un proyecto generado está en https://fabianleon.github.io/app/#/, autenticandose con google. Está pendiente hacer que el token muera. y hacer una solicitud del nuevo token. la manipulacion del token se hace desde la fabrica token en scripts/services/token.js
github-vet/rangeloop-pointer-findings
775620897
Title: cd1989/cycli: pkg/stages/render.go; 8 LoC Question: username_0: [Click here to see the code in its original context.](https://github.com/cd1989/cycli/blob/f9749115fb0f7d250217ca68bcdaa1569c85eb8c/pkg/stages/render.go#L39-L46) <details> <summary>Click here to show the 8 line(s) of Go which triggered the analyzer.</summary> ```go for _, stg := range stgs { rows = append(rows, []string{ stg.Name, stg.Namespace, getLabel(&stg, meta.LabelProjectName), stg.CreationTimestamp.Format(time.RFC3339), getAnnotation(&stg, meta.AnnotationDescription)}) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: f9749115fb0f7d250217ca68bcdaa1569c85eb8c
yasirkula/UnityIonicIntegration
354758090
Title: iOS Splash Screen Question: username_0: We have tried a variety of iOS splash screen settings, but it seems that regardless of what we do we get a stretched version of the normal app splash screen when we open the Unity portion of our app. Has anyone had any luck changing the Unity splash screen when using this plugin? Status: Issue closed Answers: username_1: Did you resolve this issue? username_0: No...we ended up giving up on doing the custom splash screen like we were hoping for.
prettier/plugin-pug
625408539
Title: Bug: A blank line following multi-line comments is indented Question: username_0: ## Info | Tool | Version | | -------- | ------------- | | Plugin | v1.4.0 | | Prettier | v2.0.5 | | Node | v12.17.0 | | OS | mac | ## Input ```pug //- a p hi ``` ## Output ```pug //- a p hi ``` White spaces are inserted at the beginning of the 3rd line. ## Expected Output ```pug //- a p hi ``` ## Additional Context<issue_closed> Status: Issue closed
acornjs/acorn
737432042
Title: Cannot parse / operator after ternary Question: username_0: **Input:** ```js weird ? true : {}/2; ``` **Current:** [AST Explorer link](https://astexplorer.net/#/gist/ff5b87d4d1b548b774e81259c837ef02/d0e72a33af518dda328e833237d498e05a799b71) ``` Unterminated regular expression (1:18) ``` **Expected:** No error is thrown. **Context:** https://github.com/babel/babel/issues/11353<issue_closed> Status: Issue closed
invertase/react-native-firebase
368527395
Title: OnNotification not being called on release mode [iOS] Question: username_0: <!--- BEFORE YOU MAKE AN ISSUE The issue list of this repo is exclusively for bug reports. 1) For feature requests, please use our Canny board: https://react-native-firebase.canny.io/feature-requests 2) For questions and support please use our Discord chat: https://discord.gg/C9aK28N or Stack Overflow: https://stackoverflow.com/questions/tagged/react-native-firebase 3) If this is a setup issue then please make sure you've correctly followed the setup guides, most setup issues such as 'duplicate dex files', 'default app has not been initialized' etc are all down to an incorrect setup as the guides haven't been correctly followed. --> ### Issue onNotification event is being called on debug mode, but when running on release it’s not being called, only getInitialNotification is working on release mode <!--- Please write your issue here, provide as much detail as you can, code snippets, key files which will help us to debug such as your `Podfile` and/or `app/build.gradle` file). --> ### Environment 1. Application Target Platform: iOS <!--- (e.g. Sierra, Windows 10) ---> 2. Development Operating System: macOS 3. Build Tools: xcode <!--- (e.g. 0.45.1) ---> 4. React Native version: 0.55.4 5. RNFirebase Version: latest <!--- (e.g. database, auth, messaging, analytics etc - or N/A if not applicable) ---> 6. Firebase Module: Notifications/ messaging <!-- Love react-native-firebase? Please consider supporting our collective: 👉 https://opencollective.com/react-native-firebase/donate --> Answers: username_1: Same issue. Notification listener `OnNotification` works in development. When I build release app on IOS I will get only once notification in Foreground. If I set the app in the background and go back to the foreground, I will not get `foreground` notification any more. But background notification works. username_2: I have the same issue in android onNotification being triggered on emulatore but once i do a build -relaese mode it never triggered on real device ! username_3: Same situation on IOS - in development notifications works, in production doesn't username_4: same problem. any solution ? username_4: @username_1 I have same issue as you. did you find any solution? username_4: same issue @username_3 @username_2 @username_1 did you find any solution ?
davidgranstrom/tidal-bootstrap
260041298
Title: Cabal deps Question: username_0: Hi this script stopped working for people a few weeks ago. One solution is here: http://lurk.org/groups/tidal/messages/topic/2kkap8lbgQrky3992ALIGx/ Another is to install core haskell only. Would it be OK to move this script / repo into the tidalcycles organisation? Answers: username_1: Hi, Thanks for letting me know about this. I think that a stack based solution would be nice (it was actually my original intention to use stack, but I wanted to stay with the then recommended approach). However, this causes an issue that you have already adressed in the thread, namely that the atom plugin settings will have to be modified. This could of course be something that the script informs the user about, maybe printing the exact steps that you outline in your post. Installing the haskell-core only in an automated fashion seems to be harder, as homebrew only seem to package the [haskell-platform](https://www.haskell.org/platform/mac.html#osx-homebrewcask) as a homebrew cask. The solutions I can see at the moment is: 1) Let the user manually install [haskell-core](https://www.haskell.org/platform/mac.html#osx-none) and remove the `ghc` dependency from the script. 2) Use a stack based approach, modify the script to use `stack` instead of `cabal` and print instructions on how to modify the atom plugin ghci path. If you think the stack based install will be the recommended approach in the future, I think option 2 would be best. Unfortunately I will not have the time to do something about this at the moment, but if anyone wants to send a PR that would of course be very welcome. And yes, I think it would be a great idea to move this repo into the org! 👍 username_1: I just confirmed that by uninstalling `haskell-platform` and installing `haskell-core` manually the script works as intended. Would like try the install procedure again with `haskell-platform` to see if/how it breaks. username_0: Now using ghcup Status: Issue closed
Nastel/tnt4j-streams
255997719
Title: Minor: Message "RUNTIME" in tnt4j-streams.log has UTC vs local timestamp as in other trace msgs Question: username_0: For consistency, use local timestamp; see below at the arrow "<===" Extract from tnt4j-streams.log: 2017-09-07 12:17:03,489 TRACE [9:WmqTraceStream:WmqActivityTraceStreamQM_A!WmqTraceParser] - Locator 'MQGACF_ACTIVITY_TRACE.MQIACF_INVALID_DEST_COUNT' resolved value: null | TRACE | 2017-09-07 16:17:03.489000 UTC | RUNTIME=9068@Howard2-PC#SERVER=Howard2-PC#NETADDR=172.16.58.3#DATACENTER=UNKNOWN#GEOADDR=0,0 <==== 2017-09-07 12:17:03,489 TRACE [9:WmqTraceStream:WmqActivityTraceStreamQM_A!ActivityInfo] - Applying field MQTrace.InvalidDestCount from: null | TRACE | 2017-09-07 16:17:03.489000 UTC | RUNTIME=9068@Howard2-PC#SERVER=Howard2-PC#NETADDR=172.16.58.3#DATACENTER=UNKNOWN#GEOADDR=0,0 2017-09-07 12:17:03,489 TRACE [9:WmqTraceStream:WmqActivityTraceStreamQM_A!ActivityInfo] - Field MQTrace.InvalidDestCount resolves to null value. Will not apply field to activity... | TRACE | Answers: username_1: Changed in TNT4J commit <PASSWORD> Status: Issue closed