repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
neo4j-contrib/neo4j-apoc-procedures | 693122259 | Title: Espace AsciiDoc attribute references to suppress warnings when building the documentation with Antora
Question:
username_0: Currently, when we build the documentation with Antora from https://github.com/neo4j-documentation/docs-refresh, Asciidoctor prints almost 300 warnings because the documentation contains unresolved attribute references.
For reference, the syntax of an attribute reference in AsciiDoc is: `{myattribute}`.
The main issue is that the description of the `apoc.*` procedures frequently contain this syntax.
For instance:
```
apoc.cypher.runTimeboxed('cypherStatement',{params}, timeout) - abort kernelTransaction after timeout ms if not finished
```
As you can see in this example, AsciiDoc will try to resolve the attribute reference `{params}`.
One way to solve this issue is to escape the attribute reference with `\`:
```
apoc.cypher.runTimeboxed('cypherStatement',\{params}, timeout) - abort kernelTransaction after timeout ms if not finished
```
We could also prevent substitution using a single plus symbol. While doing so we could also use a monospaced font formatting:
```
`+apoc.cypher.runTimeboxed('cypherStatement',{params}, timeout)+` - abort kernelTransaction after timeout ms if not finished
```
In my opinion, the result will be more readable:
**Current**
apoc.cypher.runTimeboxed('cypherStatement',{params}, timeout) - abort kernelTransaction after timeout ms if not finished
**After**
`apoc.cypher.runTimeboxed('cypherStatement',{params}, timeout)` - abort kernelTransaction after timeout ms if not finished
However, this solution is more complex to implement because we need to accurately find the start and the end of the procedure (and if we guess wrong it will produce unexpected results).
Status: Issue closed
Answers:
username_0: Thanks @username_1
I think we should also backport this change to 4.0. Should I open a new pull request or do you want to cherry-pick this commit to the 4.0 branch? |
sheredom/utf8.h | 172902516 | Title: Crash waiting to happen?
Question:
username_0: So I stumbled upon this: https://github.com/username_1/utf8.h/blob/master/utf8.h#L697
Is this a crash waiting to happen, or am I reading the logic wrong?
If both `h` and `n` are equal strings, this will read past the memory of those strings.
Answers:
username_1: Hmm you are probably right! I'll add a test for that and we'll make sure it fails, then probably have to add on a check for null terminator there too.
Good find thanks! :smile:
Status: Issue closed
|
tinygo-org/tinygo-site | 942185137 | Title: installation instructions still show v0.18.0
Question:
username_0: Installation instruction still show old 0.18.0 version
i.e. Linux:
```
wget https://github.com/tinygo-org/tinygo/releases/download/v0.18.0/tinygo_0.18.0_amd64.deb
sudo dpkg -i tinygo_0.18.0_amd64.deb
``` |
Stremio/stremio-translations | 287498419 | Title: Same string keys multiple times
Question:
username_0: I see that many string keys string keys are multiple times in `Caps`, `nocaps` and `ALLCAPS` forms when string translation is the same. Is that necessary?
Because maintenance is more easy if the same translation string is only in one string key, use only one version could be a good option.
Answers:
username_1: where is that? could be by mistake
username_0: For example in `en-US.json`. Some repeated translation strings are:
````json
"INTRO_TASTE_COMEDY": "Comedy",
"Comedy": "Comedy",
"COMEDY": "Comedy",
"INTRO_TASTE_MUSIC": "Music",
"Music": "Music",
"MUSIC" : "Music",
"ADDONS": "Add-ons",
"WEBSITE_ADD_ONS": "Add-ons",
"INTRO_TASTE_HELP_US": "Help us learn more about you.",
"HELP_LEARN_MORE": "Help us learn more about you.",
"WEBSITE_CREATE_AN_ADDON": "Create an add-on",
"WEBSITE_ADDONS_CREATE": "Create an add-on",
````
And because this is the template for all you can see in more files.
username_0: Hi @username_1 , also FAQ13 and FAQ18 are the same.
username_0: "WEBSITE_LINK_ADDON_SDK" is a duplicate object key. |
inspursoft/board | 769697987 | Title: change node group label failed
Question:
username_0: In GitLab by @yhua123 on May 31, 2018, 17:48
Can't saw deployment and service yaml file.
Answers:
username_0: In GitLab by @tmon546596046 on Nov 8, 2019, 07:51
can't reproduce, so close it
username_0: In GitLab by @sokril on Nov 8, 2019, 08:21
closed
Status: Issue closed
|
godotengine/godot | 1159077436 | Title: Project can be opened multiple times from Project Manager
Question:
username_0: ### Godot version
3f69ea4
### System information
W10
### Issue description
When you double click a project on project list, it opens the project. But if you keep clicking it until project manager closes, it will spawn multiple instances of the editor.
If the project is starting, further clicks should have no effect.
### Steps to reproduce
1. Open Project Manager
2. Click a project on the list
3. Click it
4. Click again
5. Click
6. Click
7. Keep clicking
### Minimal reproduction project
_No response_
Answers:
username_1: Not bug but feature :D
username_2: Running multiple Godot editor instances *on the same project* isn't an officially supported use case. Either way, if you really need it, you can start multiple project manager instances (which can then start multiple editor instances on the same project).
Status: Issue closed
|
CDAT/cdat | 467352839 | Title: Build with hdf5=1.0.5
Question:
username_0: The latest packages on conda-forge are build with hdf5=1.0.5, but cdat v8.1 and its dependencies are built with hdf5=1.0.4. Could a new build be made with hdv5=1.0.5 for compatibility with conda-forge?
Answers:
username_1: @username_2 can we do this?
username_2: The issue I had was netcdf 4.7 and other libraries (libcf, libcdms,...) I had to pin netcdf=4.6.2 Last time I tried to upgrade libnetcf to 4.7 some conda-forge libraries were still at 4.6.2, I can try again and see if I can make it work. Issues often arise on OSX, but not Linux due to the some version metadata.
username_0: @username_2, I believe this is the official place to see what the current conda-forge pinnings are:
https://github.com/conda-forge/conda-forge-pinning-feedstock/blob/master/recipe/conda_build_config.yaml
If you look there, they're still using libnetcdf 4.6.2:
https://github.com/conda-forge/conda-forge-pinning-feedstock/blob/master/recipe/conda_build_config.yaml#L455
But hdf5=1.10.5:
https://github.com/conda-forge/conda-forge-pinning-feedstock/blob/master/recipe/conda_build_config.yaml#L416
So I don't believe you need to update to libnetcdf 4.7 in order to use hdf5 1.10.5. conda-forge is not.
username_0: I would imagine it would make all of our lives easier if more cdat packages made their way to conda-forge. If you're interested in doing that and need another maintainer to help with the process, let me know. I'm vaguely aware that some packages are on conda-forge and others are not but I'm not quite clear on which and why.
username_1: @username_0 we have recipe for all but vtk-cdat pkg takes too long to build and is killed by conda-forge... Hence we could never push it to conda-forge. And therefore all of our graphics are NOT on conda-forge. But this is not the issue here as cdms2 is on conda-forge.
username_2: @username_0 I had to recompile CDMS2 2 weeks ago since conda-forge changed the `gcc` compiler to `clang` on OSX. The `cdat/label/nightly` channel has 1.10.5 for hdf5. I could try to make a point release on conda-forge. @username_1 is this ok with you?
username_0: @username_1, ah, that makes sense. I can see why vtk-cdat might be a problem beyond building -- it vendors a lot of other stuff from what I recall and conda-forge doesn't allow that. Thanks for the clarification.
username_0: cdms2 from conda-forge is working fine with hdf5=1.10.5. I'll try to figure out which package wasn't happy with that hdf5 version. I assume it was something in `cdat/label/v81` but I didn't figure out for sure which.
username_0: I'm not sure if either of you did anything, but I am now able to build the environment I need with `hdf5=1.10.5`. Not sure what changed or what wasn't working before. But thank you both either way!
Status: Issue closed
|
grails/grails-core | 115082757 | Title: forward does not work in Grails 3.0.9
Question:
username_0: forward command does not work in Grails 3.0.9
To reproduce the issue:
*1. create a new application with grails 3.0.9*
*2. create a controller like following:*
class ForwardController {
def index() { }
def toAim() {
forward action: 'aim'
}
def aim() {
}
}
*3. create two views like following*
**index.gsp:**
<html>
<head>
<title>INDEX</title>
</head>
<body>
<p>Hello INDEX</p>
<g:link action="toAim">toAim</g:link>
</body>
</html>
**aim.jsp:**
<html>
<head>
<title>AIM</title>
</head>
<body>
<p>Hello AIM</p>
</body>
</html>```
Answers:
username_1: I'm confused you're forwarding to an action and expecting it to go to a JSP view?
username_0: Oh, sorry. It's a .gsp view. Corresponding to the action in the controller. I change it in the issue
username_1: This works fine and we have a functional test covering this behaviour https://github.com/grails/grails3-functional-tests/blob/master/app1/src/integration-test/groovy/functionaltests/ForwardingSpec.groovy#L46
If you can come up with an example that reproduces the problem we can take another look
Status: Issue closed
username_0: @username_1
https://www.dropbox.com/s/7urqdklrrqdcf6i/testForward.zip?dl=0
check out this example... |
junegunn/fzf | 1112696083 | Title: toggle whether to show hidden files on the fly
Question:
username_0: <!-- ISSUES NOT FOLLOWING THIS TEMPLATE WILL BE CLOSED AND DELETED -->
<!-- Check all that apply [x] -->
- [x] I have read through the manual page (`man fzf`)
- [x] I have the latest version of fzf
- [x] I have searched through the existing issues
## Info
- OS
- [x] Linux
- [ ] Mac OS X
- [ ] Windows
- [ ] Etc.
- Shell
- [ ] bash
- [x] zsh
- [ ] fish
## Problem / Steps to reproduce
How can I toggle between show hidden files and not show hidden files on the fly?
I know I can set whether to show them in FZF_DEFAULT_COMMAND using `--hidden` flag (for `fd`)
But I think to toggle this on the fly is more useful.
Answers:
username_1: You can set up `reload` bindings to update the input list dynamically.
See https://github.com/username_1/fzf/blob/master/ADVANCED.md#dynamic-reloading-of-the-list
Status: Issue closed
|
prolike/prolike.io | 590988251 | Title: Write the Gigs
Question:
username_0: ### Problem
Right now we don't have any stories of our customers, we don't have anything to show for it.
### Solution
We write down in the gigs all the good stories we have of our customers, the relationships, our workflow with them and what we have accomplished.
### Value
We got an archive of good and successful stories of our customers, that we can post, share and reference/show to potential customers who is possibly looking for the same solution, whether it is IT support, a website or software.
#### Witch customers should we use?
- [ ] Copenhagen Coaching Center
- [ ] Cykling Uden Alder
- [ ] Erhvervshusene (Hvad hedder de??)
- [ ] IT System Design
- [ ] Omnihorse
- [ ] Champagnekælderen
- [ ] Oticon
- [ ] ..more?
Answers:
username_0: @username_1
Do we need permission from our customers to write these stories?
username_1: Make it whole!
As in #626
Whats that plan? Why do we do it? what are we hoping to achieve? think "Landing page" and SoMe ...and alla that jazz!
username_1: Quote template: https://docs.google.com/document/d/1-kX5RCPnM2ctLaPDpJlVYgXK_Z1Eb5pQYPP4bkE8lfY
username_1: Maybe develop a format for what a gig story is?
username_1: More Issues?
* SoMe campaign?
* Update the gig layout?
* Improve the call-to-action? |
patternfly/patternfly-org | 469251868 | Title: Fix a11y issues in Documentation/React/Wizard
Question:
username_0: https://www.patternfly.org/v4/documentation/react/components/wizard/
There is 1 type of a11y issue in this page:
1. The skip-link target should exist and be focusable
<img width="486" alt="Screen Shot 2019-07-17 at 10 32 25 AM" src="https://user-images.githubusercontent.com/20118816/61384229-303a8d00-a87e-11e9-97d0-b4757c726353.png">
Steps to reproduce the a11y errors:
1. Right click on the page and choose inspect
2. In the developer tools, choose "axe" from the top menu
3. Run the test<issue_closed>
Status: Issue closed |
dominique-mueller/angular-package-builder | 288481736 | Title: fix(compile): build error when not in debug mode
Question:
username_0: Hi,
First of all, thanks for this awesome package builder which is very useful 😍
But, I think I've found a bug / regression in the lib. All the code I present here is extracted from a demo repo I've create for this issue (https://github.com/username_0/demo-angular-package-builder)
When I use dependencies coming from Angular (or maybe from something else, I haven't test...), the build doesn't work with the following result:
```shell
iMac-de-Kevin:~/W/g/d/demo-angular-package-builder kevin $ yarn build
yarn run v1.3.2
$ angular-package-builder
Angular Package Builder
➜ Configuration
➜ Prepare (line breaks, resources)
➜ Compile TypeScript into JavaScript (ES2015, ES5)
ERROR: An error occured while trying to compile the TypeScript sources using the Angular Compiler.
[TypeScript] : Error: ENOENT: no such file or directory, realpath '/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/@angular/core/core.d.ts'
[TypeScript] at ImportDeclaration in /Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/dist-angular-package-builder/lib/index.ts:1:1
[TypeScript] at SourceFile in /Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/dist-angular-package-builder/lib/index.ts:1:1
[TypeScript] at createError (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:105:17)
[TypeScript] at throwError (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:114:11)
[TypeScript] at Volume.realpathBase (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:1098:13)
[TypeScript] at Volume.realpathSync (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:1106:21)
[TypeScript] at Object.realpath (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:3989:32)
[TypeScript] at Object.realpath (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:69063:76)
[TypeScript] at realpath (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26373:42)
[TypeScript] at tryResolve (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26360:87)
[TypeScript] at nodeModuleNameResolverWorker (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26342:68)
[TypeScript] at nodeModuleNameResolver (/Users/kevin/Workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26321:16)
error Command failed with exit code 1.
info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
```
But, if I run the process in debug mode, without any modification, no error appears and the lib are correctly built.
```
iMac-de-Kevin:~/W/g/d/demo-angular-package-builder kevin $ yarn build --debug
yarn run v1.3.2
$ angular-package-builder --debug
Angular Package Builder
➜ Configuration
➜ Prepare (line breaks, resources)
➜ Compile TypeScript into JavaScript (ES2015, ES5)
➜ Generate JavaScript bundles (ES2015, ES5, UMD)
➜ Compose package
Success! [4.15 seconds]
✨ Done in 4.35s.
```
This is the full example where I run the command one after the other

The `index.ts` of my app is very simple :
```ts
import {InjectionToken} from "@angular/core";
export const foo = new InjectionToken('Foo');
```
And, of course, the error shown during the standard build process isn't coherent, because I have the file locally on my file system.

If you need more information, let me know 😉
/cc @neonox31 @Yann29 @yakoust and @aymeric-duchein
Answers:
username_1: First, thanks for making us aware of your issue, appreciate it very much :)
I actually have a few questions: Which versions of Angular and TypeScript are you using? What NodeJS version do you have? And finally, did it ever work properly or did the bug came in later on, e.g. after updating dependencies?
username_1: I also see that you're using Yarn and MacOS, whereas I always developed with npm on Windows. Yarn shouldn't be a problem, MacOS might be though ...
username_0: Angular Package Builder
➜ Configuration
➜ Prepare (line breaks, resources)
➜ Compile TypeScript into JavaScript (ES2015, ES5)
ERROR: An error occured while trying to compile the TypeScript sources using the Angular Compiler.
[TypeScript] : Error: ENOENT: no such file or directory, realpath '/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/@angular/core/core.d.ts'
[TypeScript] at ImportDeclaration in /Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/dist-angular-package-builder/lib/index.ts:1:1
[TypeScript] at SourceFile in /Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/dist-angular-package-builder/lib/index.ts:1:1
[TypeScript] at createError (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:105:17)
[TypeScript] at throwError (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:114:11)
[TypeScript] at Volume.realpathBase (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:1098:13)
[TypeScript] at Volume.realpathSync (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/memfs/lib/volume.js:1106:21)
[TypeScript] at Object.realpath (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:3989:32)
[TypeScript] at Object.realpath (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:69063:76)
[TypeScript] at realpath (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26373:42)
[TypeScript] at tryResolve (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26360:87)
[TypeScript] at nodeModuleNameResolverWorker (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26342:68)
[TypeScript] at nodeModuleNameResolver (/Users/kevin/workspace/github.com/username_0/demo-angular-package-builder/node_modules/typescript/lib/typescript.js:26321:16)
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! [email protected] build: `angular-package-builder`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the [email protected] build script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! /Users/kevin/.npm/_logs/2018-01-15T07_22_54_835Z-debug.log
MacBook-Pro-de-Kevin:~/w/g/d/demo-angular-package-builder kevin | master $
```
And, I don't right now have any Windows machine to test it... maybe @yakoust or @neonox31 can test it for me 😀
username_2: Angular Package Builder
→ Configuration
→ Prepare (line breaks, resources)
→ Compile TypeScript into JavaScript (ES2015, ES5)
ERROR: An error occured while trying to compile the TypeScript sources using the Angular Compiler.
[TypeScript] : Error: ENOENT: no such file or directory, realpath 'C:/dev /github/demo-angular-package-builder/node_modules/@angular/core/core.d.ts'
[TypeScript] at ImportDeclaration in C:/dev/github/demo-angular-package- builder/dist-angular-package-builder/lib/index.ts:1:1
[TypeScript] at SourceFile in C:/dev/github/demo-angular-package-builder /dist-angular-package-builder/lib/index.ts:1:1
[TypeScript] at createError (C:\dev\github\demo-angular-package-build er\node_modules\memfs\lib\volume.js:105:17)
[TypeScript] at throwError (C:\dev\github\demo-angular-package-builde r\node_modules\memfs\lib\volume.js:114:11)
[TypeScript] at Volume.realpathBase (C:\dev\github\demo-angular-packa ge-builder\node_modules\memfs\lib\volume.js:1098:13)
[TypeScript] at Volume.realpathSync (C:\dev\github\demo-angular-packa ge-builder\node_modules\memfs\lib\volume.js:1106:21)
[TypeScript] at Object.realpath (C:\dev\github\demo-angular-package-b uilder\node_modules\typescript\lib\typescript.js:3989:32)
[TypeScript] at Object.realpath (C:\dev\github\demo-angular-package-b uilder\node_modules\typescript\lib\typescript.js:69063:76)
[TypeScript] at realpath (C:\dev\github\demo-angular-package-builder\ node_modules\typescript\lib\typescript.js:26373:42)
[TypeScript] at tryResolve (C:\dev\github\demo-angular-package-builde r\node_modules\typescript\lib\typescript.js:26360:87)
[TypeScript] at nodeModuleNameResolverWorker (C:\dev\github\demo-angu lar-package-builder\node_modules\typescript\lib\typescript.js:26342:68)
[TypeScript] at nodeModuleNameResolver (C:\dev\github\demo-angular-pa ckage-builder\node_modules\typescript\lib\typescript.js:26321:16)
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! [email protected] build: `angular-package-builder`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the [email protected] build script.
npm ERR! This is probably not a problem with npm. There is likely additional log ging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! C:\Users\Admin_WebFactory\AppData\Roaming\npm-cache\_logs\2018-01-1 5T07_30_59_890Z-debug.log
```
username_1: Thanks for the info!
So, it turns out that Angular 5.2 breaks but Angular 5.1 works as intended. Further analysis showed that it's probably an issue with `tsickle` - Angular 5.1 uses tsickle 0.25.6 and Angular 5.2 uses 0.26.0. Either downgrading to Angular 5.1 or setting `annotateForClosureCompiler` to `false` fix this issue, at least temporarily (or using debug mode, of course).
I will try to find a fix over the course of the next few days.
username_1: Changes: https://github.com/angular/tsickle/compare/7d20a06330e342eb72ccb729da7744171082e8ea...82992d1b5afac4948cc3b56344251d4ccbc0ded5
username_0: The fact that the `--debug` flag allows to do the compilation is still strange. Do you think the lib is well built after this or some errors are still present in the package ?
username_1: The `--debug` flag doesn't change the build chain itself, it only decides whether interim build outputs should be emitted to the disk (if enabled) - or kept in memory (if disabled). To make this work, however, I had to use some not-so-fancy techniques of which one seems to break with the latest 'tsickle' ...
So no worries, the angular-package-builder still does it's job, even with the `debug` flag enabled :)
Status: Issue closed
username_1: Fixed with https://github.com/username_1/angular-package-builder/pull/41, will also make a documentation update regarding Angular & TypeScript compatibility - and hopefully publish a new release tomorrow.
Also, the CI now covers all Angular & TypeScript combinations.

username_1: Here we go, version `1.0.1` is published! Have fun!
Changelog: https://github.com/username_1/angular-package-builder/blob/master/CHANGELOG.md |
makewebau/shipper | 270116890 | Title: Does not work with themes
Question:
username_0: 
No version number (to be expected)
Not stripping out .git (weird because it takes out .shipignore)
Status: Issue closed
Answers:
username_0: @unetics Could not replicate the failure to delete `.git` folder, but the issues with the version should now be resolved. |
frioux/DBIx-Class-DeploymentHandler | 88491250 | Title: (minor) Readme mis-generated
Question:
username_0: https://metacpan.org/source/FREW/DBIx-Class-DeploymentHandler-0.002216/README#L161
Answers:
username_1: As of https://metacpan.org/source/ETJ/DBIx-Class-DeploymentHandler-0.002223/README this is no longer the case! (Not my doing though) Thanks for the report.
Status: Issue closed
|
github/docs | 928083359 | Title: Some filenames are too long for Git for Windows
Question:
username_0: ### What article on docs.github.com is affected?
`translations/<lang>/content/github/setting-up-and-managing-your-enterprise-account/setting-policies-for-organizations-in-your-enterprise-account/configuring-the-retention-period-for-github-actions-artifacts-and-logs-in-your-enterprise-account.md`
(German, Japanese, Brazilian Portuguese, and Chinese translation)
### What part(s) of the article would you like to see updated?
Git for Windows may complain about the file paths being too long, e.g. if you attempt to checkout the latest main branch of this repo. The length of just the relative path is 244 characters, with the limit being 260, but the absolute path is what counts (so you have to include `C:\Users\<name>\Documents\...` or wherever your working copy is located at.
- Cloning the repo directly to `C:\` would shorten the path, but it's just a workaround.
- Shortening the folder names and the file name of the actual content would be possible, but alter doc URLs and would make the URL slug less self-descriptive and SEO-friendly.
- [Git supports long paths](https://stackoverflow.com/a/22575737/2044940) (apparently at the expense of compatibility), but it needs to be enabled manually, e.g. with `git config core.longpaths true` for the current repo. This worked for me. I suggest to add a remark about this to the `CONTRIBUTING.md`.
### Additional information
```
error: cannot stat 'translations/de-DE/content/github/setting-up-and-managing-your-enterprise-account/setting-policies-for-organizations-in-your-enterprise-account/configuring-the-retention-period-for-github-actions-artifacts-and-logs-in-your-enterprise-account.md': Filename too long
error: cannot stat 'translations/ja-JP/content/github/setting-up-and-managing-your-enterprise-account/setting-policies-for-organizations-in-your-enterprise-account/configuring-the-retention-period-for-github-actions-artifacts-and-logs-in-your-enterprise-account.md': Filename too long
error: cannot stat 'translations/pt-BR/content/github/setting-up-and-managing-your-enterprise-account/setting-policies-for-organizations-in-your-enterprise-account/configuring-the-retention-period-for-github-actions-artifacts-and-logs-in-your-enterprise-account.md': Filename too long
error: cannot stat 'translations/zh-CN/content/github/setting-up-and-managing-your-enterprise-account/setting-policies-for-organizations-in-your-enterprise-account/configuring-the-retention-period-for-github-actions-artifacts-and-logs-in-your-enterprise-account.md': Filename too long
```
Answers:
username_1: Thanks for the issue @username_0! I'll triage this for the team to take a look 👀
username_2: In the [Windows](https://github.com/github/docs/blob/main/CONTRIBUTING.md#windows) section possibly add a `curl` command to download the .zip file.
username_3: curl
username_4: @username_0 Thank you for reporting this and explaining some of the viable workarounds. We have added this issue to our backlog. 👍🏻
username_5: > Depending on your Windows Version
>> Steps: to fix |-- >
> Step A:
> 1. Open: cmd.exe
> 2. Run: regedit.exe
> 3. Edit: HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\FileSystem
> 4. Look: for LongPathsEnabled (If Value Missing)
> |--> Do Step B:
> 4a. Right click : "FileSystem" Folder
> 4b. Choose: New Dword Value
> 4c. Add: "LongPathsEnabled" text
> Else:
> 5. Change Value 0 To 1
> 6. Then : sound=Applause
Author: username_5 <the_kat6<EMAIL>>
AuthorDate: 2021-July-29 12:52:00 -0500
Commit: username_5 <<EMAIL>>
CommitDate: 2021-July-29 12:52:00 -0500
username_1: We've added information about this to our contributing guide! https://github.com/github/docs-internal/blob/main/CONTRIBUTING.md#windows
Status: Issue closed
|
facebook/flow | 132446891 | Title: Type narrowing on function args doesn't work as expected
Question:
username_0: ```js
function getValue(): Promise<{prop: {id: string, name: string}}> {
return Promise.resolve({prop: {id: 'whatever', name: 'whatever'}});
}
function onValue(value: {prop: {id: string}}) {
return value.prop.id;
}
getValue().then(onValue);
```
Gives me the error:
```js
function getValue(): Promise<{prop: {id: string, name: string}}> {
^^^^^^^^^^^^^^^^^^^^^^^^^^ property `name`. Property not found in
function onValue(value: {prop: {id: string}}) {
^^^^^^^^^^^^ object type
```
The type of both functions is correct, the `onValue` intentionally identifies that it is not interested in `value.prop.name`. Incidentally, this works for properties that are only one level deep.
Answers:
username_1: This seems to be related to how we type object properties.
You can simplify this further to the following (which suffers from the same errors):
```javascript
function getValue(): {prop: {id: string, name: string}} {
return {prop: {id: 'whatever', name: 'whatever'}};
}
function onValue(value: {prop: {id: string}}) {
return value.prop.id;
}
let obj: {prop: {id: string}} = getValue();
```
Status: Issue closed
username_0: This can now be handled by simply marking `prop` as covariant:
```js
function getValue(): Promise<{prop: {id: string, name: string}}> {
return Promise.resolve({prop: {id: 'whatever', name: 'whatever'}});
}
function onValue(value: {+prop: {id: string}}) {
return value.prop.id;
}
getValue().then(onValue);
```
Or in your minimal example:
```js
function getValue(): {prop: {id: string, name: string}} {
return {prop: {id: 'whatever', name: 'whatever'}};
}
let obj: {+prop: {id: string}} = getValue();
``` |
wso2/carbon-dashboards | 358734164 | Title: Report Generation / Export Feature does not work
Question:
username_0: **Description:**
Report Generation / Export Feature does not work
**Suggested Labels:**
Dashboard , PDF Generation

**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assingnees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
**OS, DB, other environment details and versions:**
**Steps to reproduce:**
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --><issue_closed>
Status: Issue closed |
ActoKids/AD440_W19_CloudPracticum | 407560754 | Title: Configure event trigger Google Calendar
Question:
username_0: Triggers will need to be set for the google calendar crawler. Please use Cloud Watch to create a trigger based on time, and on URL. You have already done something similar in #44 but this time work with Ryan to create several triggers in case he needs to break up his Google Crawler.
Answers:
username_1: Update: Now have access to project AWS account but I do not currently have permissions for Cloudwatch in order to create the trigger. I need permission to call **CloudWatch Events API** so that I can create a trigger.
username_1: I seem to have all the permissions needed to complete my task. I'm currently waiting on Ryan's crawler code to be approved so that I can create the event trigger.
username_1: Successfully created trigger for GoogleCrawler on AWS.

username_1: At this point the Trigger is created and will trigger the lambda function every 24 hours. @rberry206 is working on get his code into what is currently an empty lambda function that was just created a few minutes ago by the DevOps team. This sprint assignment may need to be extended one week in order to finish testing. Will provide updates on this thread.
username_1:  |
actor-framework/actor-framework | 552375024 | Title: CAF produces unexpected log files
Question:
username_0: It seems like CAF produces a log file when hitting a `CAF_LOG_ERROR` even if CAF was compiled with log level set to quiet.
Answers:
username_0: Already fixed in `topic/0.17.4`. The branch `issue/1024-integration` contains the same fix for `master` that applies once #1007 got merged.
Status: Issue closed
|
symengine/symengine.py | 116175529 | Title: symengine.py does not support current symengine
Question:
username_0: Building symengine.py requires a specific commit of symengine. This is undocumented.
```
git checkout 3a1152fdafc83f8d1f35004b5bcda08f355e57c3
```
Answers:
username_1: Thanks @username_0 for reporting it. Indeed, we need to document this better, currently this is only "documented" here: https://github.com/symengine/symengine.py/blob/83c1a70cd2516a597a7a8a12ef0c0ce21ed825de/.travis.yml#L111
username_2: The README does it have now https://github.com/symengine/symengine.py/blob/master/README.md#installation
Also after this PR https://github.com/symengine/symengine.py/pull/29 there will be a `symengine_version.txt` file.
username_3: I have problems understanding the issue because current symengine.py does not compile with that version mentioned, while it does with current symengine. So, is this still an issue?
Status: Issue closed
username_4: Closing the issue as this is now documented in README.md
username_4: To clarify, `symengine.py` supports only the version mentioned in README.md, but current `symengine` master may work as well. `symengine.py` will be updated from time to time to work with the latest `symengine` master |
Anaconda-Platform/anaconda-client | 906688734 | Title: API call results too large
Question:
username_0: This is a feature request for the anaconda.org API. I don't see any repository where I could make a PR, so I asked around, and @chenghlee recommended posting here so that I reach the right people.
I noticed that the anaconda.org API sometimes returns huge results for packages with a lot of release files. For example, https://api.anaconda.org/package/conda-forge/dvc is 13 MB. (This is breaking some conda-forge shields/badges due to a 10 MB limit, see conda-forge/dvc-feedstock#215 ).
Currently, to compute the total number of downloads, the whole API result is downloaded and a sum is computed over all files.
In order to address the problem, I think it would be helpful to have more endpoints like
* `/packages/{owner_login}/{package_name}/latest-version`
* `/packages/{owner_login}/{package_name}/total-downloads`
* `/packages/{owner_login}/{package_name}/platforms`
* `/packages/{owner_login}/{package_name}/license`
These endpoints would match the functionality of the badges, and would be very useful to have in general. |
huggingface/transformers | 925628111 | Title: [performance] module init w/ `from_pretrained` skip storage allocation
Question:
username_0: # 🚀 Feature request
pt-1.9.0 added `torch.nn.utils.skip_init()` which (1) skips the module init (2) doesn't allocate any memory
https://pytorch.org/tutorials/prototype/skip_param_init.html
note: `torch.nn.utils.skip_init()` itself will be in 1.9.1, but the rest of the code should be in 1.9.0
We already implemented part 1 (skipping the custom init) in https://github.com/huggingface/transformers/pull/11471.
We could further speed up the start up time and reduce CPU memory usage, by not allocating any storage for module init since `load_state_dict` will already have allocated `state_dict` from the pretrained weights (and some sub-modules that don't have pre-trained weights - will have to go through normal init). See https://pytorch.org/tutorials/prototype/skip_param_init.html#implementation-details
another note: currently deepspeed needs to have the module storage pre-allocated for its `zero.Init` gather/scatter, but if the initial model's weights aren't allocated, then we can probably get rid of `zero.Init` altogether https://github.com/huggingface/transformers/issues/12273 |
pulibrary/figgy | 276262825 | Title: Titles in show view should show up on different lines instead of joined by `and`
Question:
username_0: Figgy:
<img width="263" alt="screen shot 2017-11-22 at 8 16 35 pm" src="https://user-images.githubusercontent.com/2806645/33158819-1b5ec072-cfc2-11e7-9831-1a5f293d0e90.png">
Plum:
<img width="170" alt="screen shot 2017-11-22 at 8 16 38 pm" src="https://user-images.githubusercontent.com/2806645/33158821-21099f10-cfc2-11e7-9e76-a1b0d000e517.png">
Answers:
username_1: In the breadcrumb area, we probably want to use the first title.
In the header above, we want all the titles to display, each on its own line (to gracefully handle RTL text)
username_2: Probably a different issue, but it's weird to have a breadcrumb with no links in the case where there's not a parent.
username_2: Other places that use header:
* member lists (substitute semicolon-separated list? so translated values will be shown?) (e.g. https://github.com/pulibrary/figgy/blob/master/app/views/catalog/_members_multi_volume_work.html.erb#L15)
* https://github.com/pulibrary/figgy/blob/master/app/models/concerns/linked_data/linked_ephemera_folder.rb#L84 ? use one-line list?
* https://github.com/pulibrary/figgy/blob/master/app/services/manifest_builder.rb#L50 ? one-line list?
username_2: uses of `heading` (currently implemented as first title):
* https://github.com/pulibrary/figgy/blob/master/app/document_presenters/valkyrie_show_presenter.rb#L4 used by blacklight (catalog_helper) to set heading on record page
that's really the only one I guess; I'll change this directly to insert the newlines
username_2: Not sure how that manifest `to_s` is used; that method isn't in plum. The manifests themselves look different in the label: https://plum.princeton.edu/concern/multi_volume_works/pb5647p872/manifest vs https://figgy.princeton.edu/concern/scanned_resources/9c145976-1525-4047-80dd-49a4b27d2457/manifest will look into how that's done.
username_2: remaining instances of the sentence format that I've noticed:
- [ ] manifest label
- [ ] search results show view (not sure where this comes from; it may be that blacklight uses 'header' and I'll just need to finally change how that's formatted (in plum it's separated with a comma)
username_2: I sort of think plum was doing the manifest label wrong. it seems like it should be a string. http://iiif.io/api/presentation/2.1/#label
Status: Issue closed
|
open-telemetry/opentelemetry-java | 685605587 | Title: BatchSpanProcessor performance goals
Question:
username_0: During recent optimisations of `BatchSpanProcessor` several questions have arisen regarding further performance optimisations. But we don't currently have proper benchmarks that could clearly demonstrate the benefit of any performance gains as opposed to induced code complexity. E.g. we cannot it automatic way compare the percent of dropped spans.
We have to come up with such benchmarking setup and agree on acceptable performance characteristics of `BatchSpanProcessor`.
Answers:
username_1: Probably related to #1438 .
username_1: There is also an issue to write micro-benchmarks for the BSP: #790
Status: Issue closed
username_1: We have multiple benchmarks now for the BSP, so I'm going to close this. |
nutonomy/nuscenes-devkit | 503468487 | Title: get the yaw in global coordinates from the yaw in local coordinate
Question:
username_0: Hi,
I find strange that the cameras orientation have large pitches and rolls. How is that possible?
I am trying to get the global yaw from only the yaw in camera coordinate (and not the pitch and roll). Do you have any idea how to do that?
Thanks for your inputs
Answers:
username_1: How much is the pitch and roll? If you look at the videos I'd say they are imperceptible and therefore pretty small: https://www.youtube.com/watch?time_continue=4&v=Sipud0IonvQ
username_0: This code shows that boxes in camera frame have a large pitch and roll. This is caused by the camera (calibrated sensor) rotation which has a non zero pitch and roll.
Am I missing something?
`
nusc = NuScenes(version='v1.0-trainval', dataroot="./nuScenes/v1.0-trainval", verbose=True)
my_scene = nusc.scene[0]
first_sample_token = my_scene['first_sample_token']
sensor = 'CAM_FRONT'
my_sample = nusc.get('sample', first_sample_token)
cam_front_data = nusc.get('sample_data', my_sample['data'][sensor])
my_annotation_token = my_sample['anns'][0]
box_global = nusc.get('sample_annotation', my_annotation_token)
nusc.get('sample_annotation', my_annotation_token)
data_path, boxes, camera_intrinsic = nusc.get_sample_data(my_sample['data'][sensor])
print('box orientation in global frame: ', box_global['rotation'])
print('box orientation in camera frame: ', boxes[0].orientation.elements)
sd_record = nusc.get('sample_data', my_sample['data'][sensor])
cs_record = nusc.get('calibrated_sensor', sd_record['calibrated_sensor_token'])
print('camera rotation:',cs_record['rotation'])`
username_1: Hi, I refactored your code a bit, used the mini split and converted from quaternions to euler angles:
```
from pyquaternion import Quaternion
import numpy as np
from nuscenes import NuScenes
nusc = NuScenes(version='v1.0-mini', dataroot="/data/sets/nuscenes-mini", verbose=True)
my_scene = nusc.scene[0]
first_sample_token = my_scene['first_sample_token']
sensor = 'CAM_FRONT'
my_sample = nusc.get('sample', first_sample_token)
cam_front_data = nusc.get('sample_data', my_sample['data'][sensor])
my_annotation_token = my_sample['anns'][0]
box_global = nusc.get('sample_annotation', my_annotation_token)
data_path, boxes, camera_intrinsic = nusc.get_sample_data(my_sample['data'][sensor])
sd_record = nusc.get('sample_data', my_sample['data'][sensor])
cs_record = nusc.get('calibrated_sensor', sd_record['calibrated_sensor_token'])
angles_box_global = np.array(Quaternion(box_global['rotation']).yaw_pitch_roll) / np.pi * 180
angles_box_camera = np.array(Quaternion(boxes[0].orientation.elements).yaw_pitch_roll) / np.pi * 180
angles_camera = np.array(Quaternion(cs_record['rotation']).yaw_pitch_roll) / np.pi * 180
print('box orientation in global frame: ', angles_box_global)
print('box orientation in camera frame: ', angles_box_camera)
print('camera rotation:', angles_camera)
```
Results are:
````
box orientation in global frame: [-21.091 0. 0. ]
box orientation in camera frame: [178.79077403 1.27037324 90.88688192]
camera extrinsics: [ 44.75652071 89.54131123 -134.80356665]
````
So the annotation lives in the global frame with no pitch or roll.
When transformed to camera frame we can see that the orientations are less than 2 degrees off.
The camera extrinsics are not easy to read as ego vehicle (IMU) and camera use different coordinate frames: https://www.nuscenes.org/public/images/data.png
Status: Issue closed
username_0: Thanks a lot, I did not thought the axis orientation were different between vehicle and sensor frame. |
finos/perspective | 770185325 | Title: Defining calculated column using 'final aggregate column values'
Question:
username_0: ## Support Question
Is there a way to define a calculated column that will be calculated using aggregated columns values rather than an aggregation over itself?
Example: I'm looking at a portfolio using underlying positions data. The portfolio level values are generated using a 'group by' clause.
However, something like fill rate won't aggregate properly.
fill rate = sum(executed_qty) / sum(qty)
I would need fill rate to be calculated using aggregate values of both sum(executed_qty) and sum(qty) rather than the individual executed_qty/qty
thx.
Answers:
username_1: This can only be done explicitly in the API by _flattening_ ([Docs here](https://perspective.finos.org/docs/md/js.html#flattening-a-view-into-a-table)) an aggregated `View` into a new `Table`, then adding a computed column. For your case, I'm assuming you are using a Perspective Web Component, so you just need to create the flat `Table` and `load()` it, something like this:
```javascript
const agg_view = table.view({
row_pivots: ["portfolio"],
columns: ["executed_qty", "qty"]
});
workspace.load(worker.table(agg_view));
```
This is not currently possible through the GUI alone, starting with just `table` - you must specify the intermediate aggregate table ahead of time. The plan is to add this as the `Flatten` command in the `<perspective-workspace>` right-click context menu.
Status: Issue closed
|
DJ73/ShellCrypt | 733766220 | Title: Add more pseudo random number generators
Question:
username_0: Check `middle_square_method.py` for format
Answers:
username_1: Assign this to me
username_1: Please explain what I need to do. Your readme file is not detailed
username_0: Need to create new files like `middle_square_method.py`. The file should have one generator function that returns a value in [0, 25] and accepts a seed to generate said values
username_1: Okay, I will see to it |
yegor256/syncem | 433096298 | Title: confusion with method send()
Question:
username_0: If the method's name is `send`, we get confused.
Answers:
username_0: @username_1 release, tag is `0.1.2`
username_1: @username_0 OK, I will release it now. Please check the progress [here](http://www.username_1.com/t/19302-504747552)
username_1: @username_0 Done! FYI, the full log is [here](http://www.username_1.com/t/19302-504747552) (took me 2min)
Status: Issue closed
|
hurlbertlab/dietdatabase | 771139823 | Title: complete search for Parulidae diet data
Question:
username_0: The following species in Parulidae are missing from the `NA_avian_diet_refs.txt` file and therefore may never have been intentionally searched:
Golden-winged warbler
Black-throated gray warbler
Yellow-throated warbler
Northern Parula*
Tropical Parula
Ovenbird*
Louisiana waterthrush*
Connecticut warbler*
Mourning warbler*
Canada warbler*
For each of the above species,
1) go to https://birdsoftheworld.org/bow/home (subscription only, so you'll need to be connected to UNC's VPN), and visit the species' *Diet and Foraging* page.
2) Look up each of the references cited on that page, and add them to the `NA_avian_diet_refs.txt` and indicate whether they do or do not have any quantitative diet data. Pay attention to whether any studies are cited within those papers that might potentially contain such data.
3) On Google Scholar, put the following search terms in quotes: "[common name] AND (diet OR foraging OR food)", for example "golden-winged warbler AND (diet OR foraging OR food)". Follow up with any papers that have titles indicating they might contain quantitative data. Add all papers examined to the `NA_avian_diet_refs.txt` file regardless of whether they have data.
4) For species above with an *, some diet data exists currently in our Avian Diet Database. Be sure to check that any papers you find in your literature search are not already entered.
Answers:
username_0: Note that "golden-winged warbler" has actually been listed as its own issue, so ignore that species here. Update in this issues thread as you finish searching a species on the list above.
username_1: Finished with Black-throated Gray Warbler
Status: Issue closed
|
Radnen/spherestudio | 421686467 | Title: Null reference exception when creating script in blank project
Question:
username_0: Just opened the editor for the first time, set plugins for everything except the option for the default editor to use with unknown file types. Then created a script, and got a null reference exception.
Log:
```************** Exception Text **************
System.NullReferenceException: Object reference not set to an instance of an object.
at SphereStudio.Ide.DocumentTab..ctor(IdeWindow ide, DocumentView view, String fileName, Boolean restoreView)
at SphereStudio.Ide.IdeWindow.AddDocument(DocumentView view, String filepath, Boolean restoreView)
at SphereStudio.Ide.IdeWindow.<>c__DisplayClass52_0.<menuNew_DropDownOpening>b__3(Object s, EventArgs ea)
at System.Windows.Forms.ToolStripItem.RaiseEvent(Object key, EventArgs e)
at System.Windows.Forms.ToolStripMenuItem.OnClick(EventArgs e)
at System.Windows.Forms.ToolStripItem.HandleClick(EventArgs e)
at System.Windows.Forms.ToolStripItem.HandleMouseUp(MouseEventArgs e)
at System.Windows.Forms.ToolStrip.OnMouseUp(MouseEventArgs mea)
at System.Windows.Forms.ToolStripDropDown.OnMouseUp(MouseEventArgs mea)
at System.Windows.Forms.Control.WmMouseUp(Message& m, MouseButtons button, Int32 clicks)
at System.Windows.Forms.Control.WndProc(Message& m)
at System.Windows.Forms.ToolStrip.WndProc(Message& m)
at System.Windows.Forms.ToolStripDropDown.WndProc(Message& m)
at System.Windows.Forms.NativeWindow.Callback(IntPtr hWnd, Int32 msg, IntPtr wparam, IntPtr lparam)
************** Loaded Assemblies **************
mscorlib
Assembly Version: 4.0.0.0
Win32 Version: 4.7.3362.0 built by: NET472REL1LAST_C
CodeBase: file:///C:/Windows/Microsoft.NET/Framework64/v4.0.30319/mscorlib.dll
----------------------------------------
Sphere Studio
Assembly Version: 0.0.0.0
Win32 Version: 0.0.0.0
CodeBase: file:///C:/Program%20Files/miniSphere/ide/Sphere%20Studio.exe
----------------------------------------
System.Windows.Forms
Assembly Version: 4.0.0.0
Win32 Version: 4.7.3324.0 built by: NET472REL1LAST_C
CodeBase: file:///C:/WINDOWS/Microsoft.Net/assembly/GAC_MSIL/System.Windows.Forms/v4.0_4.0.0.0__b77a5c561934e089/System.Windows.Forms.dll
----------------------------------------
System
Assembly Version: 4.0.0.0
Win32 Version: 4.7.3362.0 built by: NET472REL1LAST_C
CodeBase: file:///C:/WINDOWS/Microsoft.Net/assembly/GAC_MSIL/System/v4.0_4.0.0.0__b77a5c561934e089/System.dll
----------------------------------------
System.Drawing
Assembly Version: 4.0.0.0
Win32 Version: 4.7.3056.0 built by: NET472REL1
CodeBase: file:///C:/WINDOWS/Microsoft.Net/assembly/GAC_MSIL/System.Drawing/v4.0_4.0.0.0__b03f5f7f11d50a3a/System.Drawing.dll
----------------------------------------
SphereStudio.Base
Assembly Version: 0.0.0.0
Win32 Version: 0.0.0.0
CodeBase: file:///C:/Program%20Files/miniSphere/ide/SphereStudio.Base.DLL
----------------------------------------
System.Core
Assembly Version: 4.0.0.0
Win32 Version: 4.7.3362.0 built by: NET472REL1LAST_C
CodeBase: file:///C:/WINDOWS/Microsoft.Net/assembly/GAC_MSIL/System.Core/v4.0_4.0.0.0__b77a5c561934e089/System.Core.dll
----------------------------------------
System.Configuration
Assembly Version: 4.0.0.0
[Truncated]
CodeBase: file:///C:/Program%20Files/miniSphere/ide/ScintillaNET.DLL
----------------------------------------
************** JIT Debugging **************
To enable just-in-time (JIT) debugging, the .config file for this
application or computer (machine.config) must have the
jitDebugging value set in the system.windows.forms section.
The application must also be compiled with debugging
enabled.
For example:
<configuration>
<system.windows.forms jitDebugging="true" />
</configuration>
When JIT debugging is enabled, any unhandled exception
will be sent to the JIT debugger registered on the computer
rather than be handled by this dialog box.
``` |
tdurieux/leboncoin-api | 372082926 | Title: TypeError: Cannot read property 'value' of undefined
Question:
username_0: Hello
When I try the example provided in main page (https://github.com/username_1/leboncoin-api), I have the following error :
```
C:\lbc\node_modules\leboncoin-api\lib\search.js:235
searchLabel = searchId[0].value;
^
TypeError: Cannot read property 'value' of undefined
at search.map.searchLabel (C:\lbc\node_modules\leboncoin-api\lib\search.js:235:39)
at Array.map (<anonymous>)
at getIdsForLabels (C:\lbc\node_modules\leboncoin-api\lib\search.js:232:21)
at _.forEach (C:\lbc\node_modules\leboncoin-api\lib\search.js:277:30)
at C:\lbc\node_modules\lodash\lodash.js:4911:15
at baseForOwn (C:\lbc\node_modules\lodash\lodash.js:2996:24)
at C:\lbc\node_modules\lodash\lodash.js:4880:18
at Function.forEach (C:\lbc\node_modules\lodash\lodash.js:9344:14)
at LodashWrapper.object.(anonymous function) [as forEach] (C:\lbc\node_modules\lodash\lodash.js:15716:25)
at getSubCategoriesExtras (C:\lbc\node_modules\leboncoin-api\lib\search.js:269:21)
at Search.getBodyParams (C:\lbc\node_modules\leboncoin-api\lib\search.js:324:22)
at Search.run (C:\lbc\node_modules\leboncoin-api\lib\search.js:369:51)
at Object.<anonymous> (C:\lbc\lbc.js:18:8)
at Module._compile (internal/modules/cjs/loader.js:688:30)
at Object.Module._extensions..js (internal/modules/cjs/loader.js:699:10)
at Module.load (internal/modules/cjs/loader.js:598:32)
```
Any idea why ?
Thanks for advance.
Sybaris
Answers:
username_1: That is strange, I never faced this error. And the example is working right now :/
I will try to fix this.
Thanks for the report.
username_2: Hello,
I have still the issue…
Any idea ?
Sybaris
username_0: Hello,
I have still the issue…
Any idea ?
Sybaris
Status: Issue closed
|
Captain-K-101/Ssrf-labs | 779390170 | Title: Could you possibly share writeup?
Question:
username_0: Hi Captain!
I am creating lab for SSRF. COuld you possibly share writeup for your tasks?
Thank you
Answers:
username_1: Hi There
yea sure , i'll be happy to but i am currently a bit busy, i'll send it in
the upcoming 2-3 days
*Regards *
<NAME>
Third Year
B.tech ,Computer Science
*Twitter <https://twitter.com/@Captainkay11> | Github
<https://github.com/username_1>*
Status: Issue closed
username_1: Hi Captain!
I am creating lab for SSRF. COuld you possibly share writeup for your tasks?
Thank you
username_2: Hi Captain,I think you forgot. I take the liberty to ask you for a writeup.
Thanks!
username_3: Can we fetch forbidden file using fopen() in php
username_1: Hey @username_2, sorry for the late response, but for writeup, i am currently swarmed with work so might not be able to make a writeup for this in the upcoming future but once free will definitely get on this .
The attacks are pretty simple to exploit basic understanding of SSRF should be enough for exploitation
for any reference materials or stuff u can ping me on twitter @Captainkay11
username_1: @username_3 if u do have control over the fopen name parameter being passed it's possible to read files using basic lfi, but again there would be certain criteria too ,ie adequate permissions be there to read etc...
username_3: Bro i ask it that can we fetch forbidden file (flag.php ) using ssrf Vulernbility because when I try to fetch it it's only give me a forbidden error :( |
nuxt/modules | 821197505 | Title: Add Agility CMS Module
Question:
username_0: Hey there, Joshua from Agility CMS here 👋
Agility is a cloud-native, API-first CMS with built-in **Page Management** & **Sitemap Control** - learn more [here](https://help.agilitycms.com/hc/en-us/articles/360002996392)!
One of our folks recently wrote a module to help integrate Agility CMS with Nuxt.js and we're interested in getting it listed into the list of modules: https://github.com/agility/agilitycms-nuxt-module
This module pairs with our [Agility CMS & Nuxt.js Blog Starter](https://github.com/agility/agility-nuxt-blog).
Let me know how we can go about getting it added!
Thank you 😁<issue_closed>
Status: Issue closed |
baidu/amis | 631434624 | Title: 按钮添加disabledOn属性后,各行的操作按钮无法对齐
Question:
username_0: 
Answers:
username_1: 我记得我好想修复了,你换个新版试试?
username_0: 刚改了下依赖,然后本地重启了项目,还是不行

username_0: 现在好了
Status: Issue closed
|
python-cmd2/cmd2 | 813779919 | Title: Cmd2 Application wide settings file via cmd2 extension.
Question:
username_0: I've been writing a few application-specific configuration file modules in different ways and haven't found something quite elegant enough. Ideally, I'd like to have a cmd2 add in module that took care of adding these configuration issues. It would work like this.
When a cmd2 application starts up you would attempt to load a configuration file (user specified json, toml, ini.. etc) if successful their would be a Cmd.Configuration object available to all of cmd and subcommands and CommandSets via self._cmd.Config.
My reasoning is that I would like to have CommandSets that are API clients for specific endpoints but each of those endpoints shouldn't be hard coded. But I cant imagine using Settables to hold these configs unless I could read them from a config file at start up in the event a url changed.
Answers:
username_1: Settables are really designed for settings you want users to be able to change at runtime.
So far we have left how to implement at-startup configuration to developers and haven't attempted to provide anything to make this easier. Typically I have seen developers use either command-line arguments via `argparse`, some sort of configuration file (JSON, YAML, ini, etc.), or a combination thereof.
But I'm open to the concept of providing something to make this easier, particularly if it integrates well with CommandSets. Any solution needs to be completely cross-platform and work well on Windows, Mac, Linux, Python 3.6+, x86, ARM, etc.
username_2: @username_0 Since `CommandSets` already have access to their parent `cmd2` object, what is blocking your application from having a configuration structure at the top level which all `CommandSets` have access to?
I'm worried a feature like this isn't generic enough to add to the cmd2 framework since applications vary greatly on how they load config settings.
username_0: 
Here's my external plugin implementation. Because I am creating the CommandSet instances and passing them into the cmd2 app constructor I lose the ability for each Commandset to simply call self.config. I fully admit that I could be thinking of this incorrectly or misunderstanding how inheritance works..
Am I doing this correctly?
username_0: I could also be easily convinced that Configuration storage is NOT a framework issue but an implementation one. Perhaps I should just look more into how other applications are implementing persistent configurations.
When I wrote CLI apps pre cmd2 I would first look to a default ini file to load variables, If they were not present or available I made the arguments mandatory.
like.. if the ini had a 'port' variable that was set to 8080 then each command would automatically assume this is true. But I could override this variable with the cli argument --port.
I'm trying to implement something similar in my cmd2 apps. I want to load a file if present at app startup that then is available to all my other commandsets.
I'll write up a quick POC and share it.
username_0: https://github.com/username_0/cmd2-app-poc-plugin
Here is the POC..
https://github.com/username_0/cmd2-app-poc-plugin/blob/main/sample_plugin/src/sample_plugin/fruits.py
This function does indeed show that a global config object is accessible via self._cmd
def do_really_like_apples(self, _:Statement):
apple_setting = self._cmd.config['apple'].getboolean('like')
but this is NOT accessible via the __init__.
Is this by design?
username_2: @username_0 Yes, it is by design. `CommandSets` have a lifecycle and various events occur within this lifecycle.
Refer to the [event handler](https://cmd2.readthedocs.io/en/stable/features/modular_commands.html#event-handlers) documentation for a description of each event.
The `cmd2` object becomes available to the `CommandSet` upon registration (`on_register()`).
Here is a link to the `CommandSet` [code](https://github.com/python-cmd2/cmd2/blob/master/cmd2/command_definition.py) as well so you can see what each event handler does. The class is at the bottom of that file.
username_2: The main reason I advocate for a common config object being implementation-based is because `cmd2` itself has no common configuration settings that aren't already individual class members. Therefore `cmd2` would have this empty `self.config` member.
Another question that arises is, what type of object is `self.config`? Your example makes it a dictionary. Other developers may prefer a `namedtuple` or `dataclass`. I would not want to enforce such a standard from the framework level.
username_0: I can agree.. Perhaps it's just better to submit an advanced example. Enforcing a standard from the framework is indeed a "bad" thing but looking at all of the different cmd2 implementations there is a lot of rope available to hang one's self ;)
Status: Issue closed
|
imixs/imixs-docker | 481060865 | Title: Can I use another postgreSQL version?
Question:
username_0: I am using postgres:11-alpine docker image in my docker-compose.yml file. But your Dockerfile is using postgres:9.6.1.
When I run docker-compose up, it shows that:
pg_dump: server version: 11.2; pg_dump version: 9.6.11
pg_dump: aborting because of server version mismatch
So I must use an older version?
Answers:
username_1: no - you can run also newer versions of Postgress - Postgres:11 is fine
username_1: note: the image just imports the JDBC Postgres Driver. This driver is based on version 9.x but it will also work fine with newer Postgress Server versions like 11. But of course you can upgrade the JDBC driver also by overwriting the image, or we can do this in this project. We just need to replace the jar file with a newer version.
Status: Issue closed
|
yoheimuta/hubot-aws | 76321702 | Title: typo: hubot ec2 run --image-id=[ami-id]
Question:
username_0: `hubot ec2 run --image_id=[ami-id]` is right.
- https://github.com/username_0/hubot-aws/blob/master/scripts/ec2/run.coffee#L17
Answers:
username_0: Fixed in [Release v0.0.5 · username_0/hubot-aws](https://github.com/username_0/hubot-aws/releases/tag/v0.0.5)
Status: Issue closed
|
telegraphic/hickle | 745808539 | Title: Dumping to io.BufferedReader Fails
Question:
username_0: I am trying to dump to in-memory bytes so that I can then compress these bytes with zlib before writing to disk.
```python
raw = io.BytesIO()
writer = io.BufferedWriter(raw)
hickle.dump(obj, writer)
```
The last line raises `AttributeError: '_io.BytesIO' object has no attribute 'name'`.
Answers:
username_1: I think the error comes from these lines in file_opener function
```
if isinstance(f, (io.TextIOWrapper, io.BufferedWriter)):
filename, mode = f.name, f.mode
f.close()
mode = mode.replace('b', '')
h5f = h5.File(filename, mode)
```
instead of passing the file like object to h5py as is or at least specifiying driver as 'fileobj'
```
h5f = h5.File(f,mode,driver = 'fileobj')
```
it is tried to get the file name form the passed in TextIOWrapper or BufferedWrite (any other filelike objects are ignored) and its mode and than the file is closed and a new hdf5 file with specified file name and node is opened. Alternatively only an already opened h5py.File object or a plain file path string are accepted any thing else causes an exception. Possibly on finalization of next minor or major release somebody shall have a look to it h5py from 2.10 on definitely is capable of handling filelike objects as long as they provide read, seek, tell and write method. (see [h5py.File](https://docs.h5py.org/en/stable/high/file.html))
username_2: Oh, I guess I missed this one.
I am not entirely sure why that does not work, as we do test for it, but I will have a look.
However, as h5py was recently updated to 3.0.0, which brought a ton of changes (that are also incompatible with hickle), I am more planning on doing a pass over the entire package to account for that.
I will add this issue to that list, but it may take a while before it will be fixed.
username_1: @username_2 @username_3 just in case it might be of any interest to you or even any help at all i wanted to let you know that:
Beeing a bit boored while waiting for @username_3 to decide upon pull request #138 tried to do some proof of concept for handling file and file like objects as supported by h5py.
The results of this trial and error can also be found in the detached concept_memp_compact_expand branch of my hickle fork.
username_1: The same here, would be included in my finalize and cleanup pullrequest after #138, and upcomming for #139 and #145.
username_2: After looking into it, I realize that this is not an error.
`hickle` can solely be used to dump to HDF5-files.
A `BufferedWriter` is not an HDF5-file, so `hickle` cannot dump to it.
Status: Issue closed
username_0: Are there plans to support writing to in-memory bytes rather than files?
username_2: Not at the moment, no.
username_3: @username_0 just to chime in here: `hickle` is indeed designed specifically for dumping to HDF5 files, and uses `h5py` as its API -- which doesn't support `BufferedWriter`. If you really wanted a HDF5 file in memory, you could try setting up a ramdisk? However I think there are probably better solutions out there for in-memory data storage...
username_1: @username_3 @username_2 not so true, according to documentation for h5py 2.10 and onward they support any file like object which is capable of reading and writing binary data and which is seekable and `io.BytesIO` exactly full fills that, one can find that example in h5py [manual](https://docs.h5py.org/en/2.9.0/high/file.html#python-file-like-objects) . Thus the questions is rather is it worth the efforts to add all the required checks if passed in file-like object conforms to requirements of h5py or not and in worst remove support for file-like objects and Python file handles from `hickle` at all. Only supporting native h5py files.
@username_0 why would you need the `io,BufferedWriter`. `io.BytesIO` already a `io,BufferedIOBase` object like `io.BufferedWriter` and `io.BufferedReader` are and thus is already buffered. So replace `io.BufferedWriter` simply by `h5py.File` to make your example work.
```
raw = io.BytesIO
writer = h5py.File(raw)
hickle.dump(obj,writer,mode='w')
```
and on read
```
reader = h5py.File(raw)
hickle.load(obj,reader,mode='r')
```
So you see no need for `io,BufferedWriter` at all or in other words `h5py.File` acts as wrapping writer and reader.
username_2: @username_1 It is still true what we are saying.
It does not matter if h5py allows writing to other filetypes, hickle does not support it.
username_1: Yes you are right, bad wording from my side. What i wanted to say, is if it does not support it it should not allow to pass file objects and file-like objects at all. As how it is done now is broken and against expectations when passing file objects, with the consequence that this will not stay the only issue related to strange or broken support of file and file like objects.###
### An example:
```
fid = open('/tmp/somefile.h5','w+b')
hickle.dump(obj,fid)
somesocket.write(fid.read())
```
But that does not work as `hickle` will just takes the filename and closes the original file or file-like object and replaces the underlying file on disk with a completely new file with the same name and the default access rights files owned by the process running `hickle` not the ones of the original file and also not necessarily with the same rights of the original file. This does not make sense at all to me . Why should i first open a file which is never used or even worse when reading for an already written hickle file the file is deleted. When i open a file beforehand i want `hickle` to place the hdf5 file content exactly in that file and nothing else and the wrapping inside 'io.BufferedWriter' or 'io.TextIOWrapper' should not make any difference here. So either and that is what i meant take the decision to properly support file and file like-objects eg from hickle >= 5.0 on and take the efforts to fix it until then or decide not to support file and file-like objects at all beyond indirect support through passed in h5py.File objects. Than remove support completely only allowing filename strings and h5py.File objects to be passed. |
mampfes/hacs_waste_collection_schedule | 1177695454 | Title: Not Able to Pull data for Auckland
Question:
username_0: Hi, What a wonderful idea to automatically update it. Great concept. I think Documentation is lacking for new users. Probably would be ideal to have more step by step guides.
I have got the following config in YAML, took ages to generate it without errors. My sensors show up but they are blank, no information is getting pulled. Any help will be much appreciated.
waste_collection_schedule:
sources:
- name: aucklandcouncil_govt_nz
args:
area_number: 12342560766
customize:
- type: brown bin
show: no
- type: Rubbish
- type: Recycling
sensor:
# next collection
- platform: waste_collection_schedule
name: next_collection
# next yard waste collection
- platform: waste_collection_schedule
name: next_yard_collection
types:
- Rubbish
# next garbage collection
- platform: waste_collection_schedule
name: next_garbage_collection
types:
- Recycling
Answers:
username_1: Try (note that yaml is case sensitive and that indentation counts!):
```yaml
waste_collection_schedule:
sources:
- name: aucklandcouncil_govt_nz
args:
area_number: 12342560766
customize:
- type: brown bin
show: no
sensor:
- platform: waste_collection_schedule
name: next_collection
- platform: waste_collection_schedule
name: next_yard_collection
types:
- rubbish
- platform: waste_collection_schedule
name: next_garbage_collection
types:
- recycle
```
Could you explain the problems you had? This would be helpful to improve to documentation. Or even better: Provide a suggestion :-)
username_0: Sorry my code was right the formatting got stuffed when i pasted here. So its exactly same as you have mentioned above, configurations check out.
But when i insert the sensor in a button , it shows as Next Garbage collection (Unknown)
it means its not picking up the values from the website, how do i test or confirm that.
I will create a video for step by step to help you out. instead of suggestions.
username_1: I had to adjust `rubbish `and `recycle` to get it working! Please check if you use exactly the same values.
username_0: Hi, Thanks for the prompt reply, I used it exactly as mentioned above but when i put the sensor to home screen it has unknown value. How to Troubleshoot ?
how do i confirm if its picked up the value from that site ? and what variables or names its using.
Below is a copy paste from my Configuration.Yaml and attached is my button
````
waste_collection_schedule:
sources:
- name: aucklandcouncil_govt_nz
args:
area_number: 12342560766
customize:
- type: brown bin
show: no
sensor:
- platform: waste_collection_schedule
name: next_collection
- platform: waste_collection_schedule
name: next_yard_collection
types:
- rubbish
- platform: waste_collection_schedule
name: next_garbage_collection
types:
- recycle
````

username_1: Ok, strange. Here is a small step-by-step instruction:
1. Check that there are no errors in the HA logs.
2. Check that the dates appear in the HA calendar. It should look like this with your configuration:

3. Open the HA Developer Tools and check the state of the 3 sensors. They should look like this:

If I use the configuration above, the entity cards look like this:

username_0: Excellent Help, At least now i know where to go and find the info.
so I have it half working. Its only pulling the rubbish and rubbish, recycle both under same sensor.
how do i fix it or do i just leave that as is.

username_0: All working now, guess took a while to pull that through.
We can close this now please.
Thanks
Status: Issue closed
username_2: I wish I'd seen this sooner @username_0 as I could have helped too. Here is what I am using myself:
In configuration.yaml:
```
waste_collection_schedule:
sources:
- name: aucklandcouncil_govt_nz
args:
area_number: 12343981102
# from https://www.aucklandcouncil.govt.nz/rubbish-recycling/rubbish-recycling-collections/Pages/rubbish-recycling-collection-days.aspx
customize:
- type: rubbish
alias: Rubbish
icon: mdi:trash-can
- type: recycle
alias: Recycling
icon: mdi:recycle
calendar_title: Auckland Council Rubbish & Recycling Collections
day_switch_time: "17:00"
```
In my sensors file (I have sensors in a separate .yaml file, but you could have these in configuration.yaml under `sensors:`)
They are all slight tweaks on the same data so I can display it in different formats for different purposes - the "this week" set is what I use the most.
```
- platform: waste_collection_schedule
name: Next Rubbish Collection Date
types:
- Rubbish
value_template: '{{value.date.strftime("%d %b")}}'
- platform: waste_collection_schedule
name: Next Recycling Collection Date
types:
- Recycling
value_template: '{{value.date.strftime("%d %b")}}'
- platform: waste_collection_schedule
name: Next Rubbish Collection Days
types:
- Rubbish
value_template: '{% if value.daysTo == 0 %}Today{% elif value.daysTo == 1 %}Tomorrow{% else %}in {{value.daysTo}} days{% endif %}'
- platform: waste_collection_schedule
name: Next Recycling Collection Days
types:
- Recycling
value_template: '{% if value.daysTo == 0 %}Today{% elif value.daysTo == 1 %}Tomorrow{% else %}in {{value.daysTo}} days{% endif %}'
- platform: waste_collection_schedule
name: Next Rubbish Collection This Week
types:
- Rubbish
value_template: '{% if value.daysTo == 0 %}Today{% elif value.daysTo == 1 %}Tomorrow{% elif value.daysTo <= 6 %}This {{value.date.strftime("%A")}}{% else %}in {{value.daysTo}} days{% endif %}'
- platform: waste_collection_schedule
name: Next Recycling Collection This Week
types:
- Recycling
value_template: '{% if value.daysTo == 0 %}Today{% elif value.daysTo == 1 %}Tomorrow{% elif value.daysTo <= 6 %}This {{value.date.strftime("%A")}}{% elif value.daysTo < 13 %}Next {{value.date.strftime("%A")}}{% else %}in {{value.daysTo}} days{% endif %}'
```
Glad you have it working now! |
Apicurio/apicurito | 984949028 | Title: Operator projects using the removed APIs in k8s 1.22 requires changes.
Question:
username_0: For further information and tips see the [comment](https://github.com/redhat-openshift-ecosystem/community-operators-prod/discussions/138#discussioncomment-1245759).
Answers:
username_1: The code for the apicurito operator has moved to Apicurio/apicurio-operators and has been updated to remove all deprecated apis. Please see [here](https://github.com/Apicurio/apicurio-operators/commit/8ffd67465a1fb1c20c249a705f9131207bffa2cf).
username_0: So, can you please publish it on the community repo https://github.com/redhat-openshift-ecosystem/community-operators-prod/tree/main/operators/apicurito ?
username_1: New version of apicurito underway:
https://github.com/redhat-openshift-ecosystem/community-operators-prod/pull/216
Status: Issue closed
|
power-assert-js/espower-typescript | 435535316 | Title: "directories" property in package.json conflicting with other npm package
Question:
username_0: Currently, "directories" property is used to specify the arbitrary test directory.
[electron-builder](https://github.com/electron-userland/electron-builder) used the same property in the past and it is deprecated now. When I try to use electron-builder with"directories" property specified in package.json, electron-builder asserts that "directories" property is deprecated and aborts.
The problem is that "directories" is overly general. If it is specified like this, it will never conflict with other packages.
```
{
"name": "your-module",
"description": "Your module",
"version": "0.0.1",
"espower-typescript": {
"directories": {
"test": "spec/"
},
}
...
}
```
Is it possible? |
hankcs/HanLP | 712912638 | Title: Which word segmentation is faster in python or java
Question:
username_0: <!--
Thank you for suggesting an idea to make HanLP better.
Please fill in the template below to bypass our spam filter.
以下必填,否则直接关闭。
-->
**Describe the feature and the current behavior/state.**
I want to try HanLP in spark in scala(in scala, i can directly use HanLP Java version 1.x) or pyspark(HanLP 2.x), the dataset is billion level, so i want to know which one is better
**Will this change the current api? How?**
No
**Who will benefit with this feature?**
No
**Are you willing to contribute it (Yes/No):**
No
**System information**
- OS Platform and Distribution : Linux Ubuntu 16.04
- Python version: 3.7
- HanLP version: 2.x in Python, 1.x in Java
- Scala version: 2.11.x
- Java version: 1.8
**Any other info**
thank you for your support, i tried to find the post button in its bbs but couldn't find it. BTW, hava a happy national day!
* [ ] I've carefully completed this form.
Answers:
username_1: [auto-reply] Thanks for your comment. However, the essential information is required. Please carefully fill out the form when open a new issue.
Status: Issue closed
|
nicothin/NTH-start-project | 437913288 | Title: Что уже должно быть установлено?
Question:
username_0: Привет. Прошу помочь разобраться. Я пытался установить зависимости на своем ноуте. Выдает ошибки что не хватает компонентов, не достаточно прав. Система Убунту, устанавливал через судо. Я новичок )
Answers:
username_1: Зависимости ставятся без sudo.
Должен быть установлен nodejs.
Без скрина ошибки точно ничего нельзя сказать об ошибке.
username_0: Приложил скриншот. Пытался сделать как пишут здесь -
https://docs.npmjs.com/resolving-eacces-permissions-errors-when-installing-packages-globally
Не
помогло (
username_1: нет.
username_0: С празником Пасхи! Пробую еще раз, спасибо за терпение )
username_0: Проблема решена установкой nvm. Спасибо. Чайники они такие чайники )
вс, 28 апр. 2019 г. в 07:12, <NAME> <<EMAIL>>:
> С празником Пасхи! Пробую еще раз, спасибо за терпение )
>
Status: Issue closed
|
opencv/opencv | 201097103 | Title: CascadeClassifier::detectMultiScale levelWeights & rejectLevels
Question:
username_0: ##### System information (version)
- OpenCV => 3.2.0
- Operating System / Platform => Ubuntu 14.04
- Compiler => gcc 4.8.4
##### Detailed description
I'm having a problem trying to implement a code that determines if an image is a face. I created a minimal [repo](https://github.com/username_0/cascade-weight-test/tree/master) to test this out. There are six face, from a good relatively fontal face to a decentered face. The output I get is strange:
```
Detection ./cropped_face1.png [109 x 109 from (2, 14)] with weight 108.512 and level: 22 and size: 1
Detection ./cropped_face2.png [111 x 111 from (9, 9)] with weight 108.411 and level: 22 and size: 1
Detection ./cropped_face3.png [109 x 109 from (2, 14)] with weight 108.512 and level: 22 and size: 1
```
The last 3 faces do not even get a single output, even if face#4 is reasonably nice:
Face#3:
Face#4:
The level always seems to be 22 (on my very little sample, sorry). Stage 22 seems to be the final stage of the classifier based on the xml file of the classifier. I am also unclear on to what the weights actually represent with such strange number. I didn't find anything in the documentation to help me. I started going through the source code, but I didn't find anything to help me yet.
I do not quite understand why face#3 gets to the final stage of the classifier, with ~108 of weight, and the face#4 has no output, it didn't even get to stage 1?
Am I doing something wrong? Is something wrong with the function? Can there be an explanation of the usage and what the weight is?
##### Steps to reproduce
See [repo](https://github.com/username_0/cascade-weight-test/tree/master)
Answers:
username_1: You should get yourself a copy of OpenCV 3 Blueprints, Chapter 5, which discusses all these parameters in the finest details. That being said, the weights are the decision level of your final stage classifier.
IN GENERAL
It seems that you are using a detector for classification, which it is not designed for. My guess, your faces are to close to the border, which makes the detector fail. Again this is designed to locate face areas in an image with background information.
username_0: So, basically this overloaded function doesn't have much use now?
Ok, it makes sense that the faces are too close to the border and that's why they aren't classified or detected.
Do you have a recommendation on how I can achieve a classification score from a detector or OpenCV?
username_2: This is what exactly the docs in [opencv](https://github.com/opencv/opencv). even the face recognizers use this method
:smiley: :smile_cat: :small_blue_diamond:
username_3: `cv::CascadeClassifier` is almost deprecated. i think this issue can be closed down.
username_1: I am sorry @username_3 but I do not agree that cascade classifiers are deprecated. Yes DL detectors work as smooth and fast and are more acurate, but these still have their applications in embedded setups :)
username_4: Citation needed.
Status: Issue closed
|
int-brain-lab/ibllib | 411410449 | Title: Ambient sensor data access
Question:
username_0: They are on the Flatiron:
```
sessions, details = one.search(dataset_types='_iblrig_ambientSensorData.raw', lab='churchlandlab')
asd = one.load(sessions[0], dataset_types='_iblrig_ambientSensorData.raw')
```
However @username_2 , is there a point in storing all those values without a timestamp (albeit imprecise) ?
Status: Issue closed
Answers:
username_1: The commands above return `asd = None` even though the jsonable loads onto disk fine
username_0: They are on the Flatiron:
```
sessions, details = one.search(dataset_types='_iblrig_ambientSensorData.raw', lab='churchlandlab')
asd = one.load(sessions[0], dataset_types='_iblrig_ambientSensorData.raw')
```
There is one temperature/pressure value per trial.
username_0: Hmm this is not reproducible here. Possible explanations:
- this happens to most recent datasets: the datasets are registered, but do not exist yet on flatIron and `none` is the expected behaviour
- it happens to all datasets: fb757883278f770b9145fbaa4b3facc20a3d7974 was implemented 19 days ago on master. You should be up to date but still worth checking
Otherwise, please send a specific session that has this behaviour, and I"ll see if I can reproduce. Which OS ?
username_2: the code above is missing a `..., details=True)`
username_1: Updating to the latest develop branch fixed the `None` problem.
Could importing `one.ibllib` throw a warning when the user is on an old version?
Also, the output returned is a cumbersome list of dicts of lists (I think?):
```
'RelativeHumidity': [32.5419921875]},
{'Temperature_C': [26.780000686645508],
'AirPressure_mb': [1011.332763671875],
'RelativeHumidity': [32.552734375]},
{'Temperature_C': [26.770000457763672],
'AirPressure_mb': [1011.30322265625],
'RelativeHumidity': [32.5634765625]},
{'Temperature_C': [26.790000915527344],
'AirPressure_mb': [1011.3099365234375],
'RelativeHumidity': [32.564453125]},
{'Temperature_C': [26.780000686645508],
'AirPressure_mb': [1011.3465576171875],
'RelativeHumidity': [32.564453125]}]]
```
So that `ambient = pd.DataFrame(asd[0])` returns
```
AirPressure_mb RelativeHumidity Temperature_C
0 [1010.5851440429688] [33.04296875] [26.219999313354492]
1 [1010.7210083007812] [33.3408203125] [25.989999771118164]
2 [1010.6853637695312] [33.3408203125] [25.979999542236328]
```
Could the dict values be numeric, rather than wrapped in a list? And what datasetType should I grab to get the timing for each of these (`stimulus_onTime`?)
Status: Issue closed
|
pyrocms/pyrocms | 242510973 | Title: [files-module] Issue with Installer when installing PyroCMS
Question:
username_0: This is a very specific issue.
The files-module throws an exception when trying to run the "install" command from a controller or a queue job to generate a new Application and the Application is configured to use a different database that the one used by the main site.
Answers:
username_0: https://github.com/anomalylabs/files-module/pull/47 |
poooi/poi | 438139989 | Title: 接7个任务后下方出现未解锁
Question:
username_0: 接6个任务的时候没有,接了第7个就出现这个未解锁了

<!-- Love poi? Please consider supporting our collective:
👉 https://opencollective.com/poi/donate -->

Answers:
username_1: 已修复等发布
Status: Issue closed
username_2: Poi 10.2.4, Win10 x64 17763仍存在该问题。 |
smorgo/3DDesigns | 417891523 | Title: Bayonet_Locks library generates locking bump incorrectly
Question:
username_0: Having changed the way in which the locking slots were generated from Alexandre's original, I haven't yet worked out the best way to generate the small 'bump' (for want of a better word) that holds the locking pin in position. |
nus-cs2030/2021-s2 | 860725261 | Title: Multiple classes/interfaces/enums found in a class
Question:
username_0: Hello! I only declared one enum in the Event class and did not declare any other interfaces/classes in my Event.java file, so does this mean I have to declare my enum separately? Did anyone face this issue before, and how did y'all go about solving it? Thanks!
<img width="585" alt="Screenshot 2021-04-19 at 1 57 01 AM" src="https://user-images.githubusercontent.com/65957285/115155433-8a4f2e80-a0b2-11eb-8ab4-f8aba08204e5.png"><issue_closed>
Status: Issue closed |
aws-amplify/amplify-js | 744539966 | Title: unauthorized users can write to local datastore
Question:
username_0: **Describe the bug**
unauthorized users in the schema.graphql file can still write data to the local datastore. A warning message is displayed in the console, but data still resides in the local datastore. Essentially, data is written to the local datastore, but then fails to sync with DynamoDB. This leads to data inconsistencies between the front and back ends.
Obviously I could write logic in the application to prevent this occurring, but the graphql auth permissions should prevent this on both front and backend.
Warning message received:
```
{localModel: Model, message: "Not Authorized to access createAsset on type Asset", operation: "Create", errorType: "Unauthorized", errorInfo: null, …}
errorInfo: null
errorType: "Unauthorized"
localModel: Model {ModelData...}
message: "Not Authorized to access createAsset on type Asset"
operation: "Create"
remoteModel: null }
```
**To Reproduce**
Ensure cloud sync is configured as per documentation:
https://docs.amplify.aws/lib/datastore/sync/q/platform/js
**schema.graphql file :**
```
type Asset
@model
@auth(
rules: [
{ allow: groups, groups: ["Admins"], queries: [get, list], mutations: [create, update, delete] }
{ allow: groups, groups: ["Users"], queries: [get, list], mutations: null }
]
) {
id: ID!
name: String!
}
```
**Perform a write to the datastore:**
```
DataStore.save(
new Asset({
name: 'assetName',
})
);
```
**Expected behavior**
If data is not permitted to be written to dynamoDB, it shouldn't be allowed to be written to the local datastore.
Amplify version:
aws-amplify: ^3.3.8 => 3.3.8
aws-amplify-angular: ^5.0.37 => 5.0.37
Note: Data sync works fine when users are permitted via the schema.graphql file.
Answers:
username_1: Hey @username_0, currently, DataStore does not enforce auth rules clientside. This is by design, as we want the AppSync back end to serve as the only source of truth for authorization rules. We recommend that customers add their own business logic around this in their applications.
Also, just want to point out that the `queries` and `mutations` arguments [are deprecated](https://docs.amplify.aws/cli/graphql-transformer/auth#definition) and are not supported with DataStore. When establishing subscriptions for the Sync Engine, DataStore will ignore those arguments.
I suggest using `operations` instead:
```graphql
type Asset
@model
@auth(
rules: [
{ allow: groups, groups: ["Admins"] operations: [create, read, update, delete] }
{ allow: groups, groups: ["Users"], operations: [read] }
]
) {
id: ID!
name: String!
}
```
username_0: I thought that might be the case, thanks @username_1
Status: Issue closed
|
glidejs/glide | 946201715 | Title: Automatic value "Perview"?
Question:
username_0: I am using multiple sliders on one page. Each slider has a different number of slides.
How do I set the "perview" value to auto? Or how to set a separate "Perview" for each slider?
Answers:
username_1: You need to initialize each instance of glide with different settings. There are no `auto` option.
Status: Issue closed
|
colin-dumitru/F.js | 53129037 | Title: Add support for event streams.
Question:
username_0: Implemented with the 0.3 release.
The event stream now follows the Promise api. So when a value is pushed into a stream, a consumer is notified. This way, we can implement a pull-push mechanism for iterable chains.
Status: Issue closed
Answers:
username_0: Implemented with the 0.3 release.
The event stream now follows the Promise api. So when a value is pushed into a stream, a consumer is notified. This way, we can implement a pull-push mechanism for iterable chains.
Status: Issue closed
|
wesselb/stheno | 1127458489 | Title: Documentation about Multi-Output Regression
Question:
username_0: Hi @username_1,
I am trying to use your example of Multi-Output Regression with some data I have. I don't understand how to correctly give them to the VGP and them make a prediction.
My data as input x_obs are not the same, so it's not exactly as the example. I have nine x observation as [x1,x2,x3,x4,x5,x6,x7,x8,x9] with their y observation as [y1,y2,y3,y4,y5,y6,y7,y8,y9].
Also, with your example provided, is it possible to optimize some hyperparameters if we had some in the VGP ?
Here are my code I was trying to use, with 3 different outputs to simulate data. Thank you in advance for your help.
```
import matplotlib.pyplot as plt
from wbml.plot import tweak
from stheno import B, Measure, GP, EQ, Delta, Matern52
class VGP:
"""A vector-valued GP."""
def __init__(self, ps):
self.ps = ps
def __add__(self, other):
return VGP([f + g for f, g in zip(self.ps, other.ps)])
def lmatmul(self, A):
m, n = A.shape
ps = [0 for _ in range(m)]
for i in range(m):
for j in range(n):
ps[i] += A[i, j] * self.ps[j]
return VGP(ps)
# Define points to predict at.
x = B.linspace(0, 10, 5)
# Create some sample data.
x1 = np.atleast_2d(np.linspace(0, 10, 5)).T
x2 = np.atleast_2d(np.linspace(0, 9, 5)).T
x3 = np.atleast_2d(np.linspace(0, 7, 5)).T
y1 = np.atleast_2d(np.linspace(0, 10, 5)).T
y2 = np.atleast_2d(np.linspace(0, 10, 5)).T
y3 = np.atleast_2d(np.linspace(0, 10, 5)).T
x_obs = [x1,x2,x3]
y_obs = [y1,y2,y3]
# Model parameters:
m = 3
p = 3
H = B.randn(p, m)
with Measure() as prior:
# Construct latent functions.
us = VGP([GP(Matern52()) for _ in range(m)])
# Construct multi-output prior.
fs = us.lmatmul(H)
# Construct noise.
e = VGP([GP(0 * Delta()) for _ in range(p)])
# Construct observation model.
ys = e + fs
# Sample a true, underlying function and observations.
[Truncated]
post = prior.condition(*((p(x_obs), y_obs) for p, y_obs, x_obs in zip(ys.ps, ys_obs, x_obs)))
preds = [post(p(x)) for p in fs.ps]
# Plot results.
def plot_prediction(x, f, pred, x_obs=None, y_obs=None):
plt.plot(x, f, label="True", style="test")
if x_obs is not None:
plt.scatter(x_obs, y_obs, label="Observations", style="train", s=20)
mean, lower, upper = pred.marginal_credible_bounds()
plt.plot(x, mean, label="Prediction", style="pred")
plt.fill_between(x, lower, upper, style="pred")
tweak()
plt.figure(figsize=(10, 6))
for i in range(3):
plt.subplot(3, 1, i + 1)
plt.title(f"Output {i + 1}")
plot_prediction(x, fs_true[i], preds[i], x_obs, ys_obs[i])
plt.show()
```
Answers:
username_1: Hi @username_0!
I've put together a simple MOGP model (not using the example) which might better suit your use case. The script uses JAX to learn hyperparameters. (You can also use another AD framework if you like.)
```python
from stheno.jax import GP, Matern52, Measure
from varz.jax import Vars, minimise_l_bfgs_b
from wbml.plot import tweak
import matplotlib.pyplot as plt
import jax.numpy as jnp
import numpy as np
x1 = np.linspace(0, 10, 30)
x2 = np.linspace(0, 9, 40)
x3 = np.linspace(0, 7, 50)
# Generate some test data.
f = GP(Matern52())
y1 = f(x1, 0.2).sample().flatten()
y2 = f(x2, 0.2).sample().flatten()
y3 = f(x3, 0.2).sample().flatten()
p = 3 # Number of outputs
m = 3 # Number of latent processes
def model(vs):
ps = vs.struct
with Measure() as prior:
# Create independent processes with learnable length scales initialised to `1`.
us = [
GP(Matern52().stretch(ps_u.scale.positive(1)))
for ps_u, _ in zip(ps.us, range(p))
]
# Mix processes together to induce correlations between the outputs.
H = ps.mixing_matrix.unbounded(shape=(p, m))
fs = [0 for _ in range(p)]
for i in range(p):
for j in range(m):
fs[i] = fs[i] + H[i, j] * us[j]
# Create learnable observation noises initialised to `0.1`
noises = ps.noises.positive(0.1, shape=(p,))
return prior, fs, noises
def objective(vs):
prior, fs, noises = model(vs)
return -prior.logpdf(
(fs[0](x1, noises[0]), y1),
(fs[1](x2, noises[1]), y2),
(fs[2](x3, noises[2]), y3),
)
# Perform learning.
[Truncated]
# Plot results.
plt.figure(figsize=(10, 6))
x_to_plot = np.linspace(0, 10, 200)
plt.subplot(3, 1, 1)
plt.title("Output 1")
plot_posterior(x_to_plot, f1_post, x1, y1)
plt.subplot(3, 1, 2)
plt.title("Output 2")
plot_posterior(x_to_plot, f2_post, x2, y2)
plt.subplot(3, 1, 3)
plt.title("Output 3")
plot_posterior(x_to_plot, f3_post, x3, y3)
plt.show()
```
The script produces the following plot:

Let me know if this suits your needs. :)
username_0: Thank you very much for your code example. It is running on my side too, and I have the same results by using my CPU.
Because the computational time is high for nine inputs by using a CPU, I wanted to use my GPU to see if it will be faster. I followed the steps to use CUDA with the Jax library and was able to link both of them. However, by using the same code as you give me, I obtained this time an error:
```
---------------------------------------------------------------------------
RuntimeError Traceback (most recent call last)
<ipython-input-4-45889ae4f67a> in <module>
55 # Perform learning.
56 vs = Vars(jnp.float64)
---> 57 minimise_l_bfgs_b(objective, vs, trace=True, jit=True)
58 vs.print() # Display learned parameters.
59
~/python-env/lib/python3.6/site-packages/varz/minimise.py in minimise_l_bfgs_b(f, vs, f_calls, iters, trace, names, jit)
77 trace=trace,
78 names=names,
---> 79 jit=jit,
80 )
81
~/python-env/lib/python3.6/site-packages/varz/minimise.py in _minimise_l_bfgs_b(f, vs, f_calls, iters, trace, names, jit)
154 # Run function once to ensure that all variables are initialised and
155 # available.
--> 156 res = convert(f(vs, *args), tuple)
157 val_init, args = res[0], res[1:]
158
<ipython-input-4-45889ae4f67a> in objective(vs)
49 (fs[0](x1, noises[0]), y1),
50 (fs[1](x2, noises[1]), y2),
---> 51 (fs[2](x3, noises[2]), y3),
52 )
53
~/python-env/lib/python3.6/site-packages/plum/function.cpython-36m-x86_64-linux-gnu.so in plum.function._BoundFunction.__call__()
~/python-env/lib/python3.6/site-packages/plum/function.cpython-36m-x86_64-linux-gnu.so in plum.function.Function.__call__()
~/python-env/lib/python3.6/site-packages/stheno/model/measure.py in logpdf(self, *pairs)
461 """
462 fdd, y = combine(*pairs)
--> 463 return self(fdd).logpdf(y)
464
465 @_dispatch
~/python-env/lib/python3.6/site-packages/stheno/random.py in logpdf(self, x)
210 B.logdet(self.var)[..., None] # Correctly line up with `iqf_diag`.
211 + B.cast(self.dtype, self.dim) * B.cast(self.dtype, B.log_2_pi)
--> 212 + B.iqf_diag(self.var, B.subtract(x, self.mean))
213 )
214 / 2
~/python-env/lib/python3.6/site-packages/plum/function.cpython-36m-x86_64-linux-gnu.so in plum.function.Function.__call__()
~/python-env/lib/python3.6/site-packages/matrix/ops/iqf_diag.py in iqf_diag(a, b)
33 @B.dispatch
34 def iqf_diag(a, b):
---> 35 return iqf_diag(a, b, b)
36
[Truncated]
608
~/python-env/lib/python3.6/site-packages/jax/interpreters/xla.py in _xla_call_impl(***failed resolving arguments***)
593 *unsafe_map(arg_spec, args))
594 try:
--> 595 return compiled_fun(*args)
596 except FloatingPointError:
597 assert config.jax_debug_nans or config.jax_debug_infs # compiled_fun can only raise in this case
~/python-env/lib/python3.6/site-packages/jax/interpreters/xla.py in _execute_compiled(compiled, avals, handlers, kept_var_idx, *args)
891 for i, x in enumerate(args)
892 if x is not token and i in kept_var_idx))
--> 893 out_bufs = compiled.execute(input_bufs)
894 check_special(xla_call_p.name, out_bufs)
895 return [handler(*bs) for handler, bs in zip(handlers, _partition_outputs(avals, out_bufs))]
RuntimeError: Internal: Unable to launch triangular solve for thunk 0x2c46c570
```
Do I need to add something to the code to make it work with a GPU ?
username_1: Ouch! That doesn't look good. Could you confirm that running other JAX code on the GPU works fine? If that's the case, I can look into this more closely to see what's going on.
username_0: I tried some examples of JAX code with my GPU (like these one: https://jax.readthedocs.io/en/latest/notebooks/quickstart.html) and it was working. I think the issue come from the library.
I have jax-0.2.17 and jaxlib-0.1.65+cuda110 install on my computer |
rebekahsong/bb_swe | 855468677 | Title: Add Ratings to Songs
Question:
username_0: Edit Django Model and Views to include ratings as an attribute of songs. Could potentially write some backend code that calculates the rating as an average of all ratings on a song.
Edit React components to include ratings in the request and the display of the responses for songs. |
exastro-suite/it-automation | 1044465971 | Title: 【エクスポート/インポート】メニューインポートのREST APIのXcommandがUPLOADとEXECUTEの時にメニューIDの形式が異なっている
Question:
username_0: メニューインポートのREST APIのXcommandがUPLOADの場合はメニューIDが「0000000001」と左0埋めで返却されるが、
XcommandがEXECUTEの場合は「1」という形で左0埋めを消して実行しないと以下のエラーになる。
メニューIDの設定値が不正です。(正規表現(/^0$|^-?[1-9][0-9]*$/s)に一致するデータを入力してください 。
EXECUTE時は「0000000001」でも「1」でもどちらの指定方法でも正常となるように修正する。<issue_closed>
Status: Issue closed |
aplpy/aplpy | 512550 | Title: Offset tick position option
Question:
username_0: Add the option to have the axes labeled in offset coordinates relative to a chosen point in the field of view. E.g. with ticks at [ -1, -0.5, 0, 0.5, 1] arcseconds for the 2.4" FOV.
Example code:
f = APYpy.FITSFigure('some_tiny_AO_image.fits')
f.show_greyscale(stretch='log')
f.set_tick_relative(True)
f.set_tick_relative_to( (200,150) )
<NAME> suggested the idea. Thanks!
Answers:
username_1: A very late bump - @jpinedaf asked for this, especially for interferometer images.
username_0: What we need to do is implement this in WCSAxes and then we'll get it for free once I merge in the changes to use WCSAxes internally.
username_1: Is there an open issue in wcsaxes for this that should be linked?
username_0: Here it is: https://github.com/username_0/wcsaxes/issues/8
username_2: Is there any chance of implementing this soon? I would really like to use it to simplify a multi-panel figure in an ApJL that I am about to submit. It would be great to have it in time for the second submission.
username_2: For others who arrive here, below is some code to do it in a kludged way:
def ra2deg(s):
token = s.split(':')
deg = 15*(float(token[0])+float(token[1])/60.+float(token[2])/3600.)
return(deg)
def dec2deg(s):
token = s.split(':')
deg = (abs(float(token[0]))+float(token[1])/60.+float(token[2])/3600.)
if (token[0].find('-') >= 0):
deg = -deg
return(deg)
target_ra=ra2deg('17:20:53.32')
target_dec=dec2deg('-35:47:00')
image_radius_arcsec=4.5
image_radius=image_radius_arcsec/3600.
if True:
# Try to draw relative tick marks
length = 0.20/3600.
cosdec = cos(radians(target_dec))
xleft = target_ra + image_radius/cosdec
xright = target_ra - image_radius/cosdec
ytop = target_dec + image_radius
ybottom = target_dec - image_radius
# x-axis ticks
for x in np.arange(-np.ceil(image_radius_arcsec),image_radius_arcsec):
xt = target_ra + x/3600./cosdec
if int(abs(np.round(x))) % 4 == 0:
lengthFactor = 1.6
else:
lengthFactor = 1.0
iline = np.array([[xt,xt],[ybottom,ybottom+length*lengthFactor]])
f.show_lines([iline], color='w')
iline = np.array([[xt,xt],[ytop,ytop-length*lengthFactor]])
f.show_lines([iline], color='w')
# y-axis ticks
for y in np.arange(-np.ceil(image_radius_arcsec),image_radius_arcsec):
if int(abs(np.round(y))) % 4 == 0:
lengthFactor = 1.6
else:
lengthFactor = 1.0
yt = target_dec + y/3600.
iline = np.array([[xleft,xleft-length*lengthFactor], [yt,yt]])
f.show_lines([iline], color='w')
iline = np.array([[xright,xright+length*lengthFactor], [yt,yt]])
f.show_lines([iline], color='w')
username_0: @username_2 - sorry for the delay, in the mean time the easiest way to do this with minimal hacking is to use this function:
```python
from astropy.wcs import WCS
from astropy.wcs.utils import skycoord_to_pixel, proj_plane_pixel_scales
def linear_offset_coords(wcs, center):
"""
Returns a locally linear offset coordinate system.
Given a 2-d celestial WCS object and a central coordinate, return a WCS
that describes an 'offset' coordinate system, assuming that the
coordinates are locally linear (that is, the grid lines of this offset
coordinate system are always aligned with the pixel coordinates, and
distortions from spherical projections and distortion terms are not taken
into account)
Parameters
----------
wcs : `~astropy.wcs.WCS`
The original WCS, which should be a 2-d celestial WCS
center : `~astropy.coordinates.SkyCoord`
The coordinates on which the offset coordinate system should be
centered.
"""
# Convert center to pixel coordinates
xp, yp = skycoord_to_pixel(center, wcs)
# Set up new WCS
new_wcs = WCS(naxis=2)
new_wcs.wcs.crpix = xp + 1, yp + 1
new_wcs.wcs.crval = 0., 0.
new_wcs.wcs.cdelt = proj_plane_pixel_scales(wcs)*3600.
new_wcs.wcs.ctype = 'XOFFSET', 'YOFFSET'
new_wcs.wcs.cunit = 'arcsec', 'arcsec'
return new_wcs
```
then you can do something like:
```python
import aplpy
from astropy.io import fits
from astropy.wcs import WCS
from offset import linear_offset_coords
from astropy import units as u
from astropy.coordinates import SkyCoord
galcen = SkyCoord(0 * u.deg, 0 * u.deg, frame='galactic')
hdu = fits.open('2MASS_k.fits')[0]
wcs = WCS(hdu.header)
wcs_offset = linear_offset_coords(wcs, galcen)
hdu_linear = hdu.copy()
hdu_linear.header = wcs_offset.to_header()
fig = aplpy.FITSFigure(hdu_linear)
fig.show_colorscale()
```
username_3: So I've been using this "linear_offset_coords" function with some success to make plots, but have run into the issue that using this method to change the axes labels makes it difficult over-plot contours from data with different wcs information. I've been using reproject.reproject_interp to put the contour data into the same pixels as the image data but this involves seriously down-sampling the contours which I'd like to avoid. Any pointers on ways to do this? I'm using wcsaxes right now.
username_4: Just posting to receive updates on this. I am attempting to use the linear_offset_coords function. It (mostly) seems to work well for simple image display, but unfortunately when I try to overlay the beam via add_beam() (they're interferometric images) or even show_ellipses(), I get the error
WCS is not celestial, cannot determine celestial pixel scale.
@username_3, can you describe how you are trying to do your overlay? I wonder if I can adapt your method for what I'm trying to do.
username_3: @username_4 For the overlay I've been using reproject.reproject_interp (http://reproject.readthedocs.io/en/stable/api/reproject.reproject_interp.html) which pretty much just rotates and rebins my contour data to the same pixel scale as the image then I can just plot them over each-other without using the wcs transformation. I've been mostly using wcsaxes not aplpy though, not 100% certain how that changes things. Using wcsaxes at least, if you can get the dimensions of the beam in pixel units you should just be able to over-plot it using the Ellipse patch in matplotlib (http://matplotlib.org/examples/pylab_examples/ellipse_demo.html).
username_5: I would LOVE to see this feature as soon as possible!
Tried the "minimal hacking" approach suggested by @username_0 , but I get the error "ValueError: 'cdelt' array is the wrong shape, must be 2"
I'm really not a pro, but guess this is because I'm working with interferometric data and there is a spectra axis as well?
What I'm trying to make is a velocity map plot, so I need my spectral axis but I still would like to show relative coordinates instead of world coordinates to show how big the structure in the images are.
username_6: would be awesome to see this implemented! |
JuliaLang/julia | 530245383 | Title: Extend at-specialize to force specialization of arguments
Question:
username_0: A trick currently used in `Base` (and probably elsewhere) to force specialization on functions is to artificially declare a type parameter, e.g. `foo(f::F) where F = f(1)`. I would bet that the majority of Julia users don't know about this and would look at this piece of code without any idea about the role of `F`, and would even fix what appears as a naive use of type parameters by simplifying it to `foo(f) = f(1)`.
What about adding some explicit means to this end, for example reusing `@specialize`: `foo(@specialize(f)) = f(1)` ?
xref: https://github.com/JuliaLang/julia/pull/33917#discussion_r349864113
Answers:
username_1: Since in `foo(f::F) where F = f(1)` the function `f` is called, it would be equivalent to `foo(f) = f(1)`, right?
username_0: I'm not the right person to answer, but this kind of prove the point that the exact rules are not obvious to everyone (talking of me!)
username_1: It is worth mentioning that #32817 added documentation on when a `Function` argument is specialized.
After thinking more about this, my opinion shifted somewhat: Having `@specialize` act in this way would not really clear up the confusion. The current syntax `foo(f::F) where F` is hacky, but so is overriding compiler heuristics. :)
username_2: But I do think the intent of `@specialize(f)` is much more transparent than `f::F ... where F`. For a person with just enough knowledge of Julia to be dangerous :smiling_imp:, the latter might induce him/her to "clean up" and remove that little annotation.
username_0: Nice, I was wondering whether it was documented somewhere, but got too lazy to check. Otherwise I agree with Tim, the current way to force specialization is too easy to miss out, and is brittle. Overrriding comiler heuristics might be hacky, but we all love to do it :D And `@inline` gets its own explicit name.
If you think `@specialize` is not the right word to clear up the confusion, would another name do it? `@force_specialize` ?
username_3: I agree we need a way to do this.
In the last year I've noticed people start to use the static parameter trick more often, and each time I have to ask the question
"Is `foo(x::T) where {T}` an attempt to force specialization, or an ugly way to get at `typeof(x)`?"
Furthermore, we're then forced to leave a comment that this is indeed to force specialization, lest someone clean up the code later.
To make matters worse, the specialization heuristics are somewhat subtle (and presumably subject to change) so I've seen people start to throw this in at the first sign of trouble in the hope that it will help.
username_3: So looking at the code which processes `@specialize`, it seems that the per-argument form is currently ignored, ie the following doesn't actually do anything:
```julia
function foo(@specialize(x))
body
end
```
See https://github.com/JuliaLang/julia/blob/52c55d7934f71c5b2d9f6e6fa98cb48817def57c/src/method.c#L554-L558
So in principle we could steal `@specialize(x)` to mean "force specialize". However this form would be inconsistent with
```julia
function foo(x)
@specialize
body
end
``` |
ghusta/android-docker-search | 808383667 | Title: Refactoring : Use ConstraintLayout
Question:
username_0: Replace Android's `RelativeLayout` by `ConstraintLayout`
See :
- https://developer.android.com/reference/androidx/constraintlayout/widget/ConstraintLayout
- https://developer.android.com/training/constraint-layout<issue_closed>
Status: Issue closed |
yii2tech/config | 213524624 | Title: Dynamic modules connection
Question:
username_0: Can I add new modules dynamicly, but not only parametrs of nested modules?
For example:
```
'NewModules' => [
'Path' => 'modules',
'Label' => new Modules',
],
```
and db row
id: `NewModules`
'value: `['user'=>'path/to/Module']`
Status: Issue closed
Answers:
username_1: It is impossible to manipulate entities, which do not exist in current application. Dynamic componenst or modules can not be added. |
platformio/platformio-vscode-ide | 435018356 | Title: Start PIO Home Server
Question:
username_0: # Description of problem
Leave a comment...
BEFORE SUBMITTING, PLEASE SEARCH FOR DUPLICATES IN
- https://github.com/platformio/platformio-vscode-ide/issues
# Configuration
VSCode: 1.33.1
PIO IDE: v1.7.0
System: Windows_NT, 10.0.17134, x64
# Exception
```
Error: Error: Detected unknown package 'contrib-pysite'
at P.Promise (C:\Users\Aj\.vscode\extensions\platformio.platformio-ide-1.7.0\node_modules\platformio-node-helpers\dist\index.js:1:14992)
at ChildProcess.h (C:\Users\Aj\.vscode\extensions\platformio.platformio-ide-1.7.0\node_modules\platformio-node-helpers\dist\index.js:1:5916)
at ChildProcess.emit (events.js:182:13)
at ChildProcess.cp.emit (C:\Users\Aj\.vscode\extensions\platformio.platformio-ide-1.7.0\node_modules\cross-spawn\lib\enoent.js:34:29)
at maybeClose (internal/child_process.js:961:16)
at Process.ChildProcess._handle.onexit (internal/child_process.js:248:5)
```
Status: Issue closed
Answers:
username_1: Duplicate of https://github.com/platformio/platformio-core/issues/2341 |
EarthSystemCoG/COG | 51328579 | Title: Add new metrics to the system users page
Question:
username_0: WHO: Sylvia
For security purposes, the following items would be useful to know about folks who have registered for cog. I propose:
1) adding these 3 metrics to the system users page:
a) Number of projects to which they are members. Looking for 0
b) Number of web URLs they have. 0 projects + a web url may be a spam bot
c) Date of last login. Eventually we may end up culling folks who never used the site.
2) changing the template for this page so that it has neither the left nor right widgets and can span the entire page.
Answers:
username_0: The number of projects is showing up as zero for all users. Will have to investigate. The number of projects is showing up correctly on individual profiles.
username_0: Ok. I determined that {{projects|length}} will not work on this page because the function that creates projects is not associated with this template now.
I created a new function that created a list of the projects|length for each user. Since it can not be associated with the user object, I had to fudge the table by breaking it it in two, drawing a separate table with the project info and then making them look seamless.
username_0: Luca pointed out a better way to do this. In testing, the search is not working on it, so we will postpone to 3.0 since 2.9 is coming out in a rush.
username_0: We realized that gathering the number of projects does not work across nodes. the information is irrelevant. Redesign this page as necessary.
username_0: Removed the number of projects value. It is not federated and is bogus.
Turned the page back on for system administrators.
Need to check the links to profiles work on dev-hydra, can't test locally because created accounts have no openid and clicking on the name fails.
username_0: verified on dev-hydra. Closing.
Status: Issue closed
|
suyashkumar/dicom | 1146132301 | Title: De-identify DICOM tags
Question:
username_0: I'm trying to update/de-identify some of the metadata tags (such as Patient ID). Is there a way to update the file directly without reading the file and writing back to the file?
Answers:
username_0: @suyashkumar @jstutters |
KIST-Iceberg/Iceberg | 288070574 | Title: Image Crop
Question:
username_0: 




Status: Issue closed
Answers:
username_0: not need to crop
https://www.kaggle.com/brassmonkey381/viewing-leak-and-machine-images |
moby/moby | 341268075 | Title: .Net ASP.Net Webapp in Container loose Primary Domain Trust randomly after some days runtime
Question:
username_0: **Reference:**
https://social.msdn.microsoft.com/Forums/en-US/542286aa-1095-4af0-91f2-43ff0c0f9469/docker-container-with-gsma-loose-trust-relationship-after-a-few-days?forum=windowscontainers
Answers:
username_0: I rebuild all my images to Server 2019 LTSC Container on new Server 2019 Docker Host VMs like descibe above.
Problem is better on a factor of 1000 :-) but still sometimes i loose trust on container starts like
username_1: Do you use a time synchronization tool to have both the container host and the DC sync their clocks? A difference in clocks can account for the behavior that you describe.
username_2: We are also experiencing this issue. @username_0 did you get it to work as expected?
username_0: No we still have this error but its not as often if we use win2019 😟
You have any success on this?
username_0: @username_2 you also reverseproxy with haproxy infront?
username_2: @username_0 Nope. We have interlock in front and it fully supports the Kerberos according to their documentation.
username_0: @memonfaheem You find anything after a near ayear?
username_3: @username_0 have you found any more information about this?
username_0: no sorry still investigating you got same Problem? and als no intel from ms or docker
username_3: Hey @username_0 , do your containers all use the GMSA account name they are running as as their hostname? This was previously a requirement, but as of Windows Server 2019 it is no longer needed. In fact, setting the container hostnames to all be the same may be what causes this issue. If you are able, try allowing the hostnames to be automatically set by docker to the container ID instead of set to the GMSA account.
username_0: Thx for your Input @username_3
We upgraded to 2019 because of the issue you mention.
All Containers got a different Containername defined by an Ansible Playbook, i need to have so the Playbook can identify the instances. All Containers share the same GMSA Account Credentials ans running on 2 VMs Maybe thats the Problem?
For each Application Containers we run them on both Docker hosts(VM). Both Dockerhosts have permission for the GMSA and the GMSA Account have the required SPNs for the Service, and we use a dedicated Container IP Adresses for each Container.
If the Problem are the IPs because of Kerberos Auth cannot validarte SPN Hostname or something they will not work after Start right?
But they work and after an unkown time they loose the trust.
Will try Random Containernames in an Testenvironment
Any other Ideas?
username_0: And each Container have different Name on both host like
App1-1,App1-2 or something, so on both dockerhosts the containernames are unique |
deviant-dev/ds-issues | 559307852 | Title: System can't find Focused Char, can't perform 'Looming'
Question:
username_0: **Describe the bug**
Playing through the demo, console shows an error that the system can't find a focused character to loom over.
**To Reproduce**
Playing through the demo with the console open, occurs near the end of goddess's chambers
**Expected behavior**
No errors to be displayed on the console
**Screenshots**

**System (please complete the following information):**
- VR Device: KB+M
- Win 10 64bit
- Version: Demo-1
**Additional context**
Screenshot taken after leaving goddess's chambers
Answers:
username_0: I'm not going to create a bug report around the 'Perform angry 50' as I'm guessing that fixing this focused character issue will likely fix any other issues around focused characters.
Status: Issue closed
|
spatie/laravel-backup | 504078022 | Title: Problems with 32bit php
Question:
username_0: Hello,
I run into some problems with a 32 bit php. I can't look up my backup list nor can I look it up in your nova tool. The main problem is described in #567 .The Fork from CharlieEtienne isn't updated so I don't want to use it. Maybe we can fix it in the real package?
Thanks for the great package and keep up the good work!
If you need more information just reply.
Answers:
username_1: I'm not running into the problem myself and don't have time to investigate. I'd accept a PR to fix it in this package.
username_0: Did a pull request for the issue. It fixed the problem for me but it needs a validation for a 64bit php system. #976
Status: Issue closed
username_1: We'll continue the conversation in #976 |
matplotlib/matplotlib | 143285472 | Title: Example embedding_in_tk.py freezes in Python3.5.1
Question:
username_0: When I try and run the [example code on matplotlib's website](http://matplotlib.org/examples/user_interfaces/embedding_in_tk.html) for embedding matplotlib in tk, the GUI never appears and the kernel crashes. I'm using Python 3.5.1, matplotlib 1.5.1, and tk 8.5.18 vc14. Everything was installed using Anaconda.
I'm pretty sure this is supposed to work, as I've seen other tutorials run it using Python 3. Interestingly, when I create a Python 2.7 environment, I'm able to run the code successfully.
Answers:
username_1: I can not reproduce this with a linux box.
Do you get any error messages from the crash?
username_2: When you say "the kernel", are you talking about the ipython kernel, or the
linux kernel? If it is the linux kernel, I doubt this would be specific to
matplotlib and wonder if it is a tk packaging issue in general?
username_0: I'm a grad student in MechE, so forgive me if I get some of the terminology wrong. I'm talking about the instance of python.
username_2: Ah, so you are on Windows? Which version?
username_0: Windows 7 Professional, SP1
username_3: I suspect this is a problem specific to the Anaconda build and/or your particular system (e.g., are there Tk versions installed in addition to the one from Anaconda?) rather then being an MPL bug. You could check with a message to the Anaconda mailing list to see whether anyone there can reproduce it, or whether there is someone with a similar configuration who *cannot* reproduce it.
username_0: I've done a fresh installed on a smiliar windows 7 machine and was able to reproduce the same issue, so I'm thinking its either an issue with anaconda or matplotlib. Is anyone else running a non-anaconda python interpreter on windows 7, and is able to reproduce the issue?
I've started a [topic on the anaconda mailing](https://groups.google.com/a/continuum.io/forum/#!topic/anaconda/f-UwlCovN_I) list to see if someone else can reproduce it. I'm moving forward using Python 2.7, but hopefully we can get this resolved.
username_0: Thanks for the help everyone!
Status: Issue closed
|
stegos/stegos | 447051502 | Title: Panic during consensus.
Question:
username_0: ```
STEGOS_LOG 2019-05-22T08:37:53.712194541+00:00 [stegos_consensus::state] DEBUG - Precommit(11450:9): collected a pre-commit: from=SecurePKey(e54df9c0de923c8972b3e7984eab8dd3261d2530fd776d2dbb29e98e7ff3
b28002954ad4ce7001f379e61323c2e8d7a2cacd45906e0677784c544c88f5c6276301)
STEGOS_LOG 2019-05-22T08:37:53.922420499+00:00 [stegos_consensus::state] DEBUG - Precommit(11450:9): collected a pre-commit: from=SecurePKey(<KEY>
240f8e7785d24b49efd49e738609aa508b15a1ea88b6c29311e60e6d3510ea13411501)
STEGOS_LOG 2019-05-22T08:37:54.123815742+00:00 [stegos_consensus::state] DEBUG - Precommit(11450:9): collected a pre-commit: from=SecurePKey(4f6adf8ff3c224c7b181a12ce48c520b0173d20fee742ad92d70f6ee3868
90cc0fc413aaa2a635401e5007538695eb01ff0ad8a697de502c0c70ab2f001b8ae600)
STEGOS_LOG 2019-05-22T08:37:54.123882127+00:00 [stegos_consensus::state] DEBUG - Precommit(11450:9) => Commit(11450:9)
STEGOS_LOG 2019-05-22T08:37:54.124543151+00:00 [stegos_node] INFO - Sent sealed block to the network: height=11450, block=80ec0bed1a1c
STEGOS_LOG 2019-05-22T08:37:54.124589928+00:00 [stegos_blockchain::validation] DEBUG - Validating a macro block: height=11450, block=80ec0bed1a1c
STEGOS_LOG 2019-05-22T08:37:54.124715373+00:00 [stegos_network::pubsub] DEBUG - Sending broadcast message: topic=block, size=1414
thread 'tokio-runtime-worker-1' panicked at 'block is validated before: BlockError(NoLeaderSignatureFound(11450, H(80ec0bed1a1c26c3cbe2716a6b4207c605ff31fb3a304e1f428f1b8416070416)))
```
looks like create propose without leader signature, this could be caused because of "round lock".
Answers:
username_1: This was a transitional problem - it has been fixed already.
Status: Issue closed
|
jlippold/tweakCompatible | 417004354 | Title: `SwipeForMore` working on iOS 12.1.1
Question:
username_0: ```
{
"packageId": "org.thebigboss.swipeformore",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "org.thebigboss.swipeformore",
"deviceId": "iPhone6,1",
"url": "http://cydia.saurik.com/package/org.thebigboss.swipeformore/",
"iOSVersion": "12.1.1",
"packageVersionIndexed": true,
"packageName": "SwipeForMore",
"category": "Tweaks",
"repository": "BigBoss",
"name": "SwipeForMore",
"installed": "1.1.7",
"packageIndexed": true,
"packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 100% with 13 working reports.",
"id": "org.thebigboss.swipeformore",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.1.2",
"shortDescription": "Manage packages via swipe.",
"latest": "1.1.7",
"author": "PoomSmart",
"packageStatus": "Working"
},
"base64": "<KEY>",
"chosenStatus": "working",
"notes": ""
}
```<issue_closed>
Status: Issue closed |
tminglei/slick-pg | 92736392 | Title: Slick 3.0 tsql type mapping
Question:
username_0: I'm trying to use the new typed sql statements in Slick 3.0 and I'm discovering that the type mapping that I have found in slick-pg doesn't seem to be working. Is this something to do with how I extend the postgres driver? Or is this something that needs direct support in slick-pg?
Answers:
username_1: Hi @username_0, do you mean those extension methods not working?
_Those extension methods are based on `JdbcType`, but slick plain sql isn't based on it. So the types supported by slick-pg plain sql hasn't those extension methods._
username_0: I'm not sure I understand your response, so I'll give a little more information.
In the normal slick methodology, I could map a case class to a table like this:
import org.joda.time._
case class Foo(name: String, date: DateTime)
class Foos(tag: Tag) extends Table[Foo](tag, "foos") {
def name = column[String]("name")
def date = column[DateTime]("date")
def * = (name, date) <> (Foo.tupled, Foo.unapply)
}
val foos = TableQuery[Foos]
With the new slick 3.0 `tsql` typed sql statements, I would like to be able to write the following:
def selectAllFoos: DBIO[Seq[(String,DateTime)]] = tsql"select * from foos"
but in this case the type mapping that I have come to expect from slick-pg for the joda DateTime class doesn't work.
Are you saying that the underlying JDBC functions are different between Slick's normal usage and its `tsql` statements? The methods used for marshalling case classes [as explained here](http://slick.typesafe.com/doc/3.0.0/sql.html#result-sets) look like they are using normal JDBC column data types.
username_1: Well, I got your means.
I just check the related codes of `slick`, and seems it requires some special customization on `slick-pg`.
I'll implement it ASAP when I have time.
_Thanks for pointing it out! :-)_
username_1: Hi @username_0, to some degree, this is a problem same as #171, which requires `slick-pg` to auto bind a pg type to a scala type.
This is not so easy, as you know, since a pg type can be bound to multiple scala types.
I need more time to construct a mechanism to help do it easily when your customized postgres driver mixing some pg type supports. So it won't be ready soon.
username_1: Hi @username_0, the implement is almost done, but we need some enhancement in `slick`, related pull request is slick/slick#1226
The working branch in `slick-pg` is tsql, you can check its codes [here](https://github.com/username_1/slick-pg/tree/tsql).
_BTW, currently I can only support bind a pg type to a scala type without type parameter.
For scala types with type parameter, I can't implement it, since it's not supported by `slick` by now._
username_1: Hi @username_0, the changes already merged into master and released in slick-pg v0.10.0-RC1.
You can help check it. Thanks! ^^
Status: Issue closed
username_0: Thanks, I'll test it out tonight. I appreciate the fast turnaround on this request! |
ThomasWest/blog20160423 | 150682391 | Title: latest endeavor 20160423
Question:
username_0: I have now complete through chapter 6 in this particular project. I have had a number of irreconcilable bugs (actually some greater success in this project) and started over, essentially, a couple of times. As always, I have many questions (re: switching to postgres, schema, data, etc and purpose and action of some "requirements in README.md), but a miniscule bit more understanding, I hope. I would hope for a face-to-face for explanations and "demonstrations" at some point. |
factor1/phoenix-design-week-2016 | 165963531 | Title: About: Revise styling on learn about MM and check out events box headlines
Question:
username_0: These headlines need to be in Proxima Nova as well—same styling as previous specs given for H2s and H3s. Basically every H-tag needs to be in Proxima Nova. Only buttons use the Bebas, and a couple subheads in the hero areas.

Status: Issue closed
Answers:
username_1: fixed with issue #41
username_0: Again, I failed to provide all of the details here: They also need to be regular weight and sentence case (I think I noted this in Issue #41 as well) |
tensorflow/tensorflow | 396732197 | Title: Support first_n in tf.print
Question:
username_0: **System information**
- TensorFlow version (you are using): 1.12
- Are you willing to contribute it (Yes/No): no
**Describe the feature and the current behavior/state.**
`tf.Print` is deprecated and `tf.print` doesn't support `first_n`. Would be good to support `first_n` in `tf.print` as well.
**Will this change the current api? How?**
Add a `first_n` parameter to tf.print.
**Who will benefit with this feature?**
Anyone who has used `first_n` in tf.Print.
**Any Other info.**
See Goomics #50.
Answers:
username_0: It seems there's already a solution to the "where to store first_n state": store it in the op just like `tf.Print` does today.
I have no experience with eager mode or autograph in TF, but it seems that the code you provided doesn't actually reproduce the behavior of `tf.Print`. First, there's just a single counter Variable - does that imply the state is shared across all calls to print_first_n? Second, it's my impression that *all* Variables are stored in checkpoints. If we had a per-print Variable, we'd have trouble restoring if someone decided to add / remove a print statement, no?
username_1: In tf2 variables are only stored in checkpoints if you add them to
checkpoints.
Re how many counters you have, this is up to you; as you can see it's easy
to control with this workaround.
And about where to put the state, the solution in tf.Print is
nondeterministic in many ways, something we do not want to reproduce in
tf.print.
username_0: Gotcha, so these are all fair points, and looks like things will be cleaned up in tf2. In the meantime, `tf.Print` is deprecated in tf1 so is there another solution I can use before we switch over to tf2 or will we just have to accept and ignore the deprecation warning?
username_2: If you're willing to get a little dirty, you can look into the silence context manager inside deprecation.py. It's not part of the official API (and it won't be), but you can use it to hackily suppress deprecation warnings. |
lvgamedev/simairport-modding | 646736693 | Title: At Baggage Scanner the trays not dissapier at the scan area
Question:
username_0: Also with right marker setting , it not works , trays are to see at the whole mod.
(i post that for marineanukotilah , because his account here not works anymore)
[RemoteBagScannerSmall.zip](https://github.com/lvgamedev/simairport-modding/files/4841078/RemoteBagScannerSmall.zip)

Answers:
username_1: Added an extra field to the bag scanner utility named "maskSprite" - set it to the name of the sprite that should be used for masking, facings and large will automatically be added.
For masking name "myMask" the large bag scanner facing to the left will use the texture "myMask_Large_Left".
Status: Issue closed
|
flutter/flutter | 705823097 | Title: Local network permission issue on iPad, crash on launch
Question:
username_0: @username_2, following up from https://github.com/flutter/flutter/issues/57225
Steps to reproduce:
1. `flutter create testproject3`
2. `cd testproject 3`
3. `flutter run -v > output_1.txt`
The app runs on the iPad, and presents the "testproject3 would like to find and connect to devices on your local network" dialog. The app then crashes (disappears from screen), leaving the dialog still there.
stderr output:
```text
testproject3$ flutter run -v > output_1.txt
[ ] Failed to attach to the observatory.
[ +2 ms] Try accepting the local network permissions popup, or enable "Settings > Privacy > Local Network" for your app.
[ ] For more information, see https://flutter.dev/docs/development/ios-14#debugging-flutter
[ ] Error launching application on Stephen’s iPad.
[ +5 ms]
#0 throwToolExit (package:flutter_tools/src/base/common.dart:14:3)
#1 RunCommand.runCommand (package:flutter_tools/src/commands/run.dart:606:7)
<asynchronous suspension>
#2 FlutterCommand.verifyThenRunCommand (package:flutter_tools/src/runner/flutter_command.dart:985:18)
#3 _rootRunUnary (dart:async/zone.dart:1198:47)
#4 _CustomZone.runUnary (dart:async/zone.dart:1100:19)
#5 _FutureListener.handleValue (dart:async/future_impl.dart:143:18)
#6 Future._propagateToListeners.handleValueCallback (dart:async/future_impl.dart:696:45)
#7 Future._propagateToListeners (dart:async/future_impl.dart:725:32)
#8 Future._completeWithValue (dart:async/future_impl.dart:529:5)
#9 _AsyncAwaitCompleter.complete (dart:async-patch/async_patch.dart:40:15)
#10 _completeOnAsyncReturn (dart:async-patch/async_patch.dart:311:13)
#11 RunCommand.usageValues (package:flutter_tools/src/commands/run.dart)
#12 _rootRunUnary (dart:async/zone.dart:1198:47)
#13 _CustomZone.runUnary (dart:async/zone.dart:1100:19)
#14 _FutureListener.handleValue (dart:async/future_impl.dart:143:18)
#15 Future._propagateToListeners.handleValueCallback (dart:async/future_impl.dart:696:45)
#16 Future._propagateToListeners (dart:async/future_impl.dart:725:32)
#17 Future._completeWithValue (dart:async/future_impl.dart:529:5)
#18 Future._asyncCompleteWithValue.<anonymous closure> (dart:async/future_impl.dart:567:7)
#19 _rootRun (dart:async/zone.dart:1190:13)
#20 _CustomZone.run (dart:async/zone.dart:1093:19)
#21 _CustomZone.runGuarded (dart:async/zone.dart:997:7)
#22 _CustomZone.bindCallbackGuarded.<anonymous closure> (dart:async/zone.dart:1037:23)
#23 _microtaskLoop (dart:async/schedule_microtask.dart:41:21)
#24 _startMicrotaskLoop (dart:async/schedule_microtask.dart:50:5)
#25 _runPendingImmediateCallback (dart:isolate-patch/isolate_patch.dart:118:13)
#26 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:169:5)
```
stdout output attached as `output_1.txt`
[output_1.txt](https://github.com/flutter/flutter/files/5257450/output_1.txt)
Other things happen if I run again, or if I erase the app from the iPad, and run again. But maybe this is enough for now?
Please let me know if I can provide anything else useful.
```text
$ flutter doctor -v
[Truncated]
• Xcode 12.0, Build version 12A7209
• CocoaPods version 1.9.3
[✓] Android Studio (version 4.0)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin version 47.1.2
• Dart plugin version 193.7361
• Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593)
[✓] VS Code (version 1.49.0)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.14.1
[✓] Connected device (1 available)
• Stephen’s iPad (mobile) • b0975f3aa3ebe98c71462a866b0356a629fc6aa3 • ios • iOS 14.0
! Error: Stephen’s iPad is busy: Copying cache files from device. Xcode will continue when Stephen’s iPad is finished. (code -10)
• No issues found!
```
Answers:
username_1: Have you tried this at all?
username_2: Oh the app is crashing, that's why the observatory isn't found. The `Try accepting the local network permissions popup` comment is just confusing in this spot.
Can you attach the crash log from the device?
username_0: It doesn't look like a crash log was produced. I'm looking in Xcode under "view device logs" in "devices and simulators". There are other crash logs here (e.g. one from Skype, looks like it was terminated for using too much CPU when backgrounded).
I tried reproducing the behaviour while viewing the console to see what's going on there. This time (starting by creating a new test project with a different name) the app launched. However, the observatory never attached.
I tried a couple more times, and I have not been able to reproduce the crash on launch.
username_2: @username_0 Now that the app doesn't crash on launch (weird there's no log), do you see `Settings > Privacy > Local Network` now? Is it checked?
username_0: @username_2 I appreciate you following up on this. Here's what I see with the latest master (1.22.0-10.0.pre.318).
Summary: default app crashes, permission setting is available but doesn't seem to do anything useful.
First, I create a new project. I'm using new names for each project to ensure there is no interference from previous state on my iPad.
```text
$ flutter create testproject8
...
All done!
[✓] Flutter: is fully installed. (Channel master, 1.22.0-10.0.pre.318, on Mac OS X 10.15.6 19G2021 x86_64, locale en-GB)
[✓] Android toolchain - develop for Android devices: is fully installed. (Android SDK version 28.0.3)
[✓] Xcode - develop for iOS and macOS: is fully installed. (Xcode 12.0)
[✓] Android Studio: is fully installed. (version 4.0)
[✓] VS Code: is fully installed. (version 1.49.1)
[✓] Connected device: is fully installed. (1 available)
```
`$ cd testproject8/`
`flutter run -v > >(tee -a stdout.log) 2> >(tee -a stderr.log >&2)`
The usual stuff happens. The app appears on the iPad homescreen. The app launches. The app displays the "In iOS14+, ..." message. Meanwhile the `flutter run` command has exited.
[stdout.log](https://github.com/flutter/flutter/files/5268463/stdout.log)
[stderr.log](https://github.com/flutter/flutter/files/5268465/stderr.log)
I open the Settings app on the iPad and look under Privacy, Local Network. No app is visible for me to allow local network access. There is no crash log for this app available via Xcode.
```text
$ flutter doctor -v
[✓] Flutter (Channel master, 1.22.0-10.0.pre.318, on Mac OS X 10.15.6 19G2021 x86_64, locale en-GB)
• Flutter version 1.22.0-10.0.pre.318 at /Users/steve/code/flutter
• Framework revision 6b39acdc53 (62 minutes ago), 2020-09-23 08:27:04 -0400
• Engine revision 3e7ca5b869
• Dart version 2.10.0 (build 2.10.0-156.0.dev)
[✓] Android toolchain - develop for Android devices (Android SDK version 28.0.3)
• Android SDK at /Users/steve/Library/Android/sdk
• Platform android-29, build-tools 28.0.3
• Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java
• Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593)
• All Android licenses accepted.
[✓] Xcode - develop for iOS and macOS (Xcode 12.0)
• Xcode at /Applications/Xcode.app/Contents/Developer
• Xcode 12.0, Build version 12A7209
• CocoaPods version 1.9.3
[✓] Android Studio (version 4.0)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin version 47.1.2
• Dart plugin version 193.7361
• Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593)
[✓] VS Code (version 1.49.1)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.14.1
[✓] Connected device (1 available)
• Stephen’s iPad (mobile) • b0975f3aa3ebe98c71462a866b0356a629fc6aa3 • ios • iOS 14.0
! Error: Stephen’s iPad is busy: Copying cache files from device. Xcode will continue when Stephen’s iPad is finished. (code -10)
• No issues found!
```
username_3: Looks like it's just not iPad, happens on iPhone as well https://github.com/flutter/flutter/issues/66408
username_4: This error message may be the cause.
`Error: Stephen’s iPad is busy: Copying cache files from device. Xcode will continue when Stephen’s iPad is finished. (code -10)`
username_2: So it's not crashing anymore, it just sits on the error message? I wouldn't expect the permissions pop up to happen when it's sitting on that message.
What happens when you run `ios/Runner.xcworkspace` from Xcode?
username_0: Running from Xcode, the app launches with a blank white screen, and the iOS dialog "testproject9 would like to find and connect to devices on your local network"
I press "OK" and then the app fully launches.
In the output window of Xcode is the following:
```text
2020-09-24 12:20:41.808173+0200 Runner[4478:1907931] Metal API Validation Enabled
2020-09-24 12:20:42.338640+0200 Runner[4478:1908359] flutter: Observatory listening on http://127.0.0.1:49305/u3r3W6zN1P8=/
2020-09-24 12:21:33.790822+0200 Runner[4478:1908352] fopen failed for data file: errno = 2 (No such file or directory)
2020-09-24 12:21:33.790937+0200 Runner[4478:1908352] Errors found! Invalidating cache...
2020-09-24 12:21:33.918937+0200 Runner[4478:1908352] fopen failed for data file: errno = 2 (No such file or directory)
2020-09-24 12:21:33.919056+0200 Runner[4478:1908352] Errors found! Invalidating cache...
```
username_0: Immediately after this, I stopped the app from Xcode, closed Xcode, and tried again with `flutter run`
I got the following terminal output, and the app launches with the "In iOS 14+, Flutter application in debug mode..." message.
```text
$ flutter run
Launching lib/main.dart on Stephen’s iPad in debug mode...
Automatically signing iOS for device deployment using specified development team in Xcode project: RPX9H7TNPD
Running Xcode build...
└─Compiling, linking and signing... 9.2s
Xcode build done. 14.7s
Failed to attach to the observatory.
Try accepting the local network permissions popup, or enable "Settings > Privacy > Local Network" for your app.
For more information, see https://flutter.dev/docs/development/ios-14#debugging-flutter
Installing and launching... 28.3s
Error launching application on Stephen’s iPad.
```
username_2: That's really odd.
```
[ +23 ms] (lldb) run
[ +85 ms] success
[ ] Application launched on the device. Waiting for observatory port.
```
`flutter` attached the debugger, so it should never show that "In iOS 14+" error message. It's not surprising the observatory isn't connecting at that point though.
Would you mind opening the Console app, selecting your app, and reproducing the problem, and attaching any `Runner` logs? Not sure why the engine thinks it can't run ptrace.
<img width="945" alt="Screen Shot 2020-09-24 at 7 08 08 PM" src="https://user-images.githubusercontent.com/682784/94218407-4936c900-fe99-11ea-9cc5-2f5939c41f93.png">
username_5: Hi All. I am having this problem too. FYI, my iPad has been busy copying files for 7 days now.... I've tried flutter clean multiple times. Anyone have any solutions or workarounds?
flutter doctor --verbose
[✓] Flutter (Channel master, 1.22.0-10.0.pre.380, on Mac OS X 10.15.6 19G2021 x86_64, locale en-US)
• Flutter version 1.22.0-10.0.pre.380 at /Users/rob/dev/flutter
• Framework revision ec40df9576 (17 hours ago), 2020-09-25 21:27:22 -0700
• Engine revision 3a73d073c8
• Dart version 2.11.0 (build 2.11.0-161.0.dev)
[✓] Android toolchain - develop for Android devices (Android SDK version 29.0.2)
• Android SDK at /Users/rob/Library/Android/sdk
• Platform android-29, build-tools 29.0.2
• Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java
• Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593)
• All Android licenses accepted.
[✓] Xcode - develop for iOS and macOS (Xcode 12.0.1)
• Xcode at /Applications/Xcode.app/Contents/Developer
• Xcode 12.0.1, Build version 12A7300
• CocoaPods version 1.9.3
[✓] Chrome - develop for the web
• Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome
[✓] Android Studio (version 4.0)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin version 45.1.1
• Dart plugin version 192.7761
• Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593)
[✓] IntelliJ IDEA Ultimate Edition (version 2019.2.4)
• IntelliJ at /Applications/IntelliJ IDEA.app
• Flutter plugin version 41.1.4
• Dart plugin version 192.7402
[✓] VS Code (version 1.49.2)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.13.2
[✓] Connected device (2 available)
• Web Server (web) • web-server • web-javascript • Flutter Tools
• Chrome (web) • chrome • web-javascript • Google Chrome 85.0.4183.121
! Error: CM2 is not connected. Xcode will continue when CM2 is connected. (code -13)
username_2: `Try accepting the local network permissions popup` was just suggesting that it could possibly be a possible reason for a failure to connect to the observatory, but people took it literally. I removed it here https://github.com/flutter/flutter/pull/66621.
username_0: I can no longer reproduce this problem with the latest Master, using a fresh project created with `flutter create`.
I first experienced problems when I was using packages that had native parts, using cocoapods. I'll try that again, and check.
```text
$ flutter doctor -v
[✓] Flutter (Channel master, 1.22.0-10.0.pre.428, on Mac OS X 10.15.6 19G2021 x86_64, locale en-GB)
• Flutter version 1.22.0-10.0.pre.428 at /Users/steve/code/flutter
• Framework revision a06f385145 (12 hours ago), 2020-09-29 21:37:04 -0400
• Engine revision d18c3ea75c
• Dart version 2.11.0 (build 2.11.0-178.0.dev)
[✓] Android toolchain - develop for Android devices (Android SDK version 28.0.3)
• Android SDK at /Users/steve/Library/Android/sdk
• Platform android-29, build-tools 28.0.3
• Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java
• Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593)
• All Android licenses accepted.
[✓] Xcode - develop for iOS and macOS (Xcode 12.0)
• Xcode at /Applications/Xcode.app/Contents/Developer
• Xcode 12.0, Build version 12A7209
• CocoaPods version 1.9.3
[✓] Android Studio (version 4.0)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin version 47.1.2
• Dart plugin version 193.7361
• Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593)
[✓] VS Code (version 1.49.1)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.14.1
[✓] Connected device (1 available)
• Stephen’s iPad (mobile) • b0975f3aa3ebe98c71462a866b0356a629fc6aa3 • ios • iOS 14.0.1
• No issues found!
```
username_6: Hello, are there any updates for this? :) |
sultan-of-rum/morrowind-scripting-for-dummies | 982701702 | Title: Add discussion on OpenMW-CS
Question:
username_0: This is a bit of a stretch goal.
With cc9cii's and unelsson's work on [OpenMW-CS](https://openmw.readthedocs.io/en/latest/manuals/openmw-cs/index.html) (example [1](https://gitlab.com/OpenMW/openmw/-/merge_requests/1052) and [2](https://gitlab.com/OpenMW/openmw/-/merge_requests/1176)), the program is rapidly approaching the point where it may compete with TESCS in dialogue and quest development workflows. Hence, it might pay to include mention of it in the book and to bring out the respective caveats for both programs.
An important thing to mention here is that OpenMW-CS does not compile script bytecode -- hence, to make the generated plugin files compatible with vanilla CS, the plugin file needs to be resaved in TESCS.
Answers:
username_1: I will indeed be mentioning the OpenMW-CS because I will mostly be using their default syntax highlighting. |
apache/pulsar | 590669167 | Title: Broker and Proxy support multiple CA certs
Question:
username_0: I'm requesting the Broker and Proxy be updated to support multiple CA certs. I have a requirement that internal communication uses an internal CA cert and we allow customers to define their own CA cert alongside.
Answers:
username_1: If you're talking about your brokers accepting multiple CAs in tlsTrustCertsFilePath I _think_ that is already supported. You should be able to cat all the ca pem files into one big file and use that as your trust cert store. I have not done it personally because we have no such requirement but I think it will work. You will get a pem file like this:
`-----BEGIN CERTIFICATE-----`
`<your CA>`
`-----END CERTIFICATE-----`
`-----BEGIN CERTIFICATE-----`
`<their CA>`
`-----END CERTIFICATE-----`
Status: Issue closed
|
w3c/csswg-drafts | 280149848 | Title: [cssom] using scroll wheel before scrollIntoView() finishes
Question:
username_0: I'm looking at [the spec](https://drafts.csswg.org/cssom-view/#dom-element-scrollintoview) and maybe I am missing it, but it doesn't seem to explain what happens when a user takes some contradictory action before the scrollIntoView() finishes, like use the scroll wheel (on their mouse), press a scroll key on keyboard (page up, page down), or click another element that triggers a different scrollIntoView().
Should the initial scrollIntoView() abort? Or should it finish the entire scroll sequence and delay the contradictory action? Or just cancel out the contradictory action? |
mrkkrp/megaparsec | 112199261 | Title: Make expression parser parse repeated prefix/postfix operator
Question:
username_0: For more information see this SO question:
https://stackoverflow.com/questions/10475337/parsec-expr-repeated-prefix-postfix-operator-not-supported
Note to self: we cannot remove chaining combinators until this is implemented.
Answers:
username_0: I think it should be implemented in 4.2.0, not in 4.3.0. Obsolete combinators of `chain` family fill be removed in 4.2.0 too. This version will be released in December probably, don't want to release it too soon because 4.1.1 should be fine for most users.
Status: Issue closed
username_1: Not sure if this is the appropriate place to ask, but how do can I parse multiple "lookups" with `Postfix`? For example, for the array lookup syntax you can find in several languages:
[1,[2,3]][1][0]
username_0: @username_1, Unfortunately I have no time right now to write working example, but see in documentation how to parse several unary operators in a row:
https://hackage.haskell.org/package/megaparsec-4.4.0/docs/Text-Megaparsec-Expr.html#v:makeExprParser
Basically, you first write a parser for `[x]` things that should actually allow nested indices and then you:
```haskell
parseManyPostfixIndices = foldr1 (.) <$> some parseOneIndex
```
This should do the trick.
username_1: I should have read the docs! Thanks a lot, and sorry for the noise.
username_2: Sorry to rehash this, but I can't for the life of me figure out goes in the `parseOneIndex` space. The type, according to ghci, is `Functor f => f (a -> a)` but it's very unclear to me if that is the expression parser, the term parser, the operator, or something else.
username_0: @username_2, `parseOneInex` should parse an index, such as `[1]` for example, and return a function that would change some `a`, which is probably what you want to return from the whole parser, i.e. a term.
For example:
```
my_array[0][1]
```
* `my_array` could on its own return something like `Var "my_array"`.
* The indices could add something like this: `Indexed (Indexed (Var "my_array") 0) 1`. So the function that `parseOneIndex` would return could be: `\x -> Indexed x n`, where `n` comes from whatever you parse in that parser, of course.
username_2: Ah, I see. So, to parse multiple prefix operators, as in the example in the haddocks, one would change
`prefix name f = Prefix (f <$ symbol name)` to
`prefix name f = Prefix $ foldr1 (.) <$> some (f <$ symbol name)`.
Thank you very much for your help, and for writing a great library!
username_3: ```
opSubscript = Postfix $ foldr1 (.) <$> some singleIndex
singleIndex = do
index < brackets expr
return $ \l -> ArrayIndex l index
```
array[1][2] is expected to be parsed as
`ArrayIndex (ArrayIndex (Var "array") 1)) 2`
However, it was
`ArrayIndex (ArrayIndex (Var "array") 2)) 1`
What is going wrong here?
username_3: [Fixed](https://www.reddit.com/r/haskell/comments/lfnd3o/issue_when_parsing_array12_using_megaparsec/) with `foldr1 (flip (.))` |
mpetroff/pannellum | 1119959999 | Title: Adsense
Question:
username_0: Is it possible to add adsense to my pano player, if people hotlink and use my popular images by throttling my resources it wouldn't be a problem if i could enable adsense, i use your amazing panorama player as a nasa nerd star guide, if one of my 20mb milkyway images becomes amazingly popular it'll slow my modest server resources. If i can enable adsense it would fund my modest resources instead of burdening my server.
Answers:
username_0: update: - oh yeah! forgot to mention with an option to hide the adsense (x) so people can gaze at my nerdy 20mb images without visible obstruction.
username_1: You can add whatever you want to the page you're using it on. You can also overlay content on the viewer, e.g., as is done in [the custom controls example](https://pannellum.org/documentation/examples/custom-controls/).
username_0: i can't find a file labelled Overlay content or Custom Controls, i'm not a coder.
.
I'll probably add an infolinks instead of my adsense cause infolinks has a neat close ability, i don't know which file to edit to add my javascript so it shows on my player.
.
Trying to add this
```
<script type="text/javascript">
var infolinks_pid = 3356228;
var infolinks_wsid = 0;
</script>
<script type="text/javascript" src="//resources.infolinks.com/js/infolinks_main.js"></script>
```
username_1: There's nothing named "overlay." The custom controls in that example are overlaid on the viewer. You can overlay other content in the same manner that that was done.
username_0: That tutorial is very limited it doesn't tell you what file to edit, i don't know which file to edit
username_1: It's an example, not a tutorial. You can overlay content on the viewer the same way you'd overlay it on any other `<div>` element.
I've pointed you in the direction of what you need, but I have neither the time nor the motivation to explain basic HTML / CSS, so the rest is up to you. Unless you formulate a concrete question specific to how such content interacts with Pannellum, there's not much more for me to do.
username_0: Okay fine, i was just a little confused with the Pannellum 50 kb version VS the 12mb version. So i uploaded both onto my server and forgot which one i was linking too in my spaghetti code. So i messed around with the 50kb version first and it seemed to work fine. I need to ask this however, i'm freelancing for a large known big brand, would i be correct on telling them that they can place advertisements on your creation or is that illegal from your optics
username_0: They're now asking for ad html placements i hope i'm not breaking your copyrights, i am sincere which is why i ask.
username_1: Your client's attorneys are free to review the license, but it allows for commercial use:
https://github.com/username_1/pannellum/blob/master/COPYING
Status: Issue closed
|
google/gapid | 297537391 | Title: Frames shown are different than output on device screen (Unity3D)
Question:
username_0: When using GAPID to debug Unity3D applications built for Android using OpenGL ES 3, the frames shown in the debugger are extremely different from the what I see on the device's screen in certain cases. It probably has something to do with transparency since it's only when I introduce alpha mask textures in my shaders that it displays this behavior. Typically, in GAPID I see what I would actually like to see on my device while my device shows something different, hence the debugging. Is there some underlying reason that there is a difference? It might help in actually debugging the erroneous behavior on the device.
Answers:
username_1: Hi @username_0,
GAPID does some translation to replay OpenGL ES on desktop GL. It is possible there's a bug in this code.
Are you willing and able to share a trace file demonstrating this issue? It would be a great help in figuring out the problem.
Cheers,
Ben
username_0: [gapid_trace.zip](https://github.com/google/gapid/files/1728862/gapid_trace.zip)
This is the trace. That is what I was expecting to see, but on my device I see a solid grey with a slightly darker row on the right. Texture 31 is supposed to be my alpha mask from which I sample the red channel. In this debug build I simply used a shader that displayed this texture full screen to see what was going on.
username_0: [gapid_trace2.zip](https://github.com/google/gapid/files/1759359/gapid_trace2.zip)
@username_1 this is another trace with the same result: What I see in GAPID is what I would want to see on my device but instead on the device all I see is a dark grey fill. One additional detail since last time is that this only happens on Nougat devices while Oreo devices work fine. Also, I get a repeating "bindTextureImage: clearing GL error: 0x502" in the adb log when running on Nougat devices.
username_1: Hi @username_0 ,
Thank you for the reports. Sorry for the delay in looking into this, I've been working on something that is particularly high priority (it's blocking us from doing any more releases). Fortunately, I'm very close to finishing this work, so I'll hopefully be taking a look into this later this week.
Thank you for your patience.
Cheers,
Ben
username_0: @username_1 Thank you for your interest. Just wanted to add another clue: by coincidence the Nougat device has a Mali G-71 (Galaxy Note 8) while the Oreo devices (Nexus 6P and OnePlus 3) both have an Adreno. Wondering if maybe the android version has nothing to do with it.
username_1: Hi @username_0,
I've finally found the time to look into this.
There are two suspicious things with these traces which are highlighted in the report tab:
From `gapit_trace.gfxtrace`:
Invalid Operation
* `30861`: Error `glBindTexture`
* `30869`: Error `glBindTexture`
Both of these are erroring because you've bound the texture as an external texture (`GL_TEXTURE_EXTERNAL_OES`), and then (where it errors) you attempt to bind it as a texture 2D (`GL_TEXTURE_2D`). This is illegal. The texture takes the kind on the first binding, so texture 31 will be an external texture.
As for why it looks different - it's probable this is just caused by undefined behavior due to the errors described above.
Does this resolve your original question?
Many thanks,
Ben
username_0: Thanks for the response but unfortunately I don't think that was it as those two keep showing up in the report even now that everything works correctly. I think those two glBindTextures fail because something is still initializing at that point but they get called every frame so they should be working.
username_1: I'm happy to take another look if you don't mind sending another trace.
username_0: Is there a way to cut frames from a trace? Trace capturing is slowing the application down quite significantly which means that the trace has to run for quite some time before the app reaches the point where everything is set up and it is executing the aforementioned alphamasking operation. The trace files way up to 1gb.
username_1: Currently, I'm afraid not. We are working on adding mid-execution capture support for OpenGL ES, which should fix this issue.
Status: Issue closed
|
gusaul/go-dynamock | 942255999 | Title: TransactWriteItems TableName expectation check for difference string references
Question:
username_0: Hello,
In my code the assignment of tableName is done when constructing the DynamoDB SDK payloads, by using a string value that is passed around the different packages of my application.
I assume most clients would rather use the String value than the pointer, and use `aws.String(myTable)` when constructing the `dynamodb.TransactWriteItemsInput` items.
The side-effect of this, is that aws.String() creates a new reference each time, and therefore the equality check fails on go-dynamock, because it compares the references and not the value.
Please find the following PR with my naive approach to fix the problem : https://github.com/gusaul/go-dynamock/pull/41 |
mageplaza/magento-2-webhook | 535695724 | Title: Webhook for customer newsletter subscription
Question:
username_0: Hello,
is there a webhook I can use when a customer subscribes to the newsletter? If not it would be nice to have to be able to integrate with some CRMs like Agile CRM for example.
Thank you!
Answers:
username_1: Hi @username_0
I’m Shin from Mageplaza.com and I will support you on this issue.
For your question, you can create a new hook with the 'subscriber' option.
Please see this screenshot:
- http://prntscr.com/q93g80
Thanks so much!
username_0: Thank you very much!
Status: Issue closed
username_0: Hello,
is there a webhook I can use to trigger when a customer subscribes to the newsletter? If not it would be nice to have in order to be able to integrate this action with some CRMs like Agile CRM for example.
Thank you!
username_0: Hello I tested it (Subscriber Webhook) on 2.3.3 (with PHP 7.2), does not work, too. There is no log. Any suggestion?
Thanks
username_0: Found the fix here: https://github.com/mageplaza/magento-2-webhook/issues/12#issuecomment-521190268
Status: Issue closed
|
TheImagingSource/tiscamera | 398929327 | Title: Raspberry PI 2 not able to access some camera properties, cannot set Trigger mode
Question:
username_0: When trying to access some properties of the camera on a Raspberry PI 2, they are not available. One example is the Trigger mode, which I can't access. When I run tcam-capture I can see the camera, the videofeed and set several options, like exposure and white balance. However on my laptop I can control many more properties, including partial readout modes and trigger modes. I compiled the code on the Pi from the newest source.
This is a message I get in the console after starting tcam-capture.
libEGL warning: DRI2: failed to authenticate
qt5ct: using qt5ct plugin
14-01-2019:15:57:09,744 WARNING [CapsDesc.py:167] Could not interpret structure. Omitting. video/x-raw, format=(string){ RGBx, xRGB, BGRx, xBGR, RGBA, ARGB, BGRA, ABGR }, width=(int)640, height=(int)480, framerate=(fraction){ 60/1, 30/1, 15/1, 5/1, 1/1 };
14-01-2019:15:57:09,746 WARNING [CapsDesc.py:167] Could not interpret structure. Omitting. video/x-raw, format=(string){ RGBx, xRGB, BGRx, xBGR, RGBA, ARGB, BGRA, ABGR }, width=(int)1920, height=(int)1080, framerate=(fraction){ 15/1, 5/1, 1/1 };
14-01-2019:15:57:09,747 WARNING [CapsDesc.py:167] Could not interpret structure. Omitting. video/x-raw, format=(string){ RGBx, xRGB, BGRx, xBGR, RGBA, ARGB, BGRA, ABGR }, width=(int)2048, height=(int)2048, framerate=(fraction){ 5/1, 1/1 };
14-01-2019:15:57:09,748 WARNING [CapsDesc.py:167] Could not interpret structure. Omitting. video/x-raw, format=(string){ RGBx, xRGB, BGRx, xBGR, RGBA, ARGB, BGRA, ABGR }, width=(int)3072, height=(int)2048, framerate=(fraction){ 5/1, 1/1 };
14-01-2019:15:57:11,351 WARNING [TcamView.py:482] get_tcam_property failed for 'Trigger Mode'
inotify_add_watch("/home/pi/.config/qt5ct") failed: "No such file or directory"
My main goal is to be able to enable Trigger mode and trigger the camera externally.
When trying to run the example Save an image on trigger in Python I get the following error:
tcambin serial="48810687" name=source ! video/x-raw,format=BGRx,width=640,height=480,framerate=30/1 ! videoconvert ! appsink name=sink
Traceback (most recent call last):
File "Program.py", line 75, in <module>
print("Gain Auto : %s " % Tis.Get_Property("Gain Auto").value)
File "/home/pi/dev/Linux-tiscamera-Programming-Samples-bak/Save an image on trigger in Python/TIS.py", line 160, in Get_Property
return CameraProperty(*self.source.get_tcam_property(PropertyName))
Other examples which use trigger work, but the Trigger mode True does not seem to apply, they just stream the frames continuously regardless of the trigger being sent.
Answers:
username_1: Hello
usually the uvc extension xml file and is installed with the "sudo make install" process. I would like you to check, whether the xml is correcly installed. A documentation about this can be found at
https://github.com/TheImagingSource/tiscamera/wiki/UVC-Extension-Units
I would like you to send me the output of
`tcam-ctrl -p 48810687`
also
`tcam-ctrl -c 48810687`
would be nice.
Stefan
username_0: Hi,
thanks for the quick reply.
The ouput of tcam-ctrl -p 48810678:
Found 7 propert(y/ies)
Brightness (int) min=0 max=4095 step=1 default=-8193 value=240 category=Exposure
Whitebalance Auto (bool) default=true value=true category=Color
Whitebalance Red (int) min=0 max=255 step=1 default=64 value=64 category=Color
Whitebalance Blue (int) min=0 max=255 step=1 default=64 value=64 category=Color
Gain (int) min=0 max=480 step=1 default=57343 value=0 category=Exposure
Exposure Auto (bool) default=false value=true category=Exposure
Exposure (int) min=100 max=60000000 step=100 default=300 value=300 category=Exposure
and of tcam-ctrl -c 48810687:
Available gstreamer-1.0 caps:
video/x-bayer, format=(string)rggb, width=(int)640, height=(int)480, framerate=(fraction){ 60/1, 30/1, 15/1, 5/1, 1/1 };
video/x-bayer, format=(string)rggb, width=(int)1920, height=(int)1080, framerate=(fraction){ 15/1, 5/1, 1/1 };
video/x-bayer, format=(string)rggb, width=(int)2048, height=(int)2048, framerate=(fraction){ 5/1, 1/1 };
video/x-bayer, format=(string)rggb, width=(int)3072, height=(int)2048, framerate=(fraction){ 5/1, 1/1 };
video/x-bayer, format=(string)rggb16, width=(int)640, height=(int)480, framerate=(fraction){ 30/1, 15/1, 5/1, 1/1 };
video/x-bayer, format=(string)rggb16, width=(int)1920, height=(int)1080, framerate=(fraction){ 7/1, 5/1, 1/1 };
video/x-bayer, format=(string)rggb16, width=(int)2048, height=(int)2048, framerate=(fraction){ 2/1, 1/1 };
video/x-bayer, format=(string)rggb16, width=(int)3072, height=(int)2048, framerate=(fraction){ 2/1, 1/1 }
I will check the documentation about the xml file.
Thanks,
Cenek
username_1: Hi Cenek
thank you. You are right, many properties are missing. I wait for your result regarding the uvc extension.
Stefan
username_0: Hi Stefan,
I tried the procedure described at https://github.com/TheImagingSource/tiscamera/wiki/UVC-Extension-Units without success. Unfortunately I cannot produce any meaningful results now, because both our cameras refuse to send any pictures due to broken connector. I need to wait for new cameras to arrive.
username_1: I already answered to you by email. I do not think, the connector is broken, because the cameras worked fine, when they left the office.
Plugin the plug into the camera, until there was faint click sound. You may need some power doing that.
Stefan
Status: Issue closed
|
OpenChemistry/tomviz | 111692885 | Title: C++11
Question:
username_0: @username_1 I'm making this issue to record our discussions about starting to use C++11 features in tomviz since it's been a few weeks and I don't recall all the details anymore.
Things we had talked about starting to use (at least in the short term):
override
std::nullptr
Unlike VTK, we only care about newer compilers. So the question is do we need to make macros for these or can we just start using them directly and turn on the cmake switch that will fail to configure on compilers that do not support them?
Answers:
username_0: Notes:
use check_cxx_compiler_flag for testing c++11 present
see [this](http://review.source.kitware.com/#/c/16780/1/CMake/vtkCompilerExtras.cmake)
windows: VS2013 is minimum supported version since we don't test anything else
linux: looks like we need minimum of debian wheezy (7) to get a new enough gcc to build with c++11
Status: Issue closed
|
InFact-coop/create-refresh | 391076478 | Title: Support for different languages
Question:
username_0: - [ ] detect user language
- [ ] Pull in content from external source in appropriate language
Answers:
username_0: @username_1 if this is something that is near the top of your priorities in the backlog then it would be a good idea to get the content ready, in case we find time to do this. If it's at the bottom of your list of priorities, then it's unlikely to get done so you don't need to worry 😄
username_1: Getting this covered early might be a bit difficult as we need to get final approval on copy (which we probably won't get until the new year) and then get that translated. What is the alternative if we didn't get this done? And what is the likelihood that you would have time for this?
username_0: @username_1 the likelihood we'll have time depends on how near the top of your priority list it is. Did you move the items in the backlog around already?
I suppose even if we don't get the final copy until we're nearly finished, we could build the external source that we pull the content in from, have all of them in English to begin with, and then you guys would have to update it with the other languages once you have translations.
username_1: @username_0 Is the backlog an actual place on Github where I can prioritise things? Sorry - still getting used to this. I would say it would make a difference in getting sign ups if we had different language options so we should make it more a priority in the backlog - although not sure where to add this!
username_0: @username_1 no worries! yes it is, if you go [here](https://github.com/InFact-coop/create-refresh/milestone/3) you can add priority labels to all of the issues, according to your feelings 😃
Hopefully this picture helps:
 |
badges/shields | 55941451 | Title: Node callback should be err, response
Question:
username_0: The accepted format is:
```js
function (error, response) {
}
```
If you don't use this format, then your API is not open to being promisified, e.g:
```js
var Promise = require('bluebird');
var badge = Promise.promisify(require('gh-badges'));
badge({..})
.then(function (svg) {
// svg here
})
.catch(function (error) {
// error here
});
```
Basically how it currently is, you can't do the above.
Answers:
username_1: That's a fair point, although this change would obviously break backward-compatibility.
I suppose we could have this for v2.0.0. Let's [make an issue][] for that.
[make an issue]: https://github.com/badges/shields/issues/362
Status: Issue closed
|
very-good-science/data-hazards | 1020891009 | Title: Create Data Hazards lesson plan materials
Question:
username_0: - [ ] Create lesson plan page
- [ ] Use explainer videos from asynchronous materials #77
- [ ] Make an explainer video for a project/projects or explain how to do it live (link to materials for workshop)
- [ ] Add timeline for lesson
- [ ] Add to project timeline on homepage
Status: Issue closed
Answers:
username_1: These resources are now on the website and reusable :) |
angular/angular | 176270794 | Title: Compiler should detect invalid property binding
Question:
username_0: **I'm submitting a ...** `[X ] feature request`
If possible, Angular2 compiler should detect and flag the following as an error:
Component A has an input property defined as follows:
`@Input() rows: number;`
The template of a parent component (hosting Component A) references the property `rows` as an attribute like this:
`<comp-a ... rows="5">`
instead of
`<comp-a ... [rows]="5">`
In Angular RC6/Typescript 2.0.2, this compiles, but the value of `this.rows` in Component A is the string `"5"` as opposed to the number `5`.
**What is the motivation / use case for changing the behavior?**
Developer productivity.
**Please tell us about your environment:**
Windows 8.1 x64, Visual Studio Code
* **Angular version:** 2.0.0-rc.6
<!-- Check whether this is still an issue in the most recent Angular version -->
* **Language:** TypeScript 2.0.2
Answers:
username_1: This may help https://github.com/angular/angular/issues/6919
username_2: It is not an error.
think `<p title="...">` it makes sense for directive to get the content of the title
Status: Issue closed
|
brittyazel/Neuron | 772008539 | Title: Zone ability not showing sometimes
Question:
username_0: As you level zone abiltiy is empty from a start but sometimes it switches on later as you quest or you get second ability. Also yesterday i noticed that when i got ventyr ability to change execute into condem -it did change execute into condem but didnt change icon for it on bar.
Answers:
username_0: Also there is no CD showing for Condemn.
username_0: Fixing bugs in Neuron lately?
username_1: I'm very busy at the moment, and am fixing things when and where I can. Neuron is not my priority at the moment, no. Once I finish the immediate deadlines in my personal life, I'll likely ramp Neuron work back up. Until then, I'll continue to do drive by fixes when I have a spare hour or two. 80+ hour work weeks make it hard to devote time to side projects, sadly.
username_0: Its not big problem . Its just hard to see if you have given up on addon or just too busy. I havent received comment on this bug from 21 dec 2020 and too many times i have seen authors give up on addon and never tell users of their addon that they dont have time/interest to maintain it any longer. And given history i had with you as authors - meaning you at least acknowledeged bugs preatty quick even i you didnt fix it that moment made me wonder whats happening with addon which i used for very long (back to Maul days ).
username_0: Any chance you looked at Condemn problem? Now every time i log in for first time on warr and try to use Condemn i get spell not learned message and i have to take spell out of spellbook again and relog for Condemn to start working.
username_1: Have you tried using the spellID?
username_0: Is that even possiable for spells? For fun i tried it but nothing happens
/cast 317349
username_1: Oh you're right! My mistake. I'll see what I can do to check this out. What class/spec is this?
username_0: warr venthyr fury
username_1: https://eu.forums.blizzard.com/en/wow/t/condemn-macro-doesnt-work/214663
username_1: Is this your issue? Seems like an issue with macros in general
username_0: Not sure if its totally same since i dont get spell disapering from spellbook but in general yes. Plus i dont get 4sec cd timer on it .Just 1 sec global cooldown spin. |
YingboMa/MaBLAS.jl | 622853994 | Title: The analytical model for blocking
Question:
username_0: ```julia
julia> using Hwloc
julia> function params(::Type{T}) where T
micro_m = LoopVectorization.mᵣ * VectorizationBase.pick_vector_width(T)
micro_n = LoopVectorization.nᵣ
topology = Hwloc.topology_load()
l3 = getdatacache(topology, :L3Cache)
l2 = getdatacache(l3, :L2Cache)
l1 = getdatacache(l2, :L1Cache)
l1associativity = l1.attr.associativity
l1associativity = l1associativity == 0 ? 8 : # assume 8 if unknown
l1associativity == -1 ? l1.attr.size ÷ l1.attr.linesize : # fully associative
l1associativity
l1sets = l1.attr.size ÷ (l1.attr.linesize * l1associativity)
ncachelines_per_panelA = floor(Int, (l1associativity - 1) / (1 + micro_n/micro_m))
cache_k = ncachelines_per_panelA * l1.attr.linesize * l1sets ÷ (micro_m * sizeof(T))
end
params (generic function with 1 method)
julia> getdatacache(topology, name) = for t in topology.children
isdefined(t, :type_) || return nothing
return t.type_ === name && t.attr.type_ in (:Unified, :Data) ? t : getdatacache(t, name)
end
getdatacache (generic function with 1 method)
julia> params(Float64)
213
julia> params(Float32)
256
```
someone may find this code useful.
Answers:
username_1: This is an, uhh, interesting choice
```julia
julia> params(Float32)
76
julia> params(Float64)
76
``` |
Azure/azure-functions-sql-extension | 679365052 | Title: Deserialization into POCO
Question:
username_0: Currently, whenever we have to deserialize the SQL rows from the user's table into a POCO (like [here](https://github.com/Azure/azure-functions-sql-extension/blob/dev/src/SqlBinding/SqlAsyncEnumerable.cs#L117) for the input binding and [here](https://github.com/Azure/azure-functions-sql-extension/blob/dev/src/SqlBinding/TriggerBinding/SqlTableWatcher.cs#L735) for the trigger binding), we first build up a dictionary mapping from column name to column value. We then serialize that dictionary into a JSON string, and then deserialize the dictionary into the POCO type. This is a pretty round-about and inefficient way to do it, so it might be worth looking into alternatives, such as building our own deserializer that uses reflection to map column names to property names. |
udistrital/generator-oas | 200680670 | Title: Finalización del proceso del token.
Question:
username_0: El login se hace mediante AngularJS-OAuth2, pero tenia problemas para manipulacion del token.
Hasta hora el token se almacena en localstorage usando usando ngStorage, a su vez es posible decriptarlo, para esto se uso https://jwt.io/ -> kjur-jsrsasign.
una version funcional de un proyecto generado está en https://fabianleon.github.io/app/#/, autenticandose con google.
Está pendiente hacer que el token muera. y hacer una solicitud del nuevo token. la manipulacion del token se hace desde la fabrica token en scripts/services/token.js |
github-vet/rangeloop-pointer-findings | 775620897 | Title: cd1989/cycli: pkg/stages/render.go; 8 LoC
Question:
username_0: [Click here to see the code in its original context.](https://github.com/cd1989/cycli/blob/f9749115fb0f7d250217ca68bcdaa1569c85eb8c/pkg/stages/render.go#L39-L46)
<details>
<summary>Click here to show the 8 line(s) of Go which triggered the analyzer.</summary>
```go
for _, stg := range stgs {
rows = append(rows, []string{
stg.Name,
stg.Namespace,
getLabel(&stg, meta.LabelProjectName),
stg.CreationTimestamp.Format(time.RFC3339),
getAnnotation(&stg, meta.AnnotationDescription)})
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: f9749115fb0f7d250217ca68bcdaa1569c85eb8c |
yasirkula/UnityIonicIntegration | 354758090 | Title: iOS Splash Screen
Question:
username_0: We have tried a variety of iOS splash screen settings, but it seems that regardless of what we do we get a stretched version of the normal app splash screen when we open the Unity portion of our app. Has anyone had any luck changing the Unity splash screen when using this plugin?
Status: Issue closed
Answers:
username_1: Did you resolve this issue?
username_0: No...we ended up giving up on doing the custom splash screen like we were hoping for. |
prettier/plugin-pug | 625408539 | Title: Bug: A blank line following multi-line comments is indented
Question:
username_0: ## Info
| Tool | Version |
| -------- | ------------- |
| Plugin | v1.4.0 |
| Prettier | v2.0.5 |
| Node | v12.17.0 |
| OS | mac |
## Input
```pug
//-
a
p hi
```
## Output
```pug
//-
a
p hi
```
White spaces are inserted at the beginning of the 3rd line.
## Expected Output
```pug
//-
a
p hi
```
## Additional Context<issue_closed>
Status: Issue closed |
acornjs/acorn | 737432042 | Title: Cannot parse / operator after ternary
Question:
username_0: **Input:**
```js
weird ? true : {}/2;
```
**Current:**
[AST Explorer link](https://astexplorer.net/#/gist/ff5b87d4d1b548b774e81259c837ef02/d0e72a33af518dda328e833237d498e05a799b71)
```
Unterminated regular expression (1:18)
```
**Expected:**
No error is thrown.
**Context:**
https://github.com/babel/babel/issues/11353<issue_closed>
Status: Issue closed |
invertase/react-native-firebase | 368527395 | Title: OnNotification not being called on release mode [iOS]
Question:
username_0: <!---
BEFORE YOU MAKE AN ISSUE
The issue list of this repo is exclusively for bug reports.
1) For feature requests, please use our Canny board: https://react-native-firebase.canny.io/feature-requests
2) For questions and support please use our Discord chat: https://discord.gg/C9aK28N or Stack Overflow: https://stackoverflow.com/questions/tagged/react-native-firebase
3) If this is a setup issue then please make sure you've correctly followed the setup guides, most setup issues such as 'duplicate dex files', 'default app has not been initialized' etc are all down to an incorrect setup as the guides haven't been correctly followed.
-->
### Issue
onNotification event is being called on debug mode, but when running on release it’s not being called, only getInitialNotification is working on release mode
<!--- Please write your issue here, provide as much detail as you can, code snippets, key files which will help us to debug such as your `Podfile` and/or `app/build.gradle` file). -->
### Environment
1. Application Target Platform: iOS
<!--- (e.g. Sierra, Windows 10) --->
2. Development Operating System: macOS
3. Build Tools: xcode
<!--- (e.g. 0.45.1) --->
4. React Native version: 0.55.4
5. RNFirebase Version: latest
<!--- (e.g. database, auth, messaging, analytics etc - or N/A if not applicable) --->
6. Firebase Module: Notifications/ messaging
<!-- Love react-native-firebase? Please consider supporting our collective:
👉 https://opencollective.com/react-native-firebase/donate -->
Answers:
username_1: Same issue. Notification listener `OnNotification` works in development. When I build release app on IOS I will get only once notification in Foreground. If I set the app in the background and go back to the foreground, I will not get `foreground` notification any more. But background notification works.
username_2: I have the same issue in android
onNotification being triggered on emulatore but once i do a build -relaese mode it never triggered on real device !
username_3: Same situation on IOS - in development notifications works, in production doesn't
username_4: same problem. any solution ?
username_4: @username_1
I have same issue as you. did you find any solution?
username_4: same issue
@username_3 @username_2 @username_1 did you find any solution ? |
davidgranstrom/tidal-bootstrap | 260041298 | Title: Cabal deps
Question:
username_0: Hi this script stopped working for people a few weeks ago. One solution is here:
http://lurk.org/groups/tidal/messages/topic/2kkap8lbgQrky3992ALIGx/
Another is to install core haskell only.
Would it be OK to move this script / repo into the tidalcycles organisation?
Answers:
username_1: Hi,
Thanks for letting me know about this.
I think that a stack based solution would be nice (it was actually my original intention to use stack, but I wanted to stay with the then recommended approach). However, this causes an issue that you have already adressed in the thread, namely that the atom plugin settings will have to be modified. This could of course be something that the script informs the user about, maybe printing the exact steps that you outline in your post.
Installing the haskell-core only in an automated fashion seems to be harder, as homebrew only seem to package the [haskell-platform](https://www.haskell.org/platform/mac.html#osx-homebrewcask) as a homebrew cask.
The solutions I can see at the moment is:
1) Let the user manually install [haskell-core](https://www.haskell.org/platform/mac.html#osx-none) and remove the `ghc` dependency from the script.
2) Use a stack based approach, modify the script to use `stack` instead of `cabal` and print instructions on how to modify the atom plugin ghci path.
If you think the stack based install will be the recommended approach in the future, I think option 2 would be best.
Unfortunately I will not have the time to do something about this at the moment, but if anyone wants to send a PR that would of course be very welcome.
And yes, I think it would be a great idea to move this repo into the org! 👍
username_1: I just confirmed that by uninstalling `haskell-platform` and installing `haskell-core` manually the script works as intended. Would like try the install procedure again with `haskell-platform` to see if/how it breaks.
username_0: Now using ghcup
Status: Issue closed
|
Nastel/tnt4j-streams | 255997719 | Title: Minor: Message "RUNTIME" in tnt4j-streams.log has UTC vs local timestamp as in other trace msgs
Question:
username_0: For consistency, use local timestamp; see below at the arrow "<==="
Extract from tnt4j-streams.log:
2017-09-07 12:17:03,489 TRACE [9:WmqTraceStream:WmqActivityTraceStreamQM_A!WmqTraceParser] - Locator 'MQGACF_ACTIVITY_TRACE.MQIACF_INVALID_DEST_COUNT' resolved value: null | TRACE |
2017-09-07 16:17:03.489000 UTC | RUNTIME=9068@Howard2-PC#SERVER=Howard2-PC#NETADDR=172.16.58.3#DATACENTER=UNKNOWN#GEOADDR=0,0 <====
2017-09-07 12:17:03,489 TRACE [9:WmqTraceStream:WmqActivityTraceStreamQM_A!ActivityInfo] - Applying field MQTrace.InvalidDestCount from: null | TRACE | 2017-09-07 16:17:03.489000 UTC | RUNTIME=9068@Howard2-PC#SERVER=Howard2-PC#NETADDR=172.16.58.3#DATACENTER=UNKNOWN#GEOADDR=0,0
2017-09-07 12:17:03,489 TRACE [9:WmqTraceStream:WmqActivityTraceStreamQM_A!ActivityInfo] - Field MQTrace.InvalidDestCount resolves to null value. Will not apply field to activity... | TRACE |
Answers:
username_1: Changed in TNT4J commit <PASSWORD>
Status: Issue closed
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.