repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
oauth2-proxy/oauth2-proxy | 1042235784 | Title: Final redirect after code redemption is not original page requested
Question:
username_0: <!--- Provide a general summary of the issue in the Title above -->
after code redemption, user is not redirected back to original page requested. The observed behavior is:
1) unauthenticated user navigates to page: https://hostname/secure/index.html
2) user is redirected to OIDC provider and authenticates
3) user is redirected back to callback URL
4) code is redeemed
5) user is redirected to: https://hostname/
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
The expected behavior is:
1) unauthenticated user navigates to page: https://hostname/secure/index.html
2) user is redirected to OIDC provider and authenticates
3) user is redirected back to callback URL
4) code is redeemed
5) user is redirected to: https://hostname/secure/index.html
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
## Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1. <!--- Step 1 --->
2. <!--- Step 2 --->
3. <!--- Step 3 --->
4. <!--- Step 4 --->
## Context
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
- Version used: v7.2.0
Answers:
username_1: Are you using oauth2-proxy as a reverse proxy, or as a "sidecar" via the nginx auth-request mechanism or similar? Can you share your configuration (excluding any sensitive data like client secrets), including on the nginx side if you're using that mode? |
otissv/react-uikit-components | 237202965 | Title: UIKit3 support
Question:
username_0: Hello,
Are you planning to add UIKit v3 support to your library?
Answers:
username_1: Hi, yes at some point. I was waiting for UIKit v3 to move out of beta first and hopefully, I would have some free time by then. But as it has been in beta for a while I don't think it will be moving soon.
Currently contemplating what to do with this project as it really needs contributors to help keep it alive especially going into react 16.
username_2: Otissv, I'm here to help if you need anything.
username_3: Hey Ottissv, I just implemented a good bit from 3.0 beta 35 into a react app I have been working on and would love to contribute. I was planning on forking this and starting to make a beta release to set the ground work for when the beta is complete.
username_1: Hi @username_3, as you can probably tell I haven't done any work on this in some time due to over commitments and that I don't use this project anymore. If you would like, I am happy to transfer the ownership to you?
Would be nice to transfer this over to someone who is actively committed to it.
username_3: @username_1 I can understand that, and would be more than happy to take it over. I will send you an email here shortly in hopes to get some of your notes!
username_3: So everyone is aware we have started a beta for Ui kit 3 which can be found [here](https://github.com/moosebot/react-uikit).
username_4: Is this still in progress?
username_3: Sorry All, I tried to hand this back to username_1 but he forked it so it wont let me hand it back. The transfer process took a few months and I moved onto other frameworks in the mean time. Willing to hand this back or transfer onto someone else who will continue to move it along. |
mickelus/tetra | 562214174 | Title: Support for Ma Enchants Mod
Question:
username_0: ## Feature Request
**Feature description**
- Support for [Ma Enchants](https://www.curseforge.com/minecraft/mc-mods/ma-enchants). It adds several new enchantments.
**How it improves the player experience**
- It would add many more tool enchantment customization options.
**Tetra synergies**
- More modded enchantments support.
Answers:
username_0: Weapons Enchantments:
- Wisdom V
- Ice Aspect II
- Lifesteal
- Faster Attack V
- Combo
Bow/Crossbow Enchantments:
- Dentonation III
- Paralysis III
- Floating III
- Quick Draw III
- True Shot
Axe Enchantments:
- Butchering V
- Lumberjack
Pickaxe Enchantments:
- Stone Mending III
- Reinforced Tip III
Tools Enchantments:
- Momentum
Weapons/Tool Enchantments:
- Curse of Butterfingers
- Curse of Breaking III
username_1: For each enchantment i need the following information to implement this:
* name
* description
* id
* max level
* applicable items
* enchantability function (typically something like this: `enchantability = base + enchantmentLevel * levelMultiplier`)
username_0: How might I find the enchantabilith function?
username_1: I found a way to generate everything so I only need the enchantment id going forward!
This is available in 2.4.0+
Status: Issue closed
username_0: Awesome, thanks! |
denoland/deno_install | 363438787 | Title: Install on macOs using homebrew?
Question:
username_0: In would be reasonable to give users the opportunity to easily install deno on macOs using brew formula.
Answers:
username_1: I would appreciate that! 👏
username_2: https://github.com/Homebrew/homebrew-core/pull/35590
username_3: Thanks for setting that up username_2!
username_3: @username_2 yeah that seems appropriate
username_2: My first binausername_3-only was closed, though we could add that independently in this repo (so that you could install via `brew install denoland/deno_install/brew.rb`).
Tusername_3ing to build with brew: https://github.com/Homebrew/homebrew-core/pull/35645
Note: there's an interesting error when testing locally - it builds fine - many linker errors and then a stack-trace test fails), will see if their CI can demonstrate the same... either way will likely report as a deno issue.
username_2: @username_3 any thoughts/pointers on these linking warnings and the source_map test failure?
https://jenkins.brew.sh/job/Homebrew%20Core%20Pull%20Requests/35808/version=high_sierra/console
username_3: @username_2 yes - it’s fixed with this PR
https://github.com/denoland/deno/pull/1439
Ill land soon.
username_2: Great! Are the linker warnings anything to worusername_3 about?
username_3: @username_2 merged. If it builds there’s nothing to worusername_3 about :)
username_3: dup https://github.com/denoland/deno/issues/1486
Status: Issue closed
|
CircleCI-Public/cimg-ruby | 1100849303 | Title: Support Ruby shell
Question:
username_0: description: >
Test ruby
steps:
- run:
shell: /usr/local/bin/ruby
command: |
puts "hello"
and the error we're receiving is something like:
/usr/local/bin/ruby: No such file or directory -- puts 'hello' (LoadError) |
apple/swift-nio | 488710520 | Title: Cannot test XCode 11 project with Swift package dependent on swift nio
Question:
username_0: ### Expected behavior
If I add a swift package dependent on swift-nio in my project using new XCode 11 swift package functionality I expect to be able to compile for testing.
### Actual behavior
When I build for testing I get the following error. Missing required module ‘CNIOAtomics’
### Steps to reproduce
1. Create new XCode project
2. Go to File -> Swift Package -> Add Package Dependency
3. Select AsyncHTTPClient
4. Add “import AsyncHTTPClient” at top of AppDelegate.swift
5. Build for testing
If I build for running it works fine.
### Version
XCode 11 beta 7
Answers:
username_1: That is a very reasonable assumption, I don’t think this is a NIO bug though.
@username_2/@neonacho anything special about building an app target for testing with Xcode Package support?
username_1: @artemredkin you seen this before?
username_2: This is a known issue that we're tracking internally with <rdar://problem/54587458>. You can work around it by explicitly passing the modulemap flag:
Add -Xcc -fmodule-map-file=$(PROJECT_TEMP_ROOT)/GeneratedModuleMaps/macosx/<missing module name>.modulemap to OTHER_SWIFT_FLAGS in the test target.
username_0: Cheers guys, I didn’t really think it was an NIO bug, but thought you guys would like to know as this impacts a lot of projects using your codebase.
@username_2 Any idea on a timeline for a fix for the issue?
username_1: @username_0 yes, thanks very much for letting us know.
username_1: I'll still close this for now and started tracking the radar Ankit mentioned
Status: Issue closed
username_3: Subscribing to this so I know when the radar is fixed. The workaround worked like a charm BTW :-)
username_2: This should be fixed in Xcode 11.2 Beta 2. Please let me know if it's still reproducing.
username_3: @username_2 any idea why Xcode 11.2 works and running `xcodebuild -scheme MyAppTests -project ./MyApp.xcodeproj -destination 'platform=iOS Simulator,name=iPhone 8' build test` not?
username_4: Still happening in Xcode 11.2
username_1: ping @username_2 ^^^
username_4: Still happening in Xcode 11.2.1 GM ~~~
username_5: Still happening in 11.2
username_6: @username_2
I am currently using Xcode 11.3.1 & tried adding
$(PROJECT_TEMP_ROOT)/GeneratedModuleMaps/iphonesimulator/RxCocoaRuntime.modulemap to OTHER_SWIFT_FLAGS in the test target.
Still it keeps giving error
username_1: Thanks @username_6 , I updated the Xcode bug report with you info.
username_2: @username_6 can you share a sample project that reproduces this issue?
username_7: I've run into this on several similar projects, with Swift Packages that contain clang submodules to expose code to the parent Swift module. I've found a workaround that fixes this, at least as of Xcode 11.3.1:
1. In your unit or UI test target, add the parent module (the Swift Package library) to the linked libraries build phase, even if that test target doesn't use it directly.
2. *Sometimes Required* If you are also using an Xcode scheme that's pointed at a unit or UI test target, make sure that the "Build" section of the scheme editor includes the application target in that list. By default, if you add a new Xcode scheme for, e.g., a UI test target, the app won't be included there and so it isn't built properly.
In all reproduced cases, the underlying error is a missing `-fmodule-map-file` flag for any file in the app target that has an `import MyLibrary` statement.
username_7: [Missing-Submodule-SPM-Error.zip](https://github.com/apple/swift-nio/files/4227554/Missing-Submodule-SPM-Error.zip)
username_7: ^^^ sample project reproducing the issue.
To reproduce, choose the UI test **scheme** and CMD+U to run the tests.
username_2: I can reproduce the issue, thanks a lot for the sample project!
username_8: Works fine for the Xcode 11.4, no need for workaround anymore
username_8: It seems to work without the workaround on XCode 11.4 @username_2
username_9: With XCode 12.0 I still see this compiler error for UI test targets as already perfectly described in an earlier comment: https://github.com/apple/swift-nio/issues/1128#issuecomment-588481616
I can also still reproduce the compiler error it with Xcode 12.0 with the example project running CMD + U with the UI test scheme from the comment: https://github.com/apple/swift-nio/issues/1128#issuecomment-588483372
Basically follow the comments from @username_7 and you can still reproduce it.
username_10: Still happening with Xcode 12.2 and Xcode 12.3 for me. Here's a test run where it appears https://github.com/TokamakUI/TokamakVapor/runs/1637655163
username_1: @username_10 I just opened your test run and from what I can tell, the failure is
```
[1639/1639] Linking TokamakVaporPackageTests
1633
objc[7127]: CLASS: class 'TokamakVaporTests.TokamakVaporTests' 0x10ab02870 small method list 0x10a932c30 is not in immutable memory
1634
Exited with signal code 6
1635
Error: Process completed with exit code 1.
```
which looks like a _different_ compiler bug, no?
FWIW, this bug (`error: Missing required module ‘CNIOAtomics’`) is supposed to be fixed in Xcode 11.3.1.
@username_10 Can you confirm that the bug you're facing is the small method list error followed by a compiler crash?
username_10: Yes, that's correct. Apologies for spamming this issue then. I think I was redirected here from somewhere else that described having "small method list" error and linked to here as something that helped in resolving that error.
username_1: @username_10 don't apologise! Do you happen to already have a Swift bug (Feedback reporter or bugs.swift.org) or Xcode bug (Feedback reporter) that I could look into?
username_10: Thanks! I've filed https://bugs.swift.org/browse/SR-14013. |
nunomaduro/phpinsights | 479862284 | Title: Colors in CI
Question:
username_0: | Q | A
| ---------------- | -----
| Bug report? | no
| Feature request? | yes
| Library version | 1.7
Hey! Thank you for your work, it's very good!
There is a [dedicated section](https://phpinsights.com/continuous-integration.html#continuous-integration) for CI which it's great but do you know how to enable colors in the Web terminal ?
I use GitLab CI, and it's all black:
<img width="639" alt="Screenshot 2019-08-12 at 23 33 41" src="https://user-images.githubusercontent.com/8252238/62900341-2f502a80-bd5a-11e9-91d2-8519f6da8d49.png">
Thank you in advance
Answers:
username_1: Could you try with `--ansi` option ?
username_0: Ok I have colors now, but it seems that Gitlab does not reset lines on `\r` so the results are still weird
cf: https://gitlab.com/username_0/todolist-backend-laravel/-/jobs/271203328
I don't think you can do anything more here
Thank you for your help!
Status: Issue closed
|
project-trident/trident-build | 392910233 | Title: chromium/chrome ld-elf.so.1: /usr/local/lib/libglib-2.0.so.0: Undefined symbol "environ"
Question:
username_0: With a freshly installed vanilla Project Trident OS
OS Version: 18.12-PRERELEASE2
Build Date: 20181219130325
put into a new VirtualBox VM and then adding only one package (chromium), trying to start chrome from the command line fails with the undefined symbol error noted in the issue title. The same symptom happens following an upgrade to PRERELEASE2 from the previous release (18.12 PRERELEASE) where the installed and working chromium is rendered inoperative.
Answers:
username_1: This took quite a bit of work, but is fixed in the upcoming 18.12-RELEASE.
Status: Issue closed
|
sangcu/em-markdown-editor | 130144257 | Title: Issue on install.
Question:
username_0: Hi, I'm having an issue installing this component.
```
version: 1.13.15
Installed packages for tooling via npm.
installing em-markdown-editor
The `ember generate <entity-name>` command requires an entity name to be specified. For more details, use `ember help`.
```<issue_closed>
Status: Issue closed |
HitFox/foxinator-generator | 90400590 | Title: Manual steps during foxinator:setup
Question:
username_0: When you run `rails g foxinator:setup`, you're prompted on screen to perform some manual steps by the devise generator:
- Ensure you have defined default url options in your environments files
- Ensure you have defined root_url
- Ensure you have flash messages in app layout
If any of these are actually necessary, can we have our setup perform them?
A lot of migrations are generated by the different steps of the `foxinator:setup` and I'm asked to run them manually. Can't we run the migrations automatically at the end of `foxinator:setup`?
Answers:
username_1: First part of your question: those are standard devise notifications, and handling them should be on a per-project basis.
Second part: I think it's possible, would need to look into it though.
Status: Issue closed
username_1: rake db:migrate and the admins:setup task now happen automatically. |
ashfurrow/TIL | 546502572 | Title: React/Native SVGs are odd
Question:
username_0: You _could_ theoretically use the `<svg>` HTML tag in React, but some of the HTML tag attributes have dashes instead of CamelCase, so React has its own `SVG` component:
https://github.com/artsy/emission/pull/2025/files#diff-3f34973345e9c7d8d5f47e5904357300R1-R13 |
actix/sockjs | 322811735 | Title: `()` doesn't implement `std::fmt::Display`
Question:
username_0: ```
error[E0277]: `()` doesn't implement `std::fmt::Display`
--> /Users/username_0/.cargo/registry/src/mirrors.ustc.edu.cn-61ef6e0cd06fb9b8/sockjs-0.3.0/src/transports/xhrsend.rs:89:49
|
89 | Err(error::ErrorNotFound(())),
| ^^^^^^^^^^^^^^^^^^^^ `()` cannot be formatted with the default formatter; try using `:?` instead if you are using a format string
|
= help: the trait `std::fmt::Display` is not implemented for `()`
= note: required by `actix_web::error::ErrorNotFound`
error[E0277]: `()` doesn't implement `std::fmt::Display`
--> /Users/username_0/.cargo/registry/src/mirrors.ustc.edu.cn-61ef6e0cd06fb9b8/sockjs-0.3.0/src/transports/jsonp.rs:271:49
|
271 | Err(error::ErrorNotFound(())),
| ^^^^^^^^^^^^^^^^^^^^ `()` cannot be formatted with the default formatter; try using `:?` instead if you are using a format string
|
= help: the trait `std::fmt::Display` is not implemented for `()`
= note: required by `actix_web::error::ErrorNotFound`
error: aborting due to 2 previous errors
For more information about this error, try `rustc --explain E0277`.
error: Could not compile `sockjs`.
``` |
networknt/light-kafka | 983834960 | Title: update sidecar producer to use config serializer
Question:
username_0: instead of the hard-coded byte array serializer.
Status: Issue closed
Answers:
username_0:  [update sidecar producer to use config serializer light-kafka master](https://trello.com/c/ihQpFUW0/1188-update-sidecar-producer-to-use-config-serializer-light-kafka-master) |
FauxFaux/zrs | 562036467 | Title: Instructions unclear
Question:
username_0: I installed the tool and have `zrs` in my path. I also ran `zrs --add-to-profile` and it added `. '/Users/nikivi/Library/Application Support/zrs/z.sh'` to my ~/.zshrc
Instructions make it sound that I can then run `z bar` to change to dirs but `z` is not part of my path.
What am I missing?
Answers:
username_1: `z` is a function, which is declared in that sourced file.
`zsh`'s built-in `which` is smart enough to know:
```
faux@astoria:~% which z
z: aliased to _z 2>&1
faux@astoria:~% which _z
_z () {
local output ret
output="$(zrs "$@")"
...
```
`z bar` will do nothing if it can't find a directory (but will set an exit code).
`z` with no arguments will show the status of the database?
username_0: ```
❯ which z
z not found
```
username_1: That looks reasonable to me. I have no idea what's going on there.
Try starting `zsh` with `-x`?
```
% zsh -x
...
+/home/faux/.zshrc:193> . /home/faux/.local/share/zrs/z.sh
+/home/faux/.local/share/zrs/z.sh:28> [ -d /home/faux/.z ']'
+/home/faux/.local/share/zrs/z.sh:61> alias 'z=_z 2>&1'
+/home/faux/.local/share/zrs/z.sh:63> [ '' ']'
+/home/faux/.local/share/zrs/z.sh:63> _Z_RESOLVE_SYMLINKS=-P
+/home/faux/.local/share/zrs/z.sh:65> type compctl
+/home/faux/.local/share/zrs/z.sh:67> [ '' ']'
+/home/faux/.local/share/zrs/z.sh:69> [ '' ']'
+/home/faux/.local/share/zrs/z.sh:78> [[ -n '' ]]
+/home/faux/.local/share/zrs/z.sh:79> precmd_functions[$(($#precmd_functions+1))]=_z_precmd
+/home/faux/.local/share/zrs/z.sh:88> compctl -U -K _z_zsh_tab_completion _z
...
```
Here I see it create the alias and the completion?
username_0: It sources too much stuff :(
I'll try to find the cause and dig into it more. |
agdsn/pycroft | 244392713 | Title: Heuristic Schema check
Question:
username_0: It would be convenient to fail starting pycroft when the given database does not have the correct schema yet.
The current use case would be that the importer has not been run yet, so a heuristic checking for the existance of all the tables would be sufficient. Check out e.g. [this SO thread](https://stackoverflow.com/questions/30428639/check-database-schema-matches-sqlalchemy-models-on-application-startup#30653553)
Answers:
username_1: What about versioning the schema and having a schema_version table with a single row containing the version?
Status: Issue closed
username_0: That sounds better and is not too difficult. Closing, since there's no benefit of putting the work into such a hack right now. |
thingsboard/thingsboard | 330088852 | Title: Installing ThingsBoard using Docker(Linux or Mac OS) Error
Question:
username_0: Get https://registry-1.docker.io/v2/thingsboard/application/manifests/2.0.2: unauthorized: incorrect username or password
Answers:
username_1: After installing docker and docker-compose:
- You just need to **clone TB repository**.
- **cd thingsboard/docker/**
- **nano tb.env** and change database to **cassandra** if you don't want to use **sql**
- **sudo docker-compose up**
Status: Issue closed
|
cpan-testers/cpantesters-deploy | 200828445 | Title: Build automated Perl upgrade workflow
Question:
username_0: When new versions of Perl are released, we need to be able to easily upgrade. This means that we need a task to install the latest version of Perl, but also that we need to ensure that each individual project can be easily deployed into the new Perl without breaking.
Ideally, this project would have a task to create a new Perl installation in the perlbrew, and each individual project would deploy to the "latest" Perl when it was next deployed.
Some possible problems to ensure are addressed are:
* Compiled modules must be recompiled for the new Perl. This is generally handled by the individual Perl project's `deploy` task, and handling it there will be fine.
* The new Perl and associated modules must not interfere with the running production environment while they are being prepared. This likely means that an individual project must keep track of what Perl it was deployed with. This goes along with the next bullet point.
* Ideally, an individual project would use the old Perl until it was `deploy`ed, at which time it would notice that there is a new Perl and try to deploy that. This probably means that each individual project must be deployed so that their `#!` line is changed to the correct Perl path so they use the right Perl that they were installed to.
* It should be possible to specify which Perl a project should deploy to in case of incompatibilities or hotfixes or otherwise. This should be done as an option to the `deploy` task.
Answers:
username_1: Couldn't we use Docker for each, I mean, one image for each application your their own perl?
username_0: We could, but we don't have enough hardware to handle the overhead of the VMs (however slight Docker makes that), we don't always get a choice of what hardware we get and what it can do, and we don't have the staff to maintain the added complexity. Docker is overkill for the problem we're trying to solve (and that I'm not too concerned about solving due to how rare the situation is). |
mahmoud/glom | 1043617229 | Title: SKIP all PathAccessErrors Recursively
Question:
username_0: How to skip all path access errors ?
```
input= {
"firstname": "satish",
"lastname": "reddy" #may or may not be present
"details" : {
"phoneno": "987654321",
"address": "xxxxx", # may or may not be present
"pincode": "xxxxx" # may or may not be present
},
"familydetails": [
{
"name": "PersonA",
"address": "Adress-A",
"phoneno": "999999999"
},
{
"name": "PersonB",
"phoneno": "999999999"
},
{
"name": "PersonC"
}
]
}
```
```
output_needed = {
"Captain": {
"FirstName": "satish",
"LastName": "reddy", # include only if it is present in input
"PersonalDetails": {
"MobileNo": "987654321",
"Address": "xxxxx", # include only if it is present in input
"Pincode": "xxxxx" # include only if it is present in input
} ,
"CaptainFamilyDetails": [
{
"Name": "PersonA",
"Address": "Adress-A", # include only if it is present in input
"MobileNo": "999999999" # include only if it is present in input
},
{
"Name": "PersonB",
"MobileNo": "999999999"
},
{
"Name": "PersonC"
}
]
}
}
```
```
spec = {
"Captain": {
"FirstName": "firstname",
[Truncated]
},
"CaptainFamilyDetails" : ("familydetails", [
{
"Name": "name",
"MobileNo": "phoneno",
"Address": "address"
}
])
}
}
```
This `spec` works only when all required paramaters are provided, but if some are missing getting a `PathAccessError`.
I tried `Coalesce` with optional fields, but it will not be viable for me, i have 25-30 more parameters to map, we exactly have no knowledge on which paramters are optional.
So How to skip all Path access error recursively ?
Answers:
username_0: @username_1 can you help me with this?
username_1: {'Captain': Or({'CaptainFamilyDetails': Or(('familydetails', [Or({'Address': Or('address', Val(Sentinel('SKIP'))), 'MobileNo': Or('phoneno', Val(Sentinel('SKIP'))), 'Name': Or('name', Val(Sentinel('SKIP')))}, Val(Sentinel('SKIP')))]), Val(Sentinel('SKIP'))), 'FirstName': Or('firstname', Val(Sentinel('SKIP'))), 'LastName': Or('lastname', Val(Sentinel('SKIP'))), 'PersonalDetails': Or({'Address': Or('details.address', Val(Sentinel('SKIP'))), 'MobileNo': Or('details.phoneno', Val(Sentinel('SKIP'))), 'Pincode': Or('details.pincode', Val(Sentinel('SKIP')))}, Val(Sentinel('SKIP')))}, Val(Sentinel('SKIP')))}
```
that way you can keep your main spec clean :-)
username_1: {'Captain': {'PersonalDetails': {}}}
```
if you only want to have certain fields be skipped if missing, then probably better to explicitly mark them with `or_skip()` rather that this meta-spec approach
username_0: Thankyou @username_1 , I will try meta_spec approach , if it works I will close the issue.
username_0: @username_1 can you explain `Auto(([Ref('spec')], tuple))` this please ?
username_1: Sure! Maybe I can get a new snippet out of this :-)
To unpack:
`Auto` -- this says switch mode(https://glom.readthedocs.io/en/latest/modes.html) back to default from `Match`
`[...]` -- we know `T` is a tuple here, so we want to iterate over each element of the tuple
`Ref(spec)` -- recurse downwards into the tuple
`( ..., tuple)` -- `[]` by default will return a list, convert it back to a tuple so that tuple in = tuple out
The reason I added this is so that `("familydetails", [` will work properly -- the recursion can "pass through" the tuple and get to the dict inside that tuple.
username_0: tq so much!!!
Status: Issue closed
|
jeffbass/imagezmq | 694544922 | Title: Looking for Advice
Question:
username_0: Jeff. Love your project. I am working on a project that uses multiple RPI 's with camera's could be webcams as well. The hub program starts and waits for the currently three sender RPI's to send an image. Each image is saved with a sequence number. The hub then sends each sender a timestamp for the next timelapse to be taken. This will be identical for each sender. The hub then does an image stitch to make a pano image and saves with sequence number then increments for next cycle. This is all working well although I had to make custom camera mounts out of foam board, small block of wood and a short dowel.
The problem I have is that the hub Must start before the senders. I want to automate the system so I was thinking of setting up a watch program on each sender that listens on another port. When the hub is started or restarted it sends each sender watch program a new python configuration file for settings eg resolution on a different port. Watch saves File as a config.py. Watch program then starts/restarts the sender program in background via subprocess Popen. sender then restarts and reads the new config.py as an import. Still working on this.
Would it be possible to send a text file and return a confirming text message between the hub and sender rather than a jpeg and text per imagezmq. Currently I was looking at just using basic zmq commands to the watch program on the senders but would be nicer if it was possible with imagezmq. Could not find this feature in imagezmq code but thought I would drop you a line.
Note I had to modify the https://github.com/ppwwyyxx/OpenPano c++ code to get it to accept an output file name for the pano jpg/png since it would only generate a fixed out.jpg named file in the same folder. My version is here https://github.com/username_0/OpenPano. Created curl bash scripts for easier install. I could have used Adrians opencv image stitching but prefer the self contained openpano approach since users would not need the latest/greatest opencv contrib version.
FYI I have attached my camera holder. When the pipano project is ready I will post on my GitHub repo. Still a work in progress. There are lots of issues to work out. For one the cropping of pano's is not consistent so doing a timelapse video would need some image stabilization during video editing. If lighting is stable then most of the pan's crop pretty consistently but low light can throw things off easily. I had to build the stands to allow accurate pointing. Don't want to use a pan/tilt because images would not be synchronized with same timestamp.
[cam-stand.pdf](https://github.com/username_1/imagezmq/files/5180357/cam-stand.pdf
Excuse me if I got a little chatty
Claude ...
Answers:
username_1: Hi Claude @username_0,
Glad you love my project. Thanks. Your project sounds really interesting. I'll definitely keep an eye on it on GitHub as you make progress.
Regarding your text sending question, I do use `imageZMQ` to send and receive text messages in a number of my projects, using the `ImageSender.send_jpg `method. The `jpg_buffer` can be any bytestring; it doesn't actually have to be a jpg. To use `imageZMQ` to send text messages only, I just set the `jpg_buffer` to a short bytestring in the sender. Then I receive and ignore that short bytestring in the hub that receives the text message. For example:
```python
# Sender that sends text messages
# in the text sending program, where sender is an imageZMQ.ImageSender instance
msg_text = 'Some message text.'
hub_reply = sender.send_jpg(msg_text, b'0') # set jpg_buffer parameter to a single byte
# Receiver that receives the text messages
# in the text receiving program, where hub is an imageZMQ.ImageHub instance
msg_text, buf = self.recv_jpg() # buf contains b'0', but just don't bother to use it
# or, another way to not use the jpg_buffer that is received.
msg_text, _ = hub.recv_jpg() # another way to receive and not use the jpg buffer
```
This code works well for me in a number of my projects. I do sometimes have to use the `text.encode` and `byte.decode` methods to avoid errors about text versus bytestrings in my message text:
```python
# convert bytes to Python 3 string:
mst_text = msg_text_in_bytes.decode('utf-8') # decode from bytes to Python 3 string
# or, converting from Python 3 string to bytes:
msg_text_in_bytes = msg_text.encode() # convert Python 3 string to bytes to make ZMQ happy
```
Your method of using config.py text files and a watch program to control startup and options of your cameras versus the hub sounds similar to what I am doing with my `imagenode` and `imagehub` programs. I have a threaded function that watches for `imagehub` restarts (or other glitches) and it then restarts the `imagenodes` accordingly. To change the `imagenode` settings (there are many of those!), I copy a new `imagenode.yaml` text file to the imagenode and restart it. I use systemctl for starting and restarting my imagenodes using an `imagenode.service` file. You can see that in my `imagenode` repository [here.](https://github.com/username_1/imagenode/blob/master/imagenode.service) My approach to the fact that hubs have to start before senders is to have all my `imagenodes` running continuously as systemctl services with the (`Restart=always, RestartSec=5`) options. Then, whenever the imagenode detects an imagehub restart, it simply exits. Then systemctl waits 5 seconds and restarts it.
Thanks for sending your RPi camera stand template. I'm going to give it a try. No need to apologize for being chatty. I love to hear about the projects that benefit from `imageZMQ`. I learn a lot from what others are doing. It is why I find GitHub so useful.
Jeff
username_0: Thanks for the quick reply. I will implement per you sample code. I was hoping I could use text instead of jpg data and now I know. I like the systemd service. I have used it on several other projects and the systemd script file can be pretty simple, just a [Unit], [Service], [Install] and a related bash script if req'd. Attached a small image of my 3 cam pano board. These are on old RPI 3's with no built in wifi but they work fine. Can be moved around. Thing1 is a sender and hub but might move hub to one of my RPI4's with mounted HD. Stitching would be faster. Also can use wireless ad-hoc network setup if in a remote location away from home wifi.

username_1: I love your DIY hardware build. But I haven't seen those RPi cases before. It looks like the case holds both the RPi and the PiCamera in the same case? Where did you get that?
username_0: I bought mine a few years ago but found similar RPI3 case on amazon at a good price.
https://www.amazon.com/Keyestudio-Supporting-Camera-Installation-Raspberry/dp/B076PQVMN2/ref=sr_1_2?dchild=1&keywords=raspberry+pi+case+with+camera+mount&qid=1599445678&sr=8-2
I live in Canada but winter in Texas. Will be delayed this year due to covid issues in US. Probably go down possibly spring or summer next year. Hope you are doing well.
RPI 4 Case with built in camera mount
https://www.amazon.com/LABISTS-Raspberry-Heatsink-Heatsinks-Supply/dp/B085GBCLYR/ref=sr_1_11?dchild=1&keywords=raspberry+pi+case+with+camera+mount&qid=1599445294&sr=8-11
https://www.amazon.com/Keyestudio-Supporting-Camera-Installation-Raspberry/dp/B076PQVMN2/ref=sr_1_2?dchild=1&keywords=raspberry+pi+case+with+camera+mount&qid=1599445294&sr=8-2
username_1: Thanks!
username_0: Jeff.
Hope you are doing well. I have the initial issue of my timelapse panorama project.
https://github.com/username_0/panopi
Got panosend.py auto start and stop by using a panowatch.py program (runs in background using panowatch.sh). uses zmq port 5556 for communications. Using a RPI4 for panohub stitching (much faster than my older machines) Still a few items to implement. I plan to read the panosend.yaml settings from panohub.yaml (eliminates panosend.yaml). Also will auto change the hub IP address in the panosend yaml settings that are sent to panosend RPI's when the panohub is started. Will use logging lib rather than print statements. So far it seems to be working OK. Needs zeroconf since I am getting IP addresses from host names sockets.
This helps if DHCP changes IP addresses. I think RPI's use zeroconf out of the box as far as I can tell.
Made install easy using curl scripts. Documentation needs work and I plan to make a YouTube video on the project.
Stay Safe
Regards Claude ...
username_0: Just an update. I now read the panosend yaml settings from the panohub.yaml file in panohub folder. The panosend_settings are read and the ZMQ_PANOHUB_IP is dynamically updated. This avoids having to manually edit panohub.yaml setting and also auto adjusts if DHCP makes changes or you change the hub to another machine. I did this recently and all I had to do was install panohub onto new machine, start panohub.py and everything just worked since panowatch was running on all the panosend RPI's. New RPI4 hub can stitch three 720p images in a few seconds.
Will relocate setup for more interesting scene and do a pano video from the pano images. I am going to do a pan tilt camera image stitching setup. I have a remote pan tilt camera setup in Texas running pi-timolo that sends images up to my google drive but it would be a lot nicer to just send one pano image instead of multiple pan timelapse images. I will also add feature to my robot since it can rotate as well as pan camera. It would be possible to get 360 + view or at least two 180 + pano's if robot rotation is a problem with stitching.
There are lots of enhancements possible but may just try to keep it simpler and do multiple simpler projects. My PI-TIMOLO and SPEED CAMERA projects grew in complexity and feature creep but lots of people like these projects.
Thanks for your assistance
Stay Safe
Claude .....
username_1: Hi Claude,
Thanks for these updates. It sounds like you are making great progress. I am following your work with interest because I plan do panoramic multi-camera wildlife tracking to my own projects in the future.
Thanks for sending links to your videos, too. I've been enjoying watching them, especially the ones about your PI-TIMOLO, MOTION_TRACK and SPEED_CAMERA projects.
Looking forward to continued updates!
Thanks,
Jeff
username_0: Thanks Jeff
I appreciate your help and support. Today I brought the PANOPI camera
setup upstairs and put it outside on the window sill. I used a powered hub
and 1 ft usb cables for a clean setup that is easy to move.
The new version reads all settings from the panohub.yaml. This is my first
project using yaml files. Had to put my thinking cap on to figure out how
to dynamically edit the *ZMQ_PANOHUB_IP* variable from the
panosend_settings and transmit the stream to panowatch.py. This means you
can easily change hub machines and not have to worry about editing the zmq
ip address.
I am thinking of adding a motion tracking feature on one of the senders
that would trigger the pano image sequence. Also added image resolution
rounding so picamera does not generate warning. Also updated my PI-TIMOLO
to use image resolution rounding and implemented change to reduce cpu usage
when only timelapse mode is used per a request by a user.
I had fun with this project. I want to do a howto YouTube video and also
generate a PANO Video from the full size pano images. Looks like there is
consistent cropping if there is good lighting and stitch points.
I want to try different resolutions. 1080p stitches well due to higher
resolution so overlap can be narrower. Tomorrow I am going to try a 1920 x
720 resolution. A combination of 1080p width and 720p height or may try
the max camera width and 720p..
Had a slight issue with two of the panosend camera's. The images were not
very good and I had to clean the camera lenses. Used a blunted tooth pick
(cut with scissors) with lens cleaning cloth over it. Might try a bit of
lense cleaner solution next time. Canned air did not work but
physical cleaning worked ok. Will try another cleaning tomorrow. Will do
some research to see if anyone has a different way. The camera pin
hole lense opening is pretty small and I would not want to damage a
perfectly good camera module but my cleaning method improved things a lot
as confirmed by my wife.
I think my speed camera needs cleaning after looking at the images. Seems
a little foggy.
Stay Safe and Healthy
Regards
Claude
. |
VirtwhoQE/virtwho-ci | 814257012 | Title: tc_1109 failed due to KeyError
Question:
username_0: 
Need to analyze if something wrong in
` if guest_uuid in data[owner].keys() and host_uuid in data[owner][guest_uuid]['guest_hypervisor']: `
Answers:
username_0: It is env issue, need to checked the `Propagate to children` option when we set the read-only account:

username_0: Checked, and the test result was fine, will close this issue.
Status: Issue closed
|
CircuitSetup/Split-Single-Phase-Energy-Meter | 445608658 | Title: Caps on schematic?
Question:
username_0: Looking at the datasheet for DVDD on atm90e32. Notes DVD should be decoupled w/ 10uF and .1 uF. Where is the 10uF?
Also, VDD18 notes only need a 10uF cap. why .1uF tied here? I'm thinking DVDD and VDD18 swapped?
Answers:
username_1: You are correct! That is something I initially left out but will be adding for the final version. It's actually not in the Application note examples, and I'm not sure why.
Status: Issue closed
username_1: This is fixed in the latest version. See here: https://github.com/username_1/Split-Single-Phase-Energy-Meter/blob/master/Hardware/energy_meter_pcb_v1.4.png |
google/automl | 865881603 | Title: Training failes: RuntimeError: Key _CHECKPOINTABLE_OBJECT_GRAPH not found in checkpoint
Question:
username_0: Hi I'm trying to retrain the efficientdet-d1 on a custom dataset.
However, when I run the main.py as follows I get an error.
The training command I use is:
python3 main.py --mode=train_and_eval --train_file_pattern=train/* --val_file_pattern=test/* --model_name=efficientdet-d1 --model_dir efficientdet-d1/ --ckpt=efficientdet-d1 --train_batch_size=2 --eval_batch_size=2 --num_epochs=10 --num_examples_per_epoch=5000 --eval_samples=100 --hparams=configs/default.yaml
I run the script on Ubuntu 18.04
python 3.6.9
with
tensorflow 2.5.0-dev20201224'
The error:
Traceback (most recent call last):
File "/home/acaris/.local/lib/python3.6/site-packages/tensorflow/python/client/session.py", line 1375, in _do_call
return fn(*args)
File "/home/acaris/.local/lib/python3.6/site-packages/tensorflow/python/client/session.py", line 1360, in _run_fn
target_list, run_metadata)
File "/home/acaris/.local/lib/python3.6/site-packages/tensorflow/python/client/session.py", line 1453, in _call_tf_sessionrun
run_metadata)
tensorflow.python.framework.errors_impl.NotFoundError: 2 root error(s) found.
(0) Not found: Key box_net/box-0-bn-3/beta/Momentum not found in checkpoint
[[{{node save/RestoreV2}}]]
(1) Not found: Key box_net/box-0-bn-3/beta/Momentum not found in checkpoint
[[{{node save/RestoreV2}}]]
[[save/RestoreV2/_4891]]
0 successful operations.
0 derived errors ignored.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/acaris/.local/lib/python3.6/site-packages/tensorflow/python/training/saver.py", line 1303, in restore
{self.saver_def.filename_tensor_name: save_path})
File "/home/acaris/.local/lib/python3.6/site-packages/tensorflow/python/client/session.py", line 968, in run
run_metadata_ptr)
File "/home/acaris/.local/lib/python3.6/site-packages/tensorflow/python/client/session.py", line 1191, in _run
feed_dict_tensor, options, run_metadata)
File "/home/acaris/.local/lib/python3.6/site-packages/tensorflow/python/client/session.py", line 1369, in _do_run
run_metadata)
File "/home/acaris/.local/lib/python3.6/site-packages/tensorflow/python/client/session.py", line 1394, in _do_call
raise type(e)(node_def, op, message)
tensorflow.python.framework.errors_impl.NotFoundError: 2 root error(s) found.
(0) Not found: Key box_net/box-0-bn-3/beta/Momentum not found in checkpoint
[[node save/RestoreV2 (defined at /home/acaris/.local/lib/python3.6/site-packages/tensorflow_estimator/python/estimator/estimator.py:1510) ]]
(1) Not found: Key box_net/box-0-bn-3/beta/Momentum not found in checkpoint
[[node save/RestoreV2 (defined at /home/acaris/.local/lib/python3.6/site-packages/tensorflow_estimator/python/estimator/estimator.py:1510) ]]
[[save/RestoreV2/_4891]]
0 successful operations.
0 derived errors ignored.
...
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/acaris/.local/lib/python3.6/site-packages/tensorflow/python/training/py_checkpoint_reader.py", line 70, in get_tensor
self, compat.as_bytes(tensor_str))
RuntimeError: Key _CHECKPOINTABLE_OBJECT_GRAPH not found in checkpoint
During handling of the above exception, another exception occurred:
[Truncated]
...
tensorflow.python.framework.errors_impl.NotFoundError: Restoring from checkpoint failed. This is most likely due to a Variable name or other graph key that is missing from the checkpoint. Please ensure that you have not altered the graph expected based on the checkpoint. Original error:
2 root error(s) found.
(0) Not found: Key box_net/box-0-bn-3/beta/Momentum not found in checkpoint
[[node save/RestoreV2 (defined at /home/acaris/.local/lib/python3.6/site-packages/tensorflow_estimator/python/estimator/estimator.py:1510) ]]
(1) Not found: Key box_net/box-0-bn-3/beta/Momentum not found in checkpoint
[[node save/RestoreV2 (defined at /home/acaris/.local/lib/python3.6/site-packages/tensorflow_estimator/python/estimator/estimator.py:1510) ]]
[[save/RestoreV2/_4891]]
0 successful operations.
0 derived errors ignored.
Original stack trace for 'save/RestoreV2':
File "main.py", line 402, in <module>
Looks like a mismatch of some sort. Any idea how to fix this?
Answers:
username_1: Hi, based on your command line, it is problematic to set "--model_dir=efficientdet-d1 --ckpt=efficientdet-d1". Because efficientdet-d1 folder is not empty, it will skip --ckpt and tried to load all variables from efficientdet-d1, causing this issue.
Please use a different value for model_dir such as efficientdet-d1-finetune. Let me know if it still doesn't work.
Status: Issue closed
username_2: what about this case too?
python -model_inspect.py --runmode=saved_model --model_name=efficientdet-d3 --ckpt_path=efficientdet-lite3 --saved_model_dir=lite3_savedmodel
I am getting :L
2 root error(s) found.
(0) Not found: Key efficientnet-b3/blocks_0/conv2d/kernel not found in checkpoint
[[node save/RestoreV2 (defined at /home/chaahm/automl_newest/efficientdet/inference.py:217) ]]
[[save/RestoreV2/_301]]
(1) Not found: Key efficientnet-b3/blocks_0/conv2d/kernel not found in checkpoint
[[node save/RestoreV2 (defined at /home/chaahm/automl_newest/efficientdet/inference.py:217) ]] |
dotnet/AspNetCore.Docs | 785797645 | Title: DbContextOptionsBuilder' does not contain a definition for 'UseInMemoryDatabase'
Question:
username_0: In the section "Register the database context"
Updating the class "Startup.c"
I am geting the error with the line " opt.UseInMemoryDatabase("TodoList")); " it can't dind the method UseInMemoryDatabase
The error message is
"Error CS1061 'DbContextOptionsBuilder' does not contain a definition for 'UseInMemoryDatabase' and no accessible extension method 'UseInMemoryDatabase' accepting a first argument of type 'DbContextOptionsBuilder' could be found (are you missing a using directive or an assembly reference?)"
What I am I missing here ?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 583ae3e7-5915-f22b-e54c-6c100cf2a7da
* Version Independent ID: 95f2549e-29f9-e31a-b4d0-432b971643ff
* Content: [Tutorial: Create a web API with ASP.NET Core](https://docs.microsoft.com/en-us/aspnet/core/tutorials/first-web-api?view=aspnetcore-5.0&tabs=visual-studio)
* Content Source: [aspnetcore/tutorials/first-web-api.md](https://github.com/dotnet/AspNetCore.Docs/blob/master/aspnetcore/tutorials/first-web-api.md)
* Product: **aspnet-core**
* Technology: **aspnetcore-tutorials**
* GitHub Login: @username_2
* Microsoft Alias: **riande**<issue_closed>
Status: Issue closed |
StefanScherer/windows-docker-machine | 1097073693 | Title: Apple Silicon Support
Question:
username_0: Hi,
Both Parallel and VMWare now support M1 chips. Is it possible for `windows-docker-machine` to work on M1 machines?
Status: Issue closed
Answers:
username_1: I don't think this will work. Parallels demonstrated that they can run Windows ARM64, but not x86_64. VMware the same, they don't run x86_64 ISO files on M1. |
red/red | 267407056 | Title: UNSETting some of the OP!s wreaks havoc on console
Question:
username_0: ```
Answers:
username_1: Same with `unset '<` or `unset '>` except `[]` does not crash the console.
username_0: Just remembered, `unset '=` behavior looks identical to `unset 'system` ;)
username_2: The core words used by the console code are not protected for now, that will come with the introduction of `protect/unprotect` functions in 0.8.0. I don't see what we can do in the meantime, besides avoiding redefining those core words.
Status: Issue closed
|
CosmosOS/Cosmos | 430126792 | Title: Issues with StreamReader / Writer
Question:
username_0: I have a performance test program to test how big files can be handled in COSMOS.
https://pastebin.com/PdcUYT54
This performance test consists of:
Creating a file of 144 x 128 pairs of coordinates 0,0 to 143,127 and writing them in a file line by line.
Reading said coordinates into an array of point objects and then displaying the X and Y values of the points on the screen.
Thing is, the whole test doesn't progress past the reading part.
Why?
It works a bit like this:
The program stalls while writing to the disk , and when I break it and look into the vmdk on VMWare I get the following result:

What I tried was to write the file on my own on windows and then just read the contents of it and display it on the screen.
The problem is that after a certain number of lines, I get garbage on the screen consisting of random characters and the <?> character.
Answers:
username_1: I think @username_2 was working on some fixes for the file system caused by the writing or reading of larger files. The issue might be caused by the bugs.
username_1: Can you check if https://github.com/CosmosOS/Cosmos/commit/b9307acfdfa7f92b46093612fac761752763588b changed anything?
username_0: @username_1 I was about to. Will try it out.
username_0: @username_1 the problem is still there with the latest commit, albeit in a different form. (no garbage chars, just empty ones) video and test file supplied.
https://files.gitter.im/CosmosOS/Cosmos/G3pY/bandicam-2019-05-12-13-12-11-250.avi
https://files.gitter.im/CosmosOS/Cosmos/lcuC/lcet10.txt
username_0: the StreamReader reads approx. 4245 characters before displaying the empty lines.
username_1: Is it possible for you to run the code as a test, and enable cosmos debug in a few files so we have some debug log to work off?
username_0: Since I don't know how to use the testrunner... Here's the code:
```
StreamReader myreader = new StreamReader(@"0:\lcet10.txt");
int lncount = 0;
int chrlength = 0;
while (!myreader.EndOfStream)
{
string tmp = myreader.ReadLine();
chrlength += tmp.ToCharArray().Length;
Console.WriteLine(tmp);
lncount++;
Console.WriteLine("Chr:" + chrlength);
if(lncount==20)
{
Console.WriteLine("Length: " + chrlength);
Console.ReadKey();
lncount = 0;
}
}
myreader.Close();
```
username_2: I'm looking in to this.
username_2: I'm not seeing an issue in the latest master. Did I miss something?

username_0: It still doesn't work with the latest master... try reading it to the end?
username_2: Are you using your own disk image or the one provided with the UK?
username_0: The one provided, just added the file there with VMWare.
username_0: Could this be an issue with memory overflow?
username_2: Can you send me your disk image? I'll see if I can duplicate that way.
username_0: https://we.tl/t-UIdoCwMqLX
Here it is.
Status: Issue closed
username_0: It works now! |
ChurchCRM/CRM | 493573764 | Title: Improper 3rd normal form
Question:
username_0: the person classifications field needs to be split, conceptually. As per your default, it contains two different classification schemes, both of which are necessary. However a person might be a member and an attender OR a member and a non-attender.
We need to be able to separate members from attenders. I don't know which would be easier, to add multiple records of classification in the person file or to separate the fields into attendance and membership.
Collected Value Title | Data
----------------------|----------------
Page Name |/churchcrm/Menu.php
Screen Size |540x960
Window Size |704x1280
Page Size |1749x1280
Platform Information | Linux host2030.hostmonster.com 4.14.135-214.ELK.el7.x86_64 #1 SMP Wed Jul 31 01:57:22 CDT 2019 x86_64
PHP Version | 7.3.9
SQL Version | 5.7.23-23
ChurchCRM Version |3.2.3
Reporting Browser |Mozilla/5.0 (Macintosh; Intel Mac OS X 10.14; rv:69.0) Gecko/20100101 Firefox/69.0
Prerequisite Status |Missing Prerequisites: ["Mcrypt"]
Integrity check status |{"status":"success"}
Answers:
username_1: For these sort of cases, we usually recommend implementing a "Custom Field" to populate as required to meet your specific needs. Please have a look at this wiki page and evaluate: https://github.com/ChurchCRM/CRM/wiki/Custom-Fields
Please let us know how you get on.
username_1: Closing due to lack of activity. Feel free to update us on progress if you think this still needs to stay open.
Status: Issue closed
|
hyb1996-guest/AutoJsIssueReport | 291911855 | Title: [163]java.lang.RuntimeException: ImageReaderContext is not initialized
Question:
username_0: Description:
---
java.lang.RuntimeException: ImageReaderContext is not initialized
at android.media.ImageReader.nativeImageSetup(Native Method)
at android.media.ImageReader.acquireNextSurfaceImage(ImageReader.java:327)
at android.media.ImageReader.acquireNextImageNoThrowISE(ImageReader.java:308)
at android.media.ImageReader.acquireLatestImage(ImageReader.java:283)
at com.stardust.autojs.runtime.api.image.ScreenCapturer$2.onImageAvailable(ScreenCapturer.java:93)
at android.media.ImageReader$ListenerHandler.handleMessage(ImageReader.java:648)
at android.os.Handler.dispatchMessage(Handler.java:111)
at android.os.Looper.loop(Looper.java:207)
at com.stardust.autojs.runtime.api.Loopers$2.run(Loopers.java:57)
at java.lang.Thread.run(Thread.java:818)
at com.stardust.lang.ThreadCompat.run(ThreadCompat.java:61)
Device info:
---
<table>
<tr><td>App version</td><td>2.0.16 Beta2</td></tr>
<tr><td>App version code</td><td>163</td></tr>
<tr><td>Android build version</td><td>1495809106</td></tr>
<tr><td>Android release version</td><td>6.0</td></tr>
<tr><td>Android SDK version</td><td>23</td></tr>
<tr><td>Android build ID</td><td>Flyme 6.1.0.0Y</td></tr>
<tr><td>Device brand</td><td>Meizu</td></tr>
<tr><td>Device manufacturer</td><td>Meizu</td></tr>
<tr><td>Device name</td><td>M1E</td></tr>
<tr><td>Device model</td><td>M1 E</td></tr>
<tr><td>Device product name</td><td>meizu_M1E</td></tr>
<tr><td>Device hardware name</td><td>mt6755</td></tr>
<tr><td>ABIs</td><td>[arm64-v8a, armeabi-v7a, armeabi]</td></tr>
<tr><td>ABIs (32bit)</td><td>[armeabi-v7a, armeabi]</td></tr>
<tr><td>ABIs (64bit)</td><td>[arm64-v8a]</td></tr>
</table> |
TNThieding/exif | 625010134 | Title: Removing all know exif values
Question:
username_0: Hey, great package 👍. I'm using it to remove all exif values and save a new file. I'm checking each attribute to be removed, and thought it would be usefull to have a method in your _image.py to erase all most common exif data, so anyone could just call that and get it clear for cases like mine.
Do you want me to do a Pull request with those changes?
Thank you!
Answers:
username_1: That sounds like a good idea! I'd gladly welcome a pull request. Active development and CI/CD is through GitLab (at https://gitlab.com/username_1/exif/), so if you could submit the changes there it would be great. If you'd prefer to make the pull request on GitHub, you can do that too and I'll copy it into GitLab myself.
username_0: Sorry I do not have a GitLab account, it is much quicker to do it here for me.
Pull request is sent
username_1: Thank you, @username_0. I'll continue our discussion on the pull request and I can move it to GitLab afterwards.
username_1: Released it in v0.11.0 of the package.
Status: Issue closed
|
llhuii/dive-into-k8s | 913226760 | Title: kubectl apply VS replace
Question:
username_0: 1. apply 使用patch 方法
2. replace 使用update方法
3. replace `--force` 使用了先delete再create方法
patch有以下策略https://kubernetes.io/zh/docs/tasks/manage-kubernetes-objects/update-api-object-kubectl-patch/:
1. json: json patch
```json
[
{
"op" : "replace" ,
"path" : "/users/0/email" ,
"value" : "<EMAIL>"
},
]
```
2. merge: json merge patch
```json
{
"a":"z",
"c": {
"f": null
}
}
```
3. strategic: strategic merge patch
```json
```
前两者参考https://erosb.github.io/post/json-patch-vs-merge-patch/
// 待更新
Answers:
username_0: 参考链接https://blog.atomist.com/kubernetes-apply-replace-patch/ |
pmattd/big-data-generator | 600307828 | Title: Generator for creating timestamps
Question:
username_0: current timestamp and random timestamp between upper and lower limits in order ( work out how many files will be created and distribute the timestamps over the files
Answers:
username_0: did it without keeping order, just random timestamps
Status: Issue closed
|
Evrey/SHC_AIV | 294613485 | Title: Nizar lacks a brain when starting with low goods
Question:
username_0: So there I just put your Nizar against a Richard Lionheart on a balanced 1 vs 1 testing map with all necessary resources available, and 2k starting gold for both. Now, I was excited to see who would win (Richard's catapult/trebuchet spam vs helpless Nizar, or Nizar's assassins vs Richard's helpless always-closed gate), but to my surprising, the game was decided already after 3 game years, and ended after exactly 12 game years, with Nizar being defeated. I noticed the reason, but still tried again and restarted the game, and watched as the events repeated exactly the same way as in the first game. Below is the end statistics of the first game.

Alone by looking at the statistics, it becomes obvious that Nizar didn't quite get up and running at the start, producing 0 food throughout the entirety of the game.
So enough blabbering, what was happening there? Nizar spent all of his starting wood and gold on building wheat farms, bakeries, breweries, mercenary post, engineers guild, houses, trading post, granary, and even woodcutter huts and an iron mine - but he didn't build a mill, essentially stalling all of his bread production.
A couple years into the game, his start food was gone and his popularity started dropping as he was out of gold, still had no mill, and had not gotten to build hops farms and inns either, and positive fear factor alone wasn't enough to keep up his popularity. 5 years into the game, Nizar then was down to 4 workers (3 wheat farmers, 1 woodcutter) and still had no mill.

Shortly after, in total 6 years into the game, Richard's first attack smashed Nizar's front wall and some of his buildings, and with Nizar still down at 0 popularity, 4 workers, 0 mills it didn't take too long afterwards until Richard hat flattened all of Nizar's buildings in and around his castle, and killed his Lord with some pitch-ditch-induced delay.
All in all, this greatly reminded me of vanilla Wazir's poor performance with low starting goods, as can be seen in detail on that AI battle video: https://youtu.be/Tu4BQgsD_kE Wazir had the same issue there, didn't place his mill, ran out of food, popularity drop, game over.
His initial starting wood was spent on other buildings (e.g. farms, houses, trading post, possibly even an iron mine) before it was the first mill's turn on the AIV's build order, and when the mill finally would have been built, Wazir (or now, Nizar) is out of wood already and the mill will be skipped for the time being.
Later, even with a bit of gold left to trade, or woodcutters bringing in more wood, the AI will usually reach lower thresholds first, e.g. 10 wood for a bakery, and build that first, being back to square one... and of course, initial soldier recruiting will make short work of any remaining start gold reserves or early income from a single iron mine or selling a bit of surplus wheat.
When food runs out before all those 10-wood buildings and other stuff has been built, popularity will drop, and the AI has no gold left to buy food (or, for Wazir, doesn't buy food in general) or set negative taxes to try and counter this...
This behaviour can be less obvious when starting with many goods and gold, as more start wood increases chances for a successful mill placement, and lots of gold gives the AI more time/chances to buy wood (and/or food) and eventually hopefully get going with their mill, or at least get going enough other industry to then have a steady income large enough to fund food and wood for a mill later. Also, ironically, on maps that lack resources so that the AI just cannot build particularly much external industry right away, this issue is less present.
At this point in time you might be full of rage about my wall of text, and about the incompetence of Crusader's AI behaviour, yet I would argue this should certainly be fixable by slightly modifying the castle's build order. Key is setting the first mill very early in the build order, so that it has good chances to be built before that iron mine and other less relevant external industry, plus leaving a number of steps AFTER the mill during which no bakeries or other wooden buildings get placed in the AIV, but rather walls, a first tower, a moat or whatever else, to "buy some time" for the mill to be built even when the starting wood wasn't sufficient for building it right away and the AI checks through the next steps if anything else can be built with less than 20 wood.
Answers:
username_1: Yupp, I know about this pathetic bug. It affects **every** AI lord making bread. Your solution for the problem is really neat. I think I'd make the first mill of every bread baker the very first building in the queue, even before the stock pile extension, as wood always comes first. Perhaps no re-ordering of other buildings will be needed this way.
Lookin' at my current gaming and hobby schedule, it is very unlikely that I'll fix this within the next months. Not impossible, but very unlikely. So, if you are faster, then you're welcome to help, otherwise be patient. =D
Nice bug report as always, btw.!
username_0: Yeah I know you're working hard to make AEI great :) Yet I am not sure I can find the time to deal with the issue in a timely manner either... having the idea of changing the build order is one thing, adjusting it on all 8 castles and then testing if it works and if necessary finetuning the build order to make sure it works fine is another thing... anyway maybe I can fit it in somewhen...
Well, I suppose it may not take that much longer than writing long bug reports after all xD With an ingame speed of 125 multiplied by a factor of "theoretical 16", AI battles run quite fast compared to regular 90 x 1 speed.
And I see you finally saw the December post for Wolf that I made and then forgot about it... :D Good.
Status: Issue closed
username_1: [This commit](https://github.com/username_1/SHC_AIV/commit/0861c910da920754472919a924b2a6068883536e) should fix it. I did what I threatened to do: I made the windmills the very first building in the queue, even before the stockpile addons. |
mjordan/islandora_workbench | 816757543 | Title: Taxonomy terms with namespaces causing errors
Question:
username_0: I know I'm missing something here, but I think I'm following the documentation correctly, specifically the Taxonomy Reference Fields section of https://username_1.github.io/islandora_workbench_docs/fields/#content-type-specific-fields. I'm happy to contribute to documentation if that turns out to be worthwhile.
When trying to ingest a record with a taxonomy term that uses a namespace (in my csv, it's "person:<NAME>"), I'm getting a couple of warning during ``--check``, and then an actual error when running Workbench. Here are the ``--check`` warnings:
WARNING - CSV field "field_subjects_name" in row 3 contains a term ("<NAME>") that is not in the referenced vocabulary ("person"). That term will be created.
WARNING - Vocabulary "family" referenced in CSV field "field_subjects_name" may not be enabled in the "Terms in vocabulary" View (please confirm it is) or may contains no terms.
Then the info and warning that come from a real job:
INFO - {'vid': [{'target_id': 'person', 'target_type': 'taxonomy_vocabulary'}], 'status': [{'value': True}], 'name': [{'value': '<NAME>'}], 'description': [{'value': '', 'format': None}], 'weight': [{'value': 0}], 'parent': [{'target_id': None}], 'default_langcode': [{'value': True}], 'path': [{'alias': None, 'pid': None, 'langcode': 'en'}]}
WARNING - Term '<NAME>' not created, HTTP response code was 422.
...and the error:
Traceback (most recent call last):
File "./workbench", line 963, in <module>
create()
File "./workbench", line 165, in create
if config['subdelimiter'] in row[custom_field]:
TypeError: argument of type 'bool' is not iterable
I'm not sure why Workbench ``--check`` comments on the "family" vocabulary, because the csv doesn't reference it. The only thing I can think of is that ``field_subjects_name`` is not the right field to be using, but it seems right according to the documentation.
Here's the csv (sorry, the relevant term is in the last field): https://gist.github.com/username_0/ae8b52c02b2bfd3467e3c15df7087b19
And then the config is:
task: create
host: "http://localhost:8000"
username: admin
password: <PASSWORD>
input_dir: testBatch_02_clean
input_csv: barneberg_forTesting_002_as-imageObjects.csv
allow_adding_terms: true
Answers:
username_1: That misleading logging is a known issue (#194 to be specific). Drupal's Views REST response returns the same result for vocabularies that are empty as it does for ones that aren't enabled in the View, hence the vague wording of the warning. I need to filter out vocabularies that are referenced by the field being checked but are not namespaced for new terms. That would eliminate most occurances of that warning message while leaving it present when it's important that the user checks that vocabulary. (I think, I logged the issue but haven't gotten to resolving it yet.)
The 422 errors I can't explain. Another user has reported them when trying to create terms before, but I have not been able to replicate or track down what is going on. Which baffles me. Coincidentally (?) they saw them with the Person vocabulary as well. I have not seen that Python `bool` error before though. Can you share your config and CSV with me so I can try to replicate what you are seeing?
username_0: thanks @username_1, let me know if I can provide any more info.
username_1: Will do. I'm going to dig into this over the weekend.
username_1: @username_0 can you export the configuration of your Person vocabulary for me and paste it into a comment here? To to that, go to `/admin/config/development/configuration/single/export` and choose "Taxonomy vocabulary" in the "Configuration type" list, and then "Person (person)" in the "Configuration name" list.
username_0: Here's the Person vocabulary configuration:
uuid: 516861d0-dcbe-40ce-b1b3-cbca5b89695e
langcode: en
status: true
dependencies: { }
_core:
default_config_hash: GAHBiBARA1hdVQoJq1P1oyK47vGxwTr3vyiPsnwqD44
name: Person
vid: person
description: 'An individual of the human species. '
weight: 0
username_1: @username_0 haven't forgotten about this, just swamped with day job and am brain dead in evenings. Will dig in over the weekend.
username_1: To try to replicate this problem, I created 1000 nodes with new Person terms as subjects (933 new terms total) without a failure. My CSV file was pretty simple however, with only an id, title, and field_subject field.
I then spun up a new Islandora 8 box and tried to ingest the same CSV and all of the new terms failed to be created.
Can you check `/admin/reports/status` to see what version of Drupal you're running? The version where the terms are failing for me is 8.8.5. The version where I can't get the terms to fail is 8.9.12.
username_1: Following up on my comment ^ from last night, I just installed the current version of the Islandora Playbook, installed the Islandora Workbench Integration module, and used Workbench to ingest (without any HTTP 422 taxonomy terms) the same set of nodes with new `person:` values in `field_subject`. That Islandora is running Drupal 9.2.0-dev.
I'm really hoping the failures are specific to the earlier version of Drupal, since otherwise I can't seem to replicate the problem. The only time I have seen it was when I used Workbench against a Drupal 8.5.5.
Would you be interested in trying to ingest your example CSV (or a subset of it since the Playbook doesn't have all your fields) into my new instance to see what happens? It's on the public web; I could provide details to you in Slack.
username_0: I'm running 8.8.5. I'm definitely open to ingesting a modified CSV into your new instance, and will keep an eye on slack for the details. Thanks!
username_1: OK to close this?
username_0: Yes! :)
Status: Issue closed
|
ant-design/ant-design | 161016419 | Title: Button禁用后Tooltip不会正常消失。
Question:
username_0: 
截图如上,其他地方Tooltip的表现是离开该组件之后会自动消失,但是如果是在禁用后的Button上的话则不会消失了。测试代码如下:
`<Tooltip title="禁用后Tooltip">
<Button disabled>禁用按钮</Button>
</Tooltip>`
实际上一开始是用的`<Tooltip title="编辑"><Button type="ghost" shape="circle" icon="edit"></Tooltip>`这种方式,后来测试的时候写成上面那种格式。
Answers:
username_1: #1816
Status: Issue closed
|
flutter/flutter-intellij | 537405222 | Title: "Wrap with Center" and "Wrap with Padding" are not active in Flutter Outline Tool panel
Question:
username_0: ## Steps to Reproduce
I try wrap Row widget with Center (or Padding) widget by button in Flutter Outline Tool panel, but this button is not active.

## Version info
C:\src\flutter\bin\flutter.bat doctor --verbose
[√] Flutter (Channel stable, v1.9.1+hotfix.6, on Microsoft Windows [Version 10.0.18363.476], locale en-US)
• Flutter version 1.9.1+hotfix.6 at C:\src\flutter
• Framework revision 68587a0916 (3 months ago), 2019-09-13 19:46:58 -0700
• Engine revision b863200c37
• Dart version 2.5.0
[√] Android toolchain - develop for Android devices (Android SDK version 29.0.2)
• Android SDK at C:\Users\xxx\AppData\Local\Android\sdk
• Android NDK location not configured (optional; useful for native profiling support)
• Platform android-29, build-tools 29.0.2
• Java binary at: C:\Program Files\Android\Android Studio\jre\bin\java
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b03)
• All Android licenses accepted.
[√] Android Studio (version 3.5)
• Android Studio at C:\Program Files\Android\Android Studio
• Flutter plugin version 42.1.1
• Dart plugin version 191.8593
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b03)
[√] Connected device (1 available)
• Google Pixel 2 • 192.168.226.102:5555 • android-x86 • Android 9 (API 28)
• No issues found!
Process finished with exit code 0
Answers:
username_1: You need to define
`factory MyWidget.forDesignTime() `
inside of your DicePage widget.
see [Live Widget Preview](https://medium.com/flutter/flutter-outline-hot-reload-and-the-implementation-of-a-live-widget-preview-69abd39aa3bb)
Status: Issue closed
|
dotnet/docs | 1149219539 | Title: Not really an issue - Spanish Version
Question:
username_0: Hi!
There is some type of collaboration to translate this pages?
Im Learning and i like to help to translate this pages.
Im from Argentina and one of the biggest problems of latam to acces this type of content is the lenguage
---
#### Detalles del documento
⚠ *No edite esta sección. Se requiere para la vinculación de problemas de docs.microsoft.com ➟ GitHub.*
* ID: 24c8353f-69e7-9068-9b18-df0198a1b122
* Version Independent ID: 823031fe-c8a4-69cd-2b1a-9faacd282aab
* Content: [Classes and objects - C# Fundamentals tutorial](https://docs.microsoft.com/es-mx/dotnet/csharp/fundamentals/tutorials/classes)
* Content Source: [docs/csharp/fundamentals/tutorials/classes.md](https://github.com/dotnet/docs/blob/main/docs/csharp/fundamentals/tutorials/classes.md)
* Product: **dotnet-csharp**
* Technology: **csharp-fundamentals**
* GitHub Login: @username_1
* Microsoft Alias: **wiwagn**
Answers:
username_1: Hi @username_0
I'm adding @username_2 and @juricgit to this thread. They work on localization and community contributors, respectively. They can help point you to how to get involved.
username_2: @username_0 Thank you for your offer. We are looking into this missing translation.
[FEEDBACK 563566](https://ceapex.visualstudio.com/CEINTL/_workitems/edit/563566)
username_2: The page is localized. |
ScoopInstaller/Main | 1040161413 | Title: lynx: SSL error: Can't find common name in certificate
Question:
username_0: ## Bug Report
**Package Name:** lynx
### Current Behaviour
Websites using HTTPS do not open at all.
### Expected Behaviour
All websites should open.
### Additional context/output
The dependency `cacert` might be the cause of the problem.
```
❯ lynx
Looking up lynx.invisible-island.net
Making HTTPS connection to lynx.invisible-island.net
lynx: Can't access startfile https://lynx.invisible-island.net/
```
### Possible Solution
<!--- Only if you have suggestions on a fix for the bug -->
### System details
**Windows version:** 10.0.19043.1288
**OS arch (32 or 64 bit):** 64
**PowerShell version:** 7.1.5
**Additional software:** none
Answers:
username_0: It has been like that for quite some time - https://github.com/ScoopInstaller/Main/pull/512#pullrequestreview-308197679 - I'm not sure it will work in stable either.
The actual problem might be this - mentioned on the Lynx homepage

The manifest currently downloads OpenSSL 1.1.1, but that is not supported it seems. And OpenSSL 1.1.0 is hard to find anywhere.
username_1: You're right. Just tested the stable release with OpenSSL 1.1.1l and it doesn't work. Looks like there are security vulnerabilities with version 1.1.0 so that's why it's hard to find - I don't recommend we go looking for it and introduce vulnerabilities for scoop users! So, it looks like it's up to lynx to add support for 1.1.1 which was released 11 September 2018. Or better yet the latest version 3.0.0 which was recently released (7 September 2021).
username_0: Maybe we can ask the author what's going on - would you be willing to file a report in Lynx? (Ref: https://lists.gnu.org/archive/html/lynx-dev)
username_1: It looks like someone asked a similar question just over a year ago: https://lists.gnu.org/archive/html/lynx-dev/2020-09/msg00005.html
And this was the author's reply: https://lists.gnu.org/archive/html/lynx-dev/2020-09/msg00006.html
Essentially, it looks like https support for Windows is not planned/being worked on at the moment. A workaround is to install it via WSL or Cygwin instead. ¯\\\_(ツ)_/¯
username_0: I see. However our issue is slightly different - that of difference in Windows OpenSSL versions. But that might very well be due to the author not being aware/having time to look at Windows builds. 😥
username_1: I can't even get the regular installer to work, it fails to install with the error that it 'Failed to copy msvcr120.dll'. That file is from the Visual C++ Redistributable Package for Visual Studio 2013 but even with both the 32bit and 64bit versions installed it still fails. 😢
username_0: I remember the installer worked when I tried it around a month ago, but yeah now all of them fail with the same error 😕
username_1: Sad times... I would be grateful if you could make a bug report about the installer on that mailing list, I'm not so comfortable posting my real name and email there!
username_0: You don't have to use your real name/email 😉. It's just for communication, they're not going to verify you. 😏
username_1: Of course, but I couldn't see the option to leave my username - the button to reply to a thread opens your mail client to respond by email! |
ualbertalib/DMP_roadmap | 325847107 | Title: Development environment setup
Question:
username_0: Based on the current docker and docker-compose work, create proper development/UAT environment for development and user acceptance test.
Requirements:
- bring it inline with Jupiter and Discovery project setup, use simplified/rails default stack as much as possible.
- Dev env have local drive mounted.
- Allow tests with ORCID sandbox, and Canadian Access Federation Sandbox (need shibboleth, and shibboleth-discovery service) in the UAT environment.
Answers:
username_0: @piyapongch
Current DMP Roadmap docker is at:
https://github.com/ualbertalib/DMP_roadmap/blob/master/Dockerfile
What needs to be achieved is:
- [ ] A docker-compose file that can be used for a full stack dev environment, and/or deploy to the UAT environment:
https://github.com/ualbertalib/jupiter/blob/master/docker-compose.yml
https://github.com/ualbertalib/jupiter/blob/master/docker-compose.deployment.yml
(Following examples from Jupiter, but might not have to have two separate versions as in Jupiter, or regular dev stack and a deployment version - Jupiter's stack is more complicated) Also, should use mysql instead of postgresql.
- [ ] Documentation on how this can be used. Also need documentation on how dev env can be set up - can follow Jupiter's documentation:
https://github.com/ualbertalib/jupiter/blob/master/README.md#docker
username_0: Could you submit your PR for review @piyapongch |
openshift/origin | 297331954 | Title: Test flakes with error No API token found for service account "default", retry after the token is automatically created
Question:
username_0: See: https://storage.googleapis.com/origin-ci-test/pr-logs/pull/18587/test_pull_request_origin_extended_conformance_gce/16274/build-log.txt
The above mentioned PR is hitting it consistently. There are 4 of the same type of failures in the attached log and all the failures are within a couple of seconds of each other.
It looks like the test creates a service account and then tries to create a pod before the API token for the service account is available. It looks like its waiting 10 seconds on that.
Relevant snippet from the logs:
```
Feb 14 21:17:26.058: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
Feb 14 21:17:30.411: INFO: About to run a Kube e2e test, ensuring namespace is privileged
Feb 14 21:17:30.595: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled.
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [k8s.io] Sysctls
/tmp/openshift/build-rpms/rpm/BUILD/origin-3.9.0/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/sysctl.go:56
[It] should support unsafe sysctls which are actually whitelisted [Suite:openshift/conformance/parallel] [Suite:k8s]
/tmp/openshift/build-rpms/rpm/BUILD/origin-3.9.0/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/sysctl.go:101
STEP: Creating a pod with the kernel.shm_rmid_forced sysctl
Feb 14 21:17:40.101: INFO: Unexpected error occurred: No API token found for service account "default", retry after the token is automatically created and added to the service account
```
Answers:
username_1: Probably introduced by #18508
/assign @username_2
username_1: Created a bug since this may break end user apps https://bugzilla.redhat.com/show_bug.cgi?id=1546466
username_2: fixed
Status: Issue closed
username_3: fixed in https://github.com/openshift/origin/pull/18661 |
nunomaduro/collision | 579912053 | Title: Improve output formatting
Question:
username_0: As discussed on Twitter, I think it would be useful to display diff results returned by the `php artisan test` command more in line with those from `phpunit`, i.e. improved formatting with line breaks and indentation.
A particular example is the `assertDatabaseHas` helper which produces the following sample output when it fails to locate a record:
<img width="709" alt="2020-03-12-phpunit-assert-database-has" src="https://user-images.githubusercontent.com/17433578/76522180-29c3db80-6434-11ea-976f-3416de9ec0d4.png">
In contrast, `php artisan test` shows:
<img width="1309" alt="2020-03-12-collision-assert-database-has" src="https://user-images.githubusercontent.com/17433578/76522223-3cd6ab80-6434-11ea-91f8-3ce537523059.png">
Answers:
username_1: @username_0 Can you try to work in a pull request to fix this?
username_0: @username_1 Apologies Nuno, I didn't notice your reply.
I'm afraid I can't commit to working on it - I simply don't have the bandwidth right now but if my schedule clears up at some point, I'll circle back to it and try to help.
Status: Issue closed
username_1: Fell free to contribute with a solution for this. |
NucleusPowered/Nucleus | 231998628 | Title: Request: Attribute/Enchantment Hiding | Direct addition of attributes
Question:
username_0: One addition the core system of Nucleus could come fit with would be the addition of two more commands, one to deal with hiding attributes of an ItemStack, and one to also apply such attributes. As for possible command aliases, I'm drawing a blank, but let's blame that on the current time of when this ticket is being posted :P |
jdf2e/nutui | 1008086824 | Title: [bug]nut-popup与picker-view一起使用样式错乱了
Question:
username_0: https://github.com/username_0/demo-taro-picker.git
<!-- https://github.com/YOUR_REPOSITORY_URL -->
## 预期的行为和实际行为
## 复现步骤,具体代码
<!-- 请提供复现步骤,错误日志以及相关配置 -->
<!-- 可以尝试不要锁版本,重新安装依赖试试先 -->
## 相关环境信息
- **操作系统**:win
- **Node 版本**:12.20.0
Answers:
username_1: picker 正在重构中,预替换小程序原生 picker
Status: Issue closed
username_1: https://nutui.jd.com/#/picker-taro |
rancher/rancher | 153498620 | Title: Duplicated gluster mount records?
Question:
username_0: Ive created convoy+glusterfs configuration and deployed ghost blog on it. Scalled it by two containers.
as you can see on screenshot there is normal record for volume called ghost-blog with two mounts (as expected) and two additional volumes.
I think this records duplicating ghost-blog mounts and should not be displayed in ui. Am I right?

---
| Useful | Info |
| :-- | :-- |
|Versions|Rancher `v1.1.0-dev1` Cattle: `v0.160.3` UI: `v1.0.7` |
|Access|`github` `admin`|
|Route|`storagepools.index`|
Answers:
username_0: It was missconfiguration. Wrong mount point in docker-compose.yml leads to this behaviour
Status: Issue closed
|
tsmaster/nanogenmo2019book | 526066357 | Title: Seanbaby's Gastrobot battle pals
Question:
username_0: Monster: Ninja with no legs
Treasure: rowboat
Treasure: Justice barrel
treasure: metal arm holding a sword
monster: billy nitro
monster: unknown traveller
treasure: wand of heat ray
treasure: wand of iceballs
treasure: wand of killer bees
monster: frugo
monster: villain on pony
treasure: flag<issue_closed>
Status: Issue closed |
marclipovsky/vscode-string-manipulation | 883697228 | Title: I hope to add an option to switch between the specified two types
Question:
username_0: I hope to add an option to switch between the specified two types, so that I can quickly toggle through the shortcut keys
Answers:
username_1: indeed I'd like to quickly toggle between all the types actually
username_0: Sorry for not being clear. If I have ten words, the original word is in camelize format, and I want to convert it to capitalize format. Now I have to call up the command window and select capitalize format when operating. Can you provide a shortcut key to quickly use the last converted format?
username_1: I have made a PR #33 which adds a context menu and the repeat last action shortcut.
username_1: <img width="1448" alt="CleanShot 2022-03-13 at 17 15 06@2x" src="https://user-images.githubusercontent.com/2998111/158052965-83d826dd-e059-4121-84fc-ca1f2fa2d537.png">
You can define a shortcut with the repeat last action command |
lavary/laravel-menu | 559208674 | Title: Menu Dropdown
Question:
username_0: Hello, i have code like this:
```
public function handle($request, Closure $next)
{
(new Menu)->make('SideBar', function ($menu) {
$menu->add(__('DashboardPage.Menu1'), 'home')->prepend('<i class="fas fa-tachometer-alt"></i>');
});
return $next($request);
}
```
and it render like bellow:
```
<li class="active"><a href="http://myapps/home"><i class="fas fa-tachometer-alt"></i>Dashboard</a></li>
```
How i can render it like bellow:
```
<li><a href="#exampledropdownDropdown" aria-expanded="true" data-toggle="collapse" class=""> <i class="icon-windows"></i>{{ __('Dashboard.Page.Menu1') }}</a>
<ul id="exampledropdownDropdown" class="list-unstyled collapse show" style="">
<li><a href="#">{{ __('Dashboard.Page.Menu2') }}</a></li>
<li><a href="#">{{ __('Dashboard.Page.Menu3') }}</a></li>
<li><a href="#">{{ __('Dashboard.Page.Menu4') }}</a></li>
</ul>
</li>
```
Thank you
Answers:
username_1: Howdy,
There are many examples. Please take a look: https://github.com/lavary/laravel-menu#url-prefixing
Thanks,
-Dustin
username_2: Howdy,
There are many examples. Please take a look: https://github.com/lavary/laravel-menu#url-prefixing
Thanks,
-Dustin
username_0: ok, thank you
Status: Issue closed
|
typescript-eslint/typescript-eslint | 823843466 | Title: Is this a false negative with [require-await]?
Question:
username_0: Just a question for now but could turn into bug report depending on the answer:
Is `require-await` supposed to trigger an error if a function returns a promise but contains no `await` keyword? I would have thought "yes", but I'm not seeing eslint errors under this circumstance (and I *am* seeing errors if the function doesn't return a promise, so I don't think this is a configuration issue).
If this is not an error, what is the criteria for determining when async is unnecessary? Note that my code works fine whether I include the `async` or not, so to me, it's unnecessary to include it.
Status: Issue closed
Answers:
username_1: No - that's the entire point of the version of the rule that we offer.
It's mentioned in the readme.
https://github.com/typescript-eslint/typescript-eslint/blob/master/packages/eslint-plugin/docs/rules/require-await.md
If you want the rule to error in this circumstance - then you can use the base rule. |
France-ioi/codecast | 301151973 | Title: audio lost just after recording
Question:
username_0: I just did a 15min Codecast, at the end I clicked pause and then stop.
It seems that the audio was lost.

https://codecast.france-ioi.org/v5/player?base=https%3A%2F%2Ffioi-recordings.s3.amazonaws.com%2Ftcom%2F1519845933519
Answers:
username_0: seems that this is reproductible:
start a recording
during the recording click pause then just after stop : the audio is lost! |
flutter/flutter | 1174492143 | Title: Fatal app crash immediately after yield from the stream inside of bloc
Question:
username_0: <!-- Thank you for using Flutter!
If you are looking for support, please check out our documentation
or consider asking a question on Stack Overflow:
* https://flutter.dev/
* https://api.flutter.dev/
* https://stackoverflow.com/questions/tagged/flutter?sort=frequent
If you have found a bug or if our documentation doesn't have an answer
to what you're looking for, then fill out the template below. Please read
our guide to filing a bug first: https://flutter.dev/docs/resources/bug-reports
-->
## Description
App crashes after small data yield just before a http request from a stream inside of bloc with _**Fatal signal 11 (SIGSEGV), code 2 (SEGV_ACCERR)**_ error on Android and **_EXC_BAD_ACCESS (SIGBUS)_** on iOS. Crash doesn't occur when using streams outside of the bloc.
What would be the reason for such a fatal error directly from a stream?
## Steps to Reproduce
<!-- Please include full steps to reproduce so that we can reproduce the problem. -->
1. Set Flutter version to the latest stable one - **2.10.3**
2. Create a new project and set dependencies to the following:
```
flutter_bloc: ^7.3.3
http: ^0.13.4
```
4. Execute `flutter run` on the code sample <!-- (see "Code sample" section below) -->
5. Press the red button which should crash the app on both Android and iOS platforms
**Expected results:** App doesn't crash fatally and data yields normally.
**Actual results:** App crashes with fatal error immediately after yield in _fetchData_ method
<details>
<summary>Code sample</summary>
``` import 'dart:convert';
import 'package:flutter/material.dart';
import 'package:flutter_bloc/flutter_bloc.dart';
import 'package:http/http.dart';
void main() {
runApp(const MyApp());
}
class MyApp extends StatelessWidget {
const MyApp({Key? key}) : super(key: key);
@override
Widget build(BuildContext context) {
return MaterialApp(
title: 'Flutter Demo',
theme: ThemeData(
primarySwatch: Colors.blue,
),
home: const MyHomePage(title: 'Stream crash reproduce'),
);
[Truncated]
[✓] HTTP Host Availability
• All required HTTP hosts are available
• No issues found!
```
</details>
## Screenshots
Bloc code causing the crash:

Error after app crash on Android:

Answers:
username_1: Hey just a note ..Have you tried using the new emit functions from the latest bloc package.Try and confirm how it performs.
username_2: @username_0
You mentioned `bloc` plugin which is third party. Does the crash occur without using `bloc` and using framework's core implementation of Stream such as `StreamBuilder` ?
username_0: Hey, I already tested this with flutter_bloc **_on_** method where states get emitted and it worked. I posted this because it’s pretty worrying for such a crash to occur in Flutter on both platforms.
This is the first time I’ve experienced a fatal crash with pure dart code without any native functions. Is this a bug in newer versions of the Flutter framework? I’ve tested it on versions 2.8 and 2.10, on both the app crashed.
username_0: This problem occurs with StreamBuilder also, you can see it in the snippet below. I really think this is a serious issue and should not happen either way.
```dart
import 'package:flutter/material.dart';
import 'package:http/http.dart';
void main() {
runApp(const MyApp());
}
class MyApp extends StatelessWidget {
const MyApp({Key? key}) : super(key: key);
@override
Widget build(BuildContext context) {
return MaterialApp(
title: 'Flutter Demo',
theme: ThemeData(
primarySwatch: Colors.blue,
),
home: const MyHomePage(title: 'Stream crash reproduce'),
);
}
}
class StreamTest {
static Stream<String> fetchData() async* {
yield '{"name":"crash"}';
// Yielding empty json object instead of the above does not crash the app
// yield '{}';
// Uncommenting 1ms delay code prevents the crash
// await Future.delayed(const Duration(milliseconds: 1));
// Create a random API call ignoring its response for this test
get(Uri.parse("https://dog-facts-api.herokuapp.com/api/v1/resources/dogs/all"));
}
static Stream<dynamic> getStream() async* {
await for(String response in fetchData()) {
yield response;
}
}
}
class MyHomePage extends StatefulWidget {
const MyHomePage({Key? key, required this.title}) : super(key: key);
final String title;
@override
State<MyHomePage> createState() => _MyHomePageState();
}
class _MyHomePageState extends State<MyHomePage> {
@override
Widget build(BuildContext context) {
return Scaffold(
appBar: AppBar(
title: Text(widget.title),
),
body: StreamBuilder<dynamic>(
stream: StreamTest.getStream(),
builder: (context, snapshot) {
if(snapshot.data == null) {
return const CircularProgressIndicator();
}
return Center(
child: Text(
'${snapshot.data}',
),
);
}
),
);
}
}
```
username_2: @username_0
Can you check this document and see if the implementation is correct for `yield` in streams ?
Also, since yield is coming from dart, I think you should open this issue in dedicated repo: https://github.com/dart-lang/sdk/issues
username_0: Implementation is correct, I'm using it as intended. This is a minimal example and it shouldn't crash like this, it really gives Flutter a sense of instability. You didn't test and reproduce this issue? I thought Flutter team handles issues like this and communicates with Dart team if it affects your framework.
username_3: could you take a look @username_4, given that you are submerged in this area of the code anyway?
username_4: I was able to reproduce the crash in Flutter/debug mode (JIT) but not Flutter/profile mode (AOT).
After rebuilding the Flutter engine with assertions enabled in the Dart VM, I got the following assertion failure:
```
[ +283 ms] E/Dart (10621): ../../third_party/dart/runtime/vm/stack_trace.cc: 150: error: expected: future.GetClassId() ==
future_impl_class.id()
[ ] E/DartVM (10621): version=2.17.0-232.0.dev (dev) (Wed Mar 23 10:45:46 2022 -0700) on "android_arm"
[ ] E/DartVM (10621): pid=10621, thread=10639, isolate_group=main(0xb9085a98), isolate=main(0xb9089228)
[ ] E/DartVM (10621): isolate_instructions=a0b2ad80, vm_instructions=a0b2ad80
[ ] E/DartVM (10621): pc 0xa17adfe9 fp 0x9f37aa70 /data/app/com.example.bug-1/lib/arm/libflutter.so+0x2431fe9
[ ] E/DartVM (10621): pc 0xa15fdd4b fp 0x9f37aa80 /data/app/com.example.bug-1/lib/arm/libflutter.so+0x2281d4b
[ ] E/DartVM (10621): pc 0xa181c617 fp 0x9f37aaa8 /data/app/com.example.bug-1/lib/arm/libflutter.so+0x24a0617
[ ] E/DartVM (10621): pc 0xa181de33 fp 0x9f37ab28 /data/app/com.example.bug-1/lib/arm/libflutter.so+0x24a1e33
[ ] E/DartVM (10621): pc 0xa181e24f fp 0x9f37ac38 /data/app/com.example.bug-1/lib/arm/libflutter.so+0x24a224f
[ ] E/DartVM (10621): pc 0xa16134ff fp 0x9f37ac78 /data/app/com.example.bug-1/lib/arm/libflutter.so+0x22974ff
[ ] E/DartVM (10621): pc 0xa16e975f fp 0x9f37acb8 /data/app/com.example.bug-1/lib/arm/libflutter.so+0x236d75f
[ ] E/DartVM (10621): -- End of DumpStackTrace
[ ] E/DartVM (10621): pc 0x00000000 fp 0x9f37acd8 sp 0x00000000 [Stub] CallBootstrapNative
[ ] E/DartVM (10621): pc 0x94912d18 fp 0x9f37acf0 sp 0x9f37ace0 [Unoptimized] StackTrace.get:current
[ ] E/DartVM (10621): pc 0x9422a8c8 fp 0x9f37ad14 sp 0x9f37acf8 [Unoptimized]
_HttpClient@14463476._getConnection@14463476
[ ] E/DartVM (10621): pc 0x94221114 fp 0x9f37ad68 sp 0x9f37ad1c [Unoptimized] _HttpClient@14463476._openUrl@14463476
[ ] E/DartVM (10621): pc 0x9421fc94 fp 0x9f37ad8c sp 0x9f37ad70 [Unoptimized] [email protected]
[ ] E/DartVM (10621): pc 0x9421ca8c fp 0x9f37adcc sp 0x9f37ad94 [Unoptimized] IOClient.send
[ ] E/DartVM (10621): pc 0x9421c484 fp 0x9f37adf4 sp 0x9f37add4 [Unoptimized] IOClient.send
[ ] E/DartVM (10621): pc 0x94213834 fp 0x9f37ae28 sp 0x9f37adfc [Unoptimized] BaseClient._sendUnstreamed@575340606
[ ] E/DartVM (10621): pc 0x94212bd0 fp 0x9f37ae84 sp 0x9f37ae30 [Unoptimized] BaseClient._sendUnstreamed@575340606
[ ] E/DartVM (10621): pc 0x94212254 fp 0x9f37aec0 sp 0x9f37ae8c [Unoptimized] BaseClient.get
[ ] E/DartVM (10621): pc 0x94211f24 fp 0x9f37aee0 sp 0x9f37aec8 [Unoptimized] get.<anonymous closure>
[ ] E/DartVM (10621): pc 0x9420e65c fp 0x9f37af18 sp 0x9f37aee8 [Unoptimized] _withClient@20447473
[ ] E/DartVM (10621): pc 0x9420e1cc fp 0x9f37af40 sp 0x9f37af20 [Unoptimized] _withClient@20447473
[ ] E/DartVM (10621): pc 0x9420db04 fp 0x9f37af74 sp 0x9f37af48 [Unoptimized] get
[ ] E/DartVM (10621): pc 0x944f313c fp 0x9f37afa0 sp 0x9f37af7c [Unoptimized] StreamTest.fetchData
[ ] E/DartVM (10621): pc 0x99f8add4 fp 0x9f37afd8 sp 0x9f37afa8 [Unoptimized] [email protected]:call
[ ] E/DartVM (10621): pc 0x944ee750 fp 0x9f37b008 sp 0x9f37afe0 [Unoptimized]
[email protected]
[ ] E/DartVM (10621): pc 0x944ee550 fp 0x9f37b024 sp 0x9f37b010 [Unoptimized]
[email protected]
[ ] E/DartVM (10621): pc 0x99fd086c fp 0x9f37b05c sp 0x9f37b02c [Unoptimized] _rootRun@4048458
[ ] E/DartVM (10621): pc 0x99fd0580 fp 0x9f37b08c sp 0x9f37b064 [Unoptimized] _rootRun@4048458
[ ] E/DartVM (10621): pc 0x99fcfa50 fp 0x9f37b0d4 sp 0x9f37b094 [Unoptimized] [email protected]
[ ] E/DartVM (10621): pc 0x99fe3cf8 fp 0x9f37b10c sp 0x9f37b0dc [Unoptimized] [email protected]
[ ] E/DartVM (10621): pc 0x99fe3b58 fp 0x9f37b12c sp 0x9f37b114 [Unoptimized]
[email protected].<anonymous closure>
[ ] E/DartVM (10621): pc 0x99fd0aac fp 0x9f37b164 sp 0x9f37b134 [Unoptimized] _rootRun@4048458
[ +1 ms] E/DartVM (10621): pc 0x99fd0580 fp 0x9f37b194 sp 0x9f37b16c [Unoptimized] _rootRun@4048458
[ ] E/DartVM (10621): pc 0x99fcfa50 fp 0x9f37b1dc sp 0x9f37b19c [Unoptimized] [email protected]
[ ] E/DartVM (10621): pc 0x99fe3cf8 fp 0x9f37b214 sp 0x9f37b1e4 [Unoptimized] [email protected]
[ ] E/DartVM (10621): pc 0x99fe3b58 fp 0x9f37b234 sp 0x9f37b21c [Unoptimized]
[email protected].<anonymous closure>
[ ] E/DartVM (10621): pc 0x95349ac0 fp 0x9f37b258 sp 0x9f37b23c [Unoptimized] _microtaskLoop@4048458
[ ] E/DartVM (10621): pc 0x95349424 fp 0x9f37b278 sp 0x9f37b260 [Unoptimized] _startMicrotaskLoop@4048458
[ ] E/DartVM (10621): pc 0x953492bc fp 0x9f37b28c sp 0x9f37b280 [Unoptimized] _startMicrotaskLoop@4048458
[ ] E/DartVM (10621): pc 0x9e4827b8 fp 0x9f37b308 sp 0x9f37b294 [Stub] InvokeDartCode
[ ] F/libc (10621): Fatal signal 6 (SIGABRT), code -6 in tid 10639 (1.ui)
[ +99 ms] *** *** *** *** *** *** *** *** *** *** *** *** *** *** *** ***
[ ] Build fingerprint: 'motorola/athene/athene:6.0.1/MPJ24.139-64/65:user/release-keys'
[ ] Revision: 'p2a0'
[ ] ABI: 'arm'
[ ] pid: 10621, tid: 10639, name: 1.ui >>> com.example.bug <<<
[ ] signal 6 (SIGABRT), code -6 (SI_TKILL), fault addr --------
[ +37 ms] Abort message: '../../third_party/dart/runtime/vm/stack_trace.cc: 150: error: expected: future.GetClassId() ==
future_impl_class.id()'
[ ] r0 00000000 r1 0000298f r2 00000006 r3 9f37b978
[ ] r4 9f37b980 r5 9f37b930 r6 00000000 r7 0000010c
[ ] r8 a1ba4290 r9 9f37abc4 sl a1ba4290 fp 9f37ab4f
[ ] ip 00000006 sp 9f37aa28 lr b6d011cd pc b6d035c8 cpsr 40070010
[ +4 ms] backtrace:
[ ] #00 pc 000445c8 /system/lib/libc.so (tgkill+12)
[ ] #01 pc 000421c9 /system/lib/libc.so (pthread_kill+32)
[ ] #02 pc 0001bb1f /system/lib/libc.so (raise+10)
[ ] #03 pc 00018cc1 /system/lib/libc.so (__libc_android_abort+34)
[ ] #04 pc 00016880 /system/lib/libc.so (abort+4)
[ ] #05 pc 00ad3d4f /data/app/com.example.bug-1/lib/arm/libflutter.so (offset 0x17ad000)
```
The assertion fails in `CallerClosureFinder::GetCallerInFutureImpl`, which is called from `CallerClosureFinder::FindCallerInAsyncGenClosure`. The `future` object prints as `{"name":"crash"}` (probably a constant String used in `yield`) and it is not a _Future instance.
@username_5 This looks related to the lazy async stack traces. Could you please take a look?
username_4: Standalone repro for the assertion in `CallerClosureFinder::GetCallerInFutureImpl`:
```dart
class StreamTest {
static Stream<String> fetchData() async* {
yield '{"name":"crash"}';
StackTrace.current;
}
static Stream<dynamic> getStream() async* {
await for(String response in fetchData()) {
yield response;
}
}
}
main() async {
await for (var x in StreamTest.getStream()) {
print(x);
}
}
```
Assertion fails on Dart SDK in debug mode when running with `out/DebugX64/dart foo.dart`.
username_5: Yeah, this looks like something wrong in the async stack code.
@username_6 to figure out if this is something someone should work on.
username_6: Today I've had a brief look at this.
Simple things first: The [_StreamIterator](https://github.com/dart-lang/sdk/blob/248b27c37f2cb5c70d22c961bb412f3c25da4fc2/sdk/lib/async/stream_impl.dart#L924) class itself does specify clearly what the value of `_stateData` is depending on the state the iterator is in:
```dart
class _StreamIterator<T> implements StreamIterator<T> {
/// ... some elaborate comment here ...
/// After calling [moveNext] but before the returned future completes:
/// The returned future.
///
/// After calling [moveNext] and the returned future has completed
/// with `true`: The value of [current].
///
/// After calling [moveNext] and the returned future has completed
/// with `false`, or after calling [cancel]: `null`.
@pragma("vm:entry-point")
Object? _stateData;
```
The VM [assumes](https://github.com/dart-lang/sdk/blob/248b27c37f2cb5c70d22c961bb412f3c25da4fc2/runtime/vm/stack_trace.cc#L203) `_stateData` is a `_Future`:
```c++
if (callback_instance_.GetClassId() == stream_iterator_class.id()) {
future_ = Instance::Cast(callback_instance_).GetField(state_data_field);
return GetCallerInFutureImpl(future_);
}
```
which according to the comment on `_StreamIterator` is not necessarily the case.
=> We should remove the assumption in the VM and be more lenient
=> Made a small [cl/239311](https://dart-review.googlesource.com/c/sdk/+/239311) to do that.
We can see that once the value arrives in the `_StreamIterator` the field changes from being a `Future` to the yielded value:
```dart
class _StreamIterator<T> ... {
...
void _onData(T data) {
...
_Future<bool> moveNextFuture = _stateData as dynamic;
_stateData = data;
_hasValue = true;
moveNextFuture._complete(true);
if (_hasValue) _subscription?.pause();
}
...
}
```
Now we can see that the future gets completed, but the `await`er will only wake up asynchronously. In the mean time the generator seemingly resumes to run (i.e. even before the `await for (...)` loop gets its value), which causes this issue.
This seems like the wrong `async*` semantics (at least different from dart2js).
=> Filed https://github.com/dart-lang/sdk/issues/48695 for that.
=> Fixing the async semantics will also fix this issue (and have correct stack trace) I believe. |
hanami/model | 922611893 | Title: Use original error in case role is missing for Postgres
Question:
username_0: First of all, I want to thank developers and maintainers for all the good work 👍
We encountered the problem described in #537 while configuring Github Actions for our tests. Pull request was closed without any further info. The problem itself was easily fixed by providing the correct database url, but it took a while to find the source of issue.
Answers:
username_1: I think we could resurrect #537. |
w3c/w3c.github.io | 187704614 | Title: redundant spec labels?
Question:
username_0: I don't understand the difference between substantive and bug. Also, all changes are presumably enhancements, so I'm not sure what the purpose of that is, and it lacks clarity on whether it means editorial or substantive change.
Answers:
username_1: For me, a `bug` is something that is broken, while an `enhancement` is an aspect that works but could/should be improved. Yes, every change is an enhancement, in some way (or you wouldn't make it) but I think that's a common convention.
And `substantive` and `bug` aren't exclusive, I think: bugs can be trivial, substantive, catastrophic… And enhancements, too.
username_0: Fair enough. Can we have slightly more explicit labels, given that there isn't any documentation in the issues system of github? Or should we just document the labels somewhere for the first time people are looking at them? (Hello maintenance)
username_1: I guess we could be more specific on [`https://w3c.github.io/spec-labels.html`](https://w3c.github.io/spec-labels.html), yes. Although for me that suffices as a general recommendation: groups end up with their own conventions, and the exact meaning of labels will vary depending on the nature of the repo, etc.
username_1: @username_2, I'd close this one when #34 is merged
username_1: I propose we close this one now.
Status: Issue closed
|
frishkorn/timeClock | 209552507 | Title: Date Roll Over on Serial Output
Question:
username_0: When putty stays connected over-night it can be confusing which day is where. At midnight have serial output the new date.
Answers:
username_0: Starting work on this issue, creating new branch.
username_0: Roll over has been implemented and this enhancement can now be closed. Recommend merging into master.
Status: Issue closed
|
borodust/libresect | 936540393 | Title: Fix msvc mangling fiasco
Question:
username_0: I keep triggering this Clang assert when parsing in MSVC mode:
```cxx
// MicrosoftMangle.cpp
void MicrosoftCXXNameMangler::mangleType(const InjectedClassNameType *,
Qualifiers, SourceRange) {
llvm_unreachable("Cannot mangle injected class name type.");
}
```
Need to do something about it. Either figure out how to avoid dropping into this method or patch it or somehow disable MSVC mangling scheme.
Answers:
username_0: ```sh
# since clang 13 this might work
-fc++-abi=itanium
``` |
typpo/textbelt | 47837780 | Title: Not receiving text in Telia Denmark
Question:
username_0: Hi,
I just tested with my number to times
`curl -X POST http://textbelt.com/intl -d number=00454215xxxx -d "message=This is a test msg from textbelt.com"`
but I didn't receive any text. My provider is Telia Denmark (Website says it supports.)
\ Mathu Mou.
Answers:
username_1: +1
username_2: Same for Vodafone Portugal.
username_3: same in switzerland
username_4: Same in Austria, tested with: 3/Hutchinson and T-Mobile.
Status: Issue closed
username_5: Hi all,
This weekend I made the decision to turn down free service on Textbelt.com. There is now an option to pay a small amount for an API key. Although I don't have a Danish/Italian/Portuguese/Swiss/Austrian test number, the paid version should be able to text carriers in all of Europe (I encourage you to try this with the `textbelt_test` key). The free, open-source version in this Github repository is still available for people to self-host.
I'm closing this issue because it applies to the old Textbelt.com, which was continually abused and therefore blocked by many cell providers. If you are able to reproduce this issue after self-hosting, please re-open it. Thank you! |
sidnarayanan/PandaCore | 244468416 | Title: Run on T2
Question:
username_0: Instead of waiting to figure out submit access, set up running on the T2. This only involves getting the stageout to work (e.g. use gfal utils)
Answers:
username_0: Progress report: it involved a little more than just getting the stageout to work. Fixed the other problems, but stageout still does not work.
Status: Issue closed
username_0: Resolved in https://github.com/username_0/PandaCore/pull/32 |
aws/aws-cli | 156860794 | Title: feature request: it would be nice to have the ability to sync buckets between aws regions where the bucket owners are different accounts
Question:
username_0: aws-cli/1.10.33 Python/2.7.10 Darwin/15.5.0 botocore/1.4.23
problem: I need to be able to backup file assets from S3 to a second environment for a variety of reasons. I want assetA in an environment 1 bucket to be synched to an environment 2 bucket and I want to use the aws cli to do it. This does not appear possible when each environment I use has different credentials (segmented due to security requirements).
what I see:
aws s3 sync s3://bucketA/file s3://bucketB/file --source-region us-west-2 --region us-east-1 --profile company-production
A client error (AccessDenied) occurred when calling the ListObjects operation: Access Denied
Completed 1 part(s) with ... file(s) remaining
seems like we should be able to do the following:
aws s3 sync s3://bucketA/file s3://bucketB/file --source-region us-west-2 --region us-east-1 --source-profile company-production --profile company-rc
we use credentials files in /.aws that contain multiple accounts (1 for each environment). It makes it hard for me to move assets between environments with the CLI. As long as I have the correct credentials for both environments, I should be able to do this.
Answers:
username_1: Makes sense. Labeling it as a feature request. Also have you considered setting up a bucket policy to grant the account for ``company-production`` to list objects owned by profile ``company-rc``?
username_0: I have. However, that would be a change to our infrastructure that would allow other services that live in environment ```company-production``` to have access to that ```company-rc``` bucket which is something we wish to avoid managing and/or regulating. Having the CLI be able to discretely access both resources using existing user/service credentials mechanisms without having to modify the bucket access security settings seems preferable.
username_2: I don't see how this could possibly work without an intermediate transfer location, somewhere that the first set of credentials could write to, and the second set of credentials could read from.
username_1: It would work because under the hood we would create a new client to list the object for cross region bucket transfers and in this client we would just use a different profile to create the client.
username_1: Actually now that I think of it specifically for the copy object. The creds for the destination must have permissions to copy objects from the source. I am not sure if this is feasible to do this. There is no notion of granting temporary permissions to copy from a source to a destination. Permissions need to be granted.
Status: Issue closed
|
fiaas/fiaas-deploy-daemon | 462737938 | Title: Update integration tests to test against updated k8s versions
Question:
username_0: Shift the kubernetes versions in e2e tests to updated versions (min 1.9.x)
Answers:
username_1: We're already providing 1.14 clusters so this one is quite desirable.
username_2: Shifting the versions in the e2e test is a one-line change, but when we get to the point where we don't support any version of Kubernetes that uses Third Party Resources, we should also remove the code for those. That would reduce the codebase significantly, and also simplify some parts of it.
username_2: Fixed in #26
Status: Issue closed
|
blockdiag/seqdiag | 320574251 | Title: Empty diagrams
Question:
username_0: When trying to render and empty diagram like this:
seqdiag {
}
It fails due to list IndexError here:
https://bitbucket.org/username_0/seqdiag/src/ff8f48be1e29/src/seqdiag/builder.py#cl-54
----------------------------------------
- Bitbucket: https://bitbucket.org/blockdiag/seqdiag/issue/10
- Originally reported by: <NAME>
- Originally created at: 2012-09-06T09:56:52.110
Status: Issue closed
Answers:
username_0: _From <NAME> on 2012-09-07 09:00:21+00:00_
Fixed on [[https://bitbucket.org/username_0/seqdiag/changeset/773b384e67e5|773b384e67e5]] |
getgrav/grav-theme-bootstrap | 109631877 | Title: Inherit from bootstrap theme KO
Question:
username_0: # INFO:
=>pluging Bootstrapper eneble version: 1.1.0
=>template bootstrap version: 1.4.0
=>new theme : mytheme with mytheme.yaml
```yaml
streams:
schemes:
theme:
type: ReadOnlyStream
prefixes:
'':
- user/themes/mytheme
- user/themes/bootstrap
```
=>system.yaml : pages: theme: mytheme
# ISSUE:
bootstrap framework is not loaded but the content is loaded.

Alternative: If you load the antimatter theme from the mytheme, it works
```yaml
streams:
schemes:
theme:
type: ReadOnlyStream
prefixes:
'':
- user/themes/mytheme
- user/themes/antimatter
```
#SOLUTION
Not identified yet
Status: Issue closed
Answers:
username_1: Copy the bootstrap.php file into yoru theme and name it `mytheme.php` and then update it appropriately:
```
<?php
namespace Grav\Theme;
use Grav\Common\Theme;
class Mytheme extends Theme
{
// Boostrap plugin will look for this class var to know it should load
public $load_bootstrapper_plugin = true;
}
```
username_0: I have forgotten this point.
thanks alot for your quick answer.
username_1: no problem! |
whong92/recommender | 436911044 | Title: Make recommender predict for new user
Question:
username_0: Assuming the set of movies stay the same, embedding vector for a new user can be provided by converting it into a linear regression problem
Answers:
username_0: Given a new user u_j and ratings X_j, we can do:
u_j = argmin u_j Loss(u_j * I - X_j)
Where I is the item matrix embedding
Status: Issue closed
|
feuery/qmapper | 493598997 | Title: Replace prop editor with something better
Question:
username_0: I was thinking about a window which lets user edit the object as pure lisp s-expr, which would be evaluated and committed back to the document when the window closes. There could also be some sort of schema which had to validate before committing<issue_closed>
Status: Issue closed |
fossasia/susi_android | 246486423 | Title: Change Layout of Action Bar (Sub-issue of issue #857)
Question:
username_0: 1. Remove Action bar and search feature as it is of no use in a personal assistant app. This feature is more inclined towards a chat/messenger application and not Personal Assistant. Remove action bar to increase space. Instead introduce a FAB with settings icon. Clicking on which opens settings activity ( If possible, use slide animation to open settings)
Reference: 
2. Instead of adding multi select and displaying share and copy icon on actionbar, display it as a dialog box. See this for reference:

Answers:
username_1: Are we removing change wallpaper settings as well here? @username_0
username_0: Yes, wallpaper and login/logout will go in settings.
Status: Issue closed
|
cartr/homebrew-qt4 | 290815821 | Title: PyQt4 --with-python 3
Question:
username_0: I am trying to build pyqt4 with python3 support, however I get the following error where is unable to locate phonon libraries:
```
Last 15 lines from /Users/al/Library/Logs/Homebrew/pyqt@4/03.make:
cd phonon/ && /Applications/Xcode.app/Contents/Developer/usr/bin/make -f Makefile
clang++ -c -pipe -stdlib=libc++ -mmacosx-version-min=10.7 -fno-exceptions -O2 -arch x86_64 -Wall -W -fPIC -DSIP_PROTECTED_IS_PUBLIC -Dprotected=public -DQT_NO_DEBUG -DQT_PLUGIN -DQT_PHONON_LIB -DQT_GUI_LIB -DQT_CORE_LIB -DQT_SHARED -I/usr/local/etc/qt4/mkspecs/unsupported/macx-clang-libc++ -I. -I/usr/local/Cellar/qt@4/4.8.7_3/lib/QtCore.framework/Versions/4/Headers -I/usr/local/Cellar/qt@4/4.8.7_3/lib/QtCore.framework/Versions/4/Headers -I/usr/local/Cellar/qt@4/4.8.7_3/lib/QtGui.framework/Versions/4/Headers -I/usr/local/Cellar/qt@4/4.8.7_3/lib/QtGui.framework/Versions/4/Headers -I/usr/local/include/phonon -I/usr/local/include -I. -I/System/Library/Frameworks/Python.framework/Versions/2.7/include/python2.7 -I/usr/local/include/phonon_compat -I. -F/usr/local/lib -o sipphononcmodule.o sipphononcmodule.cpp
clang++ -c -pipe -stdlib=libc++ -mmacosx-version-min=10.7 -fno-exceptions -O2 -arch x86_64 -Wall -W -fPIC -DSIP_PROTECTED_IS_PUBLIC -Dprotected=public -DQT_NO_DEBUG -DQT_PLUGIN -DQT_PHONON_LIB -DQT_GUI_LIB -DQT_CORE_LIB -DQT_SHARED -I/usr/local/etc/qt4/mkspecs/unsupported/macx-clang-libc++ -I. -I/usr/local/Cellar/qt@4/4.8.7_3/lib/QtCore.framework/Versions/4/Headers -I/usr/local/Cellar/qt@4/4.8.7_3/lib/QtCore.framework/Versions/4/Headers -I/usr/local/Cellar/qt@4/4.8.7_3/lib/QtGui.framework/Versions/4/Headers -I/usr/local/Cellar/qt@4/4.8.7_3/lib/QtGui.framework/Versions/4/Headers -I/usr/local/include/phonon -I/usr/local/include -I. -I/System/Library/Frameworks/Python.framework/Versions/2.7/include/python2.7 -I/usr/local/include/phonon_compat -I. -F/usr/local/lib -o sipphononPhonon.o sipphononPhonon.cpp
/private/tmp/pyqt@4-20180123-86024-75t8w7/PyQt4_gpl_mac-4.12.1/sip/phonon/abstractaudiooutput.sip:28:10: fatal error: 'abstractaudiooutput.h' file not found
#include <abstractaudiooutput.h>
^~~~~~~~~~~~~~~~~~~~~~~
1 error generated.
make[1]: *** [sipphononPhonon.o] Error 1
make[1]: *** Waiting for unfinished jobs....
/private/tmp/pyqt@4-20180123-86024-75t8w7/PyQt4_gpl_mac-4.12.1/sip/phonon/abstractaudiooutput.sip:28:10: fatal error: 'abstractaudiooutput.h' file not found
#include <abstractaudiooutput.h>
^~~~~~~~~~~~~~~~~~~~~~~
1 error generated.
make[1]: *** [sipphononcmodule.o] Error 1
make: *** [sub-phonon-make_default-ordered] Error 2
If reporting this issue please do so at (not Homebrew/brew or Homebrew/core):
https://github.com/username_1/homebrew-qt4/issues
```
Is there a way to disable phonon when building pyqt@4 to be able to compile pyqt4?
Thanks
Answers:
username_1: Sorry to hear that `pyqt@4` isn't building correctly on your system.
In theory, PyQt should detect that Phonon is unavailable and disable it automatically. That's what it does on my system, and I'm not able to reproduce the error you reported.
Could you please run `brew config` and `brew doctor`, paste the output into a Gist, and put the link to the Gist in a comment on this issue?
Status: Issue closed
username_1: Closing, as I haven't heard back from you. If you're still experiencing this problem, please let me know and I'll reopen it. |
pyecharts/pyecharts | 433301221 | Title: 1.0.0版本render_embed, Page中添加bar报错
Question:
username_0: ```
from pyecharts.custom.page import Page
def echarts():
page = Page()
...
line.add_xaxis(x)
bar.add_xaxis(x)
bar.add_yaxis(s, t)
line.add_yaxis(s, p, is_smooth=True)
page.add_chart(bar)
page.add_chart(line)
return page.render_embed()
```
报错信息如下:
Traceback (most recent call last):
File "C:\Users\xxx\AppData\Local\Programs\Python\Python37\lib\site-packages\django\core\handlers\exception.py", line 34, in inner
response = get_response(request)
File "C:\Users\xxx\AppData\Local\Programs\Python\Python37\lib\site-packages\django\core\handlers\base.py", line 115, in _get_response
response = self.process_exception_by_middleware(e, request)
File "C:\Users\xxx\AppData\Local\Programs\Python\Python37\lib\site-packages\django\core\handlers\base.py", line 113, in _get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "D:\Projects\xxxx\xxxx\views.py", line 17, in index
'Echarts': echarts()
File "D:\Projects\xxxx\xxxx\echarts.py", line 52, in TMcharts
return page.render_embed()
File "C:\Users\xxx\AppData\Local\Programs\Python\Python37\lib\site-packages\pyecharts-1.0.0-py3.7.egg\pyecharts\custom\page.py", line 105, in render_embed
return Markup("<br/> ".join([chart.render_embed() for chart in self]))
File "C:\Users\xxx\AppData\Local\Programs\Python\Python37\lib\site-packages\pyecharts-1.0.0-py3.7.egg\pyecharts\custom\page.py", line 105, in <listcomp>
return Markup("<br/> ".join([chart.render_embed() for chart in self]))
AttributeError: 'Bar' object has no attribute 'render_embed'
求帮助,谢谢。
Answers:
username_1: @username_0 你确定你用的方法没错? Page 里面没有 render_embed 的方法
username_0: 在custom里有个page,里面的Page有page_embed
0.5.11版本的page_embed正常的
username_0: @username_1 pyecharts里的base.py中的Base()该怎么使用,这个Base里有render_embed,我试了应该是初始化了一个div容器
求教,谢谢。
username_1: @username_0
* 首先,先认真确定你的 pyecharts 的版本是 1.0.0。
* 其次 1.0.0 已经没有 render_embed 的方法了。
username_0: @username_1
确定一定以及肯定是1.0.0版本:
`from pyecharts.custom.page import Page`
这里有Page的render_embed,但是当Page.add_chart后,Chart本身没有render_embed方法了。
这个功能在0.5.11里是正常的
username_1: @username_0
* 烦请再次确认版本是否为 1.0.0。
* 源码项目截图:

* Pycharm 全局截图

* Page 类结构截图

username_1: @username_0
* 只有 0.5.11 项目结构才符合你贴出来的引入路径

* ```master``` 分支下的是 ```0.5.11```;```2019``` 分支下是 ```1.0.0```;
username_0: OK, 谢谢,我回家后做下确认,代码在家里。
也有可能是当时我用源码先做过一次master分支的install,后来又切换到2019做了一次install,可能混乱了?
昨晚我还特意追踪过代码,目测跟Master的结构以及跟2019的结构混合了。。。
username_0: 
Look,我的库里真的有custom文件夹,估计是真如我所说的,先install了master分支,后又操作了2019分支,就出现这个情况,我重装下看看
username_0: 确定是这个情况了,0.5.11 build后build目录和dist目录保留了0.5.11的源文件,切换到2019再做编译的时候,两个版本的内容合起来了,打包egg的时候就有了两个版本的内容,交错了。
估计还是rmtree没删除干净
Status: Issue closed
username_0: 问题解决,可以关闭了,谢谢 |
stephy/CalendarPicker | 420987949 | Title: Controller for monthLabel
Question:
username_0: Is there any solution for adding an action such as onPressMonthLabel? Because swiping is super slow, and an user might want to go to a date 2-3 years in the past/future I would like to have a months popup or something in order to scroll through months.
Answers:
username_1: Adding `onMonthPress` should be relatively straight-forward. PRs are welcome 😄
username_2: I just created a PR for this and for a year press as well. I just need to figure out the test snapshot as I made some large changes and it no longer matches the snapshot on the test. The snapshot test is a bit new to me, but if you need this feature now feel free to branch off my branch. If you can help adjust the tests or anything feel free to help me there as well. Check out PR #151
Status: Issue closed
username_1: This has been implemented in commit 6<PASSWORD> . |
moxystudio/react-contentful-rich-text | 609105649 | Title: Override render method for a nodetype
Question:
username_0: There might be cases where a developer wants to take control on how to render a specific `nodeType`.
Using `hyperlink` as an example, component that deals with hyperlinks inside the project can already exist and that could be used instead of what is the default behavior of the package. |
SeleniumHQ/docker-selenium | 230003396 | Title: Hover state does not work with non-debug images
Question:
username_0: I have a test for hover states that use something like `moveToElement` and check that the element hidden normally is present on hovering.
This test passes fine if I use the `debug` images, but not the regular ones.
Does anyone know why this might be the case? Is there a way I can get it to pass reliably for non-debug images?
Answers:
username_1: Hi @username_0 could you rephrase the issue description and make it deterministic?
1. "something like moveToElement" -- so do you use moveToElement or what exactly? please be very specific.
2. "element hidden normally is present" -- what is "normally" can this be rephrased with concrete reproducible steps?
3. can you provide a complete test case in https://gist.github.com/ for example? that would be awesome.
username_0: Hi @username_1,
My apologies for the vague bug report.
I did mean `moveToElement`.
Here's a simple set up to see this:
```js
// test/index.js
module.exports = {
'hover': function (browser) {
browser
.url('https://lab.tridnguyen.com/demo.html')
.moveToElement('.container', 20, 10)
.assert.visible('.menu')
.end();
}
};
```
and
```json
// nightwatch.json
{
"src_folders": ["test"],
"test_settings": {
"default": {
"desiredCapabilities": {
"browserName": "firefox"
}
}
}
}
```
And then run:
```sh
# install nightwatch
:; npm install nightwatch
# run the container
:; docker run -d -p 4444:4444 selenium/standalone-firefox
# run the test
:; nightwatch
```
When running that, I get the following error:
```
[Index] Test Suite
======================
Running: hover
✖ Testing if element <.menu> is visible. - expected "true" but got: "false"
at Object.hover (/home/tri/dev/temp/test/index.js:6:12)
at _combinedTickCallback (internal/process/next_tick.js:73:7)
FAILED: 1 assertions failed (2.469s)
TEST FAILURE: 1 assertions failed, 0 passed. (2.561s)
✖ index
- hover (2.469s)
```
It works on `chrome` and `firefox-debug`.
username_1: Thanks for the great bug report @username_0 !!!
I'm able to reproduce your issue but I'm surprised that is passing on firefox-debug as this seems to be a Firefox issue, independent of the debug image, or if using Zalenium for example.
See selenium logs from [Zalenium](https://github.com/zalando/zalenium)

And I can see the error:
```java
13:53:04.457 WARN - Exception thrown
org.openqa.selenium.UnsupportedCommandException: mouseMoveTo
```
I think this might be more related to https://github.com/SeleniumHQ/selenium/issues/3808
username_0: Yeah, the part about it passing on debug made me think originally that it's something with having a graphical interface... I believe the only difference between the regular image and the debug one is the addition of a vnc server, is that right?
I subscribed to that selenium issue, thanks for the reference.
username_1: Is it possible that is passing for you in the debug image because you have an older image? older Selenium or Firefox in there? I believe this issue is independent of VNC as the error message says `org.openqa.selenium.UnsupportedCommandException: mouseMoveTo` can't imagine how VNC can be related to that error. Can you double check your versions?
username_0: Oh you're right. It doesn't work on firefox, period. The reason I thought it was working on firefox-debug was because a script automatically switched from firefox-debug to chrome-debug in our test run due to #478. I stand corrected.
username_1: I think we can agree this is not a docker-selenium issue but upstream Selenium one so would be valid to close it here, what do you think?
Status: Issue closed
username_1: Does someone knows which was the last version of Firefox & Selenium that this worked?
username_2: Confirmed. Works with Selenium 2.53.1 and FF 47.0.1 |
NuGet/docs.microsoft.com-nuget | 890320631 | Title: How are project references treated when resolving the dependency tree?
Question:
username_0: How are project references treated when resolving the dependency tree. Are they processed in "levels" just like package references or are all project references (transitive and non-transitive) treated as top-level? It would be nice to clarify this in the docs.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 8e8a945e-ee7e-cefa-282a-41da09a951b9
* Version Independent ID: c380b1cb-6e17-7777-995f-7c8b48f581de
* Content: [NuGet Package Dependency Resolution](https://docs.microsoft.com/en-us/nuget/concepts/dependency-resolution)
* Content Source: [docs/concepts/Dependency-Resolution.md](https://github.com/NuGet/docs.microsoft.com-nuget/blob/main/docs/concepts/Dependency-Resolution.md)
* Product: **nuget**
* GitHub Login: @JonDouglas
* Microsoft Alias: **jodou** |
cakephp/queue | 789760837 | Title: [InvalidArgumentException] Unknown connection scheme: 'redis'
Question:
username_0: I've been trying this plugin for mailing purpose. Somehow I've an issue connecting to redis.
When I'm using predis directly, I can connect and submit my data to the redis.
This is my stack trace
```
2021-01-20 08:33:30 Error: [InvalidArgumentException] Unknown connection scheme: 'redis'. in C:\Dev\www\emailtemplate\vendor\predis\predis\src\Connection\Factory.php on line 87
Stack Trace:
- C:\Dev\www\emailtemplate\vendor\predis\predis\src\Client.php:120
- C:\Dev\www\emailtemplate\vendor\predis\predis\src\Client.php:59
- C:\Dev\www\emailtemplate\vendor\enqueue\redis\PRedis.php:126
- C:\Dev\www\emailtemplate\vendor\enqueue\redis\RedisConnectionFactory.php:105
- C:\Dev\www\emailtemplate\vendor\enqueue\redis\RedisConnectionFactory.php:89
- call_user_func - [internal], line ??- C:\Dev\www\emailtemplate\vendor\enqueue\redis\RedisContext.php:160
- C:\Dev\www\emailtemplate\vendor\enqueue\redis\RedisProducer.php:70
- C:\Dev\www\emailtemplate\vendor\enqueue\enqueue\Client\Driver\GenericDriver.php:225
- C:\Dev\www\emailtemplate\vendor\enqueue\enqueue\Client\Driver\GenericDriver.php:62
- C:\Dev\www\emailtemplate\vendor\enqueue\enqueue\Client\Producer.php:121
- C:\Dev\www\emailtemplate\vendor\enqueue\enqueue\Client\Producer.php:53
- C:\Dev\www\emailtemplate\vendor\enqueue\simple-client\SimpleClient.php:185
- C:\Dev\www\emailtemplate\vendor\cakephp\queue\src\QueueManager.php:217
- C:\Dev\www\emailtemplate\vendor\cakephp\queue\src\Mailer\QueueTrait.php:53
- C:\Dev\www\emailtemplate\src\Controller\Api\EmailsController.php:120
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Controller\Controller.php:529
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Controller\ControllerFactory.php:79
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\BaseApplication.php:251
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\Runner.php:77
- C:\Dev\www\emailtemplate\vendor\cakephp\authorization\src\Middleware\AuthorizationMiddleware.php:129
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\Runner.php:73
- C:\Dev\www\emailtemplate\vendor\cakephp\authentication\src\Middleware\AuthenticationMiddleware.php:122
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\Runner.php:73
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Routing\Middleware\RoutingMiddleware.php:166
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\Runner.php:73
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\Middleware\BodyParserMiddleware.php:174
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\Runner.php:73
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Routing\Middleware\RoutingMiddleware.php:166
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\Runner.php:73
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Routing\Middleware\AssetMiddleware.php:68
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\Runner.php:73
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Error\Middleware\ErrorHandlerMiddleware.php:121
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\Runner.php:73
- C:\Dev\www\emailtemplate\vendor\cakephp\debug_kit\src\Middleware\DebugKitMiddleware.php:60
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\Runner.php:73
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\Runner.php:58
- C:\Dev\www\emailtemplate\vendor\cakephp\cakephp\src\Http\Server.php:90
- C:\Dev\www\emailtemplate\webroot\index.php:40
```
My configuration look like this.
```
'DebugKit' => [
'ignoreAuthorization' => true,
],
'Queue' => [
'default' => [
// A DSN for your configured backend. default: null
'url' => 'redis:',
//'url' => 'redis://127.0.0.1:6397',
// The queue that will be used for sending messages. default: default
// This can be overriden when queuing or processing messages
'queue' => 'default',
// The name of a configured logger, default: null
'logger' => 'stdout',
// The name of an event listener class to associate with the worker
//'listener' => \App\Listener\WorkerListener::class,
]
],
```
Answers:
username_1: From the snippet you provided your `url` is `redis:` which isn't going to work. I've not used predis before. What are its connection strings supposed to look like?
username_0: This is my predis code.
```
<?php
include 'vendor/autoload.php';
require 'vendor/predis/Predis/autoload.php';
Predis\Autoloader::register();
$client = new Predis\Client('tcp://127.0.0.1:6379');
$client->set('foo', 'bar');
$value = $client->get('foo');
echo $value;
?>
```
I didn't set any password on redis. 127.0.0.1:6379 should be where my redis is running. Currently I'm just following https://book.cakephp.org/queue/1/en/index.html#configuration since the docs is stating using redis.
username_2: Try setting the following as your `url` key when configuring this library:
```
redis+predis://127.0.0.1:6379/0
```
username_0: i still get the same error message.
```
'Queue' => [
'default' => [
// A DSN for your configured backend. default: null
'url' => 'redis+predis://127.0.0.1:6379/0',
//'url' => 'redis://127.0.0.1:6379',
// The queue that will be used for sending messages. default: default
// This can be overriden when queuing or processing messages
'queue' => 'default',
// The name of a configured logger, default: null
'logger' => 'stdout',
// The name of an event listener class to associate with the worker
//'listener' => \App\Listener\WorkerListener::class,
]
],
```
By the way I'm using the Queuing Mailer Actions. In case it help to be more specific. |
iferres/MLSTar | 336097480 | Title: doMLST() Error on R Console
Question:
username_0: In R version 3.3.3 on Mac OS X Mavericks 10.9.5,
using the doMLST() function by pasting into the console printed Error:
```
Running BLASTN...Error in strsplit(db, "/") : non-character argument
The process has forked and you cannot use this CoreFoundation functionality safely. You MUST exec().
Break on __THE_PROCESS_HAS_FORKED_AND_YOU_CANNOT_USE_THIS_COREFOUNDATION_FUNCTIONALITY___YOU_MUST_EXEC__() to debug.
```
It ran without any problem in R version 3.5.0 on macOS High Sierra 10.13.4. |
js-joda/js-joda | 149810157 | Title: Update DateTimeFormatterPattern documentation to indicate which conversion currently work
Question:
username_0: As mentioned in https://github.com/js-joda/js-joda/issues/34#issuecomment-212482961, the current documentation of the ofPattern method mentions all patterns but does not indicate which of the patterns do not currently work due to missing locale support
We should either remove these patterns or mark them to make it clear, that they are not currently supported<issue_closed>
Status: Issue closed |
Mashashi/javaroles | 142906663 | Title: Call to ref rigid should vary its implementation accordingly from where it is called
Question:
username_0: If the call is made from a subclass the method called should be the same where the rigid ref was defined. Otherwise it override of this method should be allowed the attribute that currently exists in the annotation is just another annoyance for the programmer. |
DIYgod/RSSHub | 415008282 | Title: 请求支持91porn视频主视频更新提醒
Question:
username_0: <!--
请确保 [文档](https://docs.rsshub.app) 和 [issue](https://github.com/DIYgod/RSSHub/issues) 中没有相关内容,并按照模版提供信息
否则 issue 将被立即关闭
-->
### 这是一个什么样的功能?
提醒91porn视频主视频更新了
### 这个功能可以解决什么问题?
推送91porn视频主更新视频的名称、链接、时间
### 额外描述
比如:http://91porn.com/uvideos.php?UID=5d4e5y0Q94bxB5Zdp0MHhWSjHqFjeI7aeAoMROZMp5V95hy0&type=public
title://*[@id="myvideo-content"]//p//a/text()
url://*[@id="myvideo-content"]//p//a/@href
date://*[@id="myvideo-content"]/div/div[2]/p/text()[3] |
vochabular/admin-backend | 430651465 | Title: Chapter Mutations: Can't insert due to mandatory field "fkBelongsToId"
Question:
username_0: When inserting (the first...) chapter, I can't since the fkBelongsToId field is mandatory, yet I can't specify one since there is none yet available and if it is specified, it violates the FK-constraint.
Mutation to reproduce:
```
mutation createChapter($input: IntroduceChapterInput!) {
createChapter(input: $input) {
chapter {
id
}
}
}
```
with variables:
```
{"input": {
"chapterData": {
"title": "Test",
"number": 1,
"description": "Hello world!",
"fkBelongsToId":
}
}}
```<issue_closed>
Status: Issue closed |
kerinin/cfrp-rs | 71398630 | Title: Broken package links
Question:
username_0: The [homepage link](http://username_1.github.io/cfrp-rs) is returning a 404 right now and I think the repository page, which is currently pointing to the same location, should point to https://github.com/username_1/cfrp-rs instead.
If you don't have any content for a homepage right now, I would suggest removing that field from the package as I don't think it gives a good impression to other people when a project links to pages that don't exist.
Answers:
username_1: Good point, updating crate.
Status: Issue closed
|
argonlefou/DemulShooter | 817937066 | Title: Lost Land Adventures Support
Question:
username_0: Is there a way to play Lost Land Adventures with DemulShooter?
Status: Issue closed
Answers:
username_1: Unfortunatelly no, and there is no plan to add such a thing.
At least, not before someone is able to fix the display and make **really** removes every aspect of the "dome" part of the game (i.e rounded dome rendering and non linear axis) |
AdguardTeam/AdguardForMac | 970968534 | Title: Secure browsing
Question:
username_0: When I had read all the documentation some month ago I have seen that Windows and Mac version of secure browsing doesn't work the same (cache on windows and real time on Mac).
So my request is :
Does the Mac version do caching of the safe browsing (like windows) if not can you add it ?
Thanks
Answers:
username_1: Hello there!
This safe browsing method will most likely be implemented in the next release.
username_0: @username_1 Thanks. |
kiwigrid/helm-charts | 504753675 | Title: [fluentd-elasticsearch] parses dates wrong when creating indices
Question:
username_0: <!-- Thanks for filing an issue!
Before hitting the button, please answer these questions.
It's helpful to search the existing GitHub issues first.
It's likely that another user has already reported the issue you're facing,
or it's a known issue that we're already aware of-->
**Is this a request for help?**:
---
**Is this a BUG REPORT or FEATURE REQUEST?** (choose one):
This might be a BUG
<!--
If this is a BUG REPORT, please:
- Fill in as much of the template below as you can. If you leave out
information, we can't help you as well.
If this is a FEATURE REQUEST, please:
- Describe *in detail* the feature/behavior/change you'd like to see.
In both cases, be ready for followup questions, and please respond in a timely
manner. If we can't reproduce a bug or think a feature already exists, we
might close your issue. If we're wrong, PLEASE feel free to reopen it and
explain why.
-->
**Version of Helm and Kubernetes**:
Kubernetes v1.11.7 - Argocd v1.2.0
**Which chart in which version**:
fluentd-elasticsearch v2.7.0
**What happened**:
elasticsearch indices get created with wrong date. Over the course of 2 minutes the following indices are created
x-logstash-2064.03.25
x-logstash-2064.03.27
x-logstash-2064.03.08
x-logstash-2019.10.09
x-logstash-1970.01.01
x-logstash-2064.03.26
Likely that datetime is being parsed wrongly
**What you expected to happen**:
One index should be created for the day with the current timestamp
x-logstash-2019.10.09 (today)
**How to reproduce it** (as minimally and precisely as possible):
Values that are applied can be accessed here:
https://gist.github.com/username_0/dd2844807df904331f23bec045697451
**Anything else we need to know**:
Answers:
username_1: Does it work with chart verion 4.9.0?
If so it might be related to: https://github.com/kiwigrid/helm-charts/pull/169
username_0: @username_1 i'm sorry I grabbed the fluentd version instead of the chart version. I downgraded from 4.9.1 to 4.9.0 and I see the same behavior as before, multiple indices get created within minutes
username_1: Please check which apps send these dates in Elasticsearch / Kibana.
After that please check the actual logs for the timestamps please.
username_0: Here is a log entry from kubernetes-external-secrets pod where timestamp translates to Thu, 10 Oct 2019 10:08:13 GMT how ever when it is inserted into elasticsearch it has a timestamp from "2064-04-22T19:23:08.000Z"
```json
{"level":50,"time":1570702172674,"pid":17,"hostname":"kubernetes-external-secrets-5cb85dbf8d-rr69n","msg":"failure while polling the secrets","v":1}
{
"_index": "x-logstash-2064.04.22",
"_type": "_doc",
"_id": "e8UotW0B-qgoVCfSdZbS",
"_version": 1,
"_score": null,
"_source": {
"stream": "stdout",
"docker": {
"container_id": "6b0292b13fcfb15a4f0050f32719e257813b1928b0c5843382663712e4e61938"
},
"kubernetes": {
"container_name": "kubernetes-external-secrets",
"namespace_name": "operations",
"pod_name": "kubernetes-external-secrets-5cb85dbf8d-rr69n",
"container_image": "godaddy/kubernetes-external-secrets:1.3.1",
"container_image_id": "docker-pullable://godaddy/kubernetes-external-secrets@sha256:a2708a12128e4129457ca1162b6cef5801fa5e38e6a5f408ac15c34c35775629",
"pod_id": "8a2d452a-df7d-11e9-b426-06887de226b0",
"host": "ip-10-252-195-177.eu-west-1.compute.internal",
"labels": {
"pod-template-hash": "1764186948",
"app_kubernetes_io/instance": "kubernetes-external-secrets",
"app_kubernetes_io/name": "kubernetes-external-secrets"
},
"master_url": "https://172.16.31.10:443/api",
"namespace_id": "76200989-de01-11e9-a941-0a20d608e880",
"namespace_labels": {
"app_kubernetes_io/instance": "namespaces"
}
},
"level": 50,
"pid": 17,
"hostname": "kubernetes-external-secrets-5cb85dbf8d-rr69n",
"msg": "failure while polling the secrets",
"v": 1,
"@timestamp": "2064-04-22T19:23:08.000000000+00:00",
"tag": "kubernetes.var.log.containers.kubernetes-external-secrets-5cb85dbf8d-rr69n_operations_kubernetes-external-secrets-6b0292b13fcfb15a4f0050f32719e257813b1928b0c5843382663712e4e61938.log"
},
"fields": {
"@timestamp": [
"2064-04-22T19:23:08.000Z"
]
},
"sort": [
2976117788000
]
}```
username_1: Hmm... weird... never seen this before.
username_0: I think I've identified the problem to some extent at least
Fluent::Plugin::ElasticsearchErrorHandler::ElasticsearchError error="400 - Rejected by Elasticsearch [error type]:
mapper_parsing_exception [reason]:
failed to parse field [fields.time] of type [long] in document
time=1970-01-01 00:33:39.000000000 +0000
"fields.time"=>"2019-10-11T14:24:33Z"
containersInputConf: source time_format %Y-%m-%dT%H:%M:%S.%NZ does not match this time_format of %Y-%m-%dT%H:%M:%SZ where subseconds are missing from the source log.
Any suggestions on how one would accomplish handling both 2019-10-11T14:24:33Z and 2019-10-11T14:24:33.001Z time fields?
username_2: I get this same error. I am almost positive its when logs (like external secrets) which I am running have their own timestamp field. That seems to cause some issues. I will look into fluentd patches.
username_3: same error here, a index with 1970.01.01 is created and timestamps are the same :
"@timestamp" : "1970-01-01T00:33:39.644558916+00:00"
i can correctly inject a time_key with parsed time (which im using as time field):
<source>
@id fluentd-containers.log
@type tail
path /var/log/containers/stuff*.log
pos_file /var/log/containers.log.pos
tag raw.kubernetes.*
read_from_head true
<parse>
@type json
time_key time
time_format "%Y-%m-%dT%H:%M:%S.%N%Z"
</parse>
</source>
username_4: I have got the same issue, a index with 1970.01.01 is created with the same timestamp ```January 1st 1970, 01:33:39.000``` along with other index. Any help how to get only the updated time!
username_3: i solved my problem, since i had a filter parser that didnt set the time key, - and i guess overwrote the timestamp and set back to default, so either parse it once, or remember to set it again/later in subsequent "parses"
`
<source>
@id fluentd-containers.log
@type tail
path /var/log/containers/vault*.log
pos_file /var/log/containers.log.pos
tag kubernetes.*
read_from_head true
<parse>
@type json
time_format "%Y-%m-%dT%H:%M:%S.%N%Z"
</parse>
</source>
<filter kubernetes.**>
@id filter_parser
@type parser
key_name log
reserve_data true
remove_key_name_field true
<parse> # <---
@type json # <--- this section will overwrite timestamps if you already set it previously
time_format "%Y-%m-%dT%H:%M:%S.%N%Z" # <--- if needed you can parse it here instead
</parse> # <---
</filter>
`
username_4: It was the same issue in my case. Thanks!
username_4: @username_3 Thanks solved my issue.
username_1: To summarize....
You all havent used the default configs but used some override.
If so, i tend to close the issue.
username_0: @username_1 I'm revisiting this issue now. Installed without any overrides and these two services where one gets parsed correctly but the other does not.
2019-12-20T13:04:32.156Z vs 2019-12-20T12:50:30Z
January 1st 1970, 00:33:39.000
```json
{
"correlationId": "0a66d68a-e815-4de3-b374-e827f46b068e",
"languageId": "",
"level": "info",
"methodName": "AddToCache",
"msg": "Done adding item to cache, key: %!(EXTRA string=health:weather)",
"responseTime": "",
"time": "2019-12-20T12:50:30Z",
"version": ""
}
```
December 20th 2019, 13:04:32.156
```json
{
"name": "micro.booking-api",
"hostname": "booking-api-794d79fc97-qjzn5",
"pid": 27,
"level": "WARN",
"msg": "msg",
"time": "2019-12-20T13:04:32.156Z",
"v": 0
}
```
username_1: Could you guys test if this pr fixes the problem for you?
https://github.com/kiwigrid/helm-charts/pull/257
Status: Issue closed
|
ImagingDataCommons/IDC-WebApp | 687491054 | Title: Selection of the series will return the error message, "ViewerMain: Cannot read property 'SeriesInstanceUID' of undefined"
Question:
username_0: On the dev tier the selection of the series description, VOI PE Segmentation thresh=70 will return a, "ViewerMain: Cannot read property 'SeriesInstanceUID' of undefined." ISPY1 is the related collection.
<img width="960" alt="2020-08-27 (2)" src="https://user-images.githubusercontent.com/16824894/91486835-d2e37e80-e87a-11ea-9033-960c1d13486f.png">
<img width="960" alt="2020-08-27 (3)" src="https://user-images.githubusercontent.com/16824894/91487082-38d00600-e87b-11ea-8b37-8e6503fd4f10.png">
Answers:
username_0: Also, the collection qin_headneck seems return an error "ViewerMain: Referenced series for SEG dataset not present."
<img width="960" alt="2020-08-27 (6)" src="https://user-images.githubusercontent.com/16824894/91490054-d6c5cf80-e87f-11ea-96d4-1a7593c1c2f6.png">
<img width="960" alt="2020-08-27 (5)" src="https://user-images.githubusercontent.com/16824894/91490042-d3cadf00-e87f-11ea-9e67-da47ec24a8b2.png">
username_1: @username_2 let us know if you think this is an OHIF issue.
username_2: ...certainly sounds like it. All we do is send along a UID, we don't really DO anything else. @wlongabaugh This isn't possibly proxy-related, is it?
username_3: @username_0 in both cases you seem to be trying to open a SEG series, not study, which is not expected to work, and that is the ultimate source of the error. So the real bug is that IDC allows this to happen, not that it happens per se.
This is because you are trying to open a SEG series, and so OHIF does not have the reference since the URL ends with something like `1.3.6.1.4.1.14519.5.2.1.6279.6001.185810436275168701789786930141?SeriesInstanceUID=1.2.276.0.7230010.3.1.3.0.89838.1553284405.300965`. This is expected, and the right way to handle it would be to pass the reference series in the URL, but that's post MVP for sure. Next level down would be to not allow opening series that cannot be opened, but my guess is that is also post MVP.
username_3: For the MVP, we should turn off the button that corresponds to the series of types SEG or RTSTRUCT.
username_3: My preference is grayed out button with tooltip "Please open at the study level to see this series"
username_4: In progress.
Status: Issue closed
username_3: Selection of non-imaging series is disabled |
pulumi/pulumi | 734686696 | Title: Support Python 3.9
Question:
username_0: We need to make sure that pulumi works correctly on Python 3.9
Answers:
username_1: Looks like grpcio is still working through Windows support https://github.com/grpc/grpc/issues/24344
username_1: `protobuf` doesn't have python 3.9 support yet, so it’s missing the equivalent of the `_message.cpython-38-darwin.so` file that is needed for the `SetAllowOversizeProtos` method. However, it appears we only use that method to override a default that is only really applicable for a k8s bug, more context in [this issue](https://github.com/pulumi/pulumi/issues/3981).
However, I think we can wrap this in a try/except since it's not always needed, and not having the try/except is causing `Getting Started` workflows for folks using python 3.9 to fail.
Status: Issue closed
|
airblade/chartjs-ror | 415004498 | Title: New maintainer?
Question:
username_0: I don't use this plugin much any more and consequently struggle to put time into it. It doesn't need much time – just a little to stay on top of occasional issues and pull requests.
Please let me know if you would like to become a maintainer. |
bogde/HX711 | 1005499929 | Title: Cryptic message in serial monitor
Question:
username_0: I am clearly doing something wrong.
what could be the reason for me getting this in the serial monitor.
I do not get anything else.

Answers:
username_1: did you check baud rate?
username_0: I have the 9600 baud in the monitor.
I have tried different rates both higher and lower, but unfortunately I get the same error.
username_0: Ok, now I change the baud rate to 9600, in the code and the serial monitor.
Thanks, I made some progress but now i got the issues of HX711 not found.
Status: Issue closed
|
CakeDC/users | 289202628 | Title: Can not overwrite messages when session timeout.
Question:
username_0: We have been config session timeout [1 minute] under cake 3.x, after session timeout we can not overwrite message [CakeDC Users] shown screen login form.
This is Bug or missing config.
Thanks.
Answers:
username_1: hi @username_0 are you related with CakeDC? I don't remember you from our company channel :open_mouth:
Regarding your session issue, keep in mind the logged in user, and the flash messages are stored in the Session. Once the session expires this reference is lost and you can't access it from the following requests...
Status: Issue closed
|
cosmir/openmic-2018 | 354013616 | Title: Sparse label schema
Question:
username_0: Sparse label data (CSV) should consist of four columns:
`(sample_key: String, instrument: String, affinity: Float [0, 1], num_responses: int)`
Most recent label data is missing `num_responses`, so that'll have to be added.<issue_closed>
Status: Issue closed |
zhushaojun/web2019aut | 522679586 | Title: 2017040056 for week 11
Question:
username_0: * Github project url
https://github.com/username_0/spacecraft-web#space-science-website
* Live demo url
https://amaziahc.github.io/spacecraft-web/
* Description of what i have done
1. project plan
* i created my project and i did a presentation
2. creating pages
* i then created my project using HTML, CSS and bootstrap
3. github
* i published my work on github
* created read me file
Answers:
username_1: ok for week 11
Status: Issue closed
|
BowlOfSoup/NormalizerBundle | 689986059 | Title: Parent method should be checked when child method that does not contain annotations
Question:
username_0: This can happen when you're normalizing a Doctrine proxy.
Child:
```
class ProxySocial extends Social implements Proxy
{
/**
* {@inheritdoc}
*/
public function __load()
{
}
/**
* {@inheritdoc}
*/
public function __isInitialized()
{
}
/**
* {@inheritdoc}
*/
public function getFacebook()
{
return parent::getFacebook();
}
}
```
Parent, which is in case of Doctrine proxies the actual class with correct annotations:
```
class Social
{
/**
* @var string
*
* @Bos\Normalize(group={"default"})
*/
private $facebook;
/**
* @return string
*
* @Bos\Normalize(name="facebook", group={"via-method"})
*/
public function getFacebook()
{
return $this->facebook;
}
}
```
If we call the normalizer it will only check the child (proxy) method for annotations.
`$normalizer->normalize($proxySocial, 'via-method')`
We should support looking up the parent if 'inheritDoc' is used.<issue_closed>
Status: Issue closed |
lovell/sharp | 174119286 | Title: Accept svg string
Question:
username_0: Apparently sharp doesn't accept a SVG as string, it only loads the string from a file of given path.
It would be very helpful if sharp can also load the SVG as string.
Status: Issue closed
Answers:
username_1: Hello, string-based input is used for filesystem paths. I'm not convinced content sniffing is the way to go here given the simplicity of a `sharp(new Buffer(svg))...` wrapper.
username_2: @username_1 suggestion worked for me
```javascript
const png = await sharp(Buffer.from(svg)).png().toBuffer();
const image = await loadImage(png);
``` |
LeftyBC/weechancrypt | 68278965 | Title: Chunk messages that are larger than the IRC line length limit
Question:
username_0: We should be able to determine the IRC line length limit and chunk messages that, when encrypted+encoded, are larger than that limit.
Answers:
username_1: Just FYI, when implementing this in an IRC bot I wrote I learned the hard way that the limit for one user is different than another use. If the IRC server is following RFC 2812, it should allow 512 chars (subtract 2 for CRLF and you've got 510) per line. But a line includes prefix, hostname, ident, channel, etc. So it can be tricky to take all the diff variables into account. I think we settled on 200 after looking at the possible max values for all the other variables, but it's definitely possible to be smarter about it (though possibly painful.) |
limpkin/mooltiapp | 236309639 | Title: mooltiapp opens TCP port 30035 externally!
Question:
username_0: Hi,
I just figured out that the mooltiapp is opening a tcp port externally.
`nmap -sT -p - 192.168.1.2
Starting Nmap 7.01 ( https://nmap.org ) at 2017-06-15 22:18 CEST
Nmap scan report for 10.0.0.2
Host is up (0.000027s latency).
Not shown: 65533 closed ports
PORT STATE SERVICE
22/tcp open ssh
**30035/tcp open unknown**
`
`netstat -lntp
(Not all processes could be identified, non-owned process info
will not be shown, you would have to be root to see it all.)
Active Internet connections (only servers)
Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name
tcp 0 0 127.0.1.1:53 0.0.0.0:* LISTEN -
tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN -
tcp 0 0 127.0.0.1:15550 0.0.0.0:* LISTEN -
tcp 0 0 127.0.0.1:8200 0.0.0.0:* LISTEN 2956/Duplicati
tcp 0 0 127.0.0.1:8300 0.0.0.0:* LISTEN 9204/Duplicati
tcp6 0 0 :::22 :::* LISTEN -
**tcp6 0 0 :::30035 :::* LISTEN 22885/preload.js --**
`
After some short digging I found the line that does this.
[app/preload.js](https://github.com/username_1/mooltiapp/search?utf8=%E2%9C%93&q=30035&type=)
It is really necessary to bind this port to a any interface rather than localhost?
Cheers,
RandomIsResistance
Answers:
username_1: Hello there,
You are indeed right, it may be better practise to open it locally.
username_1: A PR is ready, I can't approve it until we're back from holidays.
On Jun 15, 2017 20:51, "<EMAIL>" <
<EMAIL>> wrote:
> Hello there,
>
> You are indeed right, it may be better practise to open it locally.
>
>
username_0: Wow,
thats a fast reaction! :+1:
Thanks for the support and enjoy the rest of your holiday!
username_1: 2.0.80 just pushed with that fix!
Status: Issue closed
|
NCPN/upland | 145198763 | Title: Seedlings Tally Buttons Don't Work Unless Zero is Entered First
Question:
username_0: Able to get the tally buttons to work on the seedlings form if I first enter a zero. I can probably add a note about this in the documentation and call it done, unless it later becomes a problem for the crew.<issue_closed>
Status: Issue closed |
zhanghang1989/MSG-Net | 216018274 | Title: Code is not released?
Question:
username_0: ?
Answers:
username_1: Thanks for being interested! I am working on the supplementary material for the submission. I will release the demo code with pre-trained models around this weekend. Please see our expected timeline https://github.com/username_1/MSG-Net/blob/master/README.md
username_1: We have released the code with pre-trained models. Thanks!
Status: Issue closed
|
sublimehq/sublime_text | 683406851 | Title: Don't merge repeated commands within 1s of each other in the jump list
Question:
username_0: Commands that are repeated within 1s of each other are merged into the jump list (see https://github.com/NeoVintageous/NeoVintageous/commit/0499e63bb3384b3f6d1e236789a9ac4bdc586ce6#r41644216), it can be very unexpected: when you jump back/forward you get a little lost because it's hard to know what jumps will have been merged, it's hard to keeping track of a 1s span of time.
### Preferred solution
Don't merge jumps.
For API commands maybe add various options:
```
view.run_command("add_jump_record", {"selection": [(r.a, r.b) for r in view.sel()], "allow_merge": True})
``` |
godotengine/godot-proposals | 578036057 | Title: CSG MultiMesh Node
Question:
username_0: **Describe the project you are working on:**
A 3d tile based strategy game that happens to use multi-meshes for tiling since gridmaps don't work for hexagonal layouts.
Pretty sure you can define csg meshes as tiles for gridmaps but in this case i am limited. **R.I.P. my project due to** [Godot not supporting hex gridmaps](https://github.com/godotengine/godot-proposals/issues/472). 😭
**Describe the problem or limitation you are having in your project:**
I'll have to write a script that removes my multimesh tile instances as they collide with my csg subtraction mesh object that was meant to clip the floors (and subsequently the multimesh floor tiles) of buildings down to the relevant floor so that it is more easily visible to the overhead camera.
This will not look as nice and not be as simple as the subtraction mesh being able to just clip everything.
**Describe the feature / enhancement and how it helps to overcome the problem or limitation:**
It's a multimesh node that works with constructive solid geometry.
**Describe how your proposal will work, with code, pseudocode, mockups, and/or diagrams:**
Not really necessary.
I suppose there could be an issue if the multimesh instances collide with themselves due to not allowing for csg meshes that have internal geometry, maybe have the multimesh instances merge with itself before being affected by other csg objects. Maybe this is why this doesn't exist.
**If this enhancement will not be used often, can it be worked around with a few lines of script?:**
Unsure but probably not.
**Is there a reason why this should be core and not an add-on in the asset library?:**
This seems like it would be a basic function that has been overlooked if there is no reason for this not to be viable.
Answers:
username_1: I'm not sure that this is even possible.
Are you looking for a single CSGMesh that would be instanced thousands of times? Or are you looking to apply CSG operations to the thousands of instances separately?
username_0: @username_1
The former, and yeah, I'm also skeptical if it's possible.
Depends on how the multi-mesh works.
username_2: Do you need csg operations at runtime or just to create the meshes? If the latter, there's a nice bit of script that lets you save csg as mesh, which then means you can use it with the regular multimesh...
username_0: @username_2 The csg operation has to run dynamically so the former. What you are saying isn't really relevant to the multimesh though, the in-engine fuctionality of the multi-mesh doesn't accommodate for setting it's instances at specific points so that also needs to be done at runtime anyways.
username_1: @username_0 I think what @username_2 is getting at is likely the only way that MultiMesh could be combined with CSG. The nature of CSG makes it extremely difficult to use with a MultiMesh. The CSG node needs full information about the Mesh and the transformation (position, scale, rotation). However, the benefit of the MultiMesh is that it uncouples the mesh from its transformation and lets you draw thousands of copies of the same mesh just by specifying individual transformations for each.
Accordingly, the only way the CSG node could work with a MultiMesh is if you took a snapshot of the CSG node at a given time and then used that as an input to the MultiMesh. To me it sounds like Zireal s suggesting that the needed functionality you proposed is already in-engine and you just need to add it to your game using a few lines of code.
username_0: I now see how that would work if i wanted all instances of the multimesh to be affected by a csg mesh simultaneously.
What would actually be needed for me would be the reverse of that: to save/"bake" the multimesh as a single mesh so that i could assign it to a csg mesh since i'd require the csg meshes to move through this mesh dynamically.
if there was a way to do that then i could work around it. Unfortunately the csg mesh is constantly moving back and forth through the tile mesh which that suggestion doesn't allow for.
username_1: @username_0 What you are suggesting is impossible. Every instance needs to use an identical mesh. Instancing is a very particular hardware-based technique that draws one mesh thousands of times in a single draw call.
What you need to do is bake thousands of instances of a mesh using your 3D DCC (e.g. blender) and then interact with that mesh using CSG nodes.
username_0: @username_1
Ah, i see.
Ok multi meshes are off the table but would I possibly be able to merge colliding csg meshes into one uniform mesh and also bake it **at runtime**? @username_2 stated that it was already possible to take a snapshot of a csg nodes and save it as a mesh but does that have to be in-engine? If i could do that during runtime and apply the resulting mesh to a single meshInstance node, then that would work!
This is going a different direction now probably gonna change this request entirely or close it and make another since multimeshes are not relevant to this anymore.
username_2: You can do it at runtime, yes: https://github.com/godotengine/godot/issues/19232#issuecomment-476459435
Status: Issue closed
|
AyaneYamassaki/library | 972855957 | Title: DESENVOLVER A TELA DE GERENCIAMENTO PARA USUÁRIOS
Question:
username_0: Desenvolvimento de tela genrecial do CRUD para USUÁRIOS
- [ ] list-usuario.html (Listagem de registro no banco dados, modal para adição de novos registros, modal de visualização, modal de edição e modal de deleção????)
- [ ] form-usuario.html<issue_closed>
Status: Issue closed |
dynverse/babelwhale | 398221321 | Title: Cannot pull container
Question:
username_0: Hi,
I tried to pull a singularity container using this function:
```
babelwhale::pull_container(container_id="dynverse/scorpius")
```
But, It throws me an error:
```
ERROR: Unknown command 'pull'
```
It seems that it is related to the `env = c(SINGULARITY_CACHEDIR = tempcache)` argument because
when I run:
```
processx::run(command = "singularity", args = c("pull", "--name", image_file, paste0("shub://",container_id)), echo = TRUE)
```
it does pull the container. Unfortunately, it does pull it in my working directory but doesn't save it automatically in my SINGULARITY_CACHEDIR environment variable.
Any idea why setting this env is causing this error?
This is my environment:
```
R version 3.5.1 (2018-07-02)
Platform: x86_64-pc-linux-gnu (64-bit)
Running under: CentOS Linux 7 (Core)
locale:
[1] LC_CTYPE=en_US.UTF-8 LC_NUMERIC=C LC_TIME=en_US.UTF-8 LC_COLLATE=en_US.UTF-8 LC_MONETARY=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 LC_PAPER=en_US.UTF-8
[8] LC_NAME=C LC_ADDRESS=C LC_TELEPHONE=C LC_MEASUREMENT=en_US.UTF-8 LC_IDENTIFICATION=C
```
Answers:
username_1: That's strange! @username_2 and I will both have limited connectivity for a week, but I will have a look at this asap.
username_1: What happens if you run the code in this file manually https://github.com/dynverse/babelwhale/blob/master/R/pull_image.R? Where does it get stuck?
username_0: It gets stuck here:
```
# pull container
processx::run(
command = "singularity",
args = c("pull", "--name", image_file, paste0("shub://", container_id)),
env = c("SINGULARITY_CACHEDIR" = tempcache),
echo = TRUE
)
```
username_1: Ha. Sorry for the long wait. It helps a lot to be at a computer to find out what is going wrong :)
All of the docker and singularity containers have a "ti_" prefix. E.g. dynverse/ti_scorpius, dynverse/ti_monocle_ddrtree. The following command should work immediately:
```R
babelwhale::pull_container(container_id="dynverse/ti_scorpius")
```
Could you confirm whether that works?
For the singularity containers, I'm aware the singularity slingshot container is not working (whereas the exact same docker container works fine) ─ I'm trying to figure out what the problem is.
Are you trying to manually run the containers?
Robrecht
Status: Issue closed
username_2: With the changes coming up with dynwrap v2, this is issue can be closed |
giantotter/giantotter_public | 128570651 | Title: Player sees two timers after other player rejoins
Question:
username_0: Waitress starts a game, customer starts a game and joins. After cue cards, scene loads correctly for both.
Waitress quits and restarts, while customer stays on. This forces both to go back to the cue cards (good), but when the game started, the customer had two timer UI elements drawing on top of one another.<issue_closed>
Status: Issue closed |
MicrosoftDocs/azure-docs | 1095578638 | Title: What is a "viral" azure AD account?
Question:
username_0: Can someone please clarify what exactly an unmanaged "viral" account is in Azure AD as stated in this document article?
There are no accounts of type "viral" in Azure AD but this document states the following which needs to be fixed.
"Starting November 1, 2021, we'll begin rolling out a change to turn on the email one-time passcode feature for all existing tenants and enable it by default for new tenants. As part of this change, Microsoft will stop creating new, unmanaged ("viral") Azure AD accounts and tenants during B2B collaboration invitation redemption."
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 83f1378d-b047-11d1-a138-5c52a3fee6ba
* Version Independent ID: 57c51f57-90ad-e0f9-1e6b-3262f94de02c
* Content: [One-time passcode authentication for B2B guest users - Azure AD](https://docs.microsoft.com/en-us/azure/active-directory/external-identities/one-time-passcode)
* Content Source: [articles/active-directory/external-identities/one-time-passcode.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/active-directory/external-identities/one-time-passcode.md)
* Service: **active-directory**
* Sub-service: **b2b**
* GitHub Login: @msmimart
* Microsoft Alias: **mimart**
Answers:
username_1: @username_0
Thanks for your feedback! We will investigate and update as appropriate.
username_1: @username_0
When it comes to the term "viral", it's associated with self-service sign-up, in which a user signs up for a cloud service and has an identity automatically created for them in Azure AD based on their email domain (i.e. yahoo.com, gmail, outlook, etc.).
The tenant that is created during the self-service sign-up is where the unmanaged ("viral") identity is created. The term unmanaged refers to an unmanaged tenant, since it's a tenant that has no global administrator.
**For more info:**
[What is self-service sign-up for Azure Active Directory?](https://docs.microsoft.com/en-us/azure/active-directory/enterprise-users/directory-self-service-signup)
[Take over an unmanaged directory as administrator in Azure Active Directory](https://docs.microsoft.com/en-us/azure/active-directory/enterprise-users/domains-admin-takeover)
[Office 365 Acounts of UserType Viral Appear After Registering Vanity Domain](https://answers.microsoft.com/en-us/msoffice/forum/all/office-365-acounts-of-usertype-viral-appear-after/5d0dc42d-c0a3-44a0-810b-180a61df4ed7#:~:text=%22Viral%22%20accounts%20are%20associated%20with%20the%20self-service%20offerings%2C,to%20disable%20their%20use%2C%20here%3A%20https%3A%2F%2Fazure.microsoft.com%2Fen-us%2Fdocumentation%2Farticles%2Factive-directory-self-service-signup%2F%20Report%20abuse)
I've created a PR to update the documentation and it's currently waiting for the author to review and sign off.
I will now close out this issue. Please allow some time for the author to review and the changes to reflect.
Thank you for your time and patience!
Status: Issue closed
|
ekolis/EventHorizonGui | 540030384 | Title: Port to .NET Core
Question:
username_0: Why? Because.
Answers:
username_0: This will require rewriting the Eggs module, or decompiling it and recompiling it; I seem to have lost the source code for that DLL.
username_0: .NET 5 is coming in November; this might be a moot point... just upgrade to that instead and it we'll get the best of both worlds! |
watchdogpolska/feder | 349747196 | Title: dawanie maili do wiadomości
Question:
username_0: Niektóre gminy odsyłają po informację do innych jednostek. Chciałabym aby wątek tematyczny był utrzymany, w związku z tym piszę do tych jednostek z adresu biuro, ale daję do wiadomości adres z fedrowania dotyczący danej sprawy. Niestety, nie przychodzi ta odpowiedź do tej sprawy. Nic się nie pojawia.
Zastanawia mnie po pierwsze czy tak ma być, a po drugie czy nie jest to jakiś większy problem. Może gminy wysyłają odpowiedzi, a one nie dochodzą? Zaznaczam, że nie ma żadnej zwrotki.
Pozdrawiam,
kasia
Answers:
username_1: Możesz podać dane sprawy?
username_0: tak:
<EMAIL>
<EMAIL>
<EMAIL>
<EMAIL> (tu jest błąd, nie dodałam sprawy)
Mogę jakoś tu dodać zrzuty ekranu?
username_1: Możesz dodać tu screenshoty w banalny sposób. Tak jak do Worda.

username_0: <img width="322" alt="wschowa" src="https://user-images.githubusercontent.com/13284861/43994456-655e5fee-9d9d-11e8-9516-506ffc83ac65.PNG">
<img width="785" alt="nowasol" src="https://user-images.githubusercontent.com/13284861/43994457-6c20731c-9d9d-11e8-96e5-49e9a9f04a3d.PNG">
<img width="419" alt="nowasol1" src="https://user-images.githubusercontent.com/13284861/43994459-722070a0-9d9d-11e8-9374-b4d3d6e25c30.PNG">
<img width="375" alt="gubin" src="https://user-images.githubusercontent.com/13284861/43994475-8e61aad6-9d9d-11e8-8064-584f4dbd5a8a.PNG">
username_1: Od momentu przeniesienia przychodzącej poczty fedrowania do Zenbox nie wykrywamy poprawnie wiadomości, gdzie występujemy w polu "Kopia", a nie "Do". Wcześniej wykrywaliśmy nawet umieszczenie nas w polu UDW, ale teraz serwer pocztowy nie dołącza tych informacji w sposób dla nas użyteczny.
Nierozpoznane wiadomości (w tym te) są dostępne na https://fedrowanie.siecobywatelska.pl/listy/wiadomosci/dziennik/?processed=2 do ręcznego przypisania.
username_1: Hmm...Podanie Ci tego linku do nieprzetworzonych wiadomości ujawniło pewien słaby punkt fedrowania, bo nieprzetworzone wiadomości są za każdym razem przetwarzane, gdy chodzisz po stronie. Dlatego są takie wolne. To dało impuls do tego, aby opracować inny mechanizm pobierania e-maili ( https://github.com/watchdogpolska/imap-to-webhook ), który będzie wydajniejszy.
Prawidłowo wykrywać nas w polu "Kopia" będzie działać po wprowadzeniu nowego mechanizmu.
Znalazłem także sposób na ponowne rozpoznawanie wiadomości, nawet jeżeli np. jesteśmy wyłącznie w polu BCC, ale wymaga to zmian w zewnętrznej bibliotece, na co możemy chwilę poczekać ( https://github.com/SpamScope/mail-parser/pull/38 ).
Status: Issue closed
|
ConnectSDK/Connect-SDK-iOS | 82026137 | Title: Make "+ (LaunchSession *) launchSessionFromJSONObject:(NSDictionary *)json" method visible
Question:
username_0: As specified in documentation: http://connectsdk.com/docs/ios/resuming-apps/
I need
```objective-c
+ (LaunchSession *) launchSessionFromJSONObject:(NSDictionary *)json
```
visible to restore webAppSession for background / foreground switch.
Thanks.
Answers:
username_1: Added Missing API https://github.com/ConnectSDK/Connect-SDK-iOS-Core/commit/0450d568938de1a2b7d9ea10b5f5b01acf06ed4c
Status: Issue closed
|
Subsets and Splits