repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
sentinel-official/hub | 1063330887 | Title: make: *** [Makefile:43: mod-vendor] Error 127
Question:
username_0: build failed on clean ubuntu 20.04 machine.
`/src/github.com/sentinel-official/hub# make install --debug
GNU Make 4.2.1
Built for x86_64-pc-linux-gnu
Copyright (C) 1988-2016 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html>
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law.
Reading makefiles...
Updating makefiles....
Updating goal targets....
File 'install' does not exist.
File 'mod-vendor' does not exist.
File 'tools' does not exist.
Must remake target 'tools'.
Successfully remade target file 'tools'.
Must remake target 'mod-vendor'.
/bin/sh: 1: modvendor: not found
make: *** [Makefile:43: mod-vendor] Error 127`
Answers:
username_1: try the code in #159. Honestly this looks to me like a possible issue with your go environment. |
jhennycervera/Clase8-Codeforces | 615980285 | Title: Error en test 19
Question:
username_0: El error en el test 19 es un espacio de diferencia entre la , y los ...:

Output:
, ...
Answer:
, ...<issue_closed>
Status: Issue closed |
smpallen99/coherence | 275102768 | Title: Cut new release with JSON/api support?
Question:
username_0: It'd be nice to have that in the hex packages. Depending on master isn't great but I do need the json behavior based stuff.
There's some stuff that needs to be added to the `coh.install` to get the api stuff working properly but it's not complex.
Answers:
username_1: Is there any update to this? Is the master good to use in production if not?
username_2: I believe the responder feature which has been in master for a while now should give you what you need. Please confirm.
Status: Issue closed
username_2: Feature is included in hex version 0.5.1 |
moby/moby | 258995495 | Title: docker daemon appears to be stuck, docker ps, docker info hanging
Question:
username_0: <!--
If you are reporting a new issue, make sure that we do not have any duplicates
already open. You can ensure this by searching the issue list for this
repository. If there is a duplicate, please close your issue and add a comment
to the existing issue instead.
If you suspect your issue is a bug, please edit your issue description to
include the BUG REPORT INFORMATION shown below. If you fail to provide this
information within 7 days, we cannot debug your issue and will close it. We
will, however, reopen it if you later provide the information.
For more information about reporting issues, see
https://github.com/docker/docker/blob/master/CONTRIBUTING.md#reporting-other-issues
---------------------------------------------------
GENERAL SUPPORT INFORMATION
---------------------------------------------------
The GitHub issue tracker is for bug reports and feature requests.
General support can be found at the following locations:
- Docker Support Forums - https://forums.docker.com
- IRC - irc.freenode.net #docker channel
- Post a question on StackOverflow, using the Docker tag
---------------------------------------------------
BUG REPORT INFORMATION
---------------------------------------------------
Use the commands below to provide key information from your environment:
You do NOT have to include this information if this is a FEATURE REQUEST
-->
Docker daemon appears to be stuck.
Docker commands such as docker info, docker ps, docker run, docker logs all appear to get stuck.
<!--
Briefly describe the problem you are having in a few paragraphs.
-->
**Steps to reproduce the issue:**
We launch customer configured docker images, the docker daemon appears to get stuck occasionally.
**Describe the results you received:**
**Describe the results you expected:**
**Additional information you deem important (e.g. issue happens only occasionally):**
Issue happens occasionally.
**Output of `docker version`:**
```
Client:
Version: 17.03.2-ce
API version: 1.27
Go version: go1.7.5
Git commit: f<PASSWORD>
Built: Tue Jun 27 03:35:14 2017
[Truncated]
```
Command gets stuck.
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
Azure service running containers.
Detailed logs:
journalctl -u docker.service
http://jlawnotes.azurewebsites.net/daemonlogs.txt
utput of kill -SIGUSR1 on docker daemon
data:
http://jlawnotes.azurewebsites.net/daemon-data_log.txt
output of kill -SIGUSR1 on docker daemon
stacks:
http://jlawnotes.azurewebsites.net/goroutine_stacks_log.txt
The issue appears similar to https://github.com/moby/moby/issues/30437, but I checked the call stacks in the output to the one referenced, and feel that it is different. |
phpdocker-io/phpdocker.io | 393434146 | Title: PHP 7.3 support
Question:
username_0: This is to track progress of php 7.3 support:
[] CLI base image
[] FPM base image
[] CI setup
[] Docker hub setup
[] Website support
Status: Issue closed
Answers:
username_1: Not working:
[error] 9#9: *6 recv() failed (104: Connection reset by peer) while reading response header from upstream, client: 172.23.0.1, server: , request: "GET /favicon.ico HTTP/1.1", upstream: "fastcgi://172.23.0.2:9000",
username_0: So what seems to be happening here is that the favicon request is being passed on to php-fpm and it's perhaps crashing, if trying to execute the file that's there. The nginx config isn't quite right. Which option did you choose on the "framework" select?
username_2: @username_0 same problem here, i choose the symfony, laravel... Option on the framework.
502 Bad Gateway returned.
username_0: Do you guys have a repository I can check out?
username_2: Yes, check this repo: https://github.com/username_2/phpdocker-sf4-nginx
username_0: Thank you. Hmm this is interesting:
```
carry-php-fpm | [28-Jan-2019 13:17:09] WARNING: [pool www] child 12 exited on signal 11 (SIGSEGV - core dumped) after 27.732746 seconds from start
carry-php-fpm | [28-Jan-2019 13:17:09] NOTICE: [pool www] child 14 started
```
I wonder if it's a dodgy extension.
username_0: You can track https://github.com/phpdocker-io/phpdocker.io/issues/180 instead
username_2: Of course, I removed the xdebug and tested, worked fine. Thanks @username_0
username_0: no problemo, thank you for your help. If you subscribe to #180 above you'll know once you can re-enable xdebug in your project. |
kogg/InstantLogoSearch | 185055988 | Title: TypeError: global.localStorage is null
Question:
username_0: View details in Rollbar: [https://rollbar.com/instantlogosearch/instantlogosearch/items/184/](https://rollbar.com/instantlogosearch/instantlogosearch/items/184/)
```
TypeError: global.localStorage is null
File "http://instantlogosearch.com/js/main.js", line 37, in reduceToStorage/<
File "http://instantlogosearch.com/js/main.js", line 35, in handleAction/<
File "http://instantlogosearch.com/js/main.js", line 5, in reduceReducers/</<
File "http://instantlogosearch.com/js/main.js", line 5, in reduceReducers/<
File "http://instantlogosearch.com/js/main.js", line 35, in handleActions/<
File "http://instantlogosearch.com/js/main.js", line 35, in combineReducers/<
File "http://instantlogosearch.com/js/main.js", line 35, in dispatch
File "http://instantlogosearch.com/js/main.js", line 35, in createThunkMiddleware/</</<
File "http://instantlogosearch.com/js/main.js", line 1, in [10]</module.exports<.downloadedLogo
File "http://instantlogosearch.com/js/main.js", line 37, in [651]</</executeBound
File "http://instantlogosearch.com/js/main.js", line 37, in [651]</</_.partial/bound
File "http://instantlogosearch.com/js/main.js", line 31, in invokeGuardedCallback
File "http://instantlogosearch.com/js/main.js", line 28, in executeDispatch
File "http://instantlogosearch.com/js/main.js", line 28, in executeDispatchesInOrder
File "http://instantlogosearch.com/js/main.js", line 28, in [480]</executeDispatchesAndRelease
File "http://instantlogosearch.com/js/main.js", line 28, in [480]</executeDispatchesAndReleaseTopLevel
File "http://instantlogosearch.com/js/main.js", line 34, in [580]</forEachAccumulated
File "http://instantlogosearch.com/js/main.js", line 28, in [480]</EventPluginHub.processEventQueue
File "http://instantlogosearch.com/js/main.js", line 31, in runEventQueueInBatch
File "http://instantlogosearch.com/js/main.js", line 31, in [527]</ReactEventEmitterMixin.handleTopLevel
File "http://instantlogosearch.com/js/main.js", line 31, in handleTopLevelImpl
File "http://instantlogosearch.com/js/main.js", line 34, in [570]</Mixin.perform
File "http://instantlogosearch.com/js/main.js", line 31, in [521]</ReactDefaultBatchingStrategy.batchedUpdates
File "http://instantlogosearch.com/js/main.js", line 33, in batchedUpdates
File "http://instantlogosearch.com/js/main.js", line 31, in [528]</ReactEventListener.dispatchEvent
File "http://instantlogosearch.com/js/main.js", line 36, in [650]</</</x.wrap/e._wrapped
```<issue_closed>
Status: Issue closed |
decrypto-org/spider | 321893055 | Title: Large numbers of paths on the same host leads to OOM condition
Question:
username_0: We found a website that lists every single bitcoin transaction, block, and additional information for those, that ever occurred. This leads to a case, where our randomizing of the entries is not working anymore as a load distributor. The queuedRequests buffer then fills up rapidly with entries that are taken from the pool and added to the queuedRequests buffer, in order to keep them for later requests. Since in this particular case, we encountered that about 3/4 of the stored paths in the database were from this host, after a few iterations, almost the complete database was held in memory.
Suggestion: We change the way we handle the case that all concurrent requests to a host are already in use: We change the request to the database such that we exclude all the baseurls that have a max number of concurrent requests running. |
summa-tx/relays | 572344548 | Title: prep event dispatch for requests
Question:
username_0: The goal here is to allow some routing of info about validated proofs. We're planning on serving local chains as well as remote ones over IBC.
So the current thing is to add `Origin` and `Action` parameters to the request data structure
1. `Origin` should be an integer enum in the types folder. Types should be `local` or `remote`. CLI and REST instantiation should ALWAYS use `local`.
2. `Action` should be `HexBytes`. Limited to 500 bytes (for now).
Future:
When a proof is validated, if its origin is remote, dispatch the TX and the `Action` data to a handler module.<issue_closed>
Status: Issue closed |
leapdao/nervos | 741658723 | Title: Collect deposit lumos function doesn't find cell due to lagging indexer
Question:
username_0: ## Reproduce Steps
<!-- A list of specific things which should be done to deliver the bounty. These could be seen as requirements to verify/review bounty against -->
- Made a deposit
- Increased the sleep time to 10 minutes
- I see that the deposit is recorded as second transaction in block explorer
- Called `await client.getDeposits();()`
- Print the cell which is an empty array
- Pass empty array to collectDeposits
- Query bridge state but capacity is unchanged. i.e does not reflect recent deposit
## Deliverables
<!-- Artifacts produced as the result of this bounty. Something that could be verified/reviewed. Some examples: updated code, deployment made, blog post published, public event conducted etc -->
- `await client.getDeposits();` Should return the deposit cell to pass on to collectDeposits
- bridge state capacity reflects new balance
*Might be fixed with event emitter #64
Answers:
username_1: Please try again in #74. Works for me.
username_1: Actually it was a legit bug (only happens when u deploy bridge, and not set it in config). Pushing fix in #74 .
Status: Issue closed
|
microsoft/BotFramework-Composer | 560025707 | Title: Visuals | New icon for error
Question:
username_0: When there are errors in my expressions, a x (inside a circle) shows up. in order to see what the errors are, I need to click the x.
Normally, clicking an X results in closing a menu or deleting something, which makes me hestitant to click it, as I don't want it to go away (get deleted) before seeing what the errors are. Can we please use a bell or something that is typically used for notifcations? PVA uses this too.
For the page with the error list, is there another, more common icon we can use instead of the paint swatch?
<img width="318" alt="Screen Shot 2020-02-04 at 2 44 36 PM" src="https://user-images.githubusercontent.com/48503434/73794413-4007b900-475d-11ea-98c0-403529ef20f8.png">
Answers:
username_1: This will ship with the updated diagnostics panel (and we've added a Bell icon)
Status: Issue closed
|
StompMarket/helpdesk | 532799958 | Title: Error Reported By |<EMAIL>
Question:
username_0: Partner : F_PAYTM, User : <EMAIL>
Problem : Error : Http failure response for http://localhost:9000/print/pdf?auth=<KEY>cachebuster=1575477058385: 500 Internal Server Error
Message : undefined
StatusText : undefined
User Comment : undefined
Router Link : /fulfillment/fcs/288/packstations/batchpackages/1156527/orderdetail/packageId/1795814 |
ProjectComputerEngine/ApplicationRiceShopping | 813065529 | Title: Dart : HTTP request json data on wed server
Question:
username_0: ```
Future<List<dynamic>> readMessage(String messageID) async {
String url = "http://" + SERVER + "/Inbox/inbox.php?messageID=MS00000000";
var response = await get(url);
if (response.statusCode == 200) {
if (response.body.isNotEmpty) {
var data = jsonDecode(response.body);
return data;
} else {
return null;
}
} else {
return null;
}
}
``` |
zhxnlai/ZLSwipeableViewSwift | 297773746 | Title: Invalid card positioning
Question:
username_0: Steps to reproduce:
1. Launch demo.
2. Select "Default".
3. Drag a card around a bit http://take.ms/fhPQH.
4. Rotate device or simulator.
Issue:
Card is misplaced http://take.ms/7wHWW.
Answers:
username_1: 😬 doesn't seem great.
Can you try calling layoutSubviews after the device rotates to see if that fixes the issue? We'd be very open to accepting a PR that fixes the behavior.
Thanks for the report 😄
username_2: Hi, I had the same problem.
Added a pull request with the fix: https://github.com/zhxnlai/ZLSwipeableViewSwift/pull/129 |
intel/tinycbor | 387912013 | Title: Enhancement: add encoding/decoding half precision floating point functions to public API
Question:
username_0: It's uncommon to operate half precision floating point data on application level, while such format is widely used on transport level to minimize traffic. In such case one can need API for encoding/decoding half-precision floating point data AS more common single or double. Something like this:
```c
CborError cbor_encode_float_as_half_float(CborEncoder *encoder, float value);
CborError cbor_encode_double_as_half_float(CborEncoder *encoder, double value);
CborError cbor_value_get_half_float_as_float(const CborValue *value, float *result);
CborError cbor_value_get_half_float_as_double(const CborValue *value, double *result);
// etc.
```
There is some related stuff in the private area, and I think, I can make it public.
Answers:
username_0: [Partially implemented](https://github.com/username_0/tinycbor/commit/62b22ee4781a815c1b991a8215aa32587b6acebe). Is this OK. or some amendment is needed?
username_1: Yes, it looks very good. I have one overall comment and two very minor. Overall, I think the `float` API has more value than `double`. Any system capable of double-precision either already has a library function to convert single to double or the FPU implements it. But there are systems without `double` support, or at least more costly than `float`. So let's explore changing the internal function and the API to return `float` and receive `float`.
Thanks for the tests, they look nice.
The two minor comments:
1. Name the two files `cborparser_float.c` and `cborenccoder_float.c` please
2. Add the two files to tst_cpp.
username_0: [Look at the improvement made so far](https://github.com/intel/tinycbor/compare/master...username_0:get-half-as). But I'm in doubt now: I have used ldexpf function, which is not accessible in plain old C89 and some antique compilers.
username_1: Let's do bit manipulation directly. I wonder if there's a `#define` that helps us know that FP is IEEE 754. C++ has that.
username_0: ```cpp
static inline float decode_halff(unsigned short half)
{
int exp = (half >> 10) & 0x1f;
int mant = half & 0x3ff;
float mantf, expf, val;
if (exp == 0) {
mantf = mant;
expf = 1.0f / (1 << 24);
val = mantf * expf;
} else if (exp != 31) {
mantf = mant + 1024.0f;
expf = exp >= 25 ? 1 << (exp - 25) : 1.0f / (1 << (25 - exp));
val = mantf * expf;
} else {
val = mant == 0 ? INFINITY : NAN;
}
return half & 0x8000 ? -val : val;
}
```
username_1: If that produces proper results, that's fine. I was thinking of converting the `uint16_t` to `uint32_t` first, then `memcpy` from that to a `float`, instead of doing FP calculations. If we're going to do math, I'd rather use `ldexpf`, which according to the standard is part of C99.
username_0: So, there are 4 ways with their pro-s and contra-s:
1. use ldexp: standard-compliant, works on most platforms, but can be expensive;
1. use ldexpf: needs the C99 compiler, can be less expensive (but still expensive);
1. use naive math: works without std math lib (yes, sometimes linking without -lm option saves a lot of space), but error-prone and still expensive;
1. direct bit manipulations: re-shifting mantissa, re-biasing exponent... Error-prone again, cheap for platforms without hardware floating-point math. But what's about platforms with non-IEEE-754 math? Well. we can end with dropping float and double from API at all:
```c
#ifndef CBOR_NO_PLATFORM_IEEE754_FLOATING_POINT
CBOR_INLINE_API CborError cbor_value_get_float(CborEncoder *encoder, float *value);
CBOR_INLINE_API CborError cbor_encode_float(CborEncoder *encoder, float value);
// ... for half and double
#else
// Caveat lector: IEEE754 floating point representation is assumed
CBOR_INLINE_API CborError cbor_value_get_float(CborEncoder *encoder, uint32_t *value);
CBOR_INLINE_API CborError cbor_encode_float(CborEncoder *encoder, uint32_t value);
// ... for half and double
#endif
```
What is the preferable way?
username_1: CBOR is defined as carrying IEEE754 binary64, binary32 and binary16 content. So I like your idea.
I don't think we need to provide the `uintXX_t` version of the APIs for non-IEEE754 systems, though. Let people use `cbor_encode_floating_point` and `cbor_value_get_floating_point` (the latter does not exist), which operate on `void *`.
username_0: OK, but I think, this is the task for separate feature request. As for now, let's roll back to plain old doubles (ldexp, not ldexpf) both for cbor_value_get_half_float_as_float and cbor_value_get_half_float_as_double.
username_0: Well, cbor_value_get_half_float_as_float needs a pointer to float, so user should do extra effort. Hope, nobody could try something like:
```c
double x;
/* Oops... I know a guy, who can try it without any doubt */
cbor_value_get_half_float_as_float(cbor_value, (float*)&x);
```
But in name of API brevity `..._as_double` should be eliminated.
username_1: If someone wants to shoot themselves in the foot, who are we to say no? :-)
username_0: [Implemented](https://github.com/intel/tinycbor/compare/master...username_0:get-half-as). I'll squash the branch into the single commit before pull request.
username_2: I hope you find this useful. I'm finding it helpful to view CBOR as a binary format construction kit rather than a specific binary format because a lot of decisions are left to protocol designers.
username_1: Hello @username_2
we have the code to encode and decode half floats. I just didn't want to add an API that uses that.
username_2: @username_1 Oops! Deleted. Thanks for letting me know. |
rubenspgcavalcante/leaflet-ant-path | 201041613 | Title: Installation using npm
Question:
username_0: Hi,
I just tried to install this plugin using 'npm install' command.
When I test this line : 'var ant = require('leaflet-ant-path')' the build failed cause the module is not found.
After investigation :
The npm command install a built leaflet-ant-path.js in the dist directory
The package.json refer to "main":"src/plugin/main.js" which is not existing in the npm version.
At this time, i don't know if it's an issue on your side or just something that i don't understand :)
--
To make it work for testing, i replaced "main":"src/plugin/main.js" with "main": "dist/leaflet-ant-path.js" in package.json
--
Etienne<issue_closed>
Status: Issue closed |
agentcooper/react-pdf-highlighter | 579282606 | Title: Accessing the PDF content
Question:
username_0: Is there a way to send the PSD (with highlights) as base64 to a server, for example after a highlight is added or.
I don't care if the highlights are merged and won't be able to alter them after save, but so far this is the tool that would help me in my project and seems to work perfectly. Saving is the only thing i couldn't spot in your docs/code (I might be blind though)
Answers:
username_1: I didn't explore base64, but I did [some prototyping to merge the highlights into the PDF](https://github.com/username_1/pdf-annotation-service) itself so you can view them with any viewer. Maybe from there you can go to base64.
username_2: @username_0 My requirement is also the same, Have you figured out how to do it. If so please share. |
cyberark/secretless-broker | 651732411 | Title: Generic HTTP Config supports OAuth1
Question:
username_0: ## Is your feature request related to a problem? Please describe.
Currently Secretless Broker does not support any of the 3 methods of authenticating with OAuth1
## Problems with OAuth 1.0a and Secretless
OAuth 1.0a requires 7 different parameters to authenticate:
```
Authorization: oauth_consumer_key="<KEY>",
oauth_token="<KEY>",
oauth_signature_method="HMAC-SHA1",
oauth_signature="wOJIO9A2W5mFwDgiDvZbTSMK%2FPY%3D",
oauth_timestamp="137131200",
oauth_nonce="4572616e48616d6d65724c61686176",
oauth_version="1.0"
```
`consumer_key` - Constant secret
` nonce` - Dynamic, created specifically for request
`signature` - Dynamic, [generated](https://developer.twitter.com/en/docs/basics/authentication/oauth-1-0a/creating-a-signature) from all other parameters
`timestamp` - Dynamic, timestamp of request and will be denied if too old when request is sent
`token` - Constant secret
`signature_method` - Dynamic
`version` - Dynamic
There are [3 ways](https://oauth.net/core/1.0/#consumer_req_param) to pass these parameters in the HTTP request:
```
1. In the HTTP Authorization header as defined in OAuth HTTP Authorization Scheme.
2. As the HTTP POST request body with a content-type of application/x-www-form-urlencoded.
3. Added to the URLs in the query part (as defined by RFC3986 section 3).
```
### Method 1
Only 2 of the 7 parameters are constant secrets for Secretless to inject, the rest need to be created by the application when making the request.
The main problem with OAuth 1.0a and Secretless is that Secretless currently replaces the whole `Authorization` line in the header, with no way to "save" the other parameters that are passed.
Example:
Request:
```
http_proxy=localhost:8071 curl -k --request POST \
--url 'test:4443' \
--header 'authorization: OAuth oauth_consumer_key="CONSUMER_API_KEY", oauth_nonce="OAUTH_NONCE", oauth_signature="OAUTH_SIGNATURE", oauth_signature_method="HMAC-SHA1", oauth_timestamp="OAUTH_TIMESTAMP", oauth_token="ACCESS_TOKEN", oauth_version="1.0"' \
```
Secretless generic config:
```
...
config:
headers:
Authorization: ApiKey test
forceSSL: true
authenticateURLsMatching:
- ^http[s]*
```
Request sent to server:
```
Host: test:4443
User-Agent: curl/7.64.1
[Truncated]
Accept-Encoding: gzip
```
I couldn't find a way to inject the secrets that the request needs for authentication and keep the other parameters at the same time.
### Method 2
Currently modifications to request bodies are [not supported by Secretless](https://github.com/cyberark/secretless-broker/tree/master/internal/plugin/connectors/http/generic#limitations).
### Method 3
This is not currently supported either
## Recommended Solution
I think the best solution to this, would be to add a feature that allows Secretless to inject secrets into the header, while keeping the other parameters in the authorization line.
## Additional context
Stemmed from: https://github.com/cyberark/secretless-broker/issues/1264
Answers:
username_1: @username_0 after reading that I have what may end up being a silly question - but can you have your secretless config hardcode the values in the string here: `OAuth oauth_consumer_key="CONSUMER_API_KEY", oauth_nonce="OAUTH_NONCE", oauth_signature="OAUTH_SIGNATURE", oauth_signature_method="HMAC-SHA1", oauth_timestamp="OAUTH_TIMESTAMP", oauth_token="ACCESS_TOKEN", oauth_version="1.0"` or get them from the env instead of from a secret store, and then set up the header to include all of them?
ie if the header is supposed to be `Authorization: {string from above}` can you have secretless config be
```
version: 2
services:
generic-oauth1:
connector: generic_http
listenOn: tcp://0.0.0.0:8071
credentials:
consumer_key:
from: keychain
get: service#generic/consumer-key
token:
from: keychain
get: service#generic/token
nonce:
from: environment
get: OAUTH_NONCE
signature:
from: environment
get: OAUTH_SIGNATURE
signature_method: HMAC-SHA1
timestamp:
from: environment
get: OAUTH_TIMESTAMP
config:
headers:
Authorization: OAuth oauth_consumer_key="{{ .consumer_key }}", oauth_nonce="{{ .nonce }}", oauth_signature="{{ .signature }}", oauth_signature_method="{{ .signature_method }}", oauth_timestamp="{{ .timestamp }}", oauth_token="{{ .token }}", oauth_version="1.0"
forceSSL: true
authenticateURLsMatching:
- ^http[s]*
```
I’m not sure the specific syntax of the Authorization line should work, but either something very much like it should or we should enhance the generic connector. As possible enhancements, we could define functions for the `nonce`, `timestamp`, and `hash_hmac` signature hashing function that [twitter uses](https://developer.twitter.com/en/docs/basics/authentication/oauth-1-0a/creating-a-signature).
**Note**: updating to add that the signature requires info on the token, so Secretless would almost certainly have to compute this in order to get this to work.
username_2: ## Example Config
```
version: 2
services:
a-service-name:
connector: generic_http
listenOn: tcp://0.0.0.0:8021
credentials:
consumer_key:
from: keychain
get: service#generic/consumer-key
consumer_secret:
from: keychain
get: service#generic/consumer-key-secret
token:
from: keychain
get: service#generic/token
token_secret:
from: keychain
get: service#generic/token-secret
config:
headers:
# if oauth1Params exists in the config, we will add the Authorization header programatically.
Authorization: "THIS WILL BE WRITTEN OVER”
oauth1Params:
oauth_consumer_key: {{ .consumer_key }}
oauth_consumer_secret: {{ .consumer_secret }}
oauth_token: {{ .token }}
oauth_token_secret: {{ .token_secret }}”
forceSSL: true
authenticateURLsMatching:
- ^http[s]*
```
### To-Do (edited)
Status: Issue closed
username_1: I'm going to leave this open until we create a new Secretless tag (hopefully soon). For now, these changes are merged to master.
username_1: ## Is your feature request related to a problem? Please describe.
Currently Secretless Broker does not support any of the 3 methods of authenticating with OAuth1
## Problems with OAuth 1.0a and Secretless
OAuth 1.0a requires 7 different parameters to authenticate:
```
Authorization: oauth_consumer_key="0<KEY>",
oauth_token="<KEY>",
oauth_signature_method="HMAC-SHA1",
oauth_signature="wOJIO9A2W5mFwDgiDvZbTSMK%2FPY%3D",
oauth_timestamp="137131200",
oauth_nonce="4572616e48616d6d65724c61686176",
oauth_version="1.0"
```
`consumer_key` - Constant secret
` nonce` - Dynamic, created specifically for request
`signature` - Dynamic, [generated](https://developer.twitter.com/en/docs/basics/authentication/oauth-1-0a/creating-a-signature) from all other parameters
`timestamp` - Dynamic, timestamp of request and will be denied if too old when request is sent
`token` - Constant secret
`signature_method` - Dynamic
`version` - Dynamic
There are [3 ways](https://oauth.net/core/1.0/#consumer_req_param) to pass these parameters in the HTTP request:
```
1. In the HTTP Authorization header as defined in OAuth HTTP Authorization Scheme.
2. As the HTTP POST request body with a content-type of application/x-www-form-urlencoded.
3. Added to the URLs in the query part (as defined by RFC3986 section 3).
```
### Method 1
Only 2 of the 7 parameters are constant secrets for Secretless to inject, the rest need to be created by the application when making the request.
The main problem with OAuth 1.0a and Secretless is that Secretless currently replaces the whole `Authorization` line in the header, with no way to "save" the other parameters that are passed.
Example:
Request:
```
http_proxy=localhost:8071 curl -k --request POST \
--url 'test:4443' \
--header 'authorization: OAuth oauth_consumer_key="CONSUMER_API_KEY", oauth_nonce="OAUTH_NONCE", oauth_signature="OAUTH_SIGNATURE", oauth_signature_method="HMAC-SHA1", oauth_timestamp="OAUTH_TIMESTAMP", oauth_token="<PASSWORD>_<PASSWORD>", oauth_version="1.0"' \
```
Secretless generic config:
```
...
config:
headers:
Authorization: ApiKey test
forceSSL: true
authenticateURLsMatching:
- ^http[s]*
```
Request sent to server:
```
Host: test:4443
User-Agent: curl/7.64.1
[Truncated]
I couldn't find a way to inject the secrets that the request needs for authentication and keep the other parameters at the same time.
### Method 2
Currently modifications to request bodies are [not supported by Secretless](https://github.com/cyberark/secretless-broker/tree/master/internal/plugin/connectors/http/generic#limitations).
### Method 3
This is not currently supported either
## Recommended Solution
I think the best solution to this, would be to add a feature that allows Secretless to inject secrets into the header, while keeping the other parameters in the authorization line.
EDIT: This won't work for OAuth_signatures that use the tokens that Secretless keeps in the OAuth_signature base string. The best solution might be adding the ability to create all of the parameters needed for OAuth1 in Secretless and add the ability to specify the site specific signature format in the generic config
## Additional context
Stemmed from: https://github.com/cyberark/secretless-broker/issues/1264
Status: Issue closed
|
cfug/flutter.cn | 432800257 | Title: #81 - Ephemeral vs app state - 短时 (ephemeral) 和共享 (app) 状态
Question:
username_0: Markdown 文件路径:https://github.com/cfug/flutter.cn/tree/master/src/docs/development/data-and-backend/state-mgmt/ephemeral-vs-app.md
英文文档预览:https://debug.flutter.cn/docs/development/data-and-backend/state-mgmt/ephemeral-vs-app
建议完成时间:5 - 7 天
Answers:
username_0: assign to @tyisme614, thanks for the time.
username_1: 翻译已登记!提前感谢 @tyisme614 的帮忙~
username_1: 本篇翻译重新开放认领~
username_2: I do it
username_3: @username_2 抱歉,我们的发布时间临近,这篇文章的的交稿和 review 时间不够了,是否可以请帮忙到 PR 页面为已提交稿件作出评论和改进意见?
username_2: @username_3 好的。我就校对一下 https://github.com/cfug/flutter.cn/issues/167 这个吧
username_2: @username_3 校对了https://github.com/cfug/flutter.cn/pull/323这个,167有其他道友校对过了。
username_2: @username_3 167 有道友校对过,对 [https://github.com/cfug/flutter.cn/pull/323](url) 校对了一下。
username_3: 该任务已经完成,Issue closed, thanks!
Status: Issue closed
|
genesis-community/jumpbox-genesis-kit | 370751069 | Title: Add updated shield CLI to jumpbox
Question:
username_0: Using shield CLI is easier with more modern shield CLI versions like v8.x. I am not sure what version is in Jumpbox as there is no documented version argument but newer one would be cool.
Answers:
username_1: 8.0.17 of SHIELD should have the ability to just download the right binary from the SHIELD core itself. At that point, we're better off removing `shield` from the Jumpbox altogether.
Status: Issue closed
username_2: Resolved in https://github.com/genesis-community/jumpbox-genesis-kit/releases/tag/v1.0.3 |
SilasReinagel/oo.core.dotnet | 205653536 | Title: Design Basic Types
Question:
username_0: Flag/Boolean does seem very important. What is the best human name for something like that?
WholeNumber can be a type of Number.
In what scenario would someone accidentally use Text without meaning to?
Answers:
username_1: The other basic type we will need is Flag, otherwise I like those (I have decided I can't think of a single use case where you would just need Char, even if you wanted to enforce a single character Text, you could just make an implementation of Text that did that)
We should just do Number, and RealNumber will be an implementation of Number
We should not enforce the printers no getters concept, as it becomes complicated when you have unhappy path possibilities.
I can't think of a better name than Go() for generic commands, it's intuitive
We have to be careful with an abstract class like Text, because people could very easily accidentally use a class as Text when they were not meaning to
username_0: Flag/Boolean does seem very important. What is the best human name for something like that?
WholeNumber can be a type of Number.
In what scenario would someone accidentally use Text without meaning to?
username_0: What about Condition instead of Boolean/Flag?
username_1: Condition is excellent name
username_1: I also think that number should not expose AsInt() upon reflecting you only use numbers with other numbers but right when you go to consume you either pull its value use it implicitly or you can do new FloorNumber(number).AsReal()
username_0: Another important basic type will be some sort of DateTime object.
username_1: some form of TimeSpan also seems important now that you mention it.
username_0: So, it seems clear that Number will require some more initial design.
Also, I think that `TimeSpan` and `DateTimeOffset` are good enough that we can use them as default types for now. Maybe add some smart objects as needed.
username_1: That sounds like a good idea, eventually we can build a smart object that hides the static functionality that exists in them
Status: Issue closed
|
WeAreFairphone/modem_zip_generator | 382002773 | Title: Makefile doesn't work anymore since 18.10.0
Question:
username_0: Even after fitting in the new filename scheme, version and checksum the makefile halts in line 93, probably because of a changed ota-zip file topology. Haven't figured out what exactly happened (late night on a Sunday). The manual zip for Fairphone Open hasn't changed regarding its structure, however. After inspecting the relevant image files updating the modem_zip-generator should be worth it.
Answers:
username_1: Created a PR - #12
Status: Issue closed
|
Azure/azure-cosmos-dotnet-v2 | 511876133 | Title: HTTP 410 InvalidPartition each time emulator is restarted
Question:
username_0: **Describe the bug**
Each time I restart the Cosmos emulator I get an error when I perform a query and I need to delete the database to resolve it and do my work.
The HTTP code is 410 and the message on the StorageException is "InvalidPartition".
**To Reproduce**
- Start the local Cosmos DB emu with Tables API enabled.
- Use your app locally so it writes data to local Cosmos DB.
- Stop the emu, wait for it to close, then start it again as before.
- Use your app again and observe a long pause before a StorageException: InvalidPartition is thrown.
**Expected behavior**
Before yesterday, restarting the emulator was not a problem.
**Actual behavior**
Restarting the emu means having to delete my database and let my app code auto-generate it all again.
**Environment summary**
SDK Version: VS 16.3.6 and Cosmos Emulator 2.4.3.0 (50x45d91)
OS Version (e.g. Windows, Linux, MacOSX): Windows 10
**Additional context**
Yesterday, after rebooting to update Visual Studio to 16.3.6 I ran my app and received an HTTP 410 from the local Cosmos DB emulator over the Table API port.
I looked on Google and nothing, so I deleted the database in Storage Explorer and it recreated and all was okay again.
Today I reboot for a Windows Update and the same thing has happened.
exception.Message = InvalidPartition
exception.StackTrace =
at Microsoft.Azure.Cosmos.Table.RestExecutor.TableCommand.Executor.<ExecuteAsync>d__1`1.MoveNext()
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter`1.GetResult()
at Evoq.Data.TableStorage.TableStorageHelper`1.<ExecuteOperationAsync>d__27.MoveNext()
Answers:
username_0: 
Just a screenshot showing the partition does exist and my code is looking up a valid part and row key, I'm not going mad :)
Status: Issue closed
username_0: I did a Reset Data from the Systray menu and though in Windows Explorer it looked like it did nothing, when I restarted the emulator it give me a warning about the version of the emulator and the data in the folder; it wasn't giving me this before.
So to be sure, I stopped the emu and cleared the folder of everything using Windows Explorer and started the emu.
Everything is fine now. I can restart the emulator without having to delete the DB each time now. Closing. |
magefree/mage | 555878177 | Title: Hateful Eidolon bug
Question:
username_0: If Hateful Eidolon is controlled by someone other than the owner, the draw effect fails.
Status: Issue closed
Answers:
username_1: That fix do nothing (just replace source to target effect). I tested eidolon and it works fine in current version. Real bug was fixed in #6205
username_2: The issue I was having was a bit different than what's described here. Hateful Eidolon was both owned and controlled by the same person. The draw effect was failing when the creatures owned by a different player died (while being enchanted by Auras controlled by Hateful Eidolon's controller).
username_1: @username_2 in latest v3 version all problems with Hateful Eidolon was fixed.
username_2: Can confirm. Okay, I guess the top comment was just misphrased. |
axios/axios | 808898583 | Title: Confirmation - no support on sending body with get
Question:
username_0: Hi guys. I just went through [issue 787](https://github.com/axios/axios/issues/787). It is closed with no clear conclusion.
I just want to confirm, does Axios think that a get method should not have body, and not supporting so? |
napalm-automation/napalm | 1110263768 | Title: NXOS: Interface's is_up and is_enabled are flipped
Question:
username_0: ### Description of Issue/Question
NXOS devices report `is_up` as the port being `admin up` (ie: `no shutdown`) and `is_enabled` as having link.
ios, eos and mos (not 1st party) report this the other way around.
### Did you follow the steps from https://github.com/napalm-automation/napalm#faq
- [x] Yes
- [ ] No
### Setup
### napalm version
(Paste verbatim output from `pip freeze | grep napalm` between quotes below)
```
napalm==3.3.1
```
### Network operating system version
(Paste verbatim output from `show version` - or equivalent - between quotes below)
```
Software
BIOS: version 5.4.0
NXOS: version 7.0(3)I7(8)
BIOS compile time: 10/23/2019
NXOS image file is: bootflash:///nxos.7.0.3.I7.8.bin
NXOS compile time: 3/3/2020 20:00:00 [03/04/2020 05:49:49]
```
### Steps to Reproduce the Issue
Rad [NXOS implementation](https://github.com/napalm-automation/napalm/blob/develop/napalm/nxos/nxos.py#L897) vs [EOS implementation](https://github.com/napalm-automation/napalm/blob/develop/napalm/eos/eos.py#L480) |
selectize/selectize.js | 496283157 | Title: select an option with jquery
Question:
username_0: I'm using selectize this tool that let's me make a select field searchable. The thing though is I can't select an option, useful when editing existing data.
The code looks like this to instantiate selectize
```
$('#vendor').selectize({
sortField: 'text'
});
```
I can select something then find it's value on the console using `$('#vendor').val();` but after refreshing and doing `$('#vendor').val('21');` which is the value I got back, it doesn't select the option.
This normally works on a normal select field.<issue_closed>
Status: Issue closed |
tavicu/homebridge-samsung-tizen | 797417279 | Title: TvOfflineError
Question:
username_0: I always get TvOfflineError when I try to turn on my TV. Other commands are working fine.

Answers:
username_1: Not much i can help you with ... the plugin check if the TV is on by sending a ping request.
The request is blocked by your network or your ip is wrong in config.
Status: Issue closed
|
primefaces/primefaces | 350069515 | Title: Provide Maven snapshots repository
Question:
username_0: It'd be nice if snapshots built after every commit to `master` could be uploaded to a Maven repository. The upload could be automated and be be conducted from within Travis CI. This would be a continuous delivery pattern which is generally considered constructive and an improvement of the project. If repository.primefaces.org can't be used for snapshots, [Sonatype OSSRH](https://central.sonatype.org/pages/ossrh-guide.html) provides infrasturcture free of charge for FLOSS projects.
Afaik, `maven-jsf-plugin` only serves to be able to build PrimeFaces so that providing snapshots would only be useful to speed up building PrimeFaces in the CI/CD service.
I searched this tracker, the wiki and the community forums intensively and am now sure that this issue has never been discussed based on such a clean request.
Answers:
username_1: we had it long time ago into our own repo but noone reconfiguired it.
i won't invest time on it, thats up to PrimeTek.
username_2: This can be done already with JitPack. You can get the current snapshot of GitHub any time you want with this in your pom.xml.
```xml
<dependency>
<groupId>com.github.primefaces</groupId>
<artifactId>primefaces</artifactId>
<version>master-SNAPSHOT</version>
</dependency>
<repositories>
<repository>
<id>jitpack.io</id>
<url>https://jitpack.io</url>
</repository>
</repositories>
```
Status: Issue closed
|
cassiehuang/study | 395832548 | Title: js基础-apply,call
Question:
username_0: 每个Function对象都有一个apply() ,call()方法
Array.concat.apply(arr,arguments) arr调用Array的方法
apply最多有两个参数
可以用来代替另一个对象调用方法,将函数上下文从初始上下文变成指定新对象的上下文
如果没有指定新对象,则为global对象
当使用apply或者call的时候,就是调用的时候,这个时候实际上正在执行这个改变上下文,
并且完成某个功能的过程
apply的应用
1.Math.max()不支持数组参数
var max = Math.Max.apply(null,array)
apply会将数组转换为一个参数接一个参数的方式传递给方法,[arg1,arg2,arg3]
实际上就变成了Math.max(arg1,arg2,arg3)
2.两个数组的合并
Array.push.apply(arr1,arr2) |
Sequel-Ace/Sequel-Ace | 952069229 | Title: documentation for a shortcut list is out of order.
Question:
username_0: - Sequel Ace Version (including build number): Version 3.3.3 Build 3034
- Sequel Ace Source (App Store/GitHub/Homebrew):
- macOS Version: 11.4 (20F71)
- Processor Type (Intel/Apple): Apple
- MySQL Version: mysql Ver 15.1 Distrib 10.6.3-MariaDB, for osx10.16 (x86_64) using readline 5.1
- macOS Localization: English
**Is Issue Present in [Latest Beta](https://github.com/Sequel-Ace/Sequel-Ace/releases)?**
This is in the sequel-ace documentation.
**Description**
The bug is not in the application but on the web page that the application points to for Help (https://sequel-ace.com/shortcuts.html). On this page it states that the Custom query shortcut in the main tab is **⌘5** but should be **⌘6**. This also means that table triggers should be **⌘5**.
**Steps To Reproduce**
Visit https://sequel-ace.com/shortcuts.html , main tab section at towards top of page.
**Expected Behaviour**
Correct documentation
**Related Issues**
None
**Additional Context**
See attached screenshot
<img width="653" alt="Screen Shot 2021-07-24 at 8 43 19 AM" src="https://user-images.githubusercontent.com/149050/126868866-aba4099a-468f-4ef4-af35-5bfc65c02682.png"><issue_closed>
Status: Issue closed |
void-linux/void-packages | 343857293 | Title: KDE Frameworks, Applications and Plasma are outdated
Question:
username_0: https://www.kde.org/announcements/kde-frameworks-5.48.0.php
https://www.kde.org/announcements/announce-applications-18.04.3.php
https://www.kde.org/announcements/plasma-5.13.3.php
Answers:
username_1: @username_0 feel free to create a PR to update those packages ;-)
username_2: There are PRs already:
https://github.com/void-linux/void-packages/pulls?utf8=%E2%9C%93&q=is%3Apr+is%3Aopen+18.04.3
https://github.com/void-linux/void-packages/pull/891
Status: Issue closed
username_3: Should be fixed, if there are packages that need updating don't be afraid to make a pull request |
akhodakivskiy/VimFx | 179389896 | Title: gC cannot load config.js in 0.19 version
Question:
username_0: Steps to reproduce:
1. type "gC" to reload the config.js. the settings I customized in config.js cannot load, all of them failed. The same config.js worked in 0.18.1 version.
2. my config.js: https://github.com/username_0/userChromeJS/blob/master/Local/VimFx/config.js
Expected result:
Actual result:
Detailed use case example:
VimFx version: 0.19.1
Firefox version: 48.0.2
OS: Win10 X64 1511
Keyboard language: en-US
Worked before:
Example URL(s):
Answers:
username_0: hint_chars has changed to hints.chars
problem fixed
Status: Issue closed
|
maximdanilchenko/aiohttp-apispec | 430552846 | Title: 500 internal error if no data in json
Question:
username_0: if i send non json it just responds proper with missing fields
if i send legit it works normal
```
POST http://localhost:8080/reg HTTP/1.1
content-type: application/json
{
"msg":"hello",
"data":
{
"uName": "nekinalog",
"uPass": "<PASSWORD>",
"uEmail": "<EMAIL>"
}
}
```
but if i send json that has no "data" it crashes
```
POST http://localhost:8080/reg HTTP/1.1
content-type: application/json
{
"uName": "nekinalog",
"uPass": "<PASSWORD>",
"uEmail": "<EMAIL>"
}
```
```
python3 authserver.py
======== Running on http://0.0.0.0:8080 ========
(Press CTRL+C to quit)
Error handling request
Traceback (most recent call last):
File "/home/me/.local/lib/python3.7/site-packages/aiohttp/web_protocol.py", line 418, in start
resp = await task
File "/home/me/.local/lib/python3.7/site-packages/aiohttp/web_app.py", line 458, in _handle
resp = await handler(request)
File "/home/me/.local/lib/python3.7/site-packages/aiohttp/web_middlewares.py", line 119, in impl
return await handler(request)
File "/home/me/.local/lib/python3.7/site-packages/aiohttp_apispec/middlewares.py", line 40, in validation_middleware
request[request.app["_apispec_request_data_name"]] = kwargs
File "/home/me/.local/lib/python3.7/site-packages/aiohttp/web_app.py", line 160, in __getitem__
return self._state[key]
KeyError: '_apispec_request_data_name'
```
Hmm is this normal behavior or am i doing something wrong?
Here my simple server api
```python
from aiohttp_apispec import (
docs,
request_schema,
setup_aiohttp_apispec,
)
from marshmallow import Schema, fields
from aiohttp_apispec import validation_middleware
[Truncated]
uEmail = fields.Str(required=True)
ROUTES = web.RouteTableDef()
@ROUTES.post('/reg')
@request_schema(ReqRequestSchema(strict=True))
async def try_reg(request):
try: #probbaly can ignore all stuff inhere cuz it crashes in validation_midleware
data = await request.json() #before this even runs, anyway will change this after
print(data)
txt = data['uName']
except json.decoder.JSONDecodeError as json_error:
return web.json_response({'error':'recived json is bad'})
return web.Response(text=txt)
APP = web.Application(middlewares=[validation_middleware]) #error_middleware,
APP.add_routes(ROUTES)
web.run_app(APP)
```
Answers:
username_1: Hi, @username_0!
To make `validation_middleware` work you should setup your app with `setup_aiohttp_apispec` function before running your server:
```python
setup_aiohttp_apispec(app)
```
After that you can change your code to:
```python
@routes.post('/reg')
@request_schema(ReqRequestSchema(strict=True))
async def try_reg(request):
data = request['data'] # request data is already validated and loaded into Request object
txt = data['uName']
return web.Response(text=txt)
```
To catch json decode errors add your own middleware just before `validation_middleware`. Smth like:
```python
@web.middleware
async def intercept_error(request, handler):
try:
return await handler(request)
except json.decoder.JSONDecodeError:
return web.json_response({'error':'recived json is bad'}, status=400)
```
And in your app factory it can look like this:
```python
app.middlewares.extend([intercept_error, validation_middleware, authorization, database_init, e.t.c.])
```
username_0: Ty for response i got it working, main issue was i somhow missed that setup_aiohttp_apispec(app) :)
Status: Issue closed
|
pmret/papermario | 977524130 | Title: Merge jp and us source directories
Question:
username_0: At the moment jp's splat.yaml defines `src/jp/` as its root source directory. Ideally it should be `src` - all the existing segment names need to be prefixed with `jp/` to keep the existing structure, though - besides the first C file, since it's identical to `is_debug` from us.<issue_closed>
Status: Issue closed |
ruby-grape/grape | 192001210 | Title: grape 'array' type not compatible with swagger ?
Question:
username_0: here is my code:
```
params do
requires :logs, type: Array, documentation: {param_type: 'body'}
end
```
below is the result using 'grape-swagger' and swagger-ui:
```
[
{
"logs": "string"
}
]
```
obviously, it's not correct. Seems a bug?
Status: Issue closed
Answers:
username_1: This belongs in grape-swagger, you can open an issue there. |
caseydierking/wcss | 367624606 | Title: Make it look like horse is walking left or right
Question:
username_0: Due to a recent contribution, we have a left and right sprite kit that was added for the horse. It would be cool to have the horse animate and give the impression of walking left when holding the "left" arrow key. Similarly, it would be great to have the horse walk right when holding the "right" arrow key.
Answers:
username_1: I'll take this. Is that fine?
username_1: First look. Is that fine?

I'm using background-image for animation, the size is fixed so seems like we need to use `transform: scale` to make it smaller. I'll create a pull request later.
username_0: Great first stab at this. This is what I had envisioned. One thing to consider is which way the horse is left facing after holding down the arrow keys. Another consideration is the area that is determined to be the horses "mouth". And how this change affects that. The horse can eat certain food items only near its mouth.
Just a few thoughts. Great work and quick turnaround!
Status: Issue closed
|
stride-nyc/remote_retro | 411909569 | Title: welcome email for new users
Question:
username_0: As a user who has just signed up for RemoteRetro,
I want to receive an email confirming my new account,
So that I have a simple means of returning to RemoteRetro should I forget.
__By Way Of Explanation:__ email is an incredibly powerful tool that we barely utilize. (At the moment, the only emails we send are the action items emails send out at the end of retros.) With this story, we a) provide new users a paper trail for _returning_ to RemoteRetro, and, as as nice side effect, give them a simple means of *forwarding* RemoteRetro to their teams/teammates.
__Acceptance Criteria:__
- [ ] upon persisting a *new* user, we send a welcome email:
- [ ] the subject line is personalized, "[FIRST_NAME]! Welcome to RemoteRetro!"
- [ ] body reads:
- [ ]
```
"[FIRST_NAME]!
We're thrilled that you're interested in continuously improving your team with RemoteRetro. You can now visit your dashboard at https://remoteretro.org/retros, and we encourage you to invite your teams(s) to check us out!
__Helpful Links:__
- [RetroManagement](https://www.superoffice.com/blog/email-marketing-strategy/)
- [Bamboo](https://www.superoffice.com/blog/email-marketing-strategy/)
- [Email Marketing Strategy](https://www.superoffice.com/blog/email-marketing-strategy/)
Answers:
username_0: paired on this with @grantziolkowski. up and on prod!
Status: Issue closed
|
netty/netty | 187338280 | Title: IllegalArgumentException in io.netty.util.internal.PriorityQueue
Question:
username_0: 4.1.7-SNAPSHOT
It's always the same message: `e.priorityQueueIndex(): 0 (expected: -1)`. It's strange because I don't see `WeightedFairQueueByteDistributor$State#priorityQueueIndex(0)` ever being called (or the `priorityQueueIndex` field getting set).
```java
2016-11-04 13:12:14,722 [WorkerThread-2] ExceptionHandler ERROR: name=Http2ExceptionHandler, channel=[id: 0x020ee146, L:/aaa.aaa.aaa.aaa:443 - R:/bbb.bbb.bbb.bbb:63000]
java.lang.IllegalArgumentException: e.priorityQueueIndex(): 0 (expected: -1)
at io.netty.util.internal.PriorityQueue.offer(PriorityQueue.java:83) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.util.internal.PriorityQueue.offer(PriorityQueue.java:32) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.handler.codec.http2.WeightedFairQueueByteDistributor$State.offer(WeightedFairQueueByteDistributor.java:299) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.handler.codec.http2.WeightedFairQueueByteDistributor.distributeToChildren(WeightedFairQueueByteDistributor.java:198) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.handler.codec.http2.WeightedFairQueueByteDistributor.distribute(WeightedFairQueueByteDistributor.java:130) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.handler.codec.http2.DefaultHttp2RemoteFlowController$WritabilityMonitor.writePendingBytes(DefaultHttp2RemoteFlowController.java:631) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.handler.codec.http2.DefaultHttp2RemoteFlowController.writePendingBytes(DefaultHttp2RemoteFlowController.java:255) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.handler.codec.http2.Http2ConnectionHandler.flush(Http2ConnectionHandler.java:161) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.DefaultChannelPipeline.flush(DefaultChannelPipeline.java:974) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannel.flush(AbstractChannel.java:244) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at com.squarespace.echo.http2.streams.HttpToHttp2Handler.flush(HttpToHttp2Handler.java:71) [echo-http-2.17.16.jar:?]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelDuplexHandler.flush(ChannelDuplexHandler.java:117) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelDuplexHandler.flush(ChannelDuplexHandler.java:117) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelDuplexHandler.flush(ChannelDuplexHandler.java:117) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelOutboundHandlerAdapter.flush(ChannelOutboundHandlerAdapter.java:115) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelOutboundHandlerAdapter.flush(ChannelOutboundHandlerAdapter.java:115) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelOutboundHandlerAdapter.flush(ChannelOutboundHandlerAdapter.java:115) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelDuplexHandler.flush(ChannelDuplexHandler.java:117) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelDuplexHandler.flush(ChannelDuplexHandler.java:117) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelDuplexHandler.flush(ChannelDuplexHandler.java:117) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
[Truncated]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelDuplexHandler.flush(ChannelDuplexHandler.java:117) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelDuplexHandler.flush(ChannelDuplexHandler.java:117) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.flush(AbstractChannelHandlerContext.java:760) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelDuplexHandler.flush(ChannelDuplexHandler.java:117) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush0(AbstractChannelHandlerContext.java:787) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeFlush(AbstractChannelHandlerContext.java:779) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.access$1500(AbstractChannelHandlerContext.java:39) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext$16.run(AbstractChannelHandlerContext.java:767) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:418) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:454) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:873) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_102]
```
Answers:
username_1: @username_2 @nmittler can you have a look ?
username_2: sure
username_2: @username_0 - can you tell me what your priority tree looks like when flush is invoked, and also do you have a reproducer?
`priorityQueueIndex` is set by our `PriorityQueue` implementation as nodes are inserted/removed and is to avoid the traversal for removal from the queue.
username_0: @username_2 I don't have a reproducer but it happens pretty reliably in our production environment. There is an initial burst as soon as the server starts accepting connections and then one every once in a while.
How do I get the info for the priority tree? ToString or do you need something else?
username_0: Just throwing it out there but could it be a thread visibility thing because `priorityQueueIndex` is neither final nor volatile?!
username_2: codec-http2 is not thread safe. You *always* have to be on the IO thread when ever you interact with it ... is this not the case?
username_0: Checking... Possible that my `Http2ConnectionEncoder` is called from a different thread.
username_2: If you are sure you are always on the I/O thread then it would be great to see the string represetnation of the general priority tree, and the string representation of the `WeightedFairQueueByteDistributor`'s tree of queue's structure. We don't have a means to serialize these to string now ... but they can be added if necessary.
username_0: @username_2 flush() and my interactions with the encoder are on the codec's IO thread.
Below are two `toStrings()` of the WeightedFairQueueByteDistributor's queue structure. I'm also attaching the code that was used to produce it. Where do I find the general priority queue?
```java
2016-11-04 17:54:07,120 [WorkerThread-3] Http2HttpBridgeHandler ERROR: state=State{stream=0, queue=[State{stream=39, queue=[State{stream=41, queue=[State{stream=45, queue=[State{stream=47, queue=[State{stream=49, queue=[State{stream=51, queue=[State{stream=53, queue=[State{stream=55, queue=[State{stream=57, queue=[], streamableBytes=7453, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=0, pseudoTime=0, totalQueuedWeights=0, active=true}], streamableBytes=3158, activeCountForTree=2, priorityQueueIndex=0, pseudoTimeToWrite=0, pseudoTime=0, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=2, priorityQueueIndex=0, pseudoTimeToWrite=0, pseudoTime=0, totalQueuedWeights=220, active=false}], streamableBytes=6071, activeCountForTree=3, priorityQueueIndex=0, pseudoTimeToWrite=0, pseudoTime=0, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=3, priorityQueueIndex=0, pseudoTimeToWrite=0, pseudoTime=0, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=3, priorityQueueIndex=0, pseudoTimeToWrite=0, pseudoTime=0, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=3, priorityQueueIndex=0, pseudoTimeToWrite=57792, pseudoTime=0, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=3, priorityQueueIndex=0, pseudoTimeToWrite=62349, pseudoTime=62349, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=3, priorityQueueIndex=0, pseudoTimeToWrite=322177, pseudoTime=62349, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=3, priorityQueueIndex=-1, pseudoTimeToWrite=0, pseudoTime=322177, totalQueuedWeights=220, active=false}
```
```java
2016-11-04 17:54:07,121 [WorkerThread-3] ExceptionHandler ERROR: name=Http2ExceptionHandler, channel=[id: 0x6dae7b41, L:/aaa.aaa.aaa.aaa:443 - R:/bbb.bbb.bbb.bbb:54938]
java.lang.IllegalArgumentException: e.priorityQueueIndex(): 0 (expected: -1)
at io.netty.util.internal.PriorityQueue.offer(PriorityQueue.java:83) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.util.internal.PriorityQueue.offer(PriorityQueue.java:32) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.handler.codec.http2.WeightedFairQueueByteDistributor$State.offer(WeightedFairQueueByteDistributor.java:299) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.handler.codec.http2.WeightedFairQueueByteDistributor.distributeToChildren(WeightedFairQueueByteDistributor.java:198) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.handler.codec.http2.WeightedFairQueueByteDistributor.distribute(WeightedFairQueueByteDistributor.java:130) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.handler.codec.http2.DefaultHttp2RemoteFlowController$WritabilityMonitor.writePendingBytes(DefaultHttp2RemoteFlowController.java:631) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.handler.codec.http2.DefaultHttp2RemoteFlowController.writePendingBytes(DefaultHttp2RemoteFlowController.java:255) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.handler.codec.http2.Http2ConnectionHandler.flush(Http2ConnectionHandler.java:161) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at com.squarespace.echo.http2.streams.Http2HttpBridgeHandler.flush(Http2HttpBridgeHandler.java:43) ~[echo-http-2.17.30.jar:?]
at io.netty.handler.codec.http2.Http2ConnectionHandler.channelWritabilityChanged(Http2ConnectionHandler.java:385) ~[netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelWritabilityChanged(AbstractChannelHandlerContext.java:445) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelWritabilityChanged(AbstractChannelHandlerContext.java:427) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.fireChannelWritabilityChanged(AbstractChannelHandlerContext.java:420) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelInboundHandlerAdapter.channelWritabilityChanged(ChannelInboundHandlerAdapter.java:119) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelWritabilityChanged(AbstractChannelHandlerContext.java:445) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelWritabilityChanged(AbstractChannelHandlerContext.java:427) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.fireChannelWritabilityChanged(AbstractChannelHandlerContext.java:420) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelInboundHandlerAdapter.channelWritabilityChanged(ChannelInboundHandlerAdapter.java:119) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelWritabilityChanged(AbstractChannelHandlerContext.java:445) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelWritabilityChanged(AbstractChannelHandlerContext.java:427) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.fireChannelWritabilityChanged(AbstractChannelHandlerContext.java:420) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelInboundHandlerAdapter.channelWritabilityChanged(ChannelInboundHandlerAdapter.java:119) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelWritabilityChanged(AbstractChannelHandlerContext.java:445) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelWritabilityChanged(AbstractChannelHandlerContext.java:427) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.fireChannelWritabilityChanged(AbstractChannelHandlerContext.java:420) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelWritabilityChanged(DefaultChannelPipeline.java:1357) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelWritabilityChanged(AbstractChannelHandlerContext.java:445) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelWritabilityChanged(AbstractChannelHandlerContext.java:427) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.DefaultChannelPipeline.fireChannelWritabilityChanged(DefaultChannelPipeline.java:938) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelOutboundBuffer.fireChannelWritabilityChanged(ChannelOutboundBuffer.java:599) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelOutboundBuffer.setWritable(ChannelOutboundBuffer.java:565) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelOutboundBuffer.decrementPendingOutboundBytes(ChannelOutboundBuffer.java:208) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelOutboundBuffer.remove(ChannelOutboundBuffer.java:273) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.ChannelOutboundBuffer.removeBytes(ChannelOutboundBuffer.java:352) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.socket.nio.NioSocketChannel.doWrite(NioSocketChannel.java:442) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.AbstractChannel$AbstractUnsafe.flush0(AbstractChannel.java:856) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.forceFlush(AbstractNioChannel.java:368) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:645) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:574) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:488) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:450) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:873) [netty-all-4.1.7.Final-SNAPSHOT.jar:4.1.7.Final-SNAPSHOT]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_102]
```
```java
2016-11-04 18:00:29,649 [WorkerThread-5] Http2HttpBridgeHandler ERROR: state=State{stream=0, queue=[State{stream=33, queue=[State{stream=39, queue=[State{stream=41, queue=[State{stream=43, queue=[State{stream=45, queue=[State{stream=49, queue=[State{stream=51, queue=[State{stream=53, queue=[State{stream=55, queue=[State{stream=57, queue=[State{stream=59, queue=[State{stream=61, queue=[State{stream=65, queue=[State{stream=67, queue=[State{stream=69, queue=[State{stream=71, queue=[State{stream=73, queue=[State{stream=75, queue=[State{stream=77, queue=[State{stream=79, queue=[], streamableBytes=27622, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=24415, pseudoTime=0, totalQueuedWeights=0, active=true}], streamableBytes=16376, activeCountForTree=2, priorityQueueIndex=0, pseudoTimeToWrite=40801, pseudoTime=24415, totalQueuedWeights=220, active=true}], streamableBytes=52037, activeCountForTree=3, priorityQueueIndex=0, pseudoTimeToWrite=40801, pseudoTime=40801, totalQueuedWeights=220, active=true}], streamableBytes=52027, activeCountForTree=4, priorityQueueIndex=0, pseudoTimeToWrite=40811, pseudoTime=40801, totalQueuedWeights=220, active=true}], streamableBytes=52037, activeCountForTree=5, priorityQueueIndex=0, pseudoTimeToWrite=40811, pseudoTime=40811, totalQueuedWeights=220, active=true}], streamableBytes=27622, activeCountForTree=6, priorityQueueIndex=0, pseudoTimeToWrite=65226, pseudoTime=40811, totalQueuedWeights=220, active=true}], streamableBytes=51011, activeCountForTree=7, priorityQueueIndex=0, pseudoTimeToWrite=66252, pseudoTime=65226, totalQueuedWeights=220, active=true}], streamableBytes=16376, activeCountForTree=8, priorityQueueIndex=0, pseudoTimeToWrite=142704, pseudoTime=66252, totalQueuedWeights=220, active=true}], streamableBytes=52027, activeCountForTree=9, priorityQueueIndex=0, pseudoTimeToWrite=142714, pseudoTime=142704, totalQueuedWeights=220, active=true}], streamableBytes=51011, activeCountForTree=10, priorityQueueIndex=0, pseudoTimeToWrite=143740, pseudoTime=142714, totalQueuedWeights=220, active=true}], streamableBytes=52027, activeCountForTree=11, priorityQueueIndex=0, pseudoTimeToWrite=143750, pseudoTime=143740, totalQueuedWeights=220, active=true}], streamableBytes=27622, activeCountForTree=12, priorityQueueIndex=0, pseudoTimeToWrite=168165, pseudoTime=143750, totalQueuedWeights=220, active=true}], streamableBytes=52027, activeCountForTree=13, priorityQueueIndex=0, pseudoTimeToWrite=168175, pseudoTime=168165, totalQueuedWeights=220, active=true}], streamableBytes=27622, activeCountForTree=14, priorityQueueIndex=0, pseudoTimeToWrite=192590, pseudoTime=168175, totalQueuedWeights=220, active=true}], streamableBytes=16376, activeCountForTree=15, priorityQueueIndex=0, pseudoTimeToWrite=269042, pseudoTime=192590, totalQueuedWeights=220, active=true}], streamableBytes=52037, activeCountForTree=16, priorityQueueIndex=0, pseudoTimeToWrite=269042, pseudoTime=269042, totalQueuedWeights=220, active=true}], streamableBytes=51011, activeCountForTree=17, priorityQueueIndex=0, pseudoTimeToWrite=270068, pseudoTime=269042, totalQueuedWeights=220, active=true}], streamableBytes=52027, activeCountForTree=18, priorityQueueIndex=0, pseudoTimeToWrite=270078, pseudoTime=270068, totalQueuedWeights=220, active=true}], streamableBytes=27622, activeCountForTree=19, priorityQueueIndex=0, pseudoTimeToWrite=370945, pseudoTime=270078, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=19, priorityQueueIndex=0, pseudoTimeToWrite=1156657, pseudoTime=398567, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=19, priorityQueueIndex=-1, pseudoTimeToWrite=0, pseudoTime=1156657, totalQueuedWeights=220, active=false}
```
```java
2016-11-04 18:00:29,650 [WorkerThread-5] ExceptionHandler ERROR: name=Http2ExceptionHandler, channel=[id: 0x5e6e957c, L:/aaa.aaa.aaa.aaa:443 - R:/bbb.bbb.bbb.bbb:55169]
[Truncated]
private static Object connectionState(WeightedFairQueueByteDistributor distributor) {
return value(distributor, "connectionState");
}
private static Object value(Object dst, String name) {
if (dst != null) {
try {
Class<?> clazz = dst.getClass();
Field field = clazz.getDeclaredField(name);
field.setAccessible(true);
return field.get(dst);
} catch (Exception err) {
LOG.error("Exception: obj={}, name={}", dst, name, err);
}
}
return null;
}
}
```
username_2: Here is where it is created in `WeightedFairQueueByteDistributor.java`:
https://github.com/netty/netty/blob/4.1/codec-http2/src/main/java/io/netty/handler/codec/http2/WeightedFairQueueByteDistributor.java#L243
The root of the tree is here:
https://github.com/netty/netty/blob/4.1/codec-http2/src/main/java/io/netty/handler/codec/http2/WeightedFairQueueByteDistributor.java#L48
Here is the implementation:
https://github.com/netty/netty/blob/4.1/common/src/main/java/io/netty/util/internal/PriorityQueue.java
username_0: @username_2 great. That's exactly what I toString'd.
Look for the two codeblocks right before the stacktraces. They're rendered as a single line. Copy & Paste them into some Text Editor.
username_2: @username_0 - can you show me the state before?
```java
@Override
public void flush(ChannelHandlerContext ctx) throws Http2Exception {
String beforeState = null;
if (LOG.isErrorEnabled()) {
Http2ConnectionEncoder encoder = encoder();
beforeState = toString(encoder);
}
try {
super.flush(ctx);
} catch (Throwable t) {
if (LOG.isErrorEnabled()) {
Http2ConnectionEncoder encoder = encoder();
String value = toString(encoder);
LOG.error("before state={}", beforeState);
LOG.error("after state={}", value);
}
throw t;
}
}
```
username_0: ```java
016-11-04 19:00:56,334 [WorkerThread-0] Http2HttpBridgeHandler ERROR:
before=State{stream=0, queue=[State{stream=3, queue=[State{stream=11, queue=[State{stream=13, queue=[State{stream=17, queue=[State{stream=19, queue=[State{stream=23, queue=[State{stream=25, queue=[State{stream=27, queue=[State{stream=29, queue=[], streamableBytes=96568, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=32010, pseudoTime=16005, totalQueuedWeights=0, active=true}], streamableBytes=1669, activeCountForTree=2, priorityQueueIndex=0, pseudoTimeToWrite=32010, pseudoTime=32010, totalQueuedWeights=220, active=true}], streamableBytes=104854, activeCountForTree=3, priorityQueueIndex=0, pseudoTimeToWrite=32010, pseudoTime=32010, totalQueuedWeights=220, active=true}], streamableBytes=36936, activeCountForTree=4, priorityQueueIndex=0, pseudoTimeToWrite=146597, pseudoTime=32010, totalQueuedWeights=220, active=true}], streamableBytes=56007, activeCountForTree=5, priorityQueueIndex=0, pseudoTimeToWrite=146597, pseudoTime=146597, totalQueuedWeights=220, active=true}], streamableBytes=98185, activeCountForTree=6, priorityQueueIndex=0, pseudoTimeToWrite=179084, pseudoTime=146597, totalQueuedWeights=220, active=true}], streamableBytes=27101, activeCountForTree=7, priorityQueueIndex=0, pseudoTimeToWrite=179084, pseudoTime=179084, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=7, priorityQueueIndex=0, pseudoTimeToWrite=481469, pseudoTime=179084, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=7, priorityQueueIndex=0, pseudoTimeToWrite=657600, pseudoTime=528718, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=7, priorityQueueIndex=-1, pseudoTimeToWrite=0, pseudoTime=657600, totalQueuedWeights=220, active=false},
after=State{stream=0, queue=[State{stream=3, queue=[State{stream=11, queue=[State{stream=17, queue=[State{stream=19, queue=[State{stream=23, queue=[State{stream=25, queue=[State{stream=27, queue=[State{stream=29, queue=[], streamableBytes=96568, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=32010, pseudoTime=16005, totalQueuedWeights=0, active=true}], streamableBytes=1669, activeCountForTree=2, priorityQueueIndex=0, pseudoTimeToWrite=32010, pseudoTime=32010, totalQueuedWeights=220, active=true}], streamableBytes=104854, activeCountForTree=3, priorityQueueIndex=0, pseudoTimeToWrite=32010, pseudoTime=32010, totalQueuedWeights=220, active=true}], streamableBytes=36936, activeCountForTree=4, priorityQueueIndex=0, pseudoTimeToWrite=146597, pseudoTime=32010, totalQueuedWeights=220, active=true}], streamableBytes=56007, activeCountForTree=5, priorityQueueIndex=0, pseudoTimeToWrite=146597, pseudoTime=146597, totalQueuedWeights=220, active=true}], streamableBytes=98185, activeCountForTree=6, priorityQueueIndex=0, pseudoTimeToWrite=179084, pseudoTime=146597, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=6, priorityQueueIndex=0, pseudoTimeToWrite=555819, pseudoTime=206185, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=6, priorityQueueIndex=0, pseudoTimeToWrite=657600, pseudoTime=555819, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=6, priorityQueueIndex=-1, pseudoTimeToWrite=0, pseudoTime=657600, totalQueuedWeights=220, active=false}
```
```java
2016-11-04 19:05:36,067 [WorkerThread-6] Http2HttpBridgeHandler ERROR:
before=State{stream=0, queue=[State{stream=1, queue=[State{stream=3, queue=[State{stream=5, queue=[State{stream=7, queue=[State{stream=13, queue=[State{stream=15, queue=[State{stream=19, queue=[State{stream=23, queue=[], streamableBytes=113, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=339, pseudoTime=339, totalQueuedWeights=0, active=true}], streamableBytes=0, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=565, pseudoTime=452, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=452, pseudoTime=565, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=678, pseudoTime=452, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=791, pseudoTime=678, totalQueuedWeights=220, active=false}], streamableBytes=113, activeCountForTree=2, priorityQueueIndex=0, pseudoTimeToWrite=791, pseudoTime=791, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=2, priorityQueueIndex=0, pseudoTimeToWrite=791, pseudoTime=791, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=2, priorityQueueIndex=0, pseudoTimeToWrite=791, pseudoTime=791, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=2, priorityQueueIndex=-1, pseudoTimeToWrite=0, pseudoTime=791, totalQueuedWeights=220, active=false},
after=State{stream=0, queue=[State{stream=1, queue=[State{stream=3, queue=[State{stream=7, queue=[State{stream=13, queue=[State{stream=15, queue=[State{stream=19, queue=[State{stream=23, queue=[], streamableBytes=113, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=339, pseudoTime=339, totalQueuedWeights=0, active=true}], streamableBytes=0, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=565, pseudoTime=452, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=452, pseudoTime=565, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=678, pseudoTime=452, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=791, pseudoTime=678, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=904, pseudoTime=904, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=791, pseudoTime=904, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=1, priorityQueueIndex=-1, pseudoTimeToWrite=0, pseudoTime=791, totalQueuedWeights=220, active=false}
```
username_0: The first before/after "nicely" formatted:
```
before=State{stream=0,
queue=[State{stream=3,
queue=[State{stream=11,
queue=[State{stream=13,
queue=[State{stream=17,
queue=[State{stream=19,
queue=[State{stream=23,
queue=[State{stream=25,
queue=[State{stream=27,
queue=[State{stream=29,
queue=[],
streamableBytes=96568,
activeCountForTree=1,
priorityQueueIndex=0,
pseudoTimeToWrite=32010,
pseudoTime=16005,
totalQueuedWeights=0,
active=true}],
streamableBytes=1669,
activeCountForTree=2,
priorityQueueIndex=0,
pseudoTimeToWrite=32010,
pseudoTime=32010,
totalQueuedWeights=220,
active=true}],
streamableBytes=104854,
activeCountForTree=3,
priorityQueueIndex=0,
pseudoTimeToWrite=32010,
pseudoTime=32010,
totalQueuedWeights=220,
active=true}],
streamableBytes=36936,
activeCountForTree=4,
priorityQueueIndex=0,
pseudoTimeToWrite=146597,
pseudoTime=32010,
totalQueuedWeights=220,
active=true}],
streamableBytes=56007,
activeCountForTree=5,
priorityQueueIndex=0,
pseudoTimeToWrite=146597,
pseudoTime=146597,
totalQueuedWeights=220,
active=true}],
streamableBytes=98185,
activeCountForTree=6,
priorityQueueIndex=0,
pseudoTimeToWrite=179084,
pseudoTime=146597,
totalQueuedWeights=220,
active=true}],
streamableBytes=27101,
activeCountForTree=7,
priorityQueueIndex=0,
pseudoTimeToWrite=179084,
[Truncated]
priorityQueueIndex=0,
pseudoTimeToWrite=555819,
pseudoTime=206185,
totalQueuedWeights=220,
active=false}],
streamableBytes=0,
activeCountForTree=6,
priorityQueueIndex=0,
pseudoTimeToWrite=657600,
pseudoTime=555819,
totalQueuedWeights=220,
active=false}],
streamableBytes=0,
activeCountForTree=6,
priorityQueueIndex=-1,
pseudoTimeToWrite=0,
pseudoTime=657600,
totalQueuedWeights=220,
active=false}
```
username_2: @username_0 - thanks! can you also tell me the `maxBytes` value for the initial call to `distribute`? It is derived from `DefaultHttp2RemoteFlowController#writableBytes()`.
username_0: @username_2 initial call as in the very first call or the value for before/after the exception is being thrown?
username_2: @username_0 - Just the value of the argument which triggers the call to `WeightedFairQueueByteDistributor#distribute(int, Writer)` when the exception occurs (shouldn't change before and after).
username_2: @username_0 - also are you using the default value for `WeightedFairQueueByteDistributor#allocationQuantum`, and if not can you tell me what value you use?
username_0: ```
before=state{writableBytes=43398, connectionState=State{stream=0, queue=[State{stream=25, queue=[State{stream=27, queue=[State{stream=29, queue=[State{stream=33, queue=[State{stream=35, queue=[State{stream=37, queue=[State{stream=41, queue=[State{stream=43, queue=[State{stream=45, queue=[State{stream=49, queue=[State{stream=51, queue=[State{stream=53, queue=[State{stream=57, queue=[State{stream=59, queue=[State{stream=63, queue=[State{stream=71, queue=[], streamableBytes=13674, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=68440, pseudoTime=27378, totalQueuedWeights=0, active=true}], streamableBytes=10, activeCountForTree=2, priorityQueueIndex=0, pseudoTimeToWrite=82124, pseudoTime=68440, totalQueuedWeights=220, active=true}], streamableBytes=13674, activeCountForTree=3, priorityQueueIndex=0, pseudoTimeToWrite=82134, pseudoTime=82124, totalQueuedWeights=220, active=true}], streamableBytes=13674, activeCountForTree=4, priorityQueueIndex=0, pseudoTimeToWrite=95828, pseudoTime=82134, totalQueuedWeights=220, active=true}], streamableBytes=13674, activeCountForTree=5, priorityQueueIndex=0, pseudoTimeToWrite=95838, pseudoTime=95828, totalQueuedWeights=220, active=true}], streamableBytes=13674, activeCountForTree=6, priorityQueueIndex=0, pseudoTimeToWrite=95848, pseudoTime=95838, totalQueuedWeights=220, active=true}], streamableBytes=13674, activeCountForTree=7, priorityQueueIndex=0, pseudoTimeToWrite=109542, pseudoTime=95848, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=7, priorityQueueIndex=0, pseudoTimeToWrite=109552, pseudoTime=109542, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=7, priorityQueueIndex=0, pseudoTimeToWrite=109562, pseudoTime=109552, totalQueuedWeights=220, active=false}], streamableBytes=13674, activeCountForTree=8, priorityQueueIndex=0, pseudoTimeToWrite=123256, pseudoTime=109562, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=8, priorityQueueIndex=0, pseudoTimeToWrite=123266, pseudoTime=123256, totalQueuedWeights=220, active=false}], streamableBytes=1042, activeCountForTree=9, priorityQueueIndex=0, pseudoTimeToWrite=135908, pseudoTime=123266, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=9, priorityQueueIndex=0, pseudoTimeToWrite=149602, pseudoTime=135908, totalQueuedWeights=220, active=false}], streamableBytes=13674, activeCountForTree=10, priorityQueueIndex=0, pseudoTimeToWrite=149612, pseudoTime=149602, totalQueuedWeights=220, active=true}], streamableBytes=1016, activeCountForTree=11, priorityQueueIndex=0, pseudoTimeToWrite=149622, pseudoTime=149612, totalQueuedWeights=220, active=true}], streamableBytes=6404, activeCountForTree=12, priorityQueueIndex=0, pseudoTimeToWrite=326792, pseudoTime=149622, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=12, priorityQueueIndex=-1, pseudoTimeToWrite=0, pseudoTime=340466, totalQueuedWeights=220, active=false}},
```
```
after=state{writableBytes=32768, connectionState=State{stream=0, queue=[State{stream=27, queue=[State{stream=33, queue=[State{stream=35, queue=[State{stream=37, queue=[State{stream=41, queue=[State{stream=43, queue=[State{stream=45, queue=[State{stream=49, queue=[State{stream=51, queue=[State{stream=53, queue=[State{stream=57, queue=[State{stream=59, queue=[State{stream=63, queue=[State{stream=71, queue=[], streamableBytes=13674, activeCountForTree=1, priorityQueueIndex=0, pseudoTimeToWrite=68440, pseudoTime=27378, totalQueuedWeights=0, active=true}], streamableBytes=10, activeCountForTree=2, priorityQueueIndex=0, pseudoTimeToWrite=82124, pseudoTime=68440, totalQueuedWeights=220, active=true}], streamableBytes=13674, activeCountForTree=3, priorityQueueIndex=0, pseudoTimeToWrite=82134, pseudoTime=82124, totalQueuedWeights=220, active=true}], streamableBytes=13674, activeCountForTree=4, priorityQueueIndex=0, pseudoTimeToWrite=95828, pseudoTime=82134, totalQueuedWeights=220, active=true}], streamableBytes=13674, activeCountForTree=5, priorityQueueIndex=0, pseudoTimeToWrite=95838, pseudoTime=95828, totalQueuedWeights=220, active=true}], streamableBytes=13674, activeCountForTree=6, priorityQueueIndex=0, pseudoTimeToWrite=95848, pseudoTime=95838, totalQueuedWeights=220, active=true}], streamableBytes=13674, activeCountForTree=7, priorityQueueIndex=0, pseudoTimeToWrite=109542, pseudoTime=95848, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=7, priorityQueueIndex=0, pseudoTimeToWrite=109552, pseudoTime=109542, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=7, priorityQueueIndex=0, pseudoTimeToWrite=109562, pseudoTime=109552, totalQueuedWeights=220, active=false}], streamableBytes=13674, activeCountForTree=8, priorityQueueIndex=0, pseudoTimeToWrite=123256, pseudoTime=109562, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=8, priorityQueueIndex=0, pseudoTimeToWrite=123266, pseudoTime=123256, totalQueuedWeights=220, active=false}], streamableBytes=1042, activeCountForTree=9, priorityQueueIndex=0, pseudoTimeToWrite=135908, pseudoTime=123266, totalQueuedWeights=220, active=true}], streamableBytes=0, activeCountForTree=9, priorityQueueIndex=0, pseudoTimeToWrite=149612, pseudoTime=135908, totalQueuedWeights=220, active=false}], streamableBytes=0, activeCountForTree=9, priorityQueueIndex=0, pseudoTimeToWrite=361560, pseudoTime=163286, totalQueuedWeights=220, active=false}],
```
username_0: @username_2 I'm using the default value. The only non-default value I'm using is maxConcurrentStreams which is set to `1`.
username_2: @username_0 - I hacked together a unit test with one of your earlier state dumps and couldn't repo (I'm missing something), but I will try again now that you have provided the `writableBytes` and maybe it will become more clear...
username_0: @username_2 would it help if... You change the `PriorityQueue#offer()`'s exception to include the `System#indentityHashCode()` of the offending argument and I'll output the identity hashcode in the before/after dumps?!
username_2: Yes it may help to know when the exception is being thrown.
username_2: Its starting to come back to me now ... if the data frame has the `EOS` flag set, and the write completes synchronously, then we may close the stream, generate a priority event, notify the stream listeners via `onPriorityTreeParentChanging`, and WeightedFairQueueByteDistributor will adjust the queues ... so it may make sense that the queues are missing streams after and the links are still valid
username_2: @username_0 - I was able to reproduce the issue locally ... stand by for a PR.
username_2: See #6004
Status: Issue closed
|
capape/observation-manager | 738188678 | Title: Use tz code in date fields in xml files
Question:
username_0: I see that the time zone offset is now added to the time in the XML file, but it would be nice to have the offset incremented if daylight saving time is implemented. Maybe a site could include TZ code?
I think importing old files will need to assume local or UT time if the new version will manage this. Probably local, since the site has a time offset. So if the offset is 1 hour you'd need to add +01:00 to all the time fields, I think. This could be done by hand, of course. |
ant-design/ant-design-charts | 1129447164 | Title: 🧐[问题]大佬们,求教最值问题
Question:
username_0: ### 🧐 问题描述 [详细地描述问题,让大家都能理解]
### 💻 示例代码 [如果有必要,展示代码,线上示例,或仓库]

如何像echarts一样指定每条线的最大值最小值,并且自定义图标样式
### 🚑 其他信息 [如截图等其他信息可以贴在这里]
Answers:
username_1: 看下annotations , https://charts.ant.design/zh/examples/component/annotation#text-annotation1
Status: Issue closed
|
openfrontier/openldap-docker | 135823330 | Title: provide an admin GUI
Question:
username_0: It would be very convenient to have an admin GUI such as phpldapadmin integrated into the openldap server.
Answers:
username_1: https://github.com/shivshav/openldap-docker
This project is a fork of my project. They added a phpldapadmin. You may try it.
username_0: Thanks. I see that he has some useful patches on a couple of the other sub-projects as well.
Status: Issue closed
|
rabbitmq/rabbitmq-tutorials | 220524885 | Title: Disposable 1000000 message in a Queues
Question:
username_0: package main
import(
"fmt"
"log"
"github.com/streadway/amqp"
)
func failOnError(err error,msg string){
if err!=nil{
log.Fatal("%s:%s",msg,err)
panic(fmt.Sprintf("%s: %s", msg, err))
}
}
func main(){
conn,err := amqp.Dial("amqp://guest:[email protected]:5672/")
failOnError(err,"Failed to connect to RabbitMQ")
defer conn.Close()
ch,err:=conn.Channel()
failOnError(err, "Failed to open a channel")
defer ch.Close()
q,err :=ch.QueueDeclare("HelloWord",false,false,false,false,nil)
failOnError(err, "Failed to declare a queue")
body:="hello"
for i := 1;i<1000000;i++ {
err = ch.Publish(
"",
q.Name,
false,
false,
amqp.Publishing{
ContentType:"text/plain",
Body:[]byte(body),
})
}
log.Printf(" [x] Sent %s", body)
failOnError(err, "Failed to publish a message")
}
when i was run this code , there is a problem when i push 1000000 message in a Queues
/usr/local/go/bin/go run /Users/neo_qiang/Documents/goWorkSpase/src/rabbitmqText/send.go
panic: sync: inconsistent mutex state
panic: sync: inconsistent mutex state
panic: sync: inconsistent mutex state
goroutine 1 [running]:
panic(0x20ee80, 0xc8202a5840)
/usr/local/go/src/runtime/panic.go:464 +0x3e6
sync.(*Mutex).Lock(0xc82000c12c)
[Truncated]
/Users/neo_qiang/Documents/goWorkSpase/src/github.com/streadway/amqp/channel.go:243 +0x79b
github.com/streadway/amqp.(*Channel).call(0xc8200ac100, 0x18c28f0, 0xc8202a0a60, 0xc820051770, 0x1, 0x1, 0x0, 0x0)
/Users/neo_qiang/Documents/goWorkSpase/src/github.com/streadway/amqp/channel.go:145 +0x79
github.com/streadway/amqp.(*Channel).Close(0xc8200ac100, 0x0, 0x0)
/Users/neo_qiang/Documents/goWorkSpase/src/github.com/streadway/amqp/channel.go:396 +0x199
panic(0x20ee80, 0xc8202a5820)
/usr/local/go/src/runtime/panic.go:426 +0x4e9
sync.(*Mutex).Lock(0xc82000c12c)
/usr/local/go/src/sync/mutex.go:75 +0x16d
github.com/streadway/amqp.(*Connection).send(0xc82000c120, 0x1840638, 0xc8202a0a40, 0x0, 0x0)
/Users/neo_qiang/Documents/goWorkSpase/src/github.com/streadway/amqp/connection.go:347 +0xb8
github.com/streadway/amqp.(*Channel).sendOpen(0xc8200ac100, 0x1840578, 0xc8202b17a0, 0x0, 0x0)
/Users/neo_qiang/Documents/goWorkSpase/src/github.com/streadway/amqp/channel.go:235 +0x57a
github.com/streadway/amqp.(*Channel).Publish(0xc8200ac100, 0x0, 0x0, 0xc8200b8030, 0x9, 0x0, 0x0, 0x2d2550, 0xa, 0x0, ...)
/Users/neo_qiang/Documents/goWorkSpase/src/github.com/streadway/amqp/channel.go:1288 +0x42b
main.main()
/Users/neo_qiang/Documents/goWorkSpase/src/rabbitmqText/send.go:40 +0x30f
exit status 2
Process finished with exit code 1
Status: Issue closed
Answers:
username_1: Thank you for your time.
Team RabbitMQ uses GitHub issues for specific actionable items engineers can work on. This assumes we have a certain amount of information to work with. Questions, investigations, root cause analysis, discussions for potential features are all considered to be mailing list material by our team. When/if we have enough details and evidence we'd be happy to file a new issue.
Please post this to [rabbitmq-users](https://groups.google.com/forum/#!forum/rabbitmq-users). Thank you.
username_1: You are running into [a resource alarm](http://www.rabbitmq.com/alarms.html), the topic of very long queues has been discussed numerous times on the mailing list. At the very least use durable queues and publish messages as persistent. |
cloudfoundry/binary-buildpack | 393573078 | Title: MultiBuild
Question:
username_0: What version of Cloud Foundry and CF CLI are you using? (i.e. What is the output of running cf curl /v2/info && cf version)?
"name": "PCF Small Footprint",
"build": "2.3.0-build.264",
"support": "https://support.pivotal.io"
What version of the buildpack you are using?
cf version 6.40.0+07673feb9.2018-10-08
If you were attempting to accomplish a task, what was it you were attempting to do?
We are planning to add support for instrumenting PCF .NET Core Windows applications using AppDynamics using an extension buildpack. We have achieved the same for applications deployed using dotnet_core_buildpack and hwc_buildpack.
We just have to be able to copy some bits in the supply phase of our extension buildpack and set some environment variables in the finalize phase via profile.d scripts.
What did you expect to happen?
I would expect binary_buildpack to use the supply buildpacks changes.
What was the actual behavior?
Since multi buildpacks are not supported, I am getting a warning
Warning: the last buildpack is not compatible with multi-buildpack apps and cannot make use of any dependencies supplied by the buildpacks specified before it
Just wondering when the support will be added for tackling multi buildpacks.
Please confirm where necessary:
[ x] I have included a log output
[ x] My log includes an error message
[ x] I have included steps for reproduction
Please confirm where necessary:
* [ ] I have included a log output
* [ ] My log includes an error message
* [ ] I have included steps for reproduction
Answers:
username_1: @username_0 thanks for submitting an issue, the buildpacks team is currently in the process of handing the responsibility for the windows related buildpacks off to a different internal team. It looks as though they just added support for extension buildpacks [here](https://github.com/cloudfoundry/binary-buildpack/pull/30).
Status: Issue closed
|
bsolomon1124/pycld3 | 673648313 | Title: Wheel building fails for Python nightly (3.10)
Question:
username_0: Hi, Python 3.10 seems to require a different syntax for wheel building (PEP 517).
I just caught the following error on Travis, see this log (https://travis-ci.org/github/username_0/trafilatura/jobs/715201011):
ERROR: Command errored out with exit status 1:
command: /home/travis/virtualenv/python3.10-dev/bin/python -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-install-suhzcy17/pycld3/setup.py'"'"'; __file__='"'"'/tmp/pip-install-suhzcy17/pycld3/setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(__file__);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' bdist_wheel -d /tmp/pip-wheel-plzmgvyb
...
DEPRECATION: Could not build wheels for pycld3 which do not use PEP 517. pip will fall back to legacy 'setup.py install' for these. pip 21.0 will remove support for this functionality. A possible replacement is to fix the wheel build issue reported above. You can find discussion regarding this at https://github.com/pypa/pip/issues/8368.
Answers:
username_1: I have neglected to get ahead of this so I will now play catch up as soon as I'm able to this week. Thank you for the reminder.
username_1: @username_0 are you able to re-run a build using the 3.10.0 final release or point me to such a failed build?
I see that your build failure occurred with 3.10.0a0. I very much want to help out and ensure compatibility with 3.10, but the package installs OK with or without `wheel` installed for me on MacOS:
```
$ python3 -V
Python 3.10.0
brad ~/Scripts/python/envs
$ python3 -m pip install pycld3
Collecting pycld3
Downloading pycld3-0.22.tar.gz (726 kB)
|████████████████████████████████| 726 kB 3.9 MB/s
Using legacy 'setup.py install' for pycld3, since package 'wheel' is not installed.
Installing collected packages: pycld3
Running setup.py install for pycld3 ... done
Successfully installed pycld3-0.22
WARNING: You are using pip version 21.2.3; however, version 21.2.4 is available.
You should consider upgrading via the '.../bin/python3 -m pip install --upgrade pip' command.
brad ~/Scripts/python/envs
$ python3 -m pip uninstall pycld3
Found existing installation: pycld3 0.22
Uninstalling pycld3-0.22:
Would remove:
.../lib/python3.10/site-packages/cld3/*
.../lib/python3.10/site-packages/pycld3-0.22-py3.10.egg-info
Proceed (Y/n)? y
Successfully uninstalled pycld3-0.22
brad ~/Scripts/python/envs
$ python3 -m pip install wheel
Collecting wheel
Using cached wheel-0.37.0-py2.py3-none-any.whl (35 kB)
Installing collected packages: wheel
Successfully installed wheel-0.37.0
WARNING: You are using pip version 21.2.3; however, version 21.2.4 is available.
You should consider upgrading via the '.../bin/python3 -m pip install --upgrade pip' command.
brad ~/Scripts/python/envs
$ python3 -m pip install pycld3
Collecting pycld3
Using cached pycld3-0.22.tar.gz (726 kB)
Building wheels for collected packages: pycld3
Building wheel for pycld3 (setup.py) ... done
Created wheel for pycld3: filename=pycld3-0.22-cp310-cp310-macosx_10_15_x86_64.whl size=510580 sha256=ccb5474ab97b8d5d0d0178bf0f2e0d145eea2211a386aa29c65932eda25e4291
Stored in directory: /Users/brad/Library/Caches/pip/wheels/45/d5/83/9a7a38189ea5f936288b326094526bbc31fe8a4dc79587fb7a
Successfully built pycld3
Installing collected packages: pycld3
Successfully installed pycld3-0.22
WARNING: You are using pip version 21.2.3; however, version 21.2.4 is available.
You should consider upgrading via the '.../bin/python3 -m pip install --upgrade pip' command.
```
username_0: @username_1 Thanks for the prompt answer!
Here is a current example on GitHub actions, it seems to be the same problem and only affects Python 3.10:
https://github.com/username_0/trafilatura/runs/3817456324
username_0: I'm closing the issue since things work, but you might want to think about a new release targeting Python 3.10 explicitly.
Status: Issue closed
|
schemaorg/schemaorg | 66900037 | Title: Non-human characters under CreativeWork
Question:
username_0: The attribute Character under CreativeWork expects a type Person to be declared. But, as seen in many works, that is not completely true. For example, a vampire is qualified as "Person"? Maybe it has human qualities, but it's not human.
Since that, maybe it's debatable if "Person" should be human or not, but then, how about Scooby Doo? It's a major character in its show, but definitely not a "Person". So, how should I describe those non-human characters?
Answers:
username_1: We can add other types in the range for the relevant attributes
guha
username_0: Then maybe, as a proposal, it could be reworked in such way that there's a new type called Entity (or Being, that represents a living being), and Person should extend from this type. And then, Character should expect a type Entity or Person.
But well, seems that this suggestion is out of the scope of this issue.
username_2: * http://schema.org/character "Fictional person connected with a creative work."
* http://schema.org/Person is "A person (alive, dead, undead, or fictional)."
We don't explicitly tie "Person" to human, although humans would obviously be the central and dominant examples of Person. Undead is (somewhat jokingly) included, so vampires e.g. http://buffy.wikia.com/wiki/Angel
Given this, for now I'd say fine, "Scooby Doo" can be described via Person. Any application that breaks when it encounters a description of Scooby Doo as a Person will probably also have trouble with descriptions of Robin Hood, King Arthur, Angel and Father Christmas. |
budde377/Part | 57644708 | Title: Images load slowly in editor.
Question:
username_0: Create a sprite for images used in the content editor. This will ensure that every image is loaded when opening the editor for the first time.
------
Imported from Bitbucket
Originally reported by: <NAME> (2014-11-22T03:06:15.346) |
eliben/pycparser | 52296513 | Title: Parse error when parsing function declaration
Question:
username_0: When parsing a function declaration like
void f(double a[restrict][5]);
pycparser emits a parsing error. The above example was copied from section 6.7.5.3 (function declarators) of the C standard. I guess the error is due to pycparser not recognizing the restrict keyword in the array declaration.
Answers:
username_1: Can this issue be closed after merging pull request #64 ?
Status: Issue closed
|
shearer/PropCIs | 583090474 | Title: but can you recommend something similar in python ?
Question:
username_0: great package
but can you recommend something similar in python ?
especially for The Clopper–Pearson interval and bounds
Answers:
username_1: Sorry for my late reply.
Take a look at statsmodels
https://github.com/statsmodels/statsmodels
username_0: I did not find there CI for very small samples
you know it is tricky to do
username_2: This is probably a better question for StackOverflow than for the R package maintainer, so I looked there and found the following resources that directly address your question @username_0:
- https://www.statsmodels.org/devel/generated/statsmodels.stats.proportion.proportion_confint.html
- https://stackoverflow.com/questions/13059011/is-there-any-python-function-library-for-calculate-binomial-confidence-intervals/65873479
- https://stackoverflow.com/questions/51794473/calculating-confidence-interval-for-a-proportion-in-one-sample
Status: Issue closed
|
owntracks/ios | 106865176 | Title: StatusTVC.m line 117
Question:
username_0: #### in -[StatusTVC viewWillDisappear:]
* Number of crashes: 1
* Impacted devices: 1
There's a lot more information about this crash on crashlytics.com:
[https://fabric.io/owntracks/ios/apps/org.mqttitude.mqttitude/issues/55f9db31f5d3a7f76b13064f](https://fabric.io/owntracks/ios/apps/org.mqttitude.mqttitude/issues/55f9db31f5d3a7f76b13064f)
Status: Issue closed
Answers:
username_1: happened only 3 times on old 8.2.11 version. Last time on Nov 10 |
transistorsoft/react-native-background-geolocation | 389469621 | Title: locationTemplate does not seem to be working
Question:
username_0: I'm using setTemplate to define a custom template and apply changing parameters through setParams. The plugin
```
[TSLocationManager locationManager:didFailWithError:] Error Domain=kCLErrorDomain Code=0 "(null)"
```
## Your Environment
* Plugin version:2.14.2
* Platform: iOS
* OS version:
* Device manufacturer / model:
* React Native version (`react-native -v`): 0.57.7
* Plugin config
```
[TSLocationManager init] {
activityRecognitionInterval = 10000;
activityType = 1;
autoSync = 1;
autoSyncThreshold = 0;
batchSync = 0;
debug = 1;
desiredAccuracy = "-2";
desiredOdometerAccuracy = 100;
disableElasticity = 0;
disableLocationAuthorizationAlert = 0;
disableMotionActivityUpdates = 0;
disableStopDetection = 0;
distanceFilter = 0;
elasticityMultiplier = 1;
enabled = 1;
extras = {
};
geofenceInitialTriggerEntry = 1;
geofenceProximityRadius = 1000;
geofenceTemplate = "";
headers = {
authorization = "Bearer XYZ";
};
heartbeatInterval = 60;
httpRootProperty = location;
httpTimeout = 60000;
isFirstBoot = 0;
isMoving = 1;
lastLocationAuthorizationStatus = 3;
locationAuthorizationAlert = {
instructions = "You must enable 'Always' in location-services";
titleWhenNotEnabled = "You need to enable background location tracking";
titleWhenOff = "You need to enable location tracking";
};
locationAuthorizationRequest = Always;
locationTemplate = "{ id=\"UpdatePartnerVehicleMutation\", query\": \"mutation UpdatePartnerVehicleMutation(\n $input: UpdatePartnerVehicleInput!\n) {\n updatePartnerVehicle(input: $input) {\n partnerVehicle {\n id\n driver {\n id\n }\n lat\n lng\n }\n }\n}\n\", \"variables\": { \"input\": { \"partnerVehicle\": { \"id\": \"<%= vehicleId %>\", \"lat\":<%= latitude %>,\"lng\":<%= longitude %>,\"driver\": { \"id\": \"<%= driverId %>\" } } } } }";
locationTimeout = 60;
locationsOrderDirection = ASC;
logLevel = 5;
logMaxDays = 3;
maxBatchSize = "-1";
maxDaysToPersist = 1;
maxRecordsToPersist = "-1";
method = POST;
[Truncated]
║ -[TSHttpService flush:error:]
╚═══════════════════════════════════════════════════════════
╔═══════════════════════════════════════════════════════════
║ -[TSHttpService finish:error:] Success: 1
╚═══════════════════════════════════════════════════════════
ℹ️-[TSConfig persist]
ℹ️+[LocationAuthorization run:] status: 3
🎾-[TSLocationManager startMonitoringBackgroundFetch] BackgroundFetch: ON
📍<+37.33444244,-122.04534100> +/- 5.00m (speed 33.15 mps / course 269.30) @ 12/10/18, 6:06:31 PM Brasilia Summer Time
╔═══════════════════════════════════════════════════════════
║ -[TSLocationManager locationManager:didUpdateLocations:] Enabled: 1 | isMoving: 0 | df: -1.0m | age: 136.3s
╚═══════════════════════════════════════════════════════════
ℹ️-[TSLocationManager locationManager:didUpdateLocations:] Received stale motionchange location. Retrying...
⚠️-[TSLocationManager locationManager:didFailWithError:] Error Domain=kCLErrorDomain Code=0 "(null)"
🎾-[TSLocationManager startUpdatingLocation] Location-services: ON
ℹ️+[LocationAuthorization run:] status: 3
```
Answers:
username_1: If you’re changing the config provided to `#ready` during development, be sure to provide `reset: true` or your changes are ignored. Read the docs for `#ready` for more info.
username_1: Are you watching the plugin logs? Are you sure you don’t have old locations rendered with previous template sitting in the plugin’s queue that your server is rejecting? The plugin is going keep posting those locations until your server returns a 200, when the plugin will finally delete them.
username_0: Hmm, that might be the case. Is there any way to view the contents of the queue?
username_0: Thanks. I will review. A hard reset did the trick, but obviously I need to implement a way to avoid these kinds of "background failures"
username_1: Presumably your location template won't change once you finalize it. This is not a "failure". The plugin persists location-data into its database with the `locationTemplate` at the the time the location was recorded. It does **not re-render records in its database** if you change the template later.
The plugin will not delete locations in its database until your server returns http `200`. If you don't, the plugin will keep trying to post that record until your server does so.
If you want to avoid these issues, simply return `200` to the plugin, even if the provided data isn't to the server's taste.
Status: Issue closed
|
amoeba/treestats.net | 467630298 | Title: Add 'template' in views
Question:
username_0: In various views on the site, we could guess the template and put that alongside which would help differentiate characters in lists of players and also make it easy to see what kind of character you're looking at it.
Templates are mostly based on spec'd and trained skills so it'd be easy to make a map of some sort to turn a character's skills into a text string with the template name.
Example templates:
- "Arch" Mage: e.g., http://treestats.net/Morningthaw/Dolt
- Battle Mage (BM): Spec War(Void), Life
- Claw Mage (CLaW): Spec Creature, Life, War
- OG Mage (OG): Spec Creature, Life
Add more to the list and implement. |
random-archer/mkinitcpio-systemd-tool | 300706012 | Title: Failed to start Switch Root
Question:
username_0: 1. `systemd 237` changed mount behavior
* now it will `unmount` the `sysroot.mount` before `initrd-switch-root.target`
2. this produces boot error:
```
Feb 27 10:22:32 work3 systemd[1]: Starting Switch Root...
Feb 27 10:22:32 work3 systemctl[974]: Failed to switch root: Specified switch root path '/sysroot' does not seem to be an OS tree. os-release file is missing.
Feb 27 10:22:32 work3 systemd[1]: initrd-switch-root.service: Main process exited, code=exited, status=1/FAILURE
Feb 27 10:22:32 work3 systemd[1]: initrd-switch-root.service: Failed with result 'exit-code'.
Feb 27 10:22:32 work3 systemd[1]: Failed to start Switch Root.
```
3. the immediate work around is:
* provide `secret` as usual
* switch to debug console: `CTRL+ALT+F8`
* provide manual mount via: `mount /dev/mapper/root /sysroot`
* continue boot process via: `systemctl start initrd-switch-root`<issue_closed>
Status: Issue closed |
plaidml/plaidml | 462411414 | Title: Wrong output dimension with `Dot` layers in plaid ML
Question:
username_0: Here is a simple keras neural network using a `Dot` layer:
```python
from keras.layers import Input, Dense, Dot
from keras.models import Model
model_in = Input((10,), name='input')
model_out = model_in
model_out = Reshape((-1, 1), name='reshape')(model_out)
model_out = Dot(axes=2, name='dot')([model_out, model_out])
model_out = Flatten(name='flatten')(model_out)
model = Model(model_in, model_out)
model.summary()
```
if I compile this network with tensorflow 1.14.0, the output shape is 100 (which is what I expect, as 10*10 = 100):
```python
Layer (type) Output Shape Param # Connected to
==================================================================================================
input (InputLayer) (None, 10) 0
reshape (Reshape) (None, 10, 1) 0 input[0][0]
dot (Dot) (None, 10, 10) 0 reshape[0][0]
reshape[0][0]
flatten (Flatten) (None, 100) 0 dot[0][0]
==================================================================================================
Total params: 0
Trainable params: 0
Non-trainable params: 0
```
However, when I switch my keras backend to `plaidml.keras.backend`, I get an output dimension of 100, which is clearly incorrect:
```python
Layer (type) Output Shape Param # Connected to
==================================================================================================
input (InputLayer) (None, 10) 0
reshape (Reshape) (None, 10, 1) 0 input[0][0]
dot (Dot) (None, 10, 10) 0 reshape[0][0]
reshape[0][0]
flatten (Flatten) (None, 10) 0 dot[0][0]
==================================================================================================
Total params: 0
Trainable params: 0
Non-trainable params: 0
```
I'm using plaidml `0.6.1`. Note that the stated output dimension of the `Dot` layer is correct (10*10), but the output dimension of the `Flatten` layer is incorrect
Answers:
username_1: We've found some issues with the way plaidml is handling `reshape`. I'll replicate this model on my end and see whether I can trace this back to the same issue, or whether this is a new issue entirely.
username_0: Gotcha. I think the error is with the dot layer. Here's an example, where the output should be `(1,2,2)` but instead is `(1, 2)`:
```python
from keras.layers import Input, Dot
from keras.models import Model
import numpy as np
model_in = Input((2,1), name='input')
model_out = Dot(axes=2, name='dot')([model_in, model_in])
model = Model(model_in, model_out)
x = np.zeros((1,2,1))
model.predict(x)
```
Which yields `array([[0., 0.]], dtype=float32)`
What's interesting is the keras model summary gives the correct output shape for `Dot`: `dot (Dot) (None, 2, 2) ` but the prediction shape is `(None, 2)` instead.
If I change the backend to tensorflow, I get the correct shape from the prediction for the same network:
```python
array([[[0., 0.],
[0., 0.]]], dtype=float32)
```
username_1: Thanks so much for this test case! This really helps with my debug process.
You are correct in that there is an issue with the `Dot` layer. In the backend, it is calling `BatchDot`, where we had recently modified the broadcast/concatenation structure to add 2D tensor support for a separate function (`K.gradients`, see #296 if interested). When we were making that change, we were not entirely certain that our implementation was returning the correct shape in all test cases, and this confirms that we are still missing some cases.
I'm continuing to look into this and I'll let you know as soon as I know exactly what we were doing wrong here. For now, I'm targeting this into our next release (0.6.2, est. 7/12)
username_0: Thanks!
username_1: Hey @username_0 - thank you for your patience on this.
We've re-looked at the [documentation for `BatchDot`](https://www.tensorflow.org/api_docs/python/tf/keras/backend/batch_dot) as well as the [tensorflow implementation of `BatchDot`](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/python/keras/backend.py) and have determined that there's a mismatch in the description of `BatchDot` and TensorFlow's implementation. With that info, we've re-implemented `BatchDot` in a way that makes the most sense to us based on the documentation. It fixes the output shapes in both of your test cases, but the actual values within the tensor may be different than what you see in TensorFlow. Our new implementation will make it into the next release.
If you find that you're having correctness issues with the new implementation of `BatchDot`, or simply would like to have it implemented the same way that TensorFlow does, we'd love for you to write your own implementation of `BatchDot`. Our implementation is in `plaidml/keras/backend.py` if that helps you to get started.
username_0: Would it be worth opening an issue on tensorflow's github about the mismatch?
When you say "the actual values in the tensor may differ" what do you mean?
I'll try out my test case again and see if the results make sense. I'm not sure I'm up for writing my own implementation of `BatchDot`, but if I'm feeling ambitious, I'll make a PR!
username_1: It might be worthwhile to open an issue on TensorFlow's github repo. You'd probably want to have a minimal reproduction (i.e. a BatchDot with `x.shape=(1,2,6,2)` and `y.shape=(1,2,2,3)` and `axes = (3, 1)`has an output shape of `(1,2,6,3)`) whereas by the TF definition for output shape "`A tensor with shape equal to the concatenation of x's shape (less the dimension that was summed over) and y's shape (less the batch dimension and the dimension that was summed over). If the final rank is 1, we reshape it to (batch_size, 1).`" sounds like it should have an output shape of (1,2,6,2,3).
We have a testing system that compares value tolerances/shape differences between our implementations and TF's implementations, and when I tested my code against that system, I was seeing some differences in values within the tensors but the same shapes.
I haven't released the code changes yet -- I'll let you know when that happens but for now, it might make sense to file the issue against tensorflow and link us back to that so that we can see whether they are going to change their implementation and/or documentation of `BatchDot`.
username_0: I opened a documentation issue on tensorflow: https://github.com/tensorflow/tensorflow/issues/30846
If the TF devs have questions, will you help answer them?
username_1: I'd certainly be willing to jump in as needed.
Please note that I'm traveling through 7/23 and responses may be slow during this period of time.
username_1: Hey @username_0 - one of my colleagues suggested that I compare `BatchDot` results between TensorFlow and other Keras backends (namely Theano and CNTK). What I'm seeing is that the TensorFlow implementation of `BatchDot` not only mismatches the Keras documentation, but also Theano's implementation (which matches the Keras documentation).
With that being said, I'm going to hold off on any changes that are trying to replicate TensorFlow's behavior when running `BatchDot`. We may instead want to change the `BatchDot` code to match Theano's backend implementation, if TensorFlow buys in to that implementation.
username_0: That's crazy!
username_2: Interestingly, I have an alternate scenario where plaidml has different behavior from both theano and tf
```
from keras import backend as K
x = K.ones((12, 256, 1))
y = K.ones((12, 1, 256))
c = K.batch_dot(x, y)
print(K.int_shape(c))
```
- plaidml @ 0.6.3: `(12, 256)`
- tensorflow @ 1.14.0 `(12, 256, 256)`
- theano @ 1.0.4 `(12, 256, 256)`
username_1: Thanks for the example! Let me look into this.
username_3: Just one more test to demonstrate the problems:
```
import os
import plaidml
os.environ["KERAS_BACKEND"] = "plaidml.keras.backend"
import numpy as np
import keras.backend as K
import tensorflow
from keras.backend import tensorflow_backend as tf
import traceback
print("Plaidml version: %s" % plaidml.__version__)
data = [
[np.ones((32, 1, 2), dtype="float32"), np.zeros((32, 2, 2), dtype="float32"), None],
[np.ones((32, 68, 2), dtype="float32"), np.zeros((32, 2, 2), dtype="float32"), None],
[np.ones((1, 2), dtype="float32"), np.zeros((1, 3, 2), dtype="float32"), (1, 2)]
]
for params in data:
print("\n%s" % ("#"*80))
print("[?] Testing with: %s" % ", ".join(str(x.shape) if isinstance(x, np.ndarray) else str(x) for x in params))
res_tf = res_k = None
params_k = [K.constant(x, dtype=x.dtype) for x in params[:2]] + params[2:]
params_tf = [tf.constant(x, dtype=x.dtype) for x in params[:2]] + params[2:]
try:
res_k = K.batch_dot(*params_k).eval()
except Exception as e:
print("Keras exception: %s" % traceback.format_exc())
with tensorflow.Session():
res_tf = tf.batch_dot(*params_tf).eval()
if res_tf is not None:
print("[?] Tensorflow shape: %s" % str(res_tf.shape))
if res_k is not None:
print("[?] Plaidml shape: %s" % str(res_k.shape))
if res_k is not None and res_tf is not None:
if res_k.shape != res_tf.shape:
print("[-] Shape mismatch")
else:
print("[+] Shape match")
print("[?] Difference: %.5f" % np.mean(np.sqrt(np.sum(np.square(res_tf - res_k)))))
```
=>
```
Plaidml version: 0.6.3
################################################################################
[?] Testing with: (32, 1, 2), (32, 2, 2), None
[?] Tensorflow shape: (32, 1, 2)
[?] Plaidml shape: (32, 2)
[-] Shape mismatch
################################################################################
[?] Testing with: (32, 68, 2), (32, 2, 2), None
Keras exception: Traceback (most recent call last):
File "/home/nope/workspace/test/fs/batch_dot.py", line 43, in <module>
res_k = K.batch_dot(*params_k).eval()
File "/usr/lib/python3.7/site-packages/plaidml/keras/backend.py", line 765, in get_value
shape = invoker.get_output_shape('out')
File "/usr/lib/python3.7/site-packages/plaidml/__init__.py", line 1442, in get_output_shape
return _Shape(self._ctx, _lib().plaidml_alloc_invoker_output_shape(self, name.encode()))
File "/usr/lib/python3.7/site-packages/plaidml/__init__.py", line 777, in _check_err
self.raise_last_status()
File "/usr/lib/python3.7/site-packages/plaidml/library.py", line 131, in raise_last_status
raise self.last_status()
plaidml.exceptions.Unknown: Type check failed due to mismatched tensor sizes: 68.000000 != 2.000000
[?] Tensorflow shape: (32, 68, 2)
################################################################################
[?] Testing with: (1, 2), (1, 3, 2), (1, 2)
[?] Tensorflow shape: (1, 3)
[?] Plaidml shape: (1, 3)
[+] Shape match
[?] Difference: 0.00000
```
Especially the exception seems interesting to me.
username_1: Thank you @username_3 for the example.
Here's what we've decided:
We'll revert the PlaidML implementation of `BatchDot` to match the Theano implementation (and docs), and create a flag that will enable TensorFlow-*like* behavior. If TensorFlow eventually decides to merge with Theano's behavior, then we'll get rid of that code altogether. However, since we already have it, we can create the flag pretty easily.
I'll work on this within the next week and provide you with an update then.
username_1: I've merged in the fixes I mentioned above, see #425 for more details
username_0: Thanks
username_4: Any update on this?
Have tried:
os.environ["PLAIDML_BATCHDOT_TF_BEHAVIOR"] = "True"
Still no luck.
username_0: The TF team claims this was fixed "in a later version", I will see if I can verify: https://github.com/tensorflow/tensorflow/issues/30846
username_0: It looks like the TF team wants to close this issue: https://github.com/tensorflow/tensorflow/issues/30846
username_0: Now that tensorflow is the only backend supported by keras, can we make sure plaidml is using a batchdot that works with tensorflow? |
fossasia/pslab-android | 536955969 | Title: Create a dataset for each instrument
Question:
username_0: The Pocket Sciencelab has many instruments now. To make the functionalities easier to understand, please implement a test for each instrument and store it in the app.
In a second step you can implement it in a way, that these test data sets cannot be deleted. |
sequelize/sequelize | 547177051 | Title: SQLITE_CONSTRAINT: FOREIGN KEY constraint failed
Question:
username_0: I have two tables: User and Course. My User table has a oneToMany relationship with Course table with a foreign key named "userId". When I try to add a new course to the table, I get this error: SQLITE_CONSTRAINT: FOREIGN KEY constraint failed. I have included my tables as well as the POST route that is causing me problems.
**user table**
```
'use strict';
const Sequelize = require('sequelize');
module.exports = (sequelize) => {
class User extends Sequelize.Model {}
User.init({
id: {
type: Sequelize.INTEGER,
primaryKey: true,
autoIncrement: true,
},
firstName: {
type: Sequelize.STRING,
validate: {
notEmpty: {
msg: '"First Name" is required'
}
}
},
lastName: {
type: Sequelize.STRING,
validate: {
notEmpty: {
msg: '"Author" is required'
}
}
},
emailAddress: {
type: Sequelize.STRING,
validate: {
notEmpty: {
msg: '"First Name" is required'
}
}
},
password: {
type: Sequelize.STRING,
validate: {
notEmpty: {
msg: '"Author" is required'
}
}
}
}, {sequelize});
User.associate = (models) => {
User.hasMany(models.Course, {
as: 'instructor',
foreignKey: {
fieldName: 'userId',
allowNull: false,
},
});
}
[Truncated]
})
res.status(201).end();
}catch(error) {
console.error(error);
}
})
```
I am using Postman to test this. In Postman, I type the following:
```
{
"title": "Spanish for Beginners",
"description": "Learn basic Spanish syntax, phonology and conversation.",
"estimatedTime": "6 weeks",
"materialsNeeded": "none",
"userId": 1
}
```
The userId of 1 does exist in the User table so I have no idea what is causing this problem. Any help would be much appreciated. Thank you!!
Answers:
username_1: Hello, please convert your code snippets above into a [sequelize-sscce](https://github.com/username_1/sequelize-sscce), this will save my time and is very important because I have so many things to look at right now. Once you do please post here a link to it and I will reopen the issue. Thanks for understanding.
Status: Issue closed
|
top-think/think-orm | 908244491 | Title: 时间字段写入后实时获取的问题
Question:
username_0: V2.0.40 调整修改器和类型转换的处理机制,导致时间字段写入后实时获取的问题。
在使用静态方法create()或者update() 返回模型的对象实例直接输出,
此时时间字段被重复格式化显示为create_time: "1970-01-01 08:33:41"。
麻烦帮忙查看一下是否需要修复,相关可能影响重复格式化的代码:
vendor/topthink/think-orm/src/Model.php:678行 开始
// 时间戳自动写入
if ($this->autoWriteTimestamp) {
if ($this->createTime && !isset($data[$this->createTime])) {
$data[$this->createTime] = $this->autoWriteTimestamp();
$this->data[$this->createTime] = $this->getTimestampValue($data[$this->createTime]);
}
if ($this->updateTime && !isset($data[$this->updateTime])) {
$data[$this->updateTime] = $this->autoWriteTimestamp();
$this->data[$this->updateTime] = $this->getTimestampValue($data[$this->updateTime]);
}
}
vendor/topthink/think-orm/src/model/concern/Attribute.php:510行 开始
} elseif ($this->autoWriteTimestamp && in_array($fieldName, [$this->createTime, $this->updateTime])) {
$value = $this->getTimestampValue($value);<issue_closed>
Status: Issue closed |
dmlc/dgl | 735355560 | Title: Can't reproduce the results on youtube dataset
Question:
username_0: ## ❓ Questions and Help
I run the deepwalk.py with the same arguments setting on youtube datasets, but the performance is much worse than that presented in the README. So i wonder what causes this. what's the parameter setting in measuring f1-micro and f1-macro?
Answers:
username_1: Hi @username_0 , do you mean this example?
https://github.com/dmlc/dgl/tree/master/examples/pytorch/ogb/deepwalk
username_0: Right. Could i have the code for computing f1-micro and f1-macro? My e-mail address is <EMAIL>, i will really appreciate it. @username_1
username_2: The command we got the f1-micro and f1-macro in the README is here:
`python3 deepwalk.py --data_file youtube --output_emb_file emb.txt --mix --lr 0.2 --gpus 0 1 2 3 --batch_size 100 --negative 5`
It uses 4 gpus to do the training. If you are using single gpu for training, HP should be changed.
username_0: Actually, I have run this command and got the embedding, but the result is still not desirable using my own score code. So i wonder whether there is something wrong on my score code. Could i have your score code? @username_2
username_2: The evaluation methods this example use is here: https://github.com/ShawXh/Evaluate-Embedding
Status: Issue closed
|
louthy/language-ext | 184621607 | Title: language-ext.With (copy update expression)
Question:
username_0: I'm thinking about adding a language-ext.With , in order to work with Lst<T>, Map<K,V>, Set<T>, Que<T>, Stck<T>, HMap<K, V>, and HSet<T>. Basically, taking the copy updat expression code from [With](https://github.com/username_0/With) and gearing it towards language-ext. Thoughts?
Answers:
username_1: For what benefit? `with` is usually a feature of record-types; all 'mutating' of the collection types is handled by the various related functions (like `Add`).
username_0: Exactly. In order to simplify creation of record types in c#.
username_1: I am unclear what that has to do with Lst, Map, Set, Que, Stck, HMap, and HSet?
Whilst I think it's impressive what you've done with the `With` library, I do feel it's an overengineered solution and will add complexity to this library that I'd rather not maintain long-term. It doesn't really add any value, if someone wants that functionality then they can just use your library directly surely? There's no integration piece here from what I can tell. Unless I'm missing something?
And there's a prettier (but more manual) way of doing record-type With methods, which is how I do them right now:
https://stackoverflow.com/questions/38575646/general-purpose-immutable-classes-in-c-sharp/38596298#38596298
Status: Issue closed
username_0: Yes, well, the easiest and prettiest way would be to actually use f# ;) |
team2059/website | 92897667 | Title: Needed: gulp task for extracting files not managed by the repository.
Question:
username_0: Motivation
--------------
Assets such as sponsor logos or fonts have license restrictions that prevent them from being managed in this repository. I have most of them downloaded on my own system, but it would be beneficial to have an easy way for others to download them and place them in the correct directories.
Implementation
--------------------
Since we use gulp to handle other automated tasks, it would make sense to use gulp for this purpose as well. The easiest way to manage the assets would be to have them uploaded as a compressed folder on Drive. The folder can then be placed in the `static` directory and gulp can be configured to unzip it and extract its contents to the correct places. |
mikesteele/dual-captions | 742802250 | Title: On Disney+ it actually doesn't function at all unfortunately
Question:
username_0: I don't see anything other than the app's menu when I try to use it with Disney+. What can be the problem? The app detects both subtitles but still no luck...
Answers:
username_0: Anyone with the same issue? I'd appreciate any tips on working on this problem as well. I'd be happy to help elevating this issue. This extension has been extremely useful for me in learning languages.
username_1: same here
username_2: Isn't the support coming anytime soon? Why not remove "Disney Plus" from the title of the app then?! |
LeetCode-Feedback/LeetCode-Feedback | 1081179071 | Title: Missing Test Case - 1608. Special Array With X Elements Greater Than or Equal X
Question:
username_0: <!--
Note - Any content mention below in `<!-- ->` blocks are just comments
to help you fill-up the issue. It won't be visible in the actual issue after
you click on submit.
-->
#### Your LeetCode username
<!-- Your LeetCode username -->
siddhantsidd15
#### Category of the bug
- [ ] Missing Test Cases
#### Description of the bug
<!-- A clear and concise description of what the bug is. -->
This test case shows error if the logic of the code is not correct.
#### Code you used for Submit/Run operation
<!--
Please make sure you wrap your code with ``` tags.
Otherwise we may reject your request.
-->
```
public int specialArray(int[] nums) {
Arrays.sort(nums);
for(int i = 0; i<101; i++){
int count = nums.length - search(nums, i);
if(count == i)return i;
}
return -1;
}
public int search(int[] nums, int target){
int start = 0;
int end = nums.length-1;
while(start<=end){
int mid = start + (end - start)/2;
if(nums[mid]>= target){
end = mid-1;
}else{
start = mid +1;
}
}
return start;
}
```
#### Language used for code
<!-- C++ -->
Java
#### Expected behavior
<!-- A clear and concise description of what you expected to happen in
contrast with what actually happened. -->
#### Screenshots
<!-- If applicable, add screenshots to explain your issue. -->
#### Additional context
<!-- Add any other additional context about the bug. -->
Answers:
username_1: What is the test case that makes the code fail?
username_2: Hi @username_0,
Thank you for reaching out to us, could you provide the test case that fails the code?
username_0: [0,0,1,4,4,4,4]
Sorry for not mentioning the test case. Here it is.
Initially this test case was causing the code to not pass but later I improved my logic and my code was accepted.
Status: Issue closed
|
mistic100/jQuery-QueryBuilder | 91836299 | Title: Delete rules and groups from the outside
Question:
username_0: This is more a question, than a real issue.
I couldn't found a way to delete a rule or group of rules by JavaScript. How can I delete rules using jQuery before calling ```getRules()``` to avoid submitting invalid rules by trying to clean some common usage mistakes (like empty rule value fields)?
Answers:
username_1: You will need to access the builder data storage, it's available at
```js
$('#builder')[0].queryBuilder.model.root // model might be null, don't forget to test it
```
This is an instance of a **Group** (see http://username_1.github.io/jQuery-QueryBuilder/#inside_the_box)
you can iterate the group with
```js
group.each(true, function(rule) {
}, function(group) {
});
```
there are two callbacks, one for rules and one for subgroup (you might use a self-calling recursive function to traverse the whole tree). The first `true` parameter is to ask a reverse lookup order, which is required if you are going to delete items.
When you have a **Rule** or a **Group** you can access it's data (cf doc) and call `drop` to delete it.
Status: Issue closed
username_0: Thank you very much. |
tlienart/Franklin.jl | 718226020 | Title: Deployment mayhem
Question:
username_0: From fiddling around with the `deploy.yml` it seems that
1. For simple cases (single user), just don't use SSH so remove the `webfactory/` stuff and at the end use `GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}` (e.g. [cormullion's site](https://github.com/cormullion/cormullion.github.io/blob/dev/.github/workflows/deploy.yaml), [this example](https://github.com/username_02/myWebsite/blob/master/.github/workflows/deploy.yml) and [this example](https://github.com/username_02/username_02.github.io/blob/dev/.github/workflows/deploy.yml))
It's a bit unclear when that's "not enough", it might not be enough for org websites (at least I couldn't get that running for the julialang website but that might just be another issue altogether)
1. For more complex cases (org websites), it seems that the above might not work and that SSH *is* required. In that case users should imitate the [setup from julialang website](https://github.com/JuliaLang/www.julialang.org/blob/master/.github/workflows/deploy.yml) which uses SSH: https://github.com/JuliaLang/www.julialang.org/blob/master/.github/workflows/deploy.yml it seems the format of the key matters, they should use PEM and not OpenSSL it seems that the following should be enough:
```
ssh-keygen -m PEM -N "" -f franklin
```
and to copy the content of the private `franklin` to `secrets.FRANKLIN_PRIV`...
this unfortunately needs more testing.<issue_closed>
Status: Issue closed |
evansd/whitenoise | 258342796 | Title: CompressedStaticFilesMixin compresses each CSS file multiple times
Question:
username_0: I noticed as part of #148, that after modifying `CompressedStaticFilesMixin` so that it passes `quiet=False` to `Compressor`, some files are run through the compressor multiple times:
```
$ ./manage.py collectstatic --noinput -v 3 > collectstatic.log
$ grep -e 'Brotli compressed .*\.css ' collectstatic.log | sort
Brotli compressed /home/vagrant/test/test/static/admin/css/base.31652d31b392.css (15K -> 3K)
Brotli compressed /home/vagrant/test/test/static/admin/css/base.6b517d0d5813.css (15K -> 3K)
Brotli compressed /home/vagrant/test/test/static/admin/css/base.6b517d0d5813.css (15K -> 3K)
Brotli compressed /home/vagrant/test/test/static/admin/css/base.css (15K -> 3K)
Brotli compressed /home/vagrant/test/test/static/admin/css/base.css (15K -> 3K)
Brotli compressed /home/vagrant/test/test/static/admin/css/base.css (15K -> 3K)
Brotli compressed /home/vagrant/test/test/static/admin/css/changelists.css (6K -> 1K)
Brotli compressed /home/vagrant/test/test/static/admin/css/changelists.css (6K -> 1K)
Brotli compressed /home/vagrant/test/test/static/admin/css/changelists.css (6K -> 1K)
Brotli compressed /home/vagrant/test/test/static/admin/css/changelists.f6dc691f8d62.css (6K -> 1K)
Brotli compressed /home/vagrant/test/test/static/admin/css/changelists.f6dc691f8d62.css (6K -> 1K)
Brotli compressed /home/vagrant/test/test/static/admin/css/changelists.f6dc691f8d62.css (6K -> 1K)
...
```
This is due to a bug in Django's `HashedFilesMixin`, for which I've filed an upstream issue:
https://code.djangoproject.com/ticket/28607
It only affects "adjustable files" (ie files that `HashedFilesMixin` has to do a second pass on, in case they cross-reference other files that have been hashed, so might need renaming with a new hash) - which currently is just *.css. However it still has a significant impact on the number of compressions performed:
```
$ grep 'Brotli compressed' collectstatic.log | wc -l
370
$ grep 'Brotli compressed' collectstatic.log | sort -u | wc -l
276
```
Answers:
username_1: I've got a workaround for this, but the patch still needs a bit of work before I push it up.
username_1: Finally fixed via this series of commits:
https://github.com/username_1/whitenoise/compare/c3745735677483cb07982a46f3985d35aa3b7647...9e02fa57c0aabd4d8174567f2a9d2ac2e2e21046
Status: Issue closed
|
adriens/chocolatey-zoe | 946914143 | Title: Release v0.27.2
Question:
username_0: See https://github.com/adevinta/zoe/releases/tag/v0.27.2
Answers:
username_0: 
username_0: 
username_0: 
Status: Issue closed
username_0:  |
fossasia/open-event-frontend | 740624352 | Title: Improve Speed: Eliminate render-blocking resources
Question:
username_0: According to the Google Page Speed test https://developers.google.com/speed/pagespeed/insights/?url=eventyay.com eliminating render-blocking resources oould improve the speed by 2 seconds. Please improve the speed by implementing the recommended changes as outlined in this article https://web.dev/render-blocking-resources/.

Answers:
username_1: Same problem as #5575. And even worse part is that we have an SPA, so everything is loaded through JS, so we have little control over deferring the render blocking resources.
If we could use rollup or webpack (like in React or Vue), it would have been somewhat doable, but still not as flexible as a handwritten server side rendered page. But we can't even do the webpack thing, cause ember...
username_1: And also, we have a JS rendered site, so there is no render to be blocked anyway. It's just a blank page. This is also only applicable generally for sites which are rendered on servers like https://www.newsgallery.com/
Status: Issue closed
|
chobits/ngx_http_proxy_connect_module | 276755277 | Title: [TODO] support HTTP2: make CONNECT tunnel work under H2 protocol
Question:
username_0: 1. for how to handle CONNECT tunnel in HTTP protocol, see http://httpwg.org/specs/rfc7540.html#CONNECT
2. some implemention discussion in https://github.com/username_0/ngx_http_proxy_connect_module/issues/22#issuecomment-346944228
Answers:
username_1: This would be great :)
username_2: Hey @username_0 - any updates on this?
username_0: @username_2 Currently not in plan. Hope that I can have free time to complete it or Someone can pull a reqeust for this issue. Details are in first comment. |
bazelbuild/bazel | 1176101882 | Title: On timeout, notify process leader before sending SIGTERM to all processes
Question:
username_0: <!--
ATTENTION! Please read and follow:
- if this is a _question_ about how to build / test / query / deploy using Bazel, or a _discussion starter_, send it to <EMAIL>
- if this is a _bug_ or _feature request_, fill the form below as best as you can.
-->
### Description of the problem / feature request:
I'm trying to make our test suite dump its thread state upon hitting the timeout from bazel. Bazel sends a SIGTERM to all processes at once, making it difficult to do this without doing some weird subprocess+PGID hacks that don't play well with other tools (in our case, pdb).
According to https://github.com/bazelbuild/bazel/blob/09c621e4cf5b968f4c6cdf905ab142d5961f9ddc/src/main/tools/process-tools.cc#L93-L98, there is already a TODO to first notify the process leader and give it a chance to clean up before sending a signal to all child processes.
Something like:
- Send SIGTERM to process leader PID
- Wait x seconds
- Send SIGTERM to all processes
- Wait x seconds
- Send SIGKILL
would be great, or maybe skipping the second SIGTERM and going straight to SIGKILL could work too.
### Feature requests: what underlying problem are you trying to solve with this feature?
Ability to run cleanup code that examines child processes before they get terminated by bazel
### Have you found anything relevant by searching the web?
Found some mentions of Java test runner being able to print thread dumps on timeout, but not sure if it's able to grab child thread dumps before they get terminated. |
hexojs/hexo-renderer-stylus | 212949136 | Title: Publish the latest version to NPM registry?
Question:
username_0: Could you please publish the latest version (0.3.2) to the NPM registry?
And is it possible to publish the newest version to NPM registry when a new version is bumped?
Thanks.
Answers:
username_1: poke @tommy351
username_1: Updated to 0.3.3: https://www.npmjs.com/package/hexo-renderer-stylus
Status: Issue closed
|
fable-compiler/ts2fable | 412966071 | Title: (babylonjs) constants need to be attached to the static, not the instance
Question:
username_0: This doesn't work:
```f#
plane.billboardMode <- billboard.BILLBOARDMODE_ALL
```
This does:
```f#
plane.billboardMode <- 7.
```
I think the issue is that the constants are attached to the instance, when they should be attached to ``AbstractMeshStatic``:

The typescript samples use ``plane.billboardMode = BABYLON.AbstractMesh.BILLBOARDMODE_ALL;``<issue_closed>
Status: Issue closed |
DestinyItemManager/DIM | 425211272 | Title: Pull from Postmaster
Question:
username_0: If a user has 15 modulus reports and some in their postmaster, pulling from postmaster will move all consumables to the vault
Answers:
username_0: ```
fyi, I have 15 in one postmaster (main) and 4 on the other (my alt character). 15 in Inventory obviously
I'll go play for a while and report back later
```
username_0: Maybe
```
if (item.uniqueStack) {
return item.maxStackSize - this.amountOfItem(item) > 0 ? item.maxStackSize - this.amountOfItem(item) : 0;
}
```
Status: Issue closed
|
hapijs/nes | 271215406 | Title: issues when using Docker
Question:
username_0: Hello,
I am experiencing some trouble when using Nes with Docker.
I have a server running Alpine with Node installed. Port 9000 is open and mapped to port 80 on my host machine. I'm also using React and RethinkDB, and the routes i've set up with Hapi work, with the exception of the subscription. I can hit a route within React and it will pull all entries from RethinkDB, I can hit a route with a POST and add a new entry, but the websocket never seems to connect. And eventually will throw..
`Error in connection establishment: net::ERR_CONNECTION_TIMED_OUT`
Has anyone successfully used Nes within Docker yet?
Relevant code:
server
``` App.subscription('/complaints/updates');
App.start(err => {
if (err) {
throw err
}
//Setup the RethinkDB change-feed and push it to the websocket connection.
App.methods.db.setupChangefeedPush();
console.log(`Server running at ${App.info.uri}`)
});```
and front-end
```this.client = new nes.Client('ws://172.18.0.1:9000');
this.client.connect(function (err) {
if (err) {
console.log(err);
return false;
}
var handler = function (item) {
console.log('item is ', item);
App.appendComplaint(item);
};
console.log('connected');
this.client.subscribe('/', handler, function (err) {console.log('err is ', err);});
});```
Answers:
username_1: Anything running between the client and server? nginx? AWS LB? anything at all?
username_0: Nope, no Apache/nginx so far. Just running Docker locally. The app container is just alpine with Node installed.
Sent from my iPhone
Sent from my iPhone
>
username_2: You have `this.client = new nes.Client('ws://172.18.0.1:9000');` in your front end code. You mentioned that 9000 was exposed from the container as port 80 on your local machine. Could the IP and port used by nes on the client be the problem?
username_0: @username_2 you hit the nail on the head; as soon as I read your comment it started to make sense. The client connection needed to be `new nes.Client('ws://localhost:80');`
closing this ticket! thanks everyone!
Status: Issue closed
|
bartbutenaers/node-red-contrib-ui-svg | 721533049 | Title: How to inject full svg source?
Question:
username_0: How to inject a full svg source e.g.
<svg height="150" width="480">
<path d="m0 35.5l6.5-13 9.5 14.5 7-13 11.8 19.7 7.7-13.7 7.8 17 9.4-19.3 9.3 19.3 16-29.3 13.3 21.3 14.7-29.3 14.7 32.6 8.6-18.6 10.7 20.6 11.3-24 12 20 7.4-14.6 12 17.3 10-22 8 14 11.3-24 14 26 7.3-13.3 10.7 19.3 12-24.7 9.7 15 10.3-23.3 12 22.3 6.3-9.3 10.4 14 12-29.3 15.6 31.3 7-13.3 10 16.6 13.4-27.3 6.6 10.7 7.7-16.7 9 19.3 7.3-9.3 11.4 19.3 9.3-17.3 13.3 22 10.7-18 8 11.3 11.3-18 11.9 22 3.8-6.8v181.5h-480v-179.5z" fill="#175720"/>
</svg>
using the inject node from NodeRED.
How to specify the injected msg.payload?
I could not find anything in the docs about that.
Answers:
username_1: Hi Patrick,
It isn't in the documentation, because it is not possible at the moment.
When an entire SVG is loaded:
+ All previous eventhandlers, databindings... need to be cleaned up.
+ All eventhandlers/databindings... need to be applied to the new SVG elements.
Although I can see some useful use cases, I 'think' it might result in quite some redesign. But I haven't looked at this yet in detail...
Bart
username_1: I have created a new branch ***"set-entire-svg-via-msg"*** that implements this feature.
You can install it directly from my Github account from within your .node-red folder:
```
npm install username_1/node-red-contrib-svg#set-entire-svg-via-msg
```
The new feature is explained on the [readme](https://github.com/username_1/node-red-contrib-ui-svg/blob/reload-svg/README.md#set-entire-svg-via-msg) page.
It would be nice if you could test this new feature (incl. event handlers, databinding, ...) before I publish it on NPM, because I'm rather short on free time to test it all by myself at the moment ...
username_0: Thanks for immediate response. I cannot add the branch to npm:
[email protected]: Permission denied (publickey).
fatal: Could not read from remote repository.
Is this branch a private branch? Please make it public.
username_1: No it is all public. Not sure why it fails... I assume you have git installed??
Perhaps [this ](https://stackoverflow.com/a/49484324)might help?
username_1: Damn, I forgot "-ui-" in the node name. Apologies!
It should be:
```
npm install username_1/node-red-contrib-ui-svg#set-entire-svg-via-msg
```
Hopefully it works now...
username_1: Finally at my home computer, after a heavy day at work...
OMG this is so embarrassing.
The command contains not the branch name, but the id of the header id of your new section in the readme.
There is no emoji available to express my feeling at the moment ...
This one works for me:
```
npm install username_1/node-red-contrib-ui-svg#reload-svg
```
username_0: Yes, it works now.
Thank you for the time you put into this.
I will test it as soon as possible.
username_0: Hey Bart,
sorry for the late response, i'm in the middle of my final exams.
I had time for a bit of testing the last days and as far as i could see, everything seems to work.
I was able to change the color of th replaced SVG with the "update_style" command via input message. Animations and (click-)events from the ui-svg-node worked too.
I can't get the add_event command to work, even in the master branch, i don't know why. Maybe you could take a look at this small example i tried:
`[{"id":"6e219eb.05dce6","type":"tab","label":"add event","disabled":false,"info":""},{"id":"3fa0b330.9420dc","type":"ui_svg_graphics","z":"6e219eb.05dce6","group":"588a0f72.d02bb","order":0,"width":0,"height":0,"svgString":"<svg xmlns=\"http://www.w3.org/2000/svg\" height=\"720\" width=\"1280\" viewBox=\"0 0 617.6 411.21\">\n <svg xmlns=\"http://www.w3.org/2000/svg\" viewBox=\"0 0 67.25 155.46\">\n<defs>\n<style>\n</style>\n</defs>\n<g id=\"1_status\" data-name=\"1_status\" style=fill:#919191>\n<circle id=\"1_status\" cx=\"33.63\" cy=\"121.84\" r=\"33.63\"/>\n</g>\n<g data-name=\"1_info\" style= fill:#09f>\n<path id=\"1_info\" d=\"M33.63,0A33.63,33.63,0,1,0,67.25,33.63,33.63,33.63,0,0,0,33.63,0ZM30.89,10.09a3.39,3.39,0,0,1,2.59-1.2A3.61,3.61,0,0,1,36.22,10a3.73,3.73,0,0,1,1.07,2.63,4.27,4.27,0,0,1-1.1,2.91,3.48,3.48,0,0,1-2.71,1.25,3.41,3.41,0,0,1-2.59-1.19,3.84,3.84,0,0,1-1.1-2.71A4,4,0,0,1,30.89,10.09ZM41,53.27H27V51.34l1.39-.12A5.72,5.72,0,0,0,29.64,51a2.18,2.18,0,0,0,1.22-.92,3.34,3.34,0,0,0,.39-1.72V30.53a3.55,3.55,0,0,0-.42-1.71,3.69,3.69,0,0,0-1.19-1.3,5.13,5.13,0,0,0-1.61-.6,11.64,11.64,0,0,0-1.93-.31V24.68L36.29,24l.41.41V48a3.27,3.27,0,0,0,.4,1.7,2.28,2.28,0,0,0,1.23,1,10.32,10.32,0,0,0,1.25.43,7.7,7.7,0,0,0,1.37.22Z\"/>\n</g>\n</svg>\n \n</svg>","clickableShapes":[],"javascriptHandlers":[],"smilAnimations":[],"bindings":[],"showCoordinates":false,"autoFormatAfterEdit":false,"showBrowserErrors":false,"outputField":"","editorUrl":"//drawsvg.org/drawsvg.html","directory":"","panning":"disabled","zooming":"disabled","panOnlyWhenZoomed":false,"doubleClickZoomEnabled":false,"mouseWheelZoomEnabled":false,"name":"SVG graphics node","x":350,"y":60,"wires":[["4d17523e.62babc"]]},{"id":"1af025f.62d4eda","type":"inject","z":"6e219eb.05dce6","name":"add click event","props":[{"p":"payload"},{"p":"topic","vt":"str"}],"repeat":"","crontab":"","once":false,"onceDelay":0.1,"topic":"","payload":"{\"command\":\"add_event\",\"event\":\"click\",\"selector\":\"#1_info\",\"payload\":\"circle has been clicked\",\"topic\":\"inform\"}","payloadType":"json","x":160,"y":60,"wires":[["3fa0b330.9420dc"]]},{"id":"4d17523e.62babc","type":"debug","z":"6e219eb.05dce6","name":"","active":true,"tosidebar":true,"console":false,"tostatus":false,"complete":"true","targetType":"full","statusVal":"","statusType":"auto","x":510,"y":60,"wires":[]},{"id":"588a0f72.d02bb","type":"ui_group","name":"Floorplan","tab":"66b5bbaf.ebf454","order":2,"disp":true,"width":"27","collapse":false},{"id":"66b5bbaf.ebf454","type":"ui_tab","name":"Background","icon":"dashboard","order":2,"disabled":false,"hidden":false}]`
username_1: Hey Patrick,
No problem. Thanks for testing this feature!
About your ***add_event*** problem. When something doesn't work in the SVG node, you should always have a look in the browser's console log, or by activating this checkbox (e.g. for Android devices where it is hard to look in the browser log):

Then you will see this error, as soon as you inject your message:

And when you google about that you will find this:

Your html element ID's start with a number, which is not allowed by the html standard...
So put your numbers at the end of the id's.
Please close this issue when everything is solved, so I know whether I can publish the new version on NPM.
Good luck with your final exams!!!
Bart
username_0: When using the same "add_element" Code in the master branch, everything works and i get no errors
username_1: Nice catch! You found a bug that has been introduced by implementing the full svg source inject feature.
It is now solved when you install it again from Github
Thanks for testing!!!
username_0: Nice!
Thank you very much for your work on this node.
Status: Issue closed
|
hypery2k/galenframework-cli | 118352028 | Title: NPM Install is failing
Question:
username_0: Galen installation failed [Error: not found: galen] Error: not found: galen
at F (/tmp/node_modules/galenframework-cli/node_modules/which/which.js:78:19)
at E (/tmp/node_modules/galenframework-cli/node_modules/which/which.js:82:29)
at /tmp/node_modules/galenframework-cli/node_modules/which/which.js:93:16
at Object.oncomplete (fs.js:107:15)
npm ERR! Linux 3.13.0-68-generic
npm ERR! argv "node" "/usr/local/bin/npm" "install" "galenframework-cli"
npm ERR! node v0.10.2
npm ERR! npm v2.8.4
npm ERR! code ELIFECYCLE
npm ERR! [email protected] install: `node install.js`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the [email protected] install script 'node install.js'.
npm ERR! This is most likely a problem with the galenframework-cli package,
npm ERR! not with npm itself.
npm ERR! Tell the author that this fails on your system:
npm ERR! node install.js
npm ERR! You can get their info via:
npm ERR! npm owner ls galenframework-cli
npm ERR! There is likely additional logging output above.
npm ERR! Please include the following file with any support request:
npm ERR! /tmp/npm-debug.log
```
npm-debug.log
```
13203 verbose unlock done using /home/alwin/.npm/_locks/galenframework-cli-f9601f1bb8fb2270.lock for /home/alwin/git/shop-prototype/node_modules/galenframework-cli
13204 verbose stack Error: [email protected] install: `node install.js`
13204 verbose stack Exit status 1
13204 verbose stack at EventEmitter.<anonymous> (/usr/local/lib/node_modules/npm/lib/utils/lifecycle.js:213:16)
13204 verbose stack at EventEmitter.emit (events.js:98:17)
13204 verbose stack at ChildProcess.<anonymous> (/usr/local/lib/node_modules/npm/lib/utils/spawn.js:24:14)
13204 verbose stack at ChildProcess.EventEmitter.emit (events.js:98:17)
13204 verbose stack at maybeClose (child_process.js:703:16)
13204 verbose stack at Process.ChildProcess._handle.onexit (child_process.js:770:5)
13205 verbose pkgid [email protected]
13206 verbose cwd /home/alwin/git/shop-prototype
13207 error Linux 3.13.0-68-generic
13208 error argv "node" "/usr/local/bin/npm" "install"
13209 error node v0.10.2
13210 error npm v2.8.4
13211 error code ELIFECYCLE
13212 error [email protected] install: `node install.js`
13212 error Exit status 1
13213 error Failed at the [email protected] install script 'node install.js'.
13213 error This is most likely a problem with the galenframework-cli package,
13213 error not with npm itself.
13213 error Tell the author that this fails on your system:
13213 error node install.js
13213 error You can get their info via:
13213 error npm owner ls galenframework-cli
13213 error There is likely additional logging output above.
13214 verbose exit [ 1, true ]
```
I'll try to solve this problem also on my own, if I've any updates I'll let you know.
Answers:
username_1: first you didn't install it globally, e.g. `npm install -g galenframework-cli`
Do you want to use it in the command line or just in a local node build?
username_0: Yes, I want it to be executed on a ci. To define the version in my Package I use a local installation. So I've this line in my package.json:
```
"galenframework-cli": "^2.0.7",
```
username_0: Version 2.1.2 is still working as expected only 2.1.3 is broken (from my point of view)
username_1: you can use galenframework (without -cli). This just download the Galen jar for local run, see https://www.npmjs.com/package/galenframework
username_0: Thanks. It seems to be fixed, I've just to wait until the ci runs threw correctly, but I think that was the issue
Status: Issue closed
username_1: the basic idea of the split was to make a small core module which just downloads the galen jar and one which eases the command line integration with autocompletion and driver installation |
lvgl/lvgl | 658652510 | Title: Code size increase in 7.0+
Question:
username_0: In a discussion with the [Zephyr project](https://github.com/zephyrproject-rtos/zephyr/pull/26654) regarding them updating to 7.0+, it was mentioned that 7.0 has [significantly increased the Flash space requirements](https://github.com/zephyrproject-rtos/zephyr/pull/26654#pullrequestreview-446548076). In some cases the requirement has doubled or even tripled.
In the `size_reduction` branch, I've found and addressed a few config-related issues that resulted in more space being used than necessary, but I think there is still a lot more that can be done to reduce Flash requirements. Ideally, one should be able to turn off most/all of the new features in 7.0 (besides the style system) and get a code footprint similar to that of 6.1.
It appears that the `lv_draw` functions are the main culprit here: together they take up about 10K of space.
Answers:
username_1: All changes look appropriate. Have measured its impact on code size?
I also believe the drawing functions can be more optimal regarding flash usage. (I was focusing on speed rather than code size)
For an estimate how much can we win from 10k? Maybe 3k? It seems there are larges problems here if the code size was doubled.
username_0: It appears to save a few kilobytes when the appropriate options are disabled. If they're enabled then it makes no difference.
In `lv_draw_rect.c` I noticed that `draw_border` and `draw_outline` look quite similar. Are they candidates for de-duplication?
Perhaps some of the logic pertaining to masks can be made configurable? That seems to be a what a lot of the code is dealing with.
Since there's already an option to disable the `opa_scale` property, maybe that should be broadened to disable support for opacity in the whole library? That might not work so well with antialiasing though.
username_1: Yes, it would break anti-aliasing. :slightly_frowning_face: `opa_scale` can be disabled to avoid a lot of inherited style querries usually for nothing.
username_0: Would that be a problem though if `LV_ANTIALIAS` is already disabled?
username_1: It requires a deeper examination to so all the possible issues. However, at first look no opacity could work with no anti-aliasing.
We need to keep in mind to make optional only the really critical parts because too many options can make maintenance quite difficult.
username_2: Hi @username_0, @username_1,
I will check the Zephyr Kconfig files and compare them to the Kconfig on the lv_port_esp32 repo, maybe we can learn something from it or even add the Kconfig file directly on the LVGL repo as planned.
username_0: @username_2 I'm not sure what the license of the Zephyr Kconfig file is; it might be worth looking into that first.
username_3: @username_0 zephyr files are apache2 licensed
@username_2 make sure to pick the Kconfigs from the Zephyr PR mentioned above or you will end up with V6 Kconfig files
Further the zephyr Kconfig follows an approach to only enable a minimal amount of features by default so that the user can pick only the features required by his application.
username_1: I added `draw_full border` for the common parts of outline and border drawing and added the `LV_USE_FONT_SUBPX` option. It makes the code 2-3 kB smaller.
Do you know where are the settings which were used when the code size doubling was measured?
username_3: @username_1 just double checked and it looks like with v6, zephyr has theme support disabled and in the v7 PR the material theme is used. Switching to an empty theme with v7 reduces the rom usage with 4k but still there is a delta of 20k compared to v6.
A quick overview of v6 vs v7 (with empty theme):
```
lib 22233 37.80% | lib 42049 51.72%
gui 22233 37.80% | gui 42049 51.72%
lvgl 22233 37.80% | lvgl 42049 51.72%
src 22233 37.80% | src 42049 51.72%
lv_core 5636 9.58% | lv_core 9760 12.00%
lv_draw 7702 13.10% | lv_draw 20484 25.20%
lv_font 2609 4.44% | lv_font 2791 3.43%
lv_hal 482 0.82% | lv_hal 554 0.68%
lv_misc 3098 5.27% | lv_misc 5340 6.57%
lv_objx 2694 4.58% | lv_widgets 2952 3.63%
lv_themes 12 0.02% | lv_themes 168 0.21%
=========================================== | ==========================================
58812 | 81300
```
Full reports can be found at https://gist.github.com/username_3/607fa993c36558cdb4236d99e7da5f17
username_1: Thanks, @username_3!
In v6, the line an image drawing were disabled when the line and image widgets were disabled. It's not the case in v7. It means ~4K extra ROM usage.
The updates in `size_reduction` branch should reduce with an other ~5k.
Could you check the current state with this branch?
`lv_draw_arc` and `lv_draw_triangle` are not present in v7 report. Might they be ignored during linker optimization?
BTW, what optimization level was used fr the reports? If O0 could you compare with Os too?
username_3: @username_1 tested with the `size_reduction` branch and there was a reduction of 2,7k with `LV_USE_FONT_COMPRESSED` and `LV_USE_FONT_SUBPX` disabled.
Full log:
https://gist.github.com/username_3/607fa993c36558cdb4236d99e7da5f17#file-lvgl-v7-rom-report-size_reduction-eb725a3a
Further zephyr is always building with Os by default
username_1: Thank you.
It's not that much... :slightly_frowning_face:
Most of the difference comes from `lv_draw` (+10k). By excluding `lv_draw_line` and `lv_draw_img` we can win ~4k.
And with some small optimization and others maybe an other ~5k.
Not a big gain. And the new (better) features have a cost. That is we can't expect to keep the same size while adding features.
So maybe we can win a few kB here and there, but to really make a difference we should reconsider a few things. Something like making the core functionalities very lean and push some features to the user to handle in events or so.
Besides with [Simian](https://www.harukizaemon.com/simian/index.html) I measured that there are about 1000 duplicated lines in lv_core+lv_draw+lv_msic and other 1000 in lv_widgets. Looking into the could also help reduce with a few kB.
username_4: With my kind of projects, at 6.1 i have ~ 10-15K of reserve, when development still stays comfortable (i don't like to drop printf & retargeting to USB, floats and so on).
My minimal checkpoint means STM32F0 with 128K flash, and 3-4 subpiхel fonts for android-like high quality interface.
I'd suggest to land quick-fixes and wait with the rest until new breaking changes complete (layout, scrolling, true RGBA for styles, ...). This should help with codebase cleanup and may change optimization priorities too much.
---
I could suggest fonts packing into "catalogs" to reuse kerning tables, but it's too early IMHO.
username_1: `fonts` is a well-separated module and we can treat them independently from the core.
---
In #[1675](https://github.com/lvgl/lvgl/issues/1675#issuecomment-663872276) performance issues were reported and the proposed solution was to make possible disabling some rarely used features.
It will reduce the size further with another few kB.
username_1: @username_0 do you agree to merge `size reduction` branch.
username_0: Yes; sounds good.
username_5: Yes I think that is a good plan, is there a plan of action for it? I could possibly help out with some of it if there are some tasks listed. (Although I am on holiday for two weeks from this coming weekend so will probably be off line for most of that period)
Kind Regards,
Pete
username_1: Have a nice holiday :palm_tree: :relaxed:
username_1: This issue has triggered a lot of rework a lot of things (drop some widget, simplify widgets, etc.) to make LVGL smaller, even smaller than it was in v6.
The updates will be in v8 and v9.
So I close this.
Status: Issue closed
username_6: @username_1 do v8/v9 have an estimate release date?
username_1: We are still considering the features and ideas but I hope it will be ready end of this year. |
RiotGames/developer-relations | 977852355 | Title: [BUG] Unable to invite friends to join practice mode
Question:
username_0: **Bug Description**
Unable to invite friends to join practice mode
Through the client API lol-lobby/v2/lobby to create a practice mode room, invite friends to start the game, too many people return errors
This is what ROIT deliberately did, or is it an unexpected bug
Answers:
username_1: Seems so, the Practice Mode only allows for a single person, so if the lobby has one already, it won't allow for another. It seems a deliberate action, since this was recently abused for clickbait Youtube content.
username_0: Too bad. Clickbait Youtube content, when did this happen, I want to know
I really want to play practice mode with my friends
username_1: https://www.youtube.com/watch?v=BYMHpY83kRY
username_0: I understand that someone created a practice mode with the client API, recorded a video on YouTube and told everyone that he modified the game client code, did these comments have a bad effect on roit?
Don't know if my understanding is correct😂
username_1: Sure
Main takeaway is that the behaviour is correct, seems intentional, and therefore not a bug. Could you close the ticket?
Status: Issue closed
username_0: OK, thanks |
fossasia/pslab-android | 348304907 | Title: License confusion
Question:
username_0: **Actual Behaviour**
Two contradicting licenses are mentioned.
**Expected Behaviour**
A clear license.
**Steps to reproduce it**
Check the [LICENSE file](https://github.com/fossasia/pslab-android/blob/development/LICENSE) against the `README.md`
**LogCat for the issue**
n/a
**Screenshots of the issue**
n/a
**Would you like to work on the issue?**
I'd just like to know what license is the correct one: Apache-2.0 or GPL-3.0-only. While on it, and as background: Someone opened an [MR to include your app with F-Droid](https://gitlab.com/fdroid/fdroiddata/merge_requests/3531). That OK with you? I couldn't find a corresponding issue in your tracker here, so I thought to mention it.
Answers:
username_1: Hi @username_0 ! Let me make the license change in the merge request in GitLab to Apache-2.0.
About the versions, we recently changed the whole package name from the old one. So I reset it to 1 and going to increment from there at each release. :)
username_0: Ah! I see, package name changed. OK, then my fears concerning `versionCode` are unwarranted :wink: Now if you just mark the two comments at Gitlab as "resolved", some of the devs can take over for merging.
I'll leave this issue here open as the Readme [still states GPLv3](https://github.com/fossasia/pslab-android/blob/development/README.md#license). I'd suggest you close it once that line was fixed. Thanks for your fast reaction!
username_1: Sure @username_0 ! Thank you very much for the review! :)
username_0: My pleasure! Thanks for sharing your app and contributing to F-Droid!
Status: Issue closed
|
ankane/chartkick | 565247402 | Title: Feature request: autodisplay closest data to cursor on chart
Question:
username_0: Hovering on a datapoint circle on a chart can be uncomfortable.
It would be great to autodisplay the data from the closest datapoint from the mouse cursor:

Answers:
username_1: Hey @username_0, thanks for the suggestion. This is already planned as the default for Chartkick 4 (https://github.com/username_1/chartkick/issues/489) but may add an option for it earlier since there's not a specific timeline yet.
username_1: Hey @username_0, decided to increase the `hitRadius` for Chartkick 4 instead of showing the closest point by default, but you can use this code to do it: https://github.com/username_1/chartkick/pull/528#issuecomment-623200836. Thanks for the suggestion!
Status: Issue closed
|
vshaxe/vshaxe | 294337079 | Title: Build task freezes when used with haxeshim
Question:
username_0: I'm using haxeshim. It installs `haxe`, `haxe.exe`, `haxe.cmd` in npm's command directory (which is first directory in my PATH containing `haxe`).
When I run the build task, it freezes. I can see the `haxe` process in the process list. Nothing seems to be happening until I manually terminate the task. The `haxe` file itself is a shell script.
I actually do have Cygwin installed (on Windows 7), and if I run `haxe build.hxml` from Cygwin bash shell, it does work. But I think maybe `vshaxe` (or `VS Code`) doesn't know how to run this script correctly, but still tries to run it somehow (in a wrong way).
A workaround that helps is this setting:
```json
"haxe.executable": {
"path": "haxe",
"windows": {
"path": "haxe.cmd"},
}
```
One question I have is what is the point of `haxe` being the default `haxe.executable` on Windows, and not `haxe.exe`?
Answers:
username_1: Seems to work fine for me:

username_2: I can confirm both that this was an issue and that I can't reproduce it anymore ^^
username_1: Nice! :)
Status: Issue closed
|
amcharts/amcharts4 | 474496537 | Title: exporting force directed graph as JSON or CSV
Question:
username_0: Hello,
I added the export menu to a force directed graph but it doesn't seem to have the option to export the chart as a CSV or JSON file.
I followed the documentation and customized the export menu as follows :
```
chart.exporting.menu.items = [{
"label": "...",
"menu": [
{ "type": "png", "label": "PNG" },
{ "type": "json", "label": "JSON" },
{ "label": "Print", "type": "print" }
]
}];
```
however the json option did not appear. Then I used the export API
```
function exportJSON() {
chart.export.toJSON({}, function(data) {
this.download(data, this.defaults.formats.JSON.mimeType, "amCharts.json");
});
}
```
and I got the error : Property 'export' does not exist on type 'ForceDirectedTree'.
So my question is, how can I get the option to export my directed graph as a CSV or JSON file?
Answers:
username_1: Export to JSON works fine on ForceDirected chart:
https://codepen.io/team/amcharts/pen/bb574b863479c13a3583dcfdb3e4828e?editors=0010
Exporting to CSV does not make sense in hierarchical scenarios, though. Especially in those that top nodes' value is auto-calculated from child nodes. But still, it appears in export menu and works.
If you have a demo which does not work, can you post it online so that we can take a look?
Status: Issue closed
|
jlippold/tweakCompatible | 304024888 | Title: `CrashReporter` working on iOS 10.2
Question:
username_0: ```
{
"packageId": "crash-reporter",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "crash-reporter",
"deviceId": "iPhone8,2",
"url": "http://cydia.saurik.com/package/crash-reporter/",
"iOSVersion": "10.2",
"packageVersionIndexed": true,
"packageName": "CrashReporter",
"category": "System",
"repository": "BigBoss",
"name": "CrashReporter",
"packageIndexed": true,
"packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 100% with 1 working reports.",
"id": "crash-reporter",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.0.6",
"shortDescription": "Send useful crash info to developers.",
"latest": "1.15.1-1",
"author": "<NAME> (ashikase)",
"packageStatus": "Working"
},
"base64": "<KEY>
"chosenStatus": "working",
"notes": ""
}
```<issue_closed>
Status: Issue closed |
devSoyoung/yeoksam-restaurant | 503839193 | Title: [후기] 이화수전통육개장 서울역삼점
Question:
username_0: ## 1. 개요
역삼역 근처 일정이 생겨 점심식사를 위해 @username_1 님의 추천으로 해당 레포를 참고하게 되었습니다.
마침 가을이 되어 쌀쌀한 날씨를 달래기위해 점심 메뉴로 육개장(`이화수전통육개장 서울역삼점`)을 선택하게 되었습니다.
## 2. 메뉴
: <del>지옥불야채탕</del> 전설육개장
<img src="https://user-images.githubusercontent.com/26498433/66369816-ce9b4200-e9d8-11e9-8793-14616599e4ed.jpg" width=400>
## 3. 후기
- 원래 요리에 들어간 파를 잘 먹지 않는 편인데, 중간중간 식감이 일품이었습니다. <del>다 먹었습니다.</del>
- 국물은 좀 짠 편이었지만, 제가 좀 싱겁게 먹는 편이라 다른 분들은 짭짤하게 잘 드실 수 있을 것 같습니다.
- 출구에서 조금만 걸어가면 있는 곳이라 접근성은 좋았습니다.
- 약 1시 정도에 방문했을때, 자리는 충분했습니다. 만약 12시~1시(회사 점심시간)에 방문한다면 자리가 없을 것 같습니다.
<img src="https://user-images.githubusercontent.com/26498433/66370069-cc85b300-e9d9-11e9-8364-d24668403a46.jpg" width=400>
Answers:
username_1: 첨부된 사진이 본연의 모습을 담지 못하는 것 같아 아쉬운 마음에 추가합니다. 양해바랍니다.

데브소영님 덕분에 든든히 채우고 가네요 😆🥄 |
arnaud-lb/php-rdkafka | 436006490 | Title: Documentation
Question:
username_0: Dear Arnaud
I was unable to find if i can create a PR to the documentation that is right now on github.io.
If that is not possible, maybe keeping the doc in the wiki or in the repo itself would be nice,
so we can help contribute to the documentation.
I opened an issue for a possible error a while back #178
Also it would be nice for 3.1.0 to have the new functions producev (support for headers) and RdKafka\KafkaConsumer::getCommittedOffsets(array $topics, int timeout_ms) documented.
Let me know what you think.
Cheers,
Nick
Answers:
username_1: Hi Nick
The docs repository can be found at https://github.com/username_1/php-rdkafka-doc. PRs welcome :)
Status: Issue closed
|
JuliaLang/julia | 351871601 | Title: "tuple must be non-empty" when iterating two nested `Iterators.product`
Question:
username_0: The code:
using Base.Iterators: product
println(collect(product(product())))
Expected:
(ProductIterator{Tuple{}}(()),)
(or something of that sort, in the real code I have
iters = product(product(...))
for x in iters
for y in x
...
end
end
)
Got:
ERROR: LoadError: ArgumentError: tuple must be non-empty
Stacktrace:
[1] first(::Tuple{}) at ./tuple.jl:74
[2] _pisdone at ./iterators.jl:803 [inlined]
[3] isdone at ./iterators.jl:810 [inlined]
[4] _pisdone at ./iterators.jl:804 [inlined]
[5] isdone at ./iterators.jl:810 [inlined]
[6] iterate at ./iterators.jl:842 [inlined]
[7] copyto!(::Array{Tuple{Tuple{}},0}, ::Base.Iterators.ProductIterator{Tuple{Base.Iterators.ProductIterator{Tuple{}}}}) at ./abstractarray.jl:650
[8] _collect(::UnitRange{Int64}, ::Base.Iterators.ProductIterator{Tuple{Base.Iterators.ProductIterator{Tuple{}}}}, ::Base.HasEltype, ::Base.HasShape{0}) at ./array.jl:563
[9] collect(::Base.Iterators.ProductIterator{Tuple{Base.Iterators.ProductIterator{Tuple{}}}}) at ./array.jl:557
[10] top-level scope at none:0
[11] include at ./boot.jl:317 [inlined]
[12] include_relative(::Module, ::String) at ./loading.jl:1038
[13] include(::Module, ::String) at ./sysimg.jl:29
[14] exec_options(::Base.JLOptions) at ./client.jl:229
[15] _start() at ./client.jl:421
`versioninfo()`:
Julia Version 1.0.0
Commit <PASSWORD> (2018-08-08 20:58 UTC)
Platform Info:
OS: macOS (x86_64-apple-darwin14.5.0)
CPU: Intel(R) Core(TM) i7-4870HQ CPU @ 2.50GHz
WORD_SIZE: 64
LIBM: libopenlibm
LLVM: libLLVM-6.0.0 (ORCJIT, haswell) |
stopbystop/sbs-app | 185839234 | Title: Scenario 2 - iphone - 'Seattle' option not displayed in drop down.
Question:
username_0: Scenario 2
Platform - iPhone
Steps:
Enter 'Seat' in the first text box.
Search for 'Seattle' option.
Expected Result:
'Seattle' option should be displayed in the drop-down list.
Actual Result:
'Seattle' option is not displayed in the drop-down list.
Answers:
username_1: Cannot reproduce. Please reactivate it it happens again.
Status: Issue closed
|
Anouar-Gherri/Team3 | 525694615 | Title: GUI: Anzeigen des kommenden Spieltags
Question:
username_0: Erweitert eure graphische Oberfläche, damit sie immer automatisch den nächsten Spieltag anzeigt.
Dabei sollten alle Begegnungen in einer Tabelle angezeigt werden und den nächsten noch ungespielten Spieltag anzeigen.<issue_closed>
Status: Issue closed |
ash16/TYTS | 382497115 | Title: Unused-argument in mousePressEvent(event)
Question:
username_0: The problem is reported when an argument is not used in the body of its function or method. In the project code it is referred to the event and text.
Module TYTS-master.buddingWidget
W: 26,30: Unused argument 'event' (unused-argument) |
jlippold/tweakCompatible | 416041151 | Title: `AppCake` working on iOS 12.1.1
Question:
username_0: ```
{
"packageId": "cydia.com.iphonecake.appcake6",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "cydia.com.iphonecake.appcake6",
"deviceId": "iPhone10,5",
"url": "http://cydia.saurik.com/package/cydia.com.iphonecake.appcake6/",
"iOSVersion": "12.1.1",
"packageVersionIndexed": false,
"packageName": "AppCake",
"category": "Utilities",
"repository": "AppCake",
"name": "AppCake",
"installed": "6.0.8",
"packageIndexed": true,
"packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.",
"id": "cydia.com.iphonecake.appcake6",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.1.1",
"shortDescription": "The new AppCake 6! For iOS 11 & 12 Jailbreak.",
"latest": "6.0.8",
"author": "hotsjf",
"packageStatus": "Unknown"
},
"base64": "<KEY>
"chosenStatus": "working",
"notes": ""
}
``` |
MicrosoftDocs/OfficeDocs-Enterprise | 577650221 | Title: Response to #716
Question:
username_0: @username_1
Thank you for looking into this. My intention was to say that any allowed characters allowed to use other than "-" and "_" will result in getting synced as "_". It says we can contain " !, #, $, %, &, ', *, +, -, /, =, ?, ^, _, `, {, |, } and ~" but it actually only allows "-" and "_" to be synced properly. Please test this and verify.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: c60ee789-fce6-6c95-150d-be4ae895730e
* Version Independent ID: d9e96954-3b85-bf8d-c7d3-916a4698fd77
* Content: [Prepare for directory synchronization to Office 365](https://docs.microsoft.com/en-us/office365/enterprise/prepare-for-directory-synchronization)
* Content Source: [Enterprise/prepare-for-directory-synchronization.md](https://github.com/MicrosoftDocs/OfficeDocs-Enterprise/blob/live/Enterprise/prepare-for-directory-synchronization.md)
* Service: **o365-administration**
* GitHub Login: @username_2
* Microsoft Alias: **josephd**
Answers:
username_0: Somehow, "_" (underscore) has been deleted from my text, so re-posting with clarifications.
Thank you for looking into this. My intention was to say that any allowed characters allowed to use other than "- (hyphen)" and "_ (underscore)" will result in getting synced as "_ (underscore)". It says we can contain " !, #, $, %, &, ', *, +, -, /, =, ?, ^, _ (underscore), `, {, |, } and ~" but it actually only allows "-" and "_ (underscore)" to be synced properly. Please test this and verify.
username_1: @username_0 Thank you for the feedback. We will test this and update shortly.
username_1: @username_0 Thank you for your feedback.I was able to test this in multiple Instances and the document is upto date. I was able to sync the below characters without any issues
" !, #, $, %, &, ', *, +, -, /, =, ?, ^, _ , `, {, |, } and ~"

Thanks
Sri
Status: Issue closed
username_0: Than you so much for testing! However, my point only applies to "mailNickname." Please test with this attribute
username_1: @username_0 Thank you for the response. I see that there is a Uservoice for this particular attribute https://office365.uservoice.com/forums/273493-office-365-admin/suggestions/39013438-fix-the-behavior-that-the-character-set-as-the-mai
We have updated the note on the document to reflect the same. Reference PR #https://github.com/MicrosoftDocs/OfficeDocs-Enterprise/pull/768
Thanks
Sri |
Macaulay2/M2 | 33490720 | Title: Matrices are not printed correctly when the coefficient field was declared with toField
Question:
username_0: The routine that prints a matrix to the screen seems to fail when the coefficient field was constructed with "toField". Here is an example:
i1 : R = (toField (QQ[z]/(z^2+1)))[x]
i2 : matrix {{x^2+x+1}}
o2 = | 1x21x1 |
I could accept the coefficient "1" being printed, but the "+" are missing. Note that printing outside of a matrix works. Also note that when leaving away toField also works:
i3 : R = (QQ[z]/(z^2+1))[x]
i4 : matrix {{x^2+x+1}}
o4 = | x2+x+1 |
Answers:
username_1: This issue looks like it has been fixed, without closing this issue. I'm going to close it now.
Status: Issue closed
|
finos/perspective | 926664141 | Title: Apache Arrow featherV2 read function
Question:
username_0: When I read Apache Arrow featherV2, I get the following error.
Table initialization failed: Error: Abort(): Failed to read file record batch: NotImplemented: LZ4 codec support not built
It would be appreciated if you could implement Apache Arrow featherV2 read function.
Answers:
username_1: The solution here might be to bump the Arrow version. This project seems to still use 1.0.1, and LZ4 compression seems to have been enabled by default in Arrow version 4.0.0
username_2: I am certain this still won't work even with Arrow 5.0. We don't yet build LZ4 for wasm for starters, it is "enabled" by default in that the python API will default the `compression` kwarg but even this depends on the python bindings for LZ4. |
ngrx/platform | 412592890 | Title: Schematics: Default effect leads to an endless loop
Question:
username_0: The command `ng generate feature` does create an effects class with the following effect (example):
```ts
loadBooks$ = this.actions$.pipe(ofType(BookActionTypes.LoadBooks));
```
(see https://github.com/ngrx/platform/blob/master/modules/schematics/src/effect/files/__name%40dasherize%40if-flat__/__name%40dasherize__.effects.ts#L24)
If left untouched, this will lead to a [loop of death](https://angular.schule/blog/2018-06-5-useful-effects-without-actions#5-loop-of-death-) as soon as the first `LoadBooks` action is being dispatched.
## Possible solutions
* Comment out the `@Effect()` decorator by default
* don't provide a default effect after creation
* set `{ dispatch: false }`
I don't like any of them – but I ran into that the infinite loop problem multiple times now 😆
Will prepare a PR if there is consensus about how we can solve this.
## I would be willing to submit a PR to fix this issue
[x] Yes (Assistance is provided if you need help submitting a pull request)
[ ] No
Answers:
username_1: We recently merged PR #530 which resolves this issue 😅 .
It will become available in the next version.
Status: Issue closed
username_2: #1530 was released with `7.2.0` 👋
username_1: The docs aren't reflecting these changes, would it be OK to open up a new issue to add the `api` flag introduced in the PR?
username_0: Thanks for the quick reply!
However, I'm not sure https://github.com/ngrx/platform/pull/1530 actually fixes this.
With `ng g feature foobar --api` I actually get a "harmless" effect – but `ng g feature foobar` still produces this one:
```
@Injectable()
export class FoobarrrrEffects {
@Effect()
loadFoobarrrrs$ = this.actions$.pipe(ofType(FoobarrrrActionTypes.LoadFoobarrrrs));
constructor(private actions$: Actions) {}
}
```
username_2: It's easier to delete the code than to write it from scratch. I think we should at least add a note in the docs that the generated effect must be updated to return a new action.
I don't usually dispatch the action until the effects are wired up but I can see how it causes an issue.
username_1: You're right @username_0.
We discussed the possible solutions in the issues thread, https://github.com/ngrx/platform/issues/1524#issuecomment-457450857.
With @username_0's proposed solutions in mind, should we re-open this issue or leave this as is. If we re-open the issue do we still agree that commenting out the effect is the best solution?
username_2: I want to put developers on the path of continuing with minor modifications. I'd recommend we do something similar to the `--api` and `concatMap` to an EMPTY observable by default.
username_0: Sounds good to me! I could live with both ways:
(1) concatMap to EMPTY
```
@Effect()
loadFoobars$ = this.actions$.pipe(
ofType(FoobarActionTypes.LoadFoobars),
concatMap(EMPTY) // remove or replace with your own side effect execution
);
```
(2) comment out
```
// @Effect()
// loadFoobars$ = this.actions$.pipe(ofType(FoobarActionTypes.LoadFoobars));
```
----------
As I find (1) a sensible solution, should I submit a PR for this?
@username_2 @username_1
username_0: Slightly related:
`ng g feature foobar --api` injects `Actions` with the new way:
```
private actions$: Actions<FoobarActions>
```
while `ng g feature foobar` still uses the old way where we need to type `ofType` explicitly:
```
private actions$: Actions
```
Am I right assuming that this is not intended and should be changed as well? 😅
username_1: The command `ng generate feature` does create an effects class with the following effect (example):
```ts
@Effect()
loadBooks$ = this.actions$.pipe(ofType(BookActionTypes.LoadBooks));
```
(see https://github.com/ngrx/platform/blob/master/modules/schematics/src/effect/files/__name%40dasherize%40if-flat__/__name%40dasherize__.effects.ts#L24)
If left untouched, this will lead to a [loop of death](https://angular.schule/blog/2018-06-5-useful-effects-without-actions#5-loop-of-death-) as soon as the first `LoadBooks` action is being dispatched.
## Possible solutions
* Comment out the `@Effect()` decorator by default
* don't provide a default effect after creation
* set `{ dispatch: false }`
I don't like any of them – but I ran into that the infinite loop problem multiple times now 😆
Will prepare a PR if there is consensus about how we can solve this.
## I would be willing to submit a PR to fix this issue
[x] Yes (Assistance is provided if you need help submitting a pull request)
[ ] No
username_1: I would also prefer option one 😄
The reason why `ng g feature foobar` doesn't use the new way, is because we don't need to infer the payload of the action. But perhaps it would be better to still type the actions? 🤔 .
username_2: It should still be typed when generating a feature because we have an actions union.
Status: Issue closed
|
uamz-vertnet/uamz-herpetology | 130985805 | Title: Monthly VertNet data use report for 2015-12, resource uamz-herpetology
Question:
username_0: Your monthly VertNet data use report is ready!
You can see the HTML rendered version of the reports with this link:
http://tools-usagestats.vertnet-portal.appspot.com/reports/88d7437e-f762-11e1-a439-00145eb45e9a/201512/
Raw text and JSON-formatted versions of the report are also available for download from this link. In addition, a copy of the text version has been uploaded to your GitHub repository, under the "Reports" folder.
Also, a full list of all reports can be accessed here:
http://tools-usagestats.vertnet-portal.appspot.com/reports/88d7437e-f762-11e1-a439-00145eb45e9a/
You can find more information on the reporting system, along with an explanation of each metric, here: http://www.vertnet.org/resources/usagereportingguide.html
Please post any comments or questions to http://www.vertnet.org/feedback/contact.html
Thank you for being a part of VertNet. |
Chadster766/McDebian | 502970883 | Title: McDebian Stretch Upgrade to Buster 4.19.75 Beta
Question:
username_0: # McDebian Stretch Upgrade to Buster 4.19.75 Beta
## Notes:
I recommend that only users that have TTL access to their WRT routers do McDebian beta testing.
## Firmware:
```
wget --user=mcdebian --password=<PASSWORD> http://www.protechs-online.com/downloads/McDebian/firmwares/McDebian-Buster-WRT1900AC-V1-FW_VER1_kernel_4.19.75.img
wget --user=mcdebian --password=<PASSWORD> http://www.protechs-online.com/downloads/McDebian/firmwares/McDebian-Buster-WRT1900AC-V2-FW_VER1_kernel_4.19.75.img
wget --user=mcdebian --password=<PASSWORD> http://www.protechs-online.com/downloads/McDebian/firmwares/McDebian-Buster-WRT1200AC-V1-FW_VER1_kernel_4.19.75.img
wget --user=mcdebian --password=<PASSWORD> http://www.protechs-online.com/downloads/McDebian/firmwares/McDebian-Buster-WRT3200ACM-V1-FW_VER1_kernel_4.19.75.img
wget --user=mcdebian --password=<PASSWORD> http://www.protechs-online.com/downloads/McDebian/firmwares/McDebian-Buster-WRT32X-V1-FW_VER1_kernel_4.19.75.img
```
Follow "Flashing firmware from McDebian" instruction to upgrade to McDebian 4.19.75 firmware.
https://github.com/username_0/McDebian/wiki/2.-Firmware-Uploading#flashing-firmware-from-mcdebian
Below is my recommended upgrade path:
```
#McDebian Stretch Kernel 4.19.75 Updates
cd /
wget --user=mcdebian --password=<PASSWORD> http://www.protechs-online.com/downloads/McDebian/rootfs-updates/mcdebian-rootfs-kernel-modules-firmwares-updates-v4.19.75-ver1.gz
rm -rf /lib/modules/*
rm -rf /lib/firmware
tar zxfk mcdebian-rootfs-kernel-modules-firmwares-updates-v4.19.75-ver1.gz
#Upgrading McDebian Stretch to Buster
sed -i 's/stretch/buster/g' /etc/apt/sources.list
apt-get update
apt-get upgrade
#Recommended answers to questions during the upgrade process:
#Restart services during package upgrades without asking? (Answer No)
#Services to restart for PAM library upgrade: cron (Select <Ok>)
#Start the MiniSSDP daemon automatically? (Answer No)
#Interfaces to listen on for UPnP queries: (Answer Ok)
#Configuration file question I recommend "N" unless you want to compare the new against the old and take what you need to update the config file with "Z" option.
#What do you want to do about modified configuration file sshd_config? (Answer "keep the local version currently installed")
apt-get full-upgrade
apt --purge autoremove
reboot
#Disable minissdpd if you like
systemctl disable minissdpd
#Changes to hostapd server to use multiple config files
cd /lib/systemd/system
rm hostapd.service
wget
https://github.com/username_0/McDebian/raw/master/rootfs/lib/systemd/system/hostapd.service
reboot
#Confirm upgrade is complete
cat /etc/debian_version
uname -a
```
Answers:
username_0: This has been superceded by McDebian release 4.19.91:
https://github.com/username_0/McDebian/wiki/X.-McDebian-Updates
Status: Issue closed
|
mozilla/addons-server | 229028845 | Title: Search will not list any search results even if there are some suggestions
Question:
username_0: Steps to reproduce:
1. Load AMO homepage https://addons-dev.allizom.org/en-US/firefox/
2. Make a search after "423"
Expected results:
If search suggestions are displayed, after finalizing the search, results should be listed.
Actual results:
One search suggestion is displayed, but clicking on the green arrow or hitting "enter" will not display the search suggestion.
Notes/Issues:
- other example "123" - will list 1 search suggestion but the suggestion will not be listed through search results.
Verified on FF53(Win 7). Issues is reproducing on all AMO servers.
Video for this issue:

Status: Issue closed
Answers:
username_1: Only happens when you search for numbers, and it's intended - there is a shortcut built into the suggestion view (and only there, apparently) that gives you the add-on with the corresponding ID. |
desktop/desktop | 520405867 | Title: Can't loggin in Bitbucket.
Question:
username_0: Login&password is correct! But I can't to loggin. Fetchins, pulling don't work. From console I can do
this.

Status: Issue closed
Answers:
username_1: Hi @username_0, thanks for the issue. Here's the documentation for authenticating with Bitbucket: https://github.com/desktop/desktop/blob/development/docs/integrations/bitbucket.md
Let us know if that doesn't resolve it! |
ecstatic-morse/fixed-bitset | 808021575 | Title: unchecked methods
Question:
username_0: Since this kind of struct seems perfect for performance sensitive situations, it would be nice if there were also unchecked versions of insert/erase. I'm open to writing a PR if this is still maintained? |
Mailu/Mailu | 563852701 | Title: Support ARGON2(I/D) algoritm for password hashing
Question:
username_0: ARGON2(I/D) is prefereable algoritm for hashing password. And increases popularity. Better to add support of ARGON2 for comfortable import data from another systems.
Answers:
username_1: Making it the default is a difficult choice, especially due to performance given the amount of password bruteforce on the internet.
However, definitely in favor of adding the option.
username_2: Hardened hash algorithms are a bad choice for non-session based authentication because the password needs to be verified for every single request. So with such a hash (BCrypt or similar) which is designed to run slow and/or to consume lots of resources a single webmail user can saturate all resources of a server.
(example: https://www.endpoint.com/blog/2016/02/09/bonked-by-basicauth-because-bcrypt)
It's the same for all hashes of this type. Using apr1 or similar is the best choice here.
username_3: Hi There,
The `Mailu`-Project is currently in a bit of a bind! We are short on man-power, and we need to judge if it is possible for us to put in some work on this issue.
To help with that, we are currently trying to find out which issues are actively keeping users from using `Mailu`, which issues have someone who want to work on them — and which issues may be less important. These a less important ones could be discarded for the time being, until the project is in a more stable and regular state once again.
In order for us to better assess this, it would be helpful if you could put a **reaction on this post** (use the :smiley: icon to the top-right).
- 👍️ if you **need this** to be able to use Mailu. Ideally, you’d also be able to test this on your installation, and provide feedback …
- 🎉 if you find it a **nice bonus**, but no deal-breaker
- 🚀 if you want to **work on it yourself**!
We want to keep this **voting open for 2 weeks** from now, so please help out!
username_3: Hi There,
we see this issue had only little attention 😟. As much as it pains us: In order to get the current issues a bit more manageable for us, we decided to close it. 👋👋
We hope that no one feels offended by doing so. Should the issue get really pressing in the future, please feel free to re-open it.
Thank you for your patience and understanding, 🙇
- Your Mailu Team
Status: Issue closed
username_4: We now support importing argon hashes in master... they'll be converted on first use to what is current (bcrypt+sha256) |
WebJamApps/combined-front | 269343912 | Title: As a app owner, I expect that new users (not Google users) who have registered for an account are using valid email addresses
Question:
username_0: * system sends an email to their email address as part of the registration process
* sends a random generated number
* stores random gen number in user model
* sends link to frontend page where user must enter the number and click submit
* verify email form has form validation
* post checks for user where email = email and gen number = gen number
if found, it sets resetCode = ''
* if not found it sends an error response and message back to verification form (frontend)
* frontend error message is displayed if present else hide form and route to homepage
* Login form checks for verification code that is set to "" or it sends error message to user requesting that they verify their email address |
AlexGustafsson/homebridge-wol | 877711067 | Title: Host Key verification failed
Question:
username_0: Hello there, I am trying to configure the WOL plugin however I receive this error, I am using hoobs :(
Thanks!
6/5/2021, 18:32:26 [Mac] NetworkDevice awake cycle started for "Mac" (192.168.4.195)
6/5/2021, 18:32:26 [Mac] Attempting to wake up "Mac" (192.168.4.195)
6/5/2021, 18:32:26 [Mac] NetworkDevice "Mac" (192.168.4.195) went from status "Offline" to "Waking Up"
6/5/2021, 18:32:26 [Mac] Attempting to start "Mac" (192.168.4.195) by sending a WoL packet
6/5/2021, 18:32:27 [Mac] Waiting for wake grace time (10000ms) for "Mac" (192.168.4.195)
6/5/2021, 18:32:29 [HOOBS A0AC@On] The write handler for the characteristic 'On' on the accessory 'Mac' was slow to respond!
6/5/2021, 18:32:36 [HOOBS A0AC@On] The write handler for the characteristic 'On' on the accessory 'Mac' didn't respond at all!. Please check that you properly call the callback!
6/5/2021, 18:32:37 [Mac] Attempting to wake up "Mac" (192.168.4.195) using "ssh"
6/5/2021, 18:32:37 [Mac] An error occured while trying to wake "Mac" (192.168.4.195): Error: Command failed: ssh mac.local caffeinate -u -t 300
Host key verification failed.
at ChildProcess.exithandler (child_process.js:308:12)
at ChildProcess.emit (events.js:315:20)
at maybeClose (internal/child_process.js:1048:16)
at Socket.<anonymous> (internal/child_process.js:439:11)
at Socket.emit (events.js:315:20)
at Pipe.<anonymous> (net.js:673:12)
{
killed: false,
code: 255,
signal: null,
cmd: 'ssh mac.local caffeinate -u -t 300',
stdout: '',
stderr: 'Host key verification failed.\r\n'<issue_closed>
Status: Issue closed |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.