repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
jonathan-laurent/AlphaZero.jl
970880804
Title: Performance Docs Question: username_0: You should really update the docs and example launch commands to explain to people the proper settings to get the most performance. Specifically using -t <V_cpus> and adjusting batchsize to get the most gpu usage possible Answers: username_1: This is a good suggestion. I think using -t <V_cpus> is not necessary with Julia 1.6+ as it uses as many threads as there are CPU available by default. But you are right I could tell something about modifying batch size. By the way, if you find something missing in the doc, don't hesitate to have a go at adding it yourself and submit a PR! username_0: Ok once I wrap my head around this and manage to fully train I will :) Status: Issue closed
SharePoint/sp-dev-docs
604457697
Title: Are shared documents Lists? Question: username_0: ## Category - [x] Question - [ ] Typo - [ ] Additional article idea ## Question In the doc https://docs.microsoft.com/en-us/sharepoint/dev/apis/webhooks/overview-sharepoint-webhooks, it has been mentioned that webhooks can only be applied to lists. But currently we are uploading our data to the Site's root Documents (in a folder within Shared Documents). When I search for all lists in the site through a Graph Explorer, it seems to have one as Shared Documents as well. So are these shared documents a list as well and can I apply webhooks on these items? Answers: username_1: Hello @username_0 in SharePoint, the term list means document libraries and lists (e.g. custom lists). The documentation says that webhooks are only enabled for SP list items which doesn't exclude document libraries! Since Shared Documents is a document library, you can also apply webhooks on these items. I hope it helps to clarify your question. Regards, Jarbas Status: Issue closed
bazelbuild/tulsi
735871524
Title: Ease of Install: Brew Cask Question: username_0: Tulsi/Bazel folks, thanks for releasing a great tool to the world! I was wondering if you'd consider making a brew cask for Tulsi. That'd make install and update even smoother. Thanks, Chris Answers: username_0: Self editing: Maybe even better would be to just use bazel as the package manager, and to load tulsi from the WORKSPACE, like XCHammer intends to? username_0: After some more thought, perhaps importing through WORKSPACE and generating/running via a rule would be equally easy but more in the spirit of bazel? username_1: You can add it to your WORKSPACE and run it with Bazel. ``` bazel run @build_bazel_tulsi//:tulsi -- -- [flags] ```` username_0: @username_1's suggestion is totally the way to go. Thank you @username_1! As an example for anyone else reading, just add [rules_apple](https://github.com/bazelbuild/rules_apple) to your WORKSPACE, followed by a http_archive for tulsi. So at the time of writing: ``` http_archive( name = "build_bazel_tulsi", # Grabbing master for this example, so it always points to the latest. Do *not* add a sha256, or you'll break the autoupdate! urls = ["https://github.com/bazelbuild/tulsi/archive/master.zip"], strip_prefix = "tulsi-master" ) # You may well need/want to bump this version. Check https://github.com/bazelbuild/tulsi/blob/master/WORKSPACE http_archive( name = "build_bazel_rules_apple", sha256 = "734813e44eb5a2fcba5ffd45de9fe5d05325420a5aa1f6c97a3d88fe2c525b17", url = "https://github.com/bazelbuild/rules_apple/releases/download/0.21.1/rules_apple.0.21.1.tar.gz", ) load("@build_bazel_rules_apple//apple:repositories.bzl", "apple_rules_dependencies") apple_rules_dependencies() load("@build_bazel_rules_swift//swift:repositories.bzl", "swift_rules_dependencies") swift_rules_dependencies() # Needed for Tulsi. Seems to also pull in protobuf load("@build_bazel_apple_support//lib:repositories.bzl", "apple_support_dependencies") apple_support_dependencies() ``` Still think it'd be sweet if this were a more official way to install--perhaps with a starlark function to check for dependencies, rules_apple-style?
tensorflow/tensorflow
463271691
Title: TFlite conversion of Conv1D with dilation !=1 Question: username_0: **System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): Yes - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Google Colab - TensorFlow installed from (source or binary): binary - TensorFlow version (use command below): tensorflow==2.0.0-beta1 - Python version: python3 **Describe the current behavior** After converting a Conv1D op to tensorflow lite the interpreter cannot allocate tensors: ` tensorflow/lite/kernels/space_to_batch_nd.cc:96 NumDimensions(op_context.input) != kInputDimensionNum (3 != 4)Node number 0 (SPACE_TO_BATCH_ND) failed to prepare. ` **Describe the expected behavior** Tflite model should be able to load and execute. **Code to reproduce the issue** ``` !pip install -q tensorflow==2.0.0-beta1 import tensorflow as tf from tensorflow.keras.models import Model from tensorflow.keras.layers import * def get_model(): input = tf.keras.Input(shape=(10,40)) #No error when dilation rate == 1 layer = Conv1D(32, (3),dilation_rate =2, padding='same',use_bias=False) (input) layer = GlobalMaxPooling1D()(layer) output = Dense(2) (layer) model = Model(inputs=[input], outputs=[output]) return model model = get_model() converter = tf.lite.TFLiteConverter.from_keras_model(model) tflite_model = converter.convert() open("./trained_model.tflite", "wb").write(tflite_model) interpreter = tf.lite.Interpreter(model_path="./trained_model.tflite") interpreter.allocate_tensors() ``` **Other info / logs** The problem does not occur when dilation_rate ==1 Answers: username_1: This issue is solved with PRs #28410, #27867 & #28179. Thanks! Status: Issue closed username_2: Closing the issue as it has been resolved username_2: **System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): Yes - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Google Colab - TensorFlow installed from (source or binary): binary - TensorFlow version (use command below): tensorflow==2.0.0-beta1 - Python version: python3 **Describe the current behavior** After converting a Conv1D op to tensorflow lite the interpreter cannot allocate tensors: ` tensorflow/lite/kernels/space_to_batch_nd.cc:96 NumDimensions(op_context.input) != kInputDimensionNum (3 != 4)Node number 0 (SPACE_TO_BATCH_ND) failed to prepare. ` **Describe the expected behavior** Tflite model should be able to load and execute. **Code to reproduce the issue** ``` !pip install -q tensorflow==2.0.0-beta1 import tensorflow as tf from tensorflow.keras.models import Model from tensorflow.keras.layers import * def get_model(): input = tf.keras.Input(shape=(10,40)) #No error when dilation rate == 1 layer = Conv1D(32, (3),dilation_rate =2, padding='same',use_bias=False) (input) layer = GlobalMaxPooling1D()(layer) output = Dense(2) (layer) model = Model(inputs=[input], outputs=[output]) return model model = get_model() converter = tf.lite.TFLiteConverter.from_keras_model(model) tflite_model = converter.convert() open("./trained_model.tflite", "wb").write(tflite_model) interpreter = tf.lite.Interpreter(model_path="./trained_model.tflite") interpreter.allocate_tensors() ``` **Other info / logs** The problem does not occur when dilation_rate ==1 username_3: Was able to reproduce the issue with [TF v2.1](https://colab.research.google.com/gist/username_3/5216cdcb0582cc62e6a6518ea5334e2e/2-1-template.ipynb) and [TF-nightly](https://colab.research.google.com/gist/username_3/14ab31228c0693ab78b78ed73f0145d0/tf-nightly.ipynb#scrollTo=ieAW-NK5iqpf) i.e. v2.2.0-dev20200327. Please find the attached gist. Thanks! username_4: @username_5 can you take a look? username_5: @username_3 Can you share the sample code ? The example in the original issue works when i tried it. Thanks username_3: @username_5, Sure, below are the links of the gist using - TF v2.1 https://colab.research.google.com/gist/username_3/5216cdcb0582cc62e6a6518ea5334e2e/2-1-template.ipynb - TF-nightly https://colab.research.google.com/gist/username_3/14ab31228c0693ab78b78ed73f0145d0/tf-nightly.ipynb#scrollTo=ieAW-NK5iqpf username_6: @username_5 Hi, what tensorflow version were you using? username_5: Thanks @username_3 I can reproduce it on the 2.1 but works with nightly. Can you please retry. @username_6 i was using tf-nightly username_6: I can confirm that this works with ```'2.2.0-dev20200414``` username_3: @username_5, Works without any issues with the latest TF-nightly i.e. v2.2.0-dev20200415. Please find the gist [here](https://colab.research.google.com/gist/username_3/67ba5ecd35433f7e61e5102f8d74611a/30315-tf-nightly.ipynb#scrollTo=ieAW-NK5iqpf). Thanks! username_5: Thanks for confirming. I am closing the issue. Please feel free to reopen/create a new one if you have any problems. Thanks Status: Issue closed username_7: Issue still exists in TF 2.2 stable.
idris-lang/Idris-dev
224571604
Title: Resource aware types? Question: username_0: I'm wondering if it would be possible to have 'types' that are resource aware, similar to raml (resource aware ml)? Answers: username_1: Could you give a brief synopsis of what this means / what raml is? username_0: The raml website explains it a bit better: http://www.raml.co > On Apr 26, 2017, at 12:41 PM, <NAME> <<EMAIL>> wrote: > > Could you give a brief synopsis of what this means / what raml is? > > username_2: @username_0 Edwin had made a [Resource DSL](https://edwinb.wordpress.com/2011/09/16/resource-safe-systems-programming-with-embedded-domain-specific-languages/). See also, the [Uniqueness Types](http://docs.idris-lang.org/en/latest/reference/uniqueness-types.html) and Linear Types extensions. Note: Could you kindly use the mailing list or IRC next time for questions, instead of the issue tracker? Status: Issue closed
DaveGamble/cJSON
315763793
Title: cJSON is more wasteful of memory Question: username_0: char *valuestring; int valueint; double valuedouble; all above can use void *value If void * is used, support for uint64 will be better Answers: username_0: cJSON.h `/* Copyright (c) 2009-2017 <NAME> and cJSON contributors Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */ #ifndef cJSON__h #define cJSON__h #ifdef __cplusplus extern "C" { #endif /* project version */ #define CJSON_VERSION_MAJOR 1 #define CJSON_VERSION_MINOR 7 #define CJSON_VERSION_PATCH 5 #include <stddef.h> /* cJSON Types: */ #define cJSON_Invalid (0) #define cJSON_False (1 << 0) #define cJSON_True (1 << 1) #define cJSON_NULL (1 << 2) #define cJSON_INT64 (1 << 3) #define cJSON_UINT64 (1 << 4) #define cJSON_DOUBLE (1 << 5) #define cJSON_String (1 << 6) #define cJSON_Array (1 << 7) #define cJSON_Object (1 << 8) #define cJSON_Raw (1 << 9) /* raw json */ #define cJSON_IsReference (1 << 30) #define cJSON_StringIsConst (1 << 31) /* The cJSON structure: */ typedef struct cJSON { /* next/prev allow you to walk array/object chains. Alternatively, use GetArraySize/GetArrayItem/GetObjectItem */ struct cJSON *next; struct cJSON *prev; [Truncated] { return 0; } double value = 0; value = *((double *)item->value); return value; } CJSON_PUBLIC(void *) cJSON_malloc(size_t size) { return global_hooks.allocate(size); } CJSON_PUBLIC(void) cJSON_free(void *object) { global_hooks.deallocate(object); } ` username_1: If only there was a way to concisely show the differences between two versions of a file ... username_2: @username_0 I'm not entirely sure what you are trying to say. I know that the `cJSON` struct is quite wasteful, this is why I have plans on how to handle that in the future, see https://github.com/DaveGamble/cJSON/issues/63#issuecomment-364743833. That change would reduce the size to 32-48 bytes on x86_64 and even down to 20 bytes with 16 bit architectures (if I calculated that correctly). **And** it would easily allow supporting diffent types for representing numbers. And the idea to use a union is not new, see #16. Status: Issue closed
akamensky/argparse
1095590339
Title: Question about returned parameters (not a bug report) Question: username_0: Congratulations on porting the ArgParse concept from Python. ArgParse is one of the Python packages that I have used heavily so this was quite welcome to find. If I find any bugs or suggestions for enhancements, I'll write up a separate issue. 1. Since Parser.X is returning something, why return a pointer to the something? Why not simply return the something itself? Yes, argparse is consistent which is always a good thing. 2. When will argparse be part of the standard package library? i.e. ```import argparse``` without the github references. Answers: username_1: Probably never as I am not one of Google developers who decide what goes into standard library. It is better to ask them perhaps. Status: Issue closed
composer/composer
57969097
Title: HTTP Header Question: username_0: Please integrate the possibility to set the http header for individual repositories. Thx Answers: username_1: What do you mean by http header exactly? username_0: By adding an individual repository the option to set its header when curling to the packages.json file. If you want to have an authentication / user based (e.g. with an api token) packages.json file. If you could define header fields for the repository (e.g. API-TOKEN) than you could generate a user based packages.json file. Nevertheless by settings a repository with https, the header fields would be encrypted, too. That would be very useful! username_1: You can configure that per-domain using the http-basic config https://getcomposer.org/doc/04-schema.md#config for http basic authentication. username_0: Right, but there is no option to set other header fields username_1: I think you can actually configure it in the repository using stream options like in https://getcomposer.org/doc/articles/handling-private-packages-with-satis.md#security but providing http > headers instead of ssh2. Haven't tried but it should work if we don't do anything dumb later on to prevent it (which could well be..). username_0: Do you have an example / does it work? Tried things like: ```json { "type": "composer", "url": "http://ex.com/", "options": { "headers":{ "API":"asd" } } } ``` OR ```json { "type": "composer", "url": "http://ex.com/", "options": { "http": "headers":{ "API":"asd" } } } } ``` each with "headers" and "header" and it didn't work :( username_0: OK, found it :D It does already work! But please put that in your documentation! :) ```json { "type": "composer", "url": "http://ex.com/", "options": { "http":{ "header":[ "API-TOKEN: asd" ] } } } ``` Status: Issue closed
lovell/sharp
485099525
Title: Error: read ECONNRESET Question: username_0: info sharp Downloading https://github.com/username_1/sharp-libvips/releases/download/v8.8.1/libvips-8.8.1-win32-x64.tar.gz D:\NodeBB-master\node_modules\sharp\install\libvips.js:82 throw err; ^ Error: read ECONNRESET at TLSWrap.onStreamRead (internal/stream_base_commons.js:183:27) { errno: 'ECONNRESET', code: 'ECONNRESET', syscall: 'read' } D:\NodeBB-master\node_modules\sharp>if not defined npm_config_node_gyp (node "C:\Program Files\nodejs\node_modules\npm\node_modules\npm-lifecycle\node-gyp-bin\\..\..\node_modules\node-gyp\bin\node-gyp.js" rebuild ) else (node "C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\bin\node-gyp.js" rebuild ) Building the projects in this solution one at a time. To enable parallel build, please add the "/m" switch. VError.cpp VInterpolate.cpp VImage.cpp win_delay_load_hook.cc LINK : fatal error LNK1104: cannot open file '..\vendor\lib\libvips.lib' [D:\NodeBB-master\node_modules\sharp\build\lib vips-cpp.vcxproj] gyp ERR! build error gyp ERR! stack Error: `C:\Program Files (x86)\MSBuild\14.0\bin\msbuild.exe` failed with exit code: 1 gyp ERR! stack at ChildProcess.onExit (C:\Program Files\nodejs\node_modules\npm\node_modules\node-gyp\lib\build.js:262:23) gyp ERR! stack at ChildProcess.emit (events.js:200:13) gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:272:12) gyp ERR! System Windows_NT 10.0.14393 gyp ERR! command "C:\\Program Files\\nodejs\\node.exe" "C:\\Program Files\\nodejs\\node_modules\\npm\\node_modules\\node-gyp\\bin\\node-gyp.js" "rebuild" gyp ERR! cwd D:\NodeBB-master\node_modules\sharp gyp ERR! node -v v12.5.0 gyp ERR! node-gyp -v v3.8.0 gyp ERR! not ok npm WARN [email protected] requires a peer of eslint@^4.19.1 || ^5.3.0 but none is installed. You must install peer dependencies yourself. npm WARN [email protected] requires a peer of nodebb-plugin-emoji@^2.0.0 but none is installed. You must install peer dependencies yourself. npm WARN [email protected] requires a peer of textcomplete@^0.14.2 but none is installed. You must install peer dependencies yourself. npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! [email protected] install: `(node install/libvips && node install/dll-copy && prebuild-install) || (node-gyp rebuild && node install/dll-copy)` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the [email protected] install script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! C:\Users\chen29\AppData\Roaming\npm-cache\_logs\2019-08-26T07_37_06_046Z-debug.log D:\NodeBB-master>npm install sharp -> installlog.txt npm ERR! code E404 npm ERR! 404 Not Found - GET https://registry.npmjs.org/- - Not found npm ERR! 404 npm ERR! 404 '-@latest' is not in the npm registry. npm ERR! 404 You should bug the author to publish it (or use the name yourself!) npm ERR! 404 npm ERR! 404 Note that you can also install from a npm ERR! 404 tarball, folder, http url, or git url. npm ERR! A complete log of this run can be found in: npm ERR! C:\Users\chen29\AppData\Roaming\npm-cache\_logs\2019-08-26T07_45_32_153Z-debug.log Answers: username_1: `ECONNRESET` is due to a (temporary? firewall?) network problem connecting to https://github.com/. username_0: https://github.com/ is accessible in the web browser. The firewall is closed. How can the sharp be installed when the sharp and libvips are downloaded. username_0: Yes. It's related to a network problem connecting to https://github.com/ in windows command line. I test with the following command in two different computer. One is working. The server with nodebb is not working. D:\11>git clone https://github.com/username_1/sharp Cloning into 'sharp'... fatal: unable to access 'https://github.com/username_1/sharp/': Unknown SSL protocol error in connection to github.com:443 username_1: Glad you worked it out. Status: Issue closed
magma/magma
1077131900
Title: Inconsistent Node version Question: username_0: ### Your Environment - **Version:** Magma v1.6.1 - **Affected Component:** NMS - **Affected Subcomponent:** all - **Deployment Environment:** all ### Describe the Issue There are significant differences between Node versions specified in different places. This likely causes downstream bugs. For example, dependent packages may target one of the referenced versions but not the other. **To Reproduce** /magma/nms/packages/magmalte/Dockerfile: line 1: `FROM node:12.18-alpine as builder` line 20: `FROM node:10-alpine` /magma/.github/workflows/nms-workflow.yml line 99: `node-version: 16` **Expected behavior** 1. The same version of Node should be used in the build and runtime environment 2. The target version of Node should be used by developers when running tests, for example by using nvm 3. The target version of Node should be made known in developer documentation 4. Packages with version dependencies should match the target version of Node 5. The engines field in package.json should match the target version. That might include: - ./nms/package.json - ./nms/packages/magmalte/package.json - ./orc8r/cloud/package.json
OneHalf3544/Japan-crossword
335178082
Title: Scale the nonogram image to the window size Question: username_0: **Is your feature request related to a problem? Please describe.** When I choose a big nonogram, it doesn't fit into the monitor resolution, and I have to scroll the window content up and down. It is very annoying. **Describe the solution you'd like** Automatically scale a content to the main window size. Also, we can add an option to hide metadata parts of a nonogram. It will save a lot of view's space.
Delgan/loguru
939138194
Title: not able to show module name Question: username_0: ```python from loguru import logger logger.info('hello world') ``` it only shows `<module>` placeholder ```python 2021-07-08 02:07:15.378 | INFO | __main__:<module>:3 - hello world ``` Answers: username_1: The default format uses the `{name}` which is replaced with the `__name__` variable at runtime. Usually, the `__name__` represents the module name but it's not the case for your entry file, [see official Python documentation](https://docs.python.org/3/library/__main__.html) about that. The `<module>` you're seeing is supposed to represent the function from where is called the log method. As your `logger.info()` is not in any function, it is say to be in the module itself. See documentation about possible `format` fields, maybe you want to use `{file.name}` instead: [The record dict](https://loguru.readthedocs.io/en/stable/api/logger.html#record). Status: Issue closed
julianhyde/sqlline
400902797
Title: Unable to see instruction to quit or read more when run !manual through sqlline CLI Question: username_0: Observed this problem while running !manual command from sqlline CLI. As per the code https://github.com/julianhyde/sqlline/commit/f3263e95c4c981515edc02a730d472d8d6d043cd#diff-4151e9ed27b575a9e975057af55c5378R1736 , it should display the following message "[ Hit "enter" for more ("q" to exit) ]" but it is not there on the terminal. Answers: username_1: Indeed it does not. This code works only on Windows OS. In case of Linux/Mac there will be used jline3's `less` implementation which is much better. Unfortunately it does not work well under Windows that is why the old code is used there e.g. there are several issues while Windows https://github.com/jline/jline3/issues/344, https://github.com/jline/jline3/issues/304 username_1: By the way here it is a comment in the code related to the topic https://github.com/julianhyde/sqlline/blob/dc969a28f009b18a9dfb1026f96cbec0450338da/src/main/java/sqlline/Commands.java#L1704-L1709 username_0: Thanks! How do I test my changes https://github.com/julianhyde/sqlline/pull/263 with CLI that uses same function call? username_1: Could you please explain a bit more what are going to test that is not covered by tests from the mentioned issue? username_0: I wanted to see things are working fine and as expected when I run DROP/DELETE command from SQLline and confirm flag set to true. #263 username_0: Also, is this fix in the roadmap? username_1: I think there could be added a way to specify if a symbols should be printed or not after key pressed while confirmation step required. If it matches your expectations I would suggest to rename the issue.
ikedaosushi/tech-news
761856367
Title: Airflow 2.0 でDAG定義をよりシンプルに TaskFlow APIの紹介Dentsu Digital Tech Blognote Question: username_0: Airflow 2.0 &#12391;DAG&#23450;&#32681;&#12434;&#12424;&#12426;&#12471;&#12531;&#12503;&#12523;&#12395;&#65281; TaskFlow API&#12398;&#32057;&#20171;&#65372;Dentsu Digital Tech Blog&#65372;note<br> <br> https://ift.tt/33Yvpe1
izhangzhihao/intellij-rainbow-brackets
868645245
Title: [Auto Generated Report]java.lang.Throwable: Stub index points to a file without PSI: file = jrt:///usr/lib/jvm/jdk-11.0.10!/jdk.scripting.nashorn/jdk/nashorn/internal/objects/NativeJSON$Constructor.class, file type = com.intellij.ide.highlighter.JavaClassFileType@64767844, indexed file type = com.intellij.ide.highlighter.JavaClassFileType@64767844, used scope = com.intellij.psi.impl.search.JavaSourceFilterScope[Module-with-dependencies:bigc-admin-business compile-only:true include-libraries:true include-other-modules:true include-tests Question: username_0: - Plugin Name: - Plugin Version: 6.17 - OS Name: Linux - OS Version: 5.8.0-50-generic - Java Version: 11.0.10 - App Name: IDEA - App Full Name: IntelliJ IDEA - Is Snapshot: false - App Build: IU-211.6693.111 - StackTrace: ``` java.lang.Throwable: Stub index points to a file without PSI: file = jrt:///usr/lib/jvm/jdk-11.0.10!/jdk.scripting.nashorn/jdk/nashorn/internal/objects/NativeJSON$Constructor.class, file type = com.intellij.ide.highlighter.JavaClassFileType@64767844, indexed file type = com.intellij.ide.highlighter.JavaClassFileType@64767844, used scope = com.intellij.psi.impl.search.JavaSourceFilterScope[Module-with-dependencies:bigc-admin-business compile-only:true include-libraries:true include-other-modules:true include-tests:true] at com.intellij.openapi.diagnostic.Logger.error(Logger.java:161) at com.intellij.psi.stubs.StubProcessingHelperBase.processStubsInFile(StubProcessingHelperBase.java:52) at com.intellij.psi.stubs.StubIndexImpl.lambda$processElements$2(StubIndexImpl.java:284) at com.intellij.psi.stubs.StubIndexImpl.processElements(StubIndexImpl.java:330) at com.intellij.psi.impl.PsiShortNamesCacheImpl.processFieldsWithName(PsiShortNamesCacheImpl.java:185) at com.intellij.psi.impl.CompositeShortNamesCache.processFieldsWithName(CompositeShortNamesCache.java:228) at com.intellij.codeInsight.daemon.impl.quickfix.StaticImportConstantFix.getMembersToImport(StaticImportConstantFix.java:73) at com.intellij.codeInsight.daemon.impl.quickfix.StaticImportMemberFix.<init>(StaticImportMemberFix.java:52) at com.intellij.codeInsight.daemon.impl.quickfix.StaticImportConstantFix.<init>(StaticImportConstantFix.java:35) at com.intellij.codeInsight.daemon.impl.quickfix.DefaultQuickFixProvider.registerFixes(DefaultQuickFixProvider.java:40) at com.intellij.codeInsight.daemon.impl.quickfix.DefaultQuickFixProvider.registerFixes(DefaultQuickFixProvider.java:24) at com.intellij.codeInsight.quickfix.UnresolvedReferenceQuickFixProvider.registerReferenceFixes(UnresolvedReferenceQuickFixProvider.java:27) at com.intellij.codeInsight.daemon.impl.analysis.HighlightMethodUtil.checkAmbiguousMethodCallIdentifier(HighlightMethodUtil.java:800) at com.intellij.codeInsight.daemon.impl.analysis.HighlightVisitorImpl.visitReferenceExpression(HighlightVisitorImpl.java:1410) at com.intellij.psi.impl.source.tree.java.PsiReferenceExpressionImpl.accept(PsiReferenceExpressionImpl.java:778) at com.intellij.codeInsight.daemon.impl.analysis.HighlightVisitorImpl.visit(HighlightVisitorImpl.java:189) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.runVisitors(GeneralHighlightingPass.java:335) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.lambda$collectHighlights$5(GeneralHighlightingPass.java:268) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.analyzeByVisitors(GeneralHighlightingPass.java:294) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.lambda$analyzeByVisitors$6(GeneralHighlightingPass.java:297) at com.github.izhangzhihao.rainbow.brackets.visitor.RainbowHighlightVisitor.analyze(RainbowHighlightVisitor.kt:35) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.analyzeByVisitors(GeneralHighlightingPass.java:297) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.lambda$analyzeByVisitors$6(GeneralHighlightingPass.java:297) at com.intellij.codeInsight.daemon.impl.analysis.HighlightVisitorImpl.lambda$analyze$1(HighlightVisitorImpl.java:214) at com.intellij.codeInsight.daemon.impl.analysis.RefCountHolder.analyze(RefCountHolder.java:369) at com.intellij.codeInsight.daemon.impl.analysis.HighlightVisitorImpl.analyze(HighlightVisitorImpl.java:213) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.analyzeByVisitors(GeneralHighlightingPass.java:297) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.lambda$analyzeByVisitors$6(GeneralHighlightingPass.java:297) at com.intellij.codeInsight.daemon.impl.DefaultHighlightVisitor.analyze(DefaultHighlightVisitor.java:96) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.analyzeByVisitors(GeneralHighlightingPass.java:297) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.collectHighlights(GeneralHighlightingPass.java:265) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.collectInformationWithProgress(GeneralHighlightingPass.java:211) at com.intellij.codeInsight.daemon.impl.ProgressableTextEditorHighlightingPass.doCollectInformation(ProgressableTextEditorHighlightingPass.java:84) at com.intellij.codeHighlighting.TextEditorHighlightingPass.collectInformation(TextEditorHighlightingPass.java:56) at com.intellij.codeInsight.daemon.impl.PassExecutorService$ScheduledPass.lambda$doRun$1(PassExecutorService.java:400) at com.intellij.openapi.application.impl.ApplicationImpl.tryRunReadAction(ApplicationImpl.java:1096) at com.intellij.codeInsight.daemon.impl.PassExecutorService$ScheduledPass.lambda$doRun$2(PassExecutorService.java:393) at com.intellij.openapi.progress.impl.CoreProgressManager.registerIndicatorAndRun(CoreProgressManager.java:688) at com.intellij.openapi.progress.impl.CoreProgressManager.executeProcessUnderProgress(CoreProgressManager.java:634) at com.intellij.openapi.progress.impl.ProgressManagerImpl.executeProcessUnderProgress(ProgressManagerImpl.java:64) at com.intellij.codeInsight.daemon.impl.PassExecutorService$ScheduledPass.doRun(PassExecutorService.java:392) at com.intellij.codeInsight.daemon.impl.PassExecutorService$ScheduledPass.lambda$run$0(PassExecutorService.java:368) at com.intellij.openapi.application.impl.ReadMostlyRWLock.executeByImpatientReader(ReadMostlyRWLock.java:167) at com.intellij.openapi.application.impl.ApplicationImpl.executeByImpatientReader(ApplicationImpl.java:178) at com.intellij.codeInsight.daemon.impl.PassExecutorService$ScheduledPass.run(PassExecutorService.java:366) at com.intellij.concurrency.JobLauncherImpl$VoidForkJoinTask$1.exec(JobLauncherImpl.java:188) at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290) at java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020) at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656) at java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594) at java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:183) ```
TooBug/wemark
377138608
Title: 疑问:我仅仅在json配置文件中加入usingComponents说明后整个页面就全部乱了? Question: username_0: 我仅仅在json配置文件中加入usingComponents说明: "usingComponents": { "wemark": "utils/wemark/wemark" }, 然后整个页面就没有数据显示了?我都还没有实际加入wemark的控件呢,什么情况 Answers: username_1: 看一下详细的报错信息,要不然无法排查。 username_0: 感谢回复,不过应该不是你这个控件的原因,我尝试了一下其余的任何控件,只要一申明 usingComponents ,整个页面就没有任何输出了,非常奇怪,不知道是不是小程序本身的问题。 Status: Issue closed
phansible/phansible
167637070
Title: Getting errors on vagrant up Question: username_0: Hi, Trying to run a box and after vagrant up I'm getting the following errors: .... ==> default: stdin: is not a tty .... ==> default: gpg: ==> default: keyring `/tmp/tmprfhgdqk1/secring.gpg' created ==> default: gpg: ==> default: keyring `/tmp/tmprfhgdqk1/pubring.gpg' created ==> default: gpg: ==> default: requesting key 7BB9C367 from hkp server keyserver.ubuntu.com ==> default: gpg: ==> default: /tmp/tmprfhgdqk1/trustdb.gpg: trustdb created ==> default: gpg: ==> default: key 7BB9C367: public key "Launchpad PPA for Ansible, Inc." imported ==> default: gpg: ==> default: Total number processed: 1 ==> default: gpg: ==> default: imported: 1 ==> default: (RSA: 1) .... ==> default: dpkg-preconfigure: unable to re-open stdin: No such file or directory .... ==> default: cp: ==> default: cannot stat ‘/vagrant/ansible/inventories/dev’ ==> default: : No such file or directory ==> default: cat: ==> default: /vagrant/ansible/files/authorized_keys ==> default: : No such file or directory ==> default: ERROR! the playbook: /vagrant/ansible/playbook.yml could not be found The SSH command responded with a non-zero exit status. Vagrant assumes that this means the command failed. The output for this command should be in the log above. Please read the output to determine what went wrong. I usually used already pre-made vagrant boxes before and creating my own for the first time. Answers: username_1: @username_0 that's interesting, could you please share the package you downloaded? By the look of it the errors indicate you are missing most of the essential files. I'd like to make a comparison between I what I have and what you got. username_0: Sure! I've unfortunately closed the window with the settings that I chose but here are the 2 packages that I got from it: [phansible_default.zip](https://github.com/phansible/phansible/files/385858/phansible_default.zip) [phansible_default (1).zip](https://github.com/phansible/phansible/files/385859/phansible_default.1.zip) username_0: Just made another one in Firefox, here's the file and screenshots: [phansible__box_php5.4_xdebug.zip](https://github.com/phansible/phansible/files/385890/phansible__box_php5.4_xdebug.zip) ![screen shot 2016-07-27 at 12 10 42](https://cloud.githubusercontent.com/assets/17159170/17173544/53732f76-53f3-11e6-8142-f8dc7f61e807.png) ![screen shot 2016-07-27 at 12 11 05](https://cloud.githubusercontent.com/assets/17159170/17173545/5377e188-53f3-11e6-8d82-19b55acca5bd.png) ![screen shot 2016-07-27 at 12 11 19](https://cloud.githubusercontent.com/assets/17159170/17173543/5371f9c6-53f3-11e6-8e5e-bbc99944d202.png) ![screen shot 2016-07-27 at 12 11 26](https://cloud.githubusercontent.com/assets/17159170/17173542/536c7212-53f3-11e6-86ce-c3c69f8f1bf0.png) username_1: @username_0 I just tried the first one of the zips you linked and it worked for me. I'm running vagrant 1.8.5 and the VirtualBox 5.0.20 at the moment, I'm going to have a go with the latest version of VirtualBox too and report back. username_1: @username_0 it works also with VirtualBox 5.1.2 over here and the same package. Just out of curiosity what OS are your running? I'm on OSX here. username_0: @username_1 OS X 10.11.5, VirtualBox 5.0.22 I'll try updating VirtualBox later today and will see if that changes anything. :( username_1: @username_0 please keep me poste as of know it looks more like a glitch of sort mostly related with ansible itself not detecting path properly for some weird reasons. The zips you share with me contains all the required files including the one your error log sais are not found. username_0: Hm, updated VirtualBox to 5.0.26, still the same. I guess there's probably something wrong with my system? username_1: @username_0 the latest virtualbox version is 5.1.2. Yes the problem appears to be related to something else in your setup. Basically what is happening in your case is that the shared folder between your host and guest doesn't get shared or, if it does, it suddenlly disappear so that ansible can't find the files it needs. In fairness there is one thing that seems strange about your setup, although it doesn't seem to affect my test. So basically if I look at the vagrant file you generated you are sharing the following: ```shell config.vm.synced_folder "/Users/vitalibokov/GitHub/", "/vagrant", type: "nfs" ``` And the error message is reporting the following: ```shell cannot stat ‘/vagrant/ansible/inventories/dev’ ``` now the thing is that the ansible folder is not in your home right? The ansible folder is in the path from where you launch or should launch `vagrant up` and you can see that in these lines: ```shell config.vm.provision "ansible" do |ansible| ansible.playbook = "ansible/playbook.yml" ansible.inventory_path = "ansible/inventories/dev" ansible.limit = 'all' end ``` So my question is, how are you executing all this? Are the file in the right and expected location or did you accidentally moved things around? Could you also post a gist of the entire log? I'd like to see at what step you error is actually getting thrown also because I can't see that output in my terminal (I'm referring about the gpg import) username_1: @username_0 did you have any luck with this? Any progress?
alpinejs/alpine
616341171
Title: Following a coding standard Question: username_0: Continuation of #475 What coding standard are you willing to follow? Answers: username_1: Some people started a discussion on this thread #190. At the moment, no coding style is enforced. username_2: Like @username_1 said there's an existing issue about this, in addition the only person with write access to the repository (ie. Only person who can merge anything) is @calebporzio so I dunno if automated style checking or event having a style guide is useful in that context. username_0: Okay-okay. I think liberalism can kill, following a style guide is a must for principled developer. @douglascrockford leads the ways with https://github.com/douglascrockford/JSLint Although it hurts feeling which is very important in a collaborative project! Status: Issue closed username_2: Hmmm the velocity of contributions is low on the project, since everything has to be merged by Caleb. In that context, I dunno what value automated style checks bring. username_0: I think without a style guide contributions are merged easier and code quality is a bit lower.
intesar/Fx-Test-Data
317020371
Title: Fx : vault_findById Question: username_0: Project : Fx Job : Stg Env : Stg Region : FxLabs/US_WEST_1 Result : fail Status Code : 503 Headers : {Cache-Control=[no-cache], Connection=[close], Content-Type=[text/html]} Endpoint : http://stg1.fxlabs.io/api/v1/vault/ Request : Response : null Logs : Assertion [@Response.errors == false] failed, expected value [false] but found []Assertion [@StatusCode == 200] failed, expected value [200] but found [503]Assertion [@vault_create_init_Response.data.id == @Response.data.id] passed, expected [] and found []Assertion [@vault_create_init_Request.key == @Response.data.key] failed, expected value [] but found [rxoHR4] Status: Issue closed Answers: username_0: Project : Fx Job : Stg Env : Stg Region : FxLabs/US_WEST_1 Result : pass Status Code : 200 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=MmYxYzNhYmMtNTE1ZC00ZTllLTk2ZWMtYjQwNGIwYWU5YWQx; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Mon, 23 Apr 2018 23:45:37 GMT]} Endpoint : http://stg1.fxlabs.io/api/v1/vault/8a80827362f4e1660162f4e5f89b0737 Request : Response : { "requestId" : "None", "requestTime" : "2018-04-23T23:45:37.757+0000", "errors" : false, "messages" : [ ], "data" : { "id" : "8a80827362f4e1660162f4e5f89b0737", "createdBy" : "8a80829061c17be10161c18999480002", "createdDate" : "2018-04-23T23:45:34.875+0000", "modifiedBy" : "8a80829061c17be10161c18999480002", "modifiedDate" : "2018-04-23T23:45:34.875+0000", "version" : null, "inactive" : false, "org" : { "id" : "8a80829061c17be10161c189993d0001", "createdBy" : "anonymousUser", "createdDate" : "2018-02-23T07:21:15.837+0000", "modifiedBy" : "anonymousUser", "modifiedDate" : "2018-02-23T07:21:15.837+0000", "version" : null, "inactive" : false, "name" : "FxLabs" }, "key" : "h1On4i", "val" : "value", "description" : "", "visibility" : "PRIVATE" }, "totalPages" : 1, "totalElements" : 1 } Logs : Assertion [@vault_create_init_Request.key == @Response.data.key] passed, expected [h1On4i] and found [h1On4i]Assertion [@vault_create_init_Response.data.id == @Response.data.id] passed, expected [8a80827362f4e1660162f4e5f89b0737] and found [8a80827362f4e1660162f4e5f89b0737]Assertion [@Response.errors == false] passed, expected [false] and found [false]Assertion [@StatusCode == 200] passed, expected [200] and found [200]
digio/terraform-google-gitlab-runner
770909120
Title: Requested disk size cannot be smaller than the image size (20 GB) Question: username_0: ``` Error creating instance: googleapi: Error 400: Invalid value for field 'resource.disks[0].initializeParams.diskSizeGb': '10'. Requested disk size cannot be smaller than the image size (20 GB), invalid ``` Answers: username_1: Fixed in https://github.com/digio/terraform-google-gitlab-runner/commit/85e8ef21bc7d5f3864f773c462456445467c4189 Status: Issue closed
BlueMond/LifeMC
680416257
Title: Add tempban counter Question: username_0: Add tempban countdown or date/time of revival on the screen displayed when a player is denied access to the server for being dead and tempban is enabled and also when a player is kicked for having lost their last life.
LonamiWebs/Telethon
1118596032
Title: Unable to run 2 or more sessions Question: username_0: Trying to make parsing from 4 different chats using 2 different accounts - 2 chats per each account. In different programs it works but in same nothing appears: Tried multiple variants, including event loops but wasn't able to make it working. Could anyone show the code of event loops how it would work or other way ``` import asyncio from telethon import TelegramClient, sync, events import asyncio import string import sys client1 = TelegramClient('session1', api_id1, api_hash1) client2 = TelegramClient('session2', api_id1, api_hash1) #Chats block chats1 =("chat1", "chat2") chats2 =("chat3", "chat4") @client1.on(events.NewMessage(chats=tuple(chats1))) async def normal_handler1(event): user_message = event.message.to_dict()['message'] print(user_message) @client2.on(events.NewMessage(chats=tuple(chats2))) async def normal_handler2(event): user_message = event.message.to_dict()['message'] print(user_message) client1.start() client2.start() async def main(): return await asyncio.gather( client1.run_until_disconnected(), client2.run_until_disconnected() ) asyncio.run(main()) ``` Answers: username_1: What's the issue? Is there an error? Are you using the development version? username_0: The problem is there is no error it just doesn't work and that's it. There was alike issue on stackoverflow -> https://stackoverflow.com/questions/70320400/run-multiple-telethon-clients But I wasn't able to properly rewrite code for event loops, maybe you could help me to make it work ? Would be appreciated username_0: @username_1 still wasn't able to find solution, could you help with code ? I really don't know how to rewrite it with event loops to make it work username_0: @username_1 do you know how to solve it ? username_2: I just run clients in different threads.
MiguelRipoll23/homebridge-securitysystem
534313177
Title: No Response from Siren switch when Security System is disarmed. Question: username_0: When the security system is Off and you hit the Siren Switch, I keep getting a "No Response" from the switch until I Arm the system. Maybe that's because the switch is just there for the purpose of the Armed states triggering automatons, but it's really nice to have a manual toggle even when the Security System is off. Answers: username_1: This was intended behavior due to #6 (creating an automation without needing to add conditions) but it can be reverted if people in here would like to add a manual trigger button too. username_0: Manual toggle would be nice but not if it’ll undo the previous work you’ve done. username_1: A new option could be added just to ignore the disarmed state and would work fine for both sides without having to revert it, I'll get this done soon. username_1: A new version has been released that adds this option, sorry for the delay. Status: Issue closed
jupyterlab/jupyterlab
302191853
Title: pipe operator (%>%) shortcut doesn't work in jupyterlab Question: username_0: pipe operator(%>%) works when 'ctrl + shift + M' in **jupyter notebook** But, in **jupyterlab**, 'ctrl + shift + M' invokes chrome 'login' popup Answers: username_1: I implemented that in [IRkernel’s kernel.js](https://github.com/IRkernel/IRkernel/blob/dfa7f4ef2c855d289df451cb8af18b3fd0c96fda/inst/kernelspec/kernel.js) With the new JS API it needs to be done in a different way. I’d be glad about pointers in that direction! username_2: This behaviour still occurs both on chrome and MS Edge. This is quite annoying. Apparantly no solution until so far. Is a workaround available (besides just typing the characters?) username_3: See https://github.com/jupyterlab/jupyterlab/issues/4519 for a way to do this (and a simple change to jlab that needs to happen to make this easy - PRs welcome!). username_4: @username_3 it would be nice see this fixed in a stable & sensible way. All IDEs allow to customise *easily* keyboard shortcuts. The absence of a customiseable shortcut for `<-` and %>% is a put off for R users. In my humple view a way to achieve this `should be offered in jpyterlab`, not pushed down to be add-ons implementers responsibility. username_3: I think maybe there is a misunderstanding. I'm proposing what we offer in core JupyterLab (see https://github.com/jupyterlab/jupyterlab/issues/4519) is a way for any user to customize a keyboard shortcut to insert any text they want. username_4: Apologies for any misunderstanding from my part! username_3: JLab 2 offers a way for a user to customize the keyboard shortcuts to insert any text they like into editors: https://jupyterlab.readthedocs.io/en/stable/getting_started/changelog.html#user-facing-changes Status: Issue closed
tarantool/cartridge-cli
582843666
Title: drop "rocks pack support" Question: username_0: There are several problems in such sphere: - cartridge-cli is an additional level between `tarantoolctl rocks pack` and it's hard-maintainable to support opportune interface. - rocks pack supports several modes ".all.rock"/".src.rock"/... and cartridge-cli knows nothing about it. - `tarantoolctl rocks pack` could be dependent on Makefile/CMakeLists and cartridge-cli couldn't verify correctness of such files - `cartridge pack` simply has a different packaging flow. So, we have a function that is already completely broken and could return some unpredictable result after packaging. So, I suggest simply drop it - if user wants to get ".rock" (s)he should be able to call `tarantoolctl rocks pack` directly.<issue_closed> Status: Issue closed
ikedaosushi/tech-news
386834313
Title: プロジェクト計画について思うこと Question: username_0: &#12503;&#12525;&#12472;&#12455;&#12463;&#12488;&#35336;&#30011;&#12395;&#12388;&#12356;&#12390;&#24605;&#12358;&#12371;&#12392;<br> &#12371;&#12398;&#35352;&#20107;&#12399;&#12289;&#12399;&#12390;&#12394;&#12456;&#12531;&#12472;&#12491;&#12450; Advent Calendar 2018 &#12398;3&#26085;&#30446;&#12398;&#35352;&#20107;&#12391;&#12377;&#12290; &#26152;&#26085;&#12398;&#35352;&#20107;&#12399;, id:t_kyt &#12398;&#65378;time_zone&#35373;&#23450;&#12398;&#36949;&#12358;MySQL&#12398;&#12524;&#12503;&#12522;&#12465;&#12540;&#12471;&#12519;&#12531;&#12395;&#12388;&#12356;&#12390; - &#35282;&#24453;&#12385;&#12399;&#23550;&#31354;&#65379;&#12391;&#12375;&#12383;&#12290;<br> https://ift.tt/2DUm4Ym
ant-design/pro-components
982029638
Title: 🐛[BUG] EditableProTable 开启了 expandRowByClick 后, 简直是灾难 Question: username_0: ### 🐛 bug 描述 <!-- 详细地描述 bug,让大家都能理解 --> 由于开启expandRowByClick 导致在edit数据(切换焦点), 点击 保存 或 取消 时, 均会触发当前行的toggle expand ### 📷 复现步骤 官方的"有子列表格"例子, 加入expandRowByClick即可 <!-- 清晰描述复现步骤,让别人也能看到问题 --> ### 🏞 期望结果 将event透传到onCancel, onSave等等上, 可以手动stopPropagation(), 或者其他更好的办法 <!-- 描述你原本期望看到的结果 --> ### 💻 复现代码 <!-- 提供可复现的代码,仓库,或线上示例 --> ### © 版本信息 - ProComponents 版本: [e.g. 4.0.0] - umi 版本 - 浏览器环境 - 开发环境 [e.g. mac OS] ### 🚑 其他信息 <!-- 如截图等其他信息可以贴在这里 --><issue_closed> Status: Issue closed
redisson/redisson
874301730
Title: While putting value into Rmap It throws io.netty.util.IllegalReferenceCountException: refCnt: 0 Question: username_0: <!-- Сonsider Redisson PRO https://redisson.pro version for advanced features and support by SLA. --> **Expected behavior** RMap.put should succeed **Actual behavior** RMap.put throws io.netty.util.IllegalReferenceCountException: refCnt: 0 when a java bean is set as value in RMap **Steps to reproduce or test case** **Redission Initialization** ``` private RMap<String, Expression> expressionMap; this.expressionMap = this.redissonClient.getMap("ExpressionMap"); ``` **Expression Class** ``` @Data @NoArgsConstructor @AllArgsConstructor @Builder @JsonInclude(JsonInclude.Include.NON_NULL) public class Expression { private String expression; private boolean custom; } ``` **setting key and value** ``` Expression expression = new Expression(); expression.setExpression("a > 20"); expression.setCustom(false); this.expressionMap.put(uuid, expression); ``` **System Setup** My application is a spring-boot application. Tried Running on both Tomcat and Undertow embedded server but the result is the same. My SpringBoot version is 2.4.3 but have tried updating it to 2.4.2 and similarly 3.15.3 but the result is the same. **Exception Trace** ``` io.netty.util.IllegalReferenceCountException: refCnt: 0 at io.netty.buffer.AbstractByteBuf.ensureAccessible(AbstractByteBuf.java:1456) ~[netty-buffer-4.1.59.Final.jar:4.1.59.Final] at io.netty.buffer.AbstractByteBuf.ensureWritable0(AbstractByteBuf.java:291) ~[netty-buffer-4.1.59.Final.jar:4.1.59.Final] at io.netty.buffer.AbstractByteBuf.ensureWritable(AbstractByteBuf.java:282) ~[netty-buffer-4.1.59.Final.jar:4.1.59.Final] at io.netty.buffer.AbstractByteBuf.writeBytes(AbstractByteBuf.java:1075) ~[netty-buffer-4.1.59.Final.jar:4.1.59.Final] at org.redisson.codec.MarshallingCodec$ByteOutputWrapper.write(MarshallingCodec.java:129) ~[redisson-3.15.3.jar:3.15.3] at org.jboss.marshalling.SimpleDataOutput.flush(SimpleDataOutput.java:336) ~[jboss-marshalling-2.0.11.Final.jar:2.0.11.Final] at org.jboss.marshalling.SimpleDataOutput.finish(SimpleDataOutput.java:378) ~[jboss-marshalling-2.0.11.Final.jar:2.0.11.Final] at org.jboss.marshalling.AbstractMarshaller.finish(AbstractMarshaller.java:126) ~[jboss-marshalling-2.0.11.Final.jar:2.0.11.Final] at org.redisson.codec.MarshallingCodec.lambda$new$1(MarshallingCodec.java:169) ~[redisson-3.15.3.jar:3.15.3] at org.redisson.command.CommandAsyncService.encodeMapValue(CommandAsyncService.java:715) ~[redisson-3.15.3.jar:3.15.3] at org.redisson.RedissonObject.encodeMapValue(RedissonObject.java:305) ~[redisson-3.15.3.jar:3.15.3] at org.redisson.RedissonMap.putOperationAsync(RedissonMap.java:1336) ~[redisson-3.15.3.jar:3.15.3] at org.redisson.RedissonMap.putAsync(RedissonMap.java:1322) ~[redisson-3.15.3.jar:3.15.3] at org.redisson.RedissonMap.put(RedissonMap.java:644) ~[redisson-3.15.3.jar:3.15.3] ``` [Truncated] idleConnectionTimeout: 10000 connectTimeout: 10000 timeout: 3000 retryAttempts: 3 retryInterval: 1500 password: <PASSWORD> subscriptionsPerConnection: 5 clientName: null address: "redis://127.0.0.1:6379" subscriptionConnectionMinimumIdleSize: 1 subscriptionConnectionPoolSize: 50 connectionMinimumIdleSize: 10 connectionPoolSize: 20 database: 0 dnsMonitoringInterval: 5000 threads: 16 nettyThreads: 32 codec: !<org.redisson.codec.MarshallingCodec> {} transportMode: "NIO" ``` Answers: username_1: Unable to reproduce. Seems there is some issue with bytebuffer allocation. try to add these setting in jvm args: `-Dio.netty.noPreferDirect=true -Dio.netty.allocator.type=unpooled` username_0: I still see the same error after applying the above JVM arguments. ![image](https://user-images.githubusercontent.com/3108570/116855020-40b82500-ac16-11eb-905c-38a1c4a227d6.png) username_1: Can you provide sample code which I can run? username_0: Thanks @username_1 Please find the unit test code which is failing on my system [redission-issue.zip](https://github.com/redisson/redisson/files/6419130/redission-issue.zip) Junit Failure: ![image](https://user-images.githubusercontent.com/3108570/116964243-abc23400-acc8-11eb-84cc-8fb650d4d405.png) username_1: It works. Here is my output with JDK 13: ``` "C:\Program Files\Java\jdk1.8.0_201\bin\java.exe" -javaagent:C:\Devel\IntelliJ\lib\idea_rt.jar=51554:C:\Devel\IntelliJ\bin -Dfile.encoding=UTF-8 -classpath "C:\Program Files\Java\jdk1.8.0_201\jre\lib\charsets.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\deploy.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\access-bridge-64.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\cldrdata.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\dnsns.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\jaccess.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\jfxrt.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\localedata.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\nashorn.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\sunec.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\sunjce_provider.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\sunmscapi.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\sunpkcs11.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\zipfs.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\javaws.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\jce.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\jfr.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\jfxswt.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\jsse.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\management-agent.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\plugin.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\resources.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\rt.jar;C:\Downloads\issue\target\classes;C:\Users\root\.m2\repository\org\redisson\redisson-all\3.15.0\redisson-all-3.15.0.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-devtools\2.3.7.RELEASE\spring-boot-devtools-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot\2.3.7.RELEASE\spring-boot-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-core\5.2.12.RELEASE\spring-core-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-jcl\5.2.12.RELEASE\spring-jcl-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-context\5.2.12.RELEASE\spring-context-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-autoconfigure\2.3.7.RELEASE\spring-boot-autoconfigure-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\org\projectlombok\lombok\1.18.14\lombok-1.18.14.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-starter-web\2.3.7.RELEASE\spring-boot-starter-web-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-starter\2.3.7.RELEASE\spring-boot-starter-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-starter-logging\2.3.7.RELEASE\spring-boot-starter-logging-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\ch\qos\logback\logback-classic\1.2.3\logback-classic-1.2.3.jar;C:\Users\root\.m2\repository\ch\qos\logback\logback-core\1.2.3\logback-core-1.2.3.jar;C:\Users\root\.m2\repository\org\slf4j\slf4j-api\1.7.30\slf4j-api-1.7.30.jar;C:\Users\root\.m2\repository\org\apache\logging\log4j\log4j-to-slf4j\2.13.3\log4j-to-slf4j-2.13.3.jar;C:\Users\root\.m2\repository\org\apache\logging\log4j\log4j-api\2.13.3\log4j-api-2.13.3.jar;C:\Users\root\.m2\repository\org\slf4j\jul-to-slf4j\1.7.30\jul-to-slf4j-1.7.30.jar;C:\Users\root\.m2\repository\jakarta\annotation\jakarta.annotation-api\1.3.5\jakarta.annotation-api-1.3.5.jar;C:\Users\root\.m2\repository\org\yaml\snakeyaml\1.26\snakeyaml-1.26.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-starter-json\2.3.7.RELEASE\spring-boot-starter-json-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\com\fasterxml\jackson\core\jackson-databind\2.11.3\jackson-databind-2.11.3.jar;C:\Users\root\.m2\repository\com\fasterxml\jackson\core\jackson-annotations\2.11.3\jackson-annotations-2.11.3.jar;C:\Users\root\.m2\repository\com\fasterxml\jackson\core\jackson-core\2.11.3\jackson-core-2.11.3.jar;C:\Users\root\.m2\repository\com\fasterxml\jackson\datatype\jackson-datatype-jdk8\2.11.3\jackson-datatype-jdk8-2.11.3.jar;C:\Users\root\.m2\repository\com\fasterxml\jackson\datatype\jackson-datatype-jsr310\2.11.3\jackson-datatype-jsr310-2.11.3.jar;C:\Users\root\.m2\repository\com\fasterxml\jackson\module\jackson-module-parameter-names\2.11.3\jackson-module-parameter-names-2.11.3.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-starter-tomcat\2.3.7.RELEASE\spring-boot-starter-tomcat-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\org\apache\tomcat\embed\tomcat-embed-core\9.0.41\tomcat-embed-core-9.0.41.jar;C:\Users\root\.m2\repository\org\glassfish\jakarta.el\3.0.3\jakarta.el-3.0.3.jar;C:\Users\root\.m2\repository\org\apache\tomcat\embed\tomcat-embed-websocket\9.0.41\tomcat-embed-websocket-9.0.41.jar;C:\Users\root\.m2\repository\org\springframework\spring-web\5.2.12.RELEASE\spring-web-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-beans\5.2.12.RELEASE\spring-beans-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-webmvc\5.2.12.RELEASE\spring-webmvc-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-aop\5.2.12.RELEASE\spring-aop-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-expression\5.2.12.RELEASE\spring-expression-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\javax\annotation\javax.annotation-api\1.3.2\javax.annotation-api-1.3.2.jar" org.issueRedisson.springBoot.StartRedisTestBoot . ____ _ __ _ _ /\\ / ___'_ __ _ _(_)_ __ __ _ \ \ \ \ ( ( )\___ | '_ | '_| | '_ \/ _` | \ \ \ \ \\/ ___)| |_)| | | | | || (_| | ) ) ) ) ' |____| .__|_| |_|_| |_\__, | / / / / =========|_|==============|___/=/_/_/_/ :: Spring Boot :: (v2.3.7.RELEASE) 2021-05-04 09:09:14.426 INFO 4024 --- [ restartedMain] o.i.springBoot.StartRedisTestBoot : Starting StartRedisTestBoot on Computer with PID 4024 (C:\Downloads\issue\target\classes started by root in C:\Downloads\issue) 2021-05-04 09:09:14.429 INFO 4024 --- [ restartedMain] o.i.springBoot.StartRedisTestBoot : No active profile set, falling back to default profiles: default 2021-05-04 09:09:14.482 INFO 4024 --- [ restartedMain] .e.DevToolsPropertyDefaultsPostProcessor : Devtools property defaults active! Set 'spring.devtools.add-properties' to 'false' to disable 2021-05-04 09:09:14.482 INFO 4024 --- [ restartedMain] .e.DevToolsPropertyDefaultsPostProcessor : For additional web related logging consider setting the 'logging.level.web' property to 'DEBUG' 2021-05-04 09:09:15.417 INFO 4024 --- [ restartedMain] o.s.b.w.embedded.tomcat.TomcatWebServer : Tomcat initialized with port(s): 8080 (http) 2021-05-04 09:09:15.427 INFO 4024 --- [ restartedMain] o.apache.catalina.core.StandardService : Starting service [Tomcat] 2021-05-04 09:09:15.427 INFO 4024 --- [ restartedMain] org.apache.catalina.core.StandardEngine : Starting Servlet engine: [Apache Tomcat/9.0.41] 2021-05-04 09:09:15.486 INFO 4024 --- [ restartedMain] o.a.c.c.C.[Tomcat].[localhost].[/] : Initializing Spring embedded WebApplicationContext 2021-05-04 09:09:15.487 INFO 4024 --- [ restartedMain] w.s.c.ServletWebServerApplicationContext : Root WebApplicationContext: initialization completed in 1004 ms 2021-05-04 09:09:16.145 DEBUG 4024 --- [sson-netty-2-21] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@560565971 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x758d3ca6, L:/127.0.0.1:51625 - R:/127.0.0.1:6379], currentCommand=null] 2021-05-04 09:09:16.145 DEBUG 4024 --- [sson-netty-2-23] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@1710482627 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0xfb62d088, L:/127.0.0.1:51627 - R:/127.0.0.1:6379], currentCommand=null] 2021-05-04 09:09:16.145 DEBUG 4024 --- [sson-netty-2-20] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@1773270043 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x25f61d78, L:/127.0.0.1:51626 - R:/127.0.0.1:6379], currentCommand=null] 2021-05-04 09:09:16.145 DEBUG 4024 --- [sson-netty-2-16] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@1036087503 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x51d1db38, L:/127.0.0.1:51631 - R:/127.0.0.1:6379], currentCommand=null] 2021-05-04 09:09:16.145 DEBUG 4024 --- [sson-netty-2-17] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@966637890 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0xe5958ec5, L:/127.0.0.1:51633 - R:/127.0.0.1:6379], currentCommand=null] 2021-05-04 09:09:16.145 DEBUG 4024 --- [sson-netty-2-13] o.r.connection.ClientConnectionsEntry : new pubsub connection created: RedisPubSubConnection@790605312 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x0269ba46, L:/127.0.0.1:51630 - R:/127.0.0.1:6379], currentCommand=null] 2021-05-04 09:09:16.145 DEBUG 4024 --- [sson-netty-2-15] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@584908344 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x3f5fede3, L:/127.0.0.1:51629 - R:/127.0.0.1:6379], currentCommand=null] 2021-05-04 09:09:16.145 DEBUG 4024 --- [sson-netty-2-18] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@1323797976 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x05f3e864, L:/127.0.0.1:51632 - R:/127.0.0.1:6379], currentCommand=null] 2021-05-04 09:09:16.145 DEBUG 4024 --- [sson-netty-2-22] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@612699270 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0xe949bdef, L:/127.0.0.1:51628 - R:/127.0.0.1:6379], currentCommand=null] 2021-05-04 09:09:16.145 DEBUG 4024 --- [sson-netty-2-14] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@648927375 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x21914d2f, L:/127.0.0.1:51624 - R:/127.0.0.1:6379], currentCommand=null] 2021-05-04 09:09:16.145 DEBUG 4024 --- [sson-netty-2-19] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@753784165 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x5066cfc9, L:/127.0.0.1:51623 - R:/127.0.0.1:6379], currentCommand=null] 2021-05-04 09:09:16.152 INFO 4024 --- [sson-netty-2-13] o.r.c.pool.MasterPubSubConnectionPool : 1 connections initialized for /127.0.0.1:6379 2021-05-04 09:09:16.155 DEBUG 4024 --- [isson-netty-2-2] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@1752841329 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x245f4c97, L:/127.0.0.1:51634 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@3356e82a(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.155 DEBUG 4024 --- [isson-netty-2-3] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@90257875 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x23fde960, L:/127.0.0.1:51635 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@7c414c(success: PONG)], command=(PING), params=[], codec=null]] 2021-05-04 09:09:16.155 DEBUG 4024 --- [isson-netty-2-4] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@1505320775 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0xd4f119dc, L:/127.0.0.1:51636 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@459d9abb(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.159 DEBUG 4024 --- [sson-netty-2-10] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@134386209 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x18886829, L:/127.0.0.1:51638 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@728e38c3(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.159 DEBUG 4024 --- [isson-netty-2-8] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@1018666403 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x74f05e90, L:/127.0.0.1:51637 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@796491e5(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.159 DEBUG 4024 --- [isson-netty-2-9] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@30448953 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x46151cdf, L:/127.0.0.1:51639 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@1df41459(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.160 DEBUG 4024 --- [sson-netty-2-12] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@71649161 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0xd8430c00, L:/127.0.0.1:51640 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@29b716b4(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.160 DEBUG 4024 --- [sson-netty-2-13] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@1780299143 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x99a32bc5, L:/127.0.0.1:51641 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@71caa1f1(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.162 DEBUG 4024 --- [sson-netty-2-14] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@743814143 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x6c085cde, L:/127.0.0.1:51642 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@69572811(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.162 DEBUG 4024 --- [sson-netty-2-16] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@926489535 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x64a35433, L:/127.0.0.1:51644 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@723b833c(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.162 DEBUG 4024 --- [sson-netty-2-15] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@673486152 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x4f10008b, L:/127.0.0.1:51647 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@6d260e28(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.162 DEBUG 4024 --- [sson-netty-2-17] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@1635158518 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0xe6b0611d, L:/127.0.0.1:51646 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@4d293cac(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.163 DEBUG 4024 --- [sson-netty-2-18] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@975442831 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0xe99fd6c6, L:/127.0.0.1:51645 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@485fd05e(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.163 DEBUG 4024 --- [sson-netty-2-20] o.r.connection.ClientConnectionsEntry : new connection created: RedisConnection@1665769323 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x9ba6a183, L:/127.0.0.1:51643 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@264b0d66(incomplete)], command=(PING), params=[], codec=org.redisson.client.codec.StringCodec]] 2021-05-04 09:09:16.163 INFO 4024 --- [sson-netty-2-20] o.r.c.pool.MasterConnectionPool : 24 connections initialized for /127.0.0.1:6379 2021-05-04 09:09:16.165 DEBUG 4024 --- [ restartedMain] org.redisson.connection.DNSMonitor : DNS monitoring enabled; Current masters: {redis://127.0.0.1:6379=/127.0.0.1:6379}, slaves: {} publishing... 2021-05-04 09:09:16.216 DEBUG 4024 --- [ restartedMain] org.redisson.command.RedisExecutor : acquired connection for command (PUBLISH) and params [xxx, PooledUnsafeDirectByteBuf(ridx: 0, widx: 91, cap: 256)] from slot NodeSource [slot=0, addr=null, redisClient=null, redirect=null, entry=null] using node /127.0.0.1:6379... RedisConnection@560565971 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x758d3ca6, L:/127.0.0.1:51625 - R:/127.0.0.1:6379], currentCommand=null] 2021-05-04 09:09:16.218 DEBUG 4024 --- [sson-netty-2-11] org.redisson.command.RedisExecutor : connection released for command (PUBLISH) and params [xxx, PooledUnsafeDirectByteBuf(ridx: 0, widx: 91, cap: 256)] from slot NodeSource [slot=0, addr=null, redisClient=null, redirect=null, entry=null] using connection RedisConnection@560565971 [redisClient=[addr=redis://127.0.0.1:6379], channel=[id: 0x758d3ca6, L:/127.0.0.1:51625 - R:/127.0.0.1:6379], currentCommand=CommandData [promise=RedissonPromise [promise=ImmediateEventExecutor$ImmediatePromise@1c689972(success: 1)], command=(PUBLISH), params=[xxx, PooledUnsafeDirectByteBuf(ridx: 0, widx: 91, cap: 256)], codec=org.redisson.client.codec.StringCodec]] published... 2021-05-04 09:09:16.404 INFO 4024 --- [ restartedMain] o.s.s.concurrent.ThreadPoolTaskExecutor : Initializing ExecutorService 'applicationTaskExecutor' 2021-05-04 09:09:16.532 INFO 4024 --- [ restartedMain] o.s.b.d.a.OptionalLiveReloadServer : LiveReload server is running on port 35729 2021-05-04 09:09:16.550 INFO 4024 --- [ restartedMain] o.s.b.w.embedded.tomcat.TomcatWebServer : Tomcat started on port(s): 8080 (http) with context path '' 2021-05-04 09:09:16.559 INFO 4024 --- [ restartedMain] o.i.springBoot.StartRedisTestBoot : Started StartRedisTestBoot in 2.429 seconds (JVM running for 2.803) 2021-05-04 09:09:21.165 DEBUG 4024 --- [sson-netty-2-21] org.redisson.connection.DNSMonitor : Request sent to resolve ip address for master host: 127.0.0.1 2021-05-04 09:09:21.167 DEBUG 4024 --- [sson-netty-2-19] org.redisson.connection.DNSMonitor : Resolved ip: /127.0.0.1 for master host: 127.0.0.1 2021-05-04 09:09:26.167 DEBUG 4024 --- [sson-netty-2-22] org.redisson.connection.DNSMonitor : Request sent to resolve ip address for master host: 127.0.0.1 2021-05-04 09:09:26.167 DEBUG 4024 --- [sson-netty-2-19] org.redisson.connection.DNSMonitor : Resolved ip: /127.0.0.1 for master host: 127.0.0.1 ``` username_1: did you try to run it with openjdk? username_0: I am using Oracle jdk-11.0.10. Do you advise checking this in Open JDK? The spring version which you are using is 2.3.7, whereas I am using 2.4.3. It seems you are running on JDK 8 from the logs. `"C:\Program Files\Java\jdk1.8.0_201\bin\java.exe" -javaagent:C:\Devel\IntelliJ\lib\idea_rt.jar=51554:C:\Devel\IntelliJ\bin -Dfile.encoding=UTF-8 -classpath "C:\Program Files\Java\jdk1.8.0_201\jre\lib\charsets.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\deploy.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\access-bridge-64.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\cldrdata.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\dnsns.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\jaccess.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\jfxrt.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\localedata.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\nashorn.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\sunec.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\sunjce_provider.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\sunmscapi.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\sunpkcs11.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\ext\zipfs.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\javaws.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\jce.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\jfr.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\jfxswt.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\jsse.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\management-agent.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\plugin.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\resources.jar;C:\Program Files\Java\jdk1.8.0_201\jre\lib\rt.jar;C:\Downloads\issue\target\classes;C:\Users\root\.m2\repository\org\redisson\redisson-all\3.15.0\redisson-all-3.15.0.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-devtools\2.3.7.RELEASE\spring-boot-devtools-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot\2.3.7.RELEASE\spring-boot-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-core\5.2.12.RELEASE\spring-core-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-jcl\5.2.12.RELEASE\spring-jcl-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-context\5.2.12.RELEASE\spring-context-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-autoconfigure\2.3.7.RELEASE\spring-boot-autoconfigure-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\org\projectlombok\lombok\1.18.14\lombok-1.18.14.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-starter-web\2.3.7.RELEASE\spring-boot-starter-web-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-starter\2.3.7.RELEASE\spring-boot-starter-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-starter-logging\2.3.7.RELEASE\spring-boot-starter-logging-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\ch\qos\logback\logback-classic\1.2.3\logback-classic-1.2.3.jar;C:\Users\root\.m2\repository\ch\qos\logback\logback-core\1.2.3\logback-core-1.2.3.jar;C:\Users\root\.m2\repository\org\slf4j\slf4j-api\1.7.30\slf4j-api-1.7.30.jar;C:\Users\root\.m2\repository\org\apache\logging\log4j\log4j-to-slf4j\2.13.3\log4j-to-slf4j-2.13.3.jar;C:\Users\root\.m2\repository\org\apache\logging\log4j\log4j-api\2.13.3\log4j-api-2.13.3.jar;C:\Users\root\.m2\repository\org\slf4j\jul-to-slf4j\1.7.30\jul-to-slf4j-1.7.30.jar;C:\Users\root\.m2\repository\jakarta\annotation\jakarta.annotation-api\1.3.5\jakarta.annotation-api-1.3.5.jar;C:\Users\root\.m2\repository\org\yaml\snakeyaml\1.26\snakeyaml-1.26.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-starter-json\2.3.7.RELEASE\spring-boot-starter-json-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\com\fasterxml\jackson\core\jackson-databind\2.11.3\jackson-databind-2.11.3.jar;C:\Users\root\.m2\repository\com\fasterxml\jackson\core\jackson-annotations\2.11.3\jackson-annotations-2.11.3.jar;C:\Users\root\.m2\repository\com\fasterxml\jackson\core\jackson-core\2.11.3\jackson-core-2.11.3.jar;C:\Users\root\.m2\repository\com\fasterxml\jackson\datatype\jackson-datatype-jdk8\2.11.3\jackson-datatype-jdk8-2.11.3.jar;C:\Users\root\.m2\repository\com\fasterxml\jackson\datatype\jackson-datatype-jsr310\2.11.3\jackson-datatype-jsr310-2.11.3.jar;C:\Users\root\.m2\repository\com\fasterxml\jackson\module\jackson-module-parameter-names\2.11.3\jackson-module-parameter-names-2.11.3.jar;C:\Users\root\.m2\repository\org\springframework\boot\spring-boot-starter-tomcat\2.3.7.RELEASE\spring-boot-starter-tomcat-2.3.7.RELEASE.jar;C:\Users\root\.m2\repository\org\apache\tomcat\embed\tomcat-embed-core\9.0.41\tomcat-embed-core-9.0.41.jar;C:\Users\root\.m2\repository\org\glassfish\jakarta.el\3.0.3\jakarta.el-3.0.3.jar;C:\Users\root\.m2\repository\org\apache\tomcat\embed\tomcat-embed-websocket\9.0.41\tomcat-embed-websocket-9.0.41.jar;C:\Users\root\.m2\repository\org\springframework\spring-web\5.2.12.RELEASE\spring-web-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-beans\5.2.12.RELEASE\spring-beans-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-webmvc\5.2.12.RELEASE\spring-webmvc-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-aop\5.2.12.RELEASE\spring-aop-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\org\springframework\spring-expression\5.2.12.RELEASE\spring-expression-5.2.12.RELEASE.jar;C:\Users\root\.m2\repository\javax\annotation\javax.annotation-api\1.3.2\javax.annotation-api-1.3.2.jar" org.issueRedisson.springBoot.StartRedisTestBoot` username_0: I did try testing with Open JDK and still facing the same issue. ![image](https://user-images.githubusercontent.com/3108570/116974610-06b15680-acdc-11eb-881a-4ecf7d1a9295.png) username_1: It works fine with Spring Boot 2.4.3. username_1: Try to store String value instead of expression object. username_0: That works. I have serialized the Expression Object to a JSON String and have been using it for now. username_1: As option you can use JsonJacksonCodec instead username_0: Thanks @username_1 Yes it works with JsonJacksonCodec. However, I still feel that it fails with MarshallingCodec. Will continue with JsonJacksonCodec for now. username_1: Can you update test project? So I could reproduce it username_0: Sure. Please find the attached spring boot project [redission-test.zip](https://github.com/redisson/redisson/files/6445458/redission-test.zip). Once you import the project you can run the test class RedissionTestApplicationTests.java ![redission-test-class](https://user-images.githubusercontent.com/3108570/117531464-3758ff00-b000-11eb-825a-d64bb763e535.jpg) username_2: Hi am am also encountering this. Java Version: ``` openjdk 172.16.58.3 2020-11-04 OpenJDK Runtime Environment AdoptOpenJDK (build 172.16.58.3+1) OpenJDK 64-Bit Server VM AdoptOpenJDK (build 172.16.58.3+1, mixed mode) ``` The RMap (I'm using the ConcurrentMap interface portion of it): ``` private final ConcurrentMap<String, Object> shutdownDisabledMap; ... static final Object PRESENT = new Object(); ``` And then whenever this executes (foo and the id are never null.): ``` shutdownDisabledMap.put(String.valueOf(foo.get_id()), PRESENT) ``` I get: ``` io.netty.util.IllegalReferenceCountException: refCnt: 0 at io.netty.buffer.AbstractByteBuf.ensureAccessible(AbstractByteBuf.java:1454) at io.netty.buffer.AbstractByteBuf.ensureWritable0(AbstractByteBuf.java:289) at io.netty.buffer.AbstractByteBuf.ensureWritable(AbstractByteBuf.java:280) at io.netty.buffer.AbstractByteBuf.writeBytes(AbstractByteBuf.java:1073) at org.redisson.codec.MarshallingCodec$ByteOutputWrapper.write(MarshallingCodec.java:129) at org.jboss.marshalling.SimpleDataOutput.flush(SimpleDataOutput.java:336) at org.jboss.marshalling.SimpleDataOutput.finish(SimpleDataOutput.java:378) at org.jboss.marshalling.AbstractMarshaller.finish(AbstractMarshaller.java:126) at org.redisson.codec.MarshallingCodec.lambda$new$1(MarshallingCodec.java:169) at org.redisson.command.CommandAsyncService.encodeMapValue(CommandAsyncService.java:715) at org.redisson.RedissonObject.encodeMapValue(RedissonObject.java:305) at org.redisson.RedissonMap.putOperationAsync(RedissonMap.java:1336) at org.redisson.RedissonMap.putAsync(RedissonMap.java:1322) at org.redisson.RedissonMap.put(RedissonMap.java:644) ``` And the docker compose is using: ``` redis: image: redis:6.0.5-alpine ```
trinodb/trino
905499095
Title: from_iso8601_timestamp_nanos is not consistent with from_iso8601_timestamp or its own docs Question: username_0: The behavior of `from_iso8601_timestamp_nanos` differs both from its documentation and the related `from_iso8601_timestamp`. `from_iso8601_timestamp_nanos`... - requires the time component (contrary to its documentation) - does not support ordinal (day-of-year) and week (-of-year) formats (`'2021-148T00:00'`, `'2021-W21-5T00:00'`) - does support political time zones (not in ISO 8601) - and can combine them with offsets, which is just bizarre (`'2000-01-01T12:00+01:00[Etc/GMT-6]'`) - does not reject timestamps that are invalid in the system zone (like 2011-12-30 in Pacific/Apia) This issue is caused by the use of `java.time`'s parsing tools instead of Joda's in [the implementation of `from_iso8601_timestamp_nanos`](https://github.com/trinodb/trino/blob/fe608f2723842037ff620d612a706900e79c52c8/core/trino-main/src/main/java/io/trino/operator/scalar/DateTimeFunctions.java#L217-L244). Answers: username_1: Could you fix this for 358 @username_0 and then we merge docs and code PRs at the same time? Requested by @electrum username_0: I've determined that there isn't a simple and clean way to make `from_iso8601_timestamp_nanos` match the behavior of `from_iso8601_timestamp` perfectly. The easiest near-complete solution probably is to use Joda to parse out the date and time, and to manually extract the fractional seconds. This would leave out a few of the formats allowed by ISO 8601 (namely fractional hours and minutes), but I don't think they are especially important cases. username_2: Were we using java.time here because Joda only supports millisecond precision? cc: @martint username_1: Okay ... as discussed with @electrum I will therefore update the documentation PR to the current behaviour. Once we fix the code .. we have to just update the docs as well.
ruby-china/homeland
197528925
Title: 邮件缺少配置项 Question: username_0: 今天测试了一天发现无法发送邮件(QQ企业邮箱),我看了一下homeland的配置 ` address: '<%= ENV['mailer_options.address'] %>' port: <%= ENV['mailer_options.port'] %> domain: '<%= ENV['mailer_options.domain'] %>' user_name: '<%= ENV['mailer_options.user_name'] %>' password: '<%= ENV['<PASSWORD>'] %>' authentication: '<%= ENV['mailer_options.authentication'] %>' enable_starttls_auto: <%= ENV['mailer_options.enable_starttls_auto'] %>` 然后下载了mail包,看了一下源码: `class SMTP include Mail::CheckDeliveryParams def initialize(values) self.settings = { :address => "localhost", :port => 25, :domain => 'localhost.localdomain', :user_name => nil, :password => nil, :authentication => nil, :enable_starttls_auto => true, :openssl_verify_mode => nil, :ssl => nil, :tls => nil }.merge!(values) end attr_accessor :settings` 当ssl为true时可以发送邮件。 也就是说,homeland没有这个配置选项,当第三方强制要求使用SSL时,就会导致发送邮件失败 Answers: username_1: `mailer_options` 是可以支持任意 key 的,只是文档没有写出来 username_2: wow! 刚遇见这个坑,就碰见有人提了。。。太好了,强烈建议 @username_1 在gethomeland文档上标识一下,说不定很多像我这样的小白解决不了 Status: Issue closed username_3: 使用 qq邮箱的时候 还需要设置tls true
ccxt/ccxt
288516642
Title: New Exchange: bitsane Question: username_0: api doc: https://bitsane.com/info-api web url: https://bitsane.com/ priority: low avg volume: https://coinmarketcap.com/exchanges/bitsane/ Answers: username_0: this exchange is in the list but its not too important as it doesnt have major volume in this exchange username_1: I could take a shot at this one too. username_2: @username_1 that would be awesome! I'm currently merging the cobinhood request. Status: Issue closed
getkirby/getkirby.com
414071822
Title: Caching - option ignore confusion Question: username_0: Hey, in the [caching guide](https://getkirby.com/docs/guide/cache) the example `ignore` function doesn't fit the option description. #### `ignore` function: ```php return [ 'cache' => [ 'pages' => [ 'active' => true, 'ignore' => function ($page) { return $page->title()->value() !== 'Do not cache me'; } ] ] ]; ``` #### ignore description Option | Value -- | -- ignore | **A function that returns a list of pages to ignore.** My guess is that the description should be something along the lines of: "A function that returns true for pages that should be ignored and false for pages that should be cached." Status: Issue closed Answers: username_1: Thanks a lot for your suggestion. https://github.com/getkirby/getkirby.com/commit/1343157b739ac1450cb34234d88a0f5c47690b03
neerajwahi/pairjam
32783538
Title: Getting Started Information Question: username_0: The Readme has a good writeup on the client and server stack, and github integration, but little on how to get setup locally. I tried for a bit and had a couple questions: 1. not sure how to fill out this form for the api key: http://cl.ly/VKgu 2. not sure what to run to start the app 3. not sure if there's anything else I should do to setup dependencies other than `npm install` in client and server 4. not sure what the testing story is. I see some tests in client, how are they run what are you trying to test, what would you like other devs to test for contributions I had a couple thoughts 1. it wasn't obvious to look for `package.json` in the server and client. 2. it might be nice to have a stubbed `github_api_secret` file in the repo. That way, I don't have to think about whether it should be in the server repo or copy the contents :) Answers: username_1: Which version of React is your project running on? Cause `React.addons.CSSTransitionGroup` seems deprecated. I'm running `[email protected]` (Since, the `package.json` file mentions the version of `react` to be 'latest'.
imgix/ember-cli-imgix
341655221
Title: Old browser support? (failures on Safari 9.1.2 July 2016) Question: username_0: Right now, we support all the browsers that the latest ember version supports. [This file](https://github.com/emberjs/ember.js/blob/54cb3b075d1173787e492cc7837e7d54e75c816a/testem.dist.js) outlines the versions that ember core auto-runs on browserstack. Specifically, we use [window.Url](https://developer.mozilla.org/en-US/docs/Web/API/URL/searchParams]) to parse the url and its params. This api is not available on older browsers (Safari 9.1.2 from July 2016). My question is, How far back should we support? My feeling is that, We should support all browsers and browser versions that the latest ember supports. If we support only what ember supports then the use of the `window.URL.searchParams` api is totally legit... and this can be closed. We should probably add a note to the readme, something like "We support the browsers that the latest Ember supports" If we DO want to support older browser that the latest ember does not technically support (but actually do work), We should use something like this: https://github.com/defunctzombie/node-url ^ We might be able to find something smaller as we only need the param parsing. ^ Would be a near trivial import using https://github.com/ef4/ember-auto-import ^ We could also use that lib in fastboot if we wanted to add better fastboot support. Here is a screenshot of that api failing on safari 9.1.2 (over 2 years old) ![image](https://user-images.githubusercontent.com/1612785/42779367-139f54f4-88f5-11e8-9941-36c9ddd22042.png) This update will be trivial and I need to do the work anyways as TheDyrt's users seem to have old computers. Thoughts? Answers: username_1: Hey @username_0. Thanks for the detailed report. My hunch is the same as yours - that we should only support the browsers that the latest version of ember supports. I think it's dangerous to say that we support browsers that ember doesn't support because there could be non-obvious, awkward edge cases that may be hard to handle. I would be against using `node-url` as a) it's another dependency, and we should try to limit dependencies, and b) it's not a very large package (only has 100 stars on github). In the past we've used `uri.js` to handle URLs, so if we *had* to add a dependency back, I'd like to use `uri.js` over `node-url` as it has more stars and is more supported. Overall, my first impression is that we should leave the code how it is now. How does this affect your work? username_1: In saying this, I've just found that [window.URL is not supported on IE 11](https://caniuse.com/#search=URL) - so we should probably fix this. Looking at `node-url` again, it does have *quite a lot* of downloads, so I'm more happy to use it now. I'd be wary of it potentially not working in the browser, so I would also like to know how it stacks up against https://github.com/github/url-polyfill. username_0: After talking with my team over here, I think this is worth fixing. Additionally, I would like to add this in there: https://github.com/imgix/ember-cli-imgix/issues/40 We will need a Url parsing mechanism during fastboot in order to pull that ^ off. I guess the only question now is which lib should we use? I like node-url only because I trust that author (a heavy hitter from the node community) But honestly, all we need to do with this lib is params parsing... Does https://github.com/imgix/imgix-core-js Have url parsing that we could easily expose? How are people handling this in React land? username_1: Agreed. This should be fixed. `node-url` looks to be well supported, but `url-polyfill` comes from the Web Components team, who are also heavy hitters in the web space. `imgix-core-js` doesn't have anything that can be exposed, and is also not supported in the browser. In React we use `jsuri` to handle query parameters. Re #40: We are planning to implement responsiveness using `srcset` and `sizes` soon (#39), which will probably interfere with #40. We'd love your thoughts on using `srcset` etc over there if you have any. username_0: Aight. As I work on this, some browsers will need a polyfill for the basic implementation of `window.URL`, and some, like old safari, only need a polyfill for the `searchParams` portion of the spec. [This lib url-polyfill](https://github.com/github/url-polyfill/blob/master/url.js) does not polyfill the searchParams portion. Only the basic `window.URL` api. So, I can use something like this, which seems to polyfill the entire spec: https://github.com/lifaon74/url-polyfill OR we can use a completely diff implementation like `jsuri` My gut, and what Im gonna go with unless I head otherwise, is to use the polyfill as, eventually 🤞... we can gut rip out that polyfill. username_0: Ok, lol. Gonna scrap that polyfill and just switch to `jsuri` We need this functionality: https://github.com/lifaon74/url-polyfill/issues/27 username_1: Do either of these polyfills work? https://github.com/WebReflection/url-search-params, https://github.com/jerrybendy/url-search-params-polyfill username_0: Yeah, but then we will need to manually manage the synchronization of the searchParams and the top level search property. In a working full implementation of `window.URL` the searchParams property, when modified, will reflect its modified values in the `search` property. We will have to manage that if we polyfill them separately. FWIW... Switching to `jsuri` was a breeze and is almost done... All tests pass. username_0: Actually, I am slightly confused now after reading through the react implementation. It seems as though the final URL is never passed to the imgix core lib, only parsed by jsuri then shoved into a src set. Is imix-core even useful in this case? username_1: The react implementation doesn't use `imgix-core-js`, but I'm not sure this is a 'feature' - it's just the state it's in right now. username_0: Is there any advantage to using imgix-core? I have everything ready to go, but I have to serialize, then deserialize params in order to pass everything to imgix-core-js correctly. We could clean up some logic if we dropped it. Imgix-core-js will always prepend the host to image paths even if the path already has a host. We should probably rip it out, or teach it how to accept a path with a host already on it. Status: Issue closed
openshift/openshift-ansible
176369132
Title: flanneld service fails to start with ose3.3 Question: username_0: When using flannel on ose3.3, openshift-ansible fails with error: TASK [flannel : Enable flanneld] *********************************************** task path: /usr/share/ansible/openshift-ansible/roles/flannel/tasks/main.yml:31 <flannel33-openshift-master-0.example.com> ESTABLISH SSH CONNECTION FOR USER: cloud-user <flannel33-openshift-master-0.example.com> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=600s -o StrictHostKeyChecking=no -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=cloud-user -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r flannel33-openshift-master-0.example.com '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-dklhtbzhphpyitayqetzzxkwfpgsbrlx; LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python'"'"'"'"'"'"'"'"' && sleep 0'"'"'' <flannel33-openshift-node-s6e22h8m.example.com> ESTABLISH SSH CONNECTION FOR USER: cloud-user <flannel33-openshift-node-s6e22h8m.example.com> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=600s -o StrictHostKeyChecking=no -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=cloud-user -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/%h-%r flannel33-openshift-node-s6e22h8m.example.com '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-jzpdgisowludpowwmyieklkvmqfncysn; LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python'"'"'"'"'"'"'"'"' && sleep 0'"'"'' fatal: [flannel33-openshift-master-0.example.com]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"arguments": "", "enabled": true, "name": "flanneld", "pattern": null, "runlevel": "default", "sleep": null, "state": "started"}, "module_name": "service"}, "msg": "Job for flanneld.service failed because a timeout was exceeded. See \"systemctl status flanneld.service\" and \"journalctl -xe\" for details.\n"} fatal: [flannel33-openshift-node-s6e22h8m.example.com]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"arguments": "", "enabled": true, "name": "flanneld", "pattern": null, "runlevel": "default", "sleep": null, "state": "started"}, "module_name": "service"}, "msg": "Job for flanneld.service failed because a timeout was exceeded. See \"systemctl status flanneld.service\" and \"journalctl -xe\" for details.\n"} NO MORE HOSTS LEFT ************************************************************* to retry, use: --limit @/var/lib/ansible/playbooks/main.retry PLAY RECAP ********************************************************************* flannel33-openshift-master-0.example.com : ok=510 changed=125 unreachable=0 failed=1 flannel33-openshift-node-s6e22h8m.example.com : ok=145 changed=37 unreachable=0 failed=1 localhost : ok=37 changed=14 unreachable=0 failed=0 On the lannel33-openshift-master-0.example.com flanneld failed probably because of this error message: Sep 12 08:58:40 flannel33-openshift-master-0.example.com flanneld[49454]: E0912 08:58:40.569847 49454 network.go:53] Failed to retrieve network config: client: etcd cluster is unavailable or misconfigured Used RPMs: openshift-ansible-lookup-plugins-3.3.2-1.git.0.1223e95.el7.noarch openshift-ansible-3.3.2-1.git.0.1223e95.el7.noarch openshift-ansible-filter-plugins-3.3.2-1.git.0.1223e95.el7.noarch openshift-ansible-roles-3.3.2-1.git.0.1223e95.el7.noarch openshift-ansible-docs-3.3.2-1.git.0.1223e95.el7.noarch openshift-ansible-playbooks-3.3.2-1.git.0.1223e95.el7.noarch Answers: username_1: I hate to ask, but the version you're using was tagged nearly two months ago, can you try a more recent version there were a few fixes to flannel support since then. username_0: Ouch, sorry, my fault. I "fixed" an old puddle build instead of "latest" in my repo url :(, I'll close this if not reproducible on the latest one. username_1: Ok, I think the other issue you opened recently on flannel is in the same boat. username_0: Hitting the same issue both with newest 3.2 and 3.3 versions: ansible-2.2.0-0.5.prerelease.el7.noarch openshift-ansible-filter-plugins-3.2.28-1.git.0.5a85fc5.el7.noarch openshift-ansible-docs-3.2.28-1.git.0.5a85fc5.el7.noarch openshift-ansible-lookup-plugins-3.2.28-1.git.0.5a85fc5.el7.noarch openshift-ansible-roles-3.2.28-1.git.0.5a85fc5.el7.noarch openshift-ansible-3.2.28-1.git.0.5a85fc5.el7.noarch openshift-ansible-playbooks-3.2.28-1.git.0.5a85fc5.el7.noarch and: ansible-2.2.0-0.5.prerelease.el7.noarch openshift-ansible-lookup-plugins-3.3.22-1.git.0.6c888c2.el7.noarch openshift-ansible-3.3.22-1.git.0.6c888c2.el7.noarch openshift-ansible-filter-plugins-3.3.22-1.git.0.6c888c2.el7.noarch openshift-ansible-roles-3.3.22-1.git.0.6c888c2.el7.noarch openshift-ansible-docs-3.3.22-1.git.0.6c888c2.el7.noarch openshift-ansible-playbooks-3.3.22-1.git.0.6c888c2.el7.noarch username_0: inventory: [OSv3:children] infra masters nodes etcd [infra] localhost [masters] flannel-puddle33-openshift-master-0.example.com [etcd] flannel-puddle33-openshift-master-0.example.com [nodes] flannel-puddle33-openshift-master-0.example.com openshift_node_labels="{'region': 'infra', 'zone': 'default'}" flannel-puddle33-openshift-node-6jbt26sb.example.com openshift_node_labels="{'region': 'primary', 'zone': 'default'}" [dns] localhost [extradnsitems] loadbalancer vars: [root@flannel-puddle33-infra ansible]# find group_vars/ group_vars/ group_vars/nodes.yml group_vars/OSv3.yml group_vars/masters.yml [root@flannel-puddle33-infra ansible]# cat group_vars/masters.yml num_infra: 1 router_vip: 192.168.0.4 openshift_schedulable: true openshift_master_api_port: 8443 openshift_hosted_router_replicas: 0 [root@flannel-puddle33-infra ansible]# cat group_vars/OSv3.yml ansible_first_run: true infra_instance_id: c886cdb3-7405-4964-9e86-66a39e1d5b5f ansible_ssh_user: cloud-user ansible_sudo: true ansible_become: true deployment_type: openshift-enterprise # deployment type valid values are origin, online and openshif-enterprise osm_default_subdomain: cloudapps.example.com # default subdomain to use for exposed routes openshift_override_hostname_check: true openshift_use_openshift_sdn: false openshift_use_flannel: true openshift_use_dnsmasq: false flannel_interface: eth1 openshift_master_cluster_hostname: flannel-puddle33-lb.example.com openshift_master_cluster_public_hostname: flannel-puddle33-lb.example.com openshift_master_identity_providers: - name: htpasswd_auth login: true challenge: true kind: HTPasswdPasswordIdentityProvider filename: /etc/origin/openshift-passwd openshift_cloudprovider_kind: openstack openshift_cloudprovider_openstack_auth_url: http://10.16.184.104:5000/v2.0 openshift_cloudprovider_openstack_username: admin openshift_cloudprovider_openstack_password: xxx openshift_cloudprovider_openstack_tenant_name: xxx openshift_cloudprovider_openstack_region: RegionOne openshift_hosted_manage_registry: false openshift_hosted_manage_router: false Status: Issue closed
angular/angular
805470042
Title: strictDomEventTypes uses incorrect types for keyup.Enter style outputs Question: username_0: <!--🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 Oh hi there! 😄 To expedite issue processing please search open and closed issues before submitting a new one. Existing issues often contain information about workarounds, resolution, or progress updates. 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅--> # 🐞 bug report ### Affected Package <!-- Can you pin-point one or more @angular/* packages as the source of the bug? --> <!-- ✍️edit: --> The issue is caused by package @angular/compiler ### Is this a regression? <!-- Did this behavior use to work in the previous version? --> No ### Description When setting `"strictDomEventTypes": true` output events like `keyup.Enter` don't use the correct typing ## 🔬 Minimal Reproduction <!-- Please create and share minimal reproduction of the issue starting with this template: https://stackblitz.com/fork/angular-ivy --> [https://stackblitz.com/edit/angular-ivy-jaygka?file=src%2Fapp%2Fapp.component.html](https://stackblitz.com/edit/angular-ivy-jaygka?file=src%2Fapp%2Fapp.component.html) <!-- If StackBlitz is not suitable for reproduction of your issue, please create a minimal GitHub repository with the reproduction of the issue. A good way to make a minimal reproduction is to create a new app via `ng new repro-app` and add the minimum possible code to show the problem. Share the link to the repo below along with step-by-step instructions to reproduce the problem, as well as expected and actual behavior. Issues that don't have enough info and can't be reproduced will be closed. You can read more about issue submission guidelines here: https://github.com/angular/angular/blob/master/CONTRIBUTING.md#-submitting-an-issue --> ## 🔥 Exception or Error <pre><code>Argument of type 'Event' is not assignable to parameter of type 'KeyboardEvent'. </code></pre> But the type should be inferred as `KeyboardEvent` ## 🌍 Your Environment **Angular Version:** <pre><code> Angular CLI: 11.1.3 Node: 14.13.1 OS: darwin x64 Angular: 11.1.2 ... animations, common, compiler, compiler-cli, core, elements ... forms, language-service, localize, platform-browser [Truncated] Package Version --------------------------------------------------------- @angular-devkit/architect 0.1100.6 @angular-devkit/build-angular 0.1101.3 @angular-devkit/core 11.0.6 @angular-devkit/schematics 11.0.6 @angular/cdk 9.0.1 @angular/cli 11.1.3 @angular/google-maps 11.0.3 @schematics/angular 11.0.6 @schematics/update 0.1101.3 ng-packagr 11.0.3 rxjs 6.6.3 typescript 4.1.3 </code></pre> **Anything else relevant?** <!-- ✍️Is this a browser specific issue? If so, please specify the browser and version. --> <!-- ✍️Do any of these matter: operating system, IDE, package manager, HTTP server, ...? If so, please mention it below. --> Answers: username_1: I didn't even know this was a thing; the problem here is the event plugin mechanism that parses the `keyup.Enter` event and processes it in a special way; the compiler is unaware of such plugins and their behavior. This is related to #40553 which requests the ability to make the compiler aware of these event plugins. The request is targeting the language service, but this would ultimately be the compiler's responsibility. username_2: @username_1 would it make sense to potentially move the filtered keyboard events to a first-class compiler feature since they're effectively considered part of the framework's core behavior? username_1: Actually, I realized that `strictDomEventTypes` is piggy-backing on `lib.dom.d.ts` and it can be extended using global augmentation/declaration merging: ```typescript declare global { // Maps a specific event name to an event type interface GlobalEventHandlersEventMap { "keydown.enter": KeyboardEvent; } // More generic way that uses template literal types (introduced in TS 4.1) to infer an event type for a set of event names. interface Document { addEventListener(type: `keyup.${string}`, listener: (ev: KeyboardEvent) => any, options?: boolean | AddEventListenerOptions): void; } } ``` See the resulting types [in this TS playground](https://www.typescriptlang.org/play?#code/CYUwxgNghgTiAEBzCB7ARlC8DeAoe8AlgHYAuIMAZlGAgOKoYQCiAbiGQBJTHAQUBnNh1IBZKAAcc+AvABEAaxABPYCgDuxAHQiKcgFzwA0irQpYwYWQDcMgL64ZJclRoIAIijABXALYjpWXgoYEt2MgAZQgFyYgoAClJlCRBDAAMlZW8JLQASbBiYEkQ7NIAaeAho2IpDeJBWQxNlMwsrUgBKeABeAD5g4mUKlAlSQhRiAQB+QzMUfh54AB94AEFQ9qiYjgoAeVHxyY7DVhRCYFsCBwdcNR9-Mi0QsJEtmph4gHJM7J0yCk+FQaAT6gXgwJsuDsHUcdz8IieG3CpDeOw+3xUak0fxcgPByJ6-TwBAhpFs0McIAAHhIUDBSDg7NYgA). The question is what API we want to offer, as the above may not be the most ergonomic. username_1: That would be an interesting idea as it would probably allow for the core event plugin to be replaced by tree-shakable instructions. One potential difficulty could be that custom event plugins also have the ability to intercept those events and that is only known at runtime. So any compiled event handler would still need to consider the event plugin registry. username_2: Are you talking about the general case for event plugins or specifically for keyboard events like `keydown.enter`? I had only meant the latter, since they're DOM events and a first-party API rather than a plugin for a third-party API like HammerJS. If we only care about those keyboard events, is the custom event registry relevant? username_1: @username_2 I actually also meant the latter (so only handling keyboard events like `keydown.enter` by the compiler). The reason why I think the event plugin registry may still be relevant is that other plugins could in theory also handle those keyboard events, thereby overriding the default keyboard plugin. Since this isn't known at compile time, any compiled key handler instruction would still need to consult with the plugin registry to allow such overrides to take over. I am not sure to what extend this is just a theoretic example or whether this is actually done in real projects. username_2: We could try to get more info from the community as to whether people actually add event plugins that would take over native event bindings like this. I suspect this would be either exceptionally rare or nonexistent; it seems like we don't even document custom event plugins anywhere. Is that's the case, the best thing for the most users would most likely be to have these keyboard events supported and type-checked based on their native behavior. username_3: We use custom event plugins heavily with our components library Taiga UI: https://github.com/TinkoffCreditSystems/taiga-ui It relies on our event plugins library: https://github.com/TinkoffCreditSystems/ng-event-plugins Which is why I created the mentioned issue. I would vote that we won't find anybody using a custom event manager plugin to intercept built-in combined key events. I'm probably the biggest user of plugins system in the wild and the way I write custom plugins is — they react to their specific modifier, wrap handler with additional actions, remove modifier and return event handling back to the manager so there's as little interference with internal Angular processing as possible. So adding a special treatment to key events in the compiler seems like a valid option to me, but I would also love custom events and custom event plugins issue resolved one day 🙂
hashgraph/hedera-sdk-java
567802222
Title: getRecord not working Question: username_0: getRecord is throwing an Exception in thread "main" java.lang.NoSuchFieldError: TRANSFER_ACCOUNT_SAME_AS_DELETE_ACCOUNT. This also affects getReceipt in Smart Contracts. -> Run example CreateStatefulContract on TESTNET Exception in thread "main" java.lang.NoSuchFieldError: TRANSFER_ACCOUNT_SAME_AS_DELETE_ACCOUNT at com.hedera.hashgraph.sdk.Status.<clinit>(Status.java:115) at com.hedera.hashgraph.sdk.TransactionReceipt.<init>(TransactionReceipt.java:30) at com.hedera.hashgraph.sdk.HederaReceiptStatusException.throwIfExceptional(HederaReceiptStatusException.java:39) at com.hedera.hashgraph.sdk.QueryBuilder.mapResponse(QueryBuilder.java:280) at com.hedera.hashgraph.sdk.QueryBuilder.mapResponse(QueryBuilder.java:27) at com.hedera.hashgraph.sdk.HederaCall.lambda$execute$0(HederaCall.java:55) at com.hedera.hashgraph.sdk.Backoff.tryWhile(Backoff.java:37) at com.hedera.hashgraph.sdk.HederaCall.execute(HederaCall.java:58) at com.hedera.hashgraph.sdk.QueryBuilder.execute(QueryBuilder.java:189) at com.hedera.hashgraph.sdk.HederaCall.execute(HederaCall.java:44) at com.hedera.hashgraph.sdk.TransactionId.getReceipt(TransactionId.java:148) at com.hedera.hashgraph.sdk.TransactionId.waitForConsensus(TransactionId.java:113) at com.hedera.hashgraph.sdk.TransactionId.getRecord(TransactionId.java:170) at CreateTransfer.main(CreateTransfer.java:54) Answers: username_1: What is the transaction you are trying to execute? --- That exception just looks like a bug in how the error is hit. Not that `getRecord` isn't working. username_0: Inside the CreateStatefulContract, there is a TransactionReceipt contractReceipt = contractTxId.getReceipt(client); in line 80. That call generates an uncaught error. Same process happens inside the Get Record, since the (waitForConsensus(client, null);) includes a getReceipt that fails. username_0: Try running the CreateStatefulContract of the examples. username_2: @username_0 are you trying to build the SDK locally? You should try `mvn clean install` in the SDK before running the example. That error suggests that the compiled bytecode of the generated `ResponseCodeEnum` class is out of date. username_0: Yes. I have a local copy of the protobuf that needed update. All is working fine now. Status: Issue closed
reactor/reactor
41729138
Title: TPED - Occasional Concurrent Modification Exceptions Question: username_0: This [Spring Integration Sample](https://github.com/spring-projects/spring-integration-samples/blob/master/intermediate/async-gateway/src/test/java/org/springframework/integration/samples/async/gateway/PromiseTest.java) occasionally raises these exceptions: ``` Exception in thread "threadPoolExecutor-3" java.lang.RuntimeException: Possible corruption through unsynchronized concurrent modification. at com.gs.collections.impl.map.mutable.UnifiedMap$2.equals(UnifiedMap.java:99) at com.gs.collections.impl.map.mutable.UnifiedMap.nonNullTableObjectEquals(UnifiedMap.java:2596) at com.gs.collections.impl.map.mutable.UnifiedMap.chainedGetIfAbsentPut(UnifiedMap.java:600) at com.gs.collections.impl.map.mutable.UnifiedMap.getIfAbsentPut(UnifiedMap.java:577) at reactor.util.PartitionedReferencePile.get(PartitionedReferencePile.java:64) at reactor.event.dispatch.AbstractMultiThreadDispatcher.allocateRecursiveTask(AbstractMultiThreadDispatcher.java:45) at reactor.event.dispatch.AbstractLifecycleDispatcher.dispatch(AbstractLifecycleDispatcher.java:131) at reactor.core.Reactor.notify(Reactor.java:242) at reactor.core.Reactor.notify(Reactor.java:249) at reactor.core.Reactor.notify(Reactor.java:57) at reactor.core.composable.Promise.valueAccepted(Promise.java:560) at reactor.core.composable.Promise$1.accept(Promise.java:91) at reactor.core.composable.Promise$1.accept(Promise.java:88) at reactor.core.action.CallbackEventAction.doAccept(CallbackEventAction.java:36) at reactor.core.action.Action.accept(Action.java:52) at reactor.core.action.Action.accept(Action.java:32) at reactor.event.routing.ArgumentConvertingConsumerInvoker.invoke(ArgumentConvertingConsumerInvoker.java:73) at reactor.event.routing.ConsumerFilteringEventRouter.route(ConsumerFilteringEventRouter.java:78) at reactor.event.dispatch.AbstractLifecycleDispatcher.route(AbstractLifecycleDispatcher.java:64) at reactor.event.dispatch.AbstractMultiThreadDispatcher$MultiThreadTask.run(AbstractMultiThreadDispatcher.java:58) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) java.lang.RuntimeException: Possible corruption through unsynchronized concurrent modification. at com.gs.collections.impl.map.mutable.UnifiedMap$2.equals(UnifiedMap.java:99) at com.gs.collections.impl.map.mutable.UnifiedMap.nonNullTableObjectEquals(UnifiedMap.java:2596) at com.gs.collections.impl.map.mutable.UnifiedMap.chainedGetIfAbsentPut(UnifiedMap.java:600) at com.gs.collections.impl.map.mutable.UnifiedMap.getIfAbsentPut(UnifiedMap.java:577) at reactor.util.PartitionedReferencePile.iteratorFor(PartitionedReferencePile.java:111) at reactor.util.PartitionedReferencePile.iterator(PartitionedReferencePile.java:89) at reactor.event.dispatch.AbstractMultiThreadDispatcher$MultiThreadTask.run(AbstractMultiThreadDispatcher.java:57) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) ``` The test uses a [`ThreadPoolExecutorDispatcher`](https://github.com/spring-projects/spring-integration-samples/blob/master/intermediate/async-gateway/src/main/resources/META-INF/reactor/default.properties); [the `Promise` is created in the `GatewayProxyFactoryBean`](https://github.com/spring-projects/spring-integration/blob/master/spring-integration-core/src/main/java/org/springframework/integration/gateway/GatewayProxyFactoryBean.java#L361) and [flushed here](https://github.com/spring-projects/spring-integration-samples/blob/master/intermediate/async-gateway/src/test/java/org/springframework/integration/samples/async/gateway/PromiseTest.java#L82).<issue_closed> Status: Issue closed
onedesign/generator-one-base
232981959
Title: Make listing errors less ignorable Question: username_0: I feel like I opened this before, but either way. Right now it's super easy to ignore linting errors, we should make those harder to ignore. I did something like this on JLLIPT where the errors were injected into the page as HTML markup. It worked pretty well as far as the "not ignoring" them goes. Answers: username_1: I agree, but I'm also not sure of the best way to solve it. Ideally, it would hook into the same browsersync-based notifications system that we currently use to display JS/CSS compilation errors in the browser, but the tricky part is making sure that they both show up, or at the very least that displaying compilation errors takes precedence over displaying linting errors. Status: Issue closed
readium/readium-lcp-server
203614406
Title: HTTP REST API, RAML spec Question: username_0: TODO: document the API using: http://raml.org Answers: username_1: Go implementation in https://github.com/Jumpscale/go-raml. username_2: I made this https://github.com/username_2/readium-lcp-server-api-doc a couple of years back and almost forgot, I've fixed a few things lately. The raml schema is handwritten, I've not found a good solution for automatic schema generation from Go http handlers. I described almost everything (with the same text) as is on [lcp wiki](https://github.com/readium/readium-lcp-server/wiki) There is also a [action](https://github.com/username_2/readium-lcp-server-api-doc/blob/master/.github/workflows/node.js.yml) building and publishing on github pages, result here: https://username_2.github.io/readium-lcp-server-api-doc/ It's still incomplete, some things needs to be tweaked, mainly errors and example responses (for example wiki pages are linking some [broke content ](https://readium.org/licensed-content-protection/error/notfound)). If you think it is correct we can revise things together, than you can fork or include raml in main lcp repo?
jlippold/tweakCompatible
340403585
Title: `Artsy` working on iOS 11.3.1 Question: username_0: ``` { "packageId": "ch.mdaus.artsy", "action": "working", "userInfo": { "arch32": false, "packageId": "ch.mdaus.artsy", "deviceId": "iPhone10,6", "url": "http://cydia.saurik.com/package/ch.mdaus.artsy/", "iOSVersion": "11.3.1", "packageVersionIndexed": true, "packageName": "Artsy", "category": "Tweaks", "repository": "Maxwell Dausch's Repo", "name": "Artsy", "packageIndexed": true, "packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 98% with 54 working reports.", "id": "ch.mdaus.artsy", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.0.7", "shortDescription": "Modify the Lockscreen, CC, Spotify, and Music app now playing views based on the album artwork.", "latest": "0.0.4-8+debug", "author": "<NAME>", "packageStatus": "Working" }, "base64": "<KEY> "chosenStatus": "working", "notes": "" } ```
ioof-holdings/redux-dynostore
732127170
Title: Static reducers in nested path disappears after attaching dynamic reducer. Question: username_0: <!--- Thanks for contributing to this redux-dynostore! Before you submit, please read the following: Search open/closed issues before submitting since someone might have asked the same thing before! --> ### Is it a bug, feature request or question? Question (or bug) ### Which package(s) does this involve? [@redux-dynostore/core] ### Input Code ```js store.attachReducers({ 'some.path.dynamicReducer1': dynamicReducerFunction }); ``` ### Current Behavior Hi! I'm trying to attach dynamic reducer to the store to a nested path. The issue is that in this path I already have static reducers and I'd like to attach a dynamic one. But after attaching static reducers disappears from the state. ```js some: { path: { staticReducer1: reducerFuction1, staticReducer2: reducerFuction2 } } ``` after calling `store.attachReducers({ 'some.path.dynamicReducer1': reducerFuction });` static reducers disappears and I have ```js some: { path: { dynamicReducer1: reducerFuction, } } ``` Expected result: ```js some: { path: { staticReducer1: reducerFuction1, staticReducer2: reducerFuction2, dynamicReducer1: reducerFuction, } } ``` ### Your Setup "redux": "^4.0.5", "react-redux": "^7.1.3", "@redux-dynostore/core": "^3.1.0", Answers: username_1: I've also encountered this issue. I've also tried specifying the state handler when attaching a reducer, but the custom one doesn't appear to get used. Configuring the deep state handler in the reducer enhancement worked for me. ```javascript dynostore(dynamicReducers({ stateHandler: deepStateHandler })) ``` username_1: @username_2 The following reproduces this in the integrations spec, though I'm not sure how to fix it. It has something to do with options on this line taking hierarchal precedence. https://github.com/ioof-holdings/redux-dynostore/blob/c2b4da08a772ff3a2a17aa76ec5ac2197154b63a/packages/redux-dynostore-core/src/reducers/createDynamicReducer.js#L15 ```javascript test('should attach dynamic reducer to static key with state handler override', () => { const reducer = combineReducers({ group1: combineReducers({ group2: combineReducers({ key1: makeTestReducer('key1') }), key3: makeTestReducer('key3') }), key5: makeTestReducer('key5') }) const store = createStore(reducer, dynostore(dynamicReducers())) store.attachReducers({ group1: { group2: { key2: makeTestReducer('key2') }, key4: makeTestReducer('key4') } }, { stateHandler: deepStateHandler }) expect(store.getState()).toEqual({ group1: { group2: { key1: 'key1 - initialValue', key2: 'key2 - initialValue' }, key3: 'key3 - initialValue', key4: 'key4 - initialValue' }, key5: 'key5 - initialValue' }) }) ``` username_2: Thanks for raising @username_0 and investigating a bit @username_1. In general, attaching dynamic reducers to a static reducer node is not allowed with the default state handler (shallow). As @username_1 discovered, the deep state handler is supposed to allow this, but there are some known issues with using it that have never been resolved and we currently advise against it. Honestly, I think the concept was flawed from the get go and I don't have any ideas for resolving it. If possible, I'd recommend separating the static reducers from the dynamic reducers at the root of the store, or if that is not possible, creating the store with an empty reducer and attach the static reducers immediately after creation. I'm aware this is not ideal. username_3: Any improvement? I am also facing the same problem in this regard. username_2: No changes for this @username_3. Honestly, I'm not sure how to solve this one in general. username_3: So how can I access a different subspace state from different subspace. username_3: I'm asking not for this topic but for a quick response. So how can I access a different subspace state from different subspace. username_2: @username_3 the general answer is that you don't and if you do need it then perhaps you need to reconsider the boundaries of your subspaces or the use of this library all together. There are some exceptions, namely global state, which is [covered in the `redux-subspace` docs](https://ioof-holdings.github.io/redux-subspace/docs/advanced/GlobalState.html) Please raise a new issue if you want to discuss this further. username_2: Closing. Please see #484 for details. Status: Issue closed
LiuuY/Blog
239957969
Title: Angular4 Best Practices Question: username_0: 1. Measure - Smaller Bundles - [source-map-explorer](https://github.com/danvk/source-map-explorer) 2. AOT - [--aot](https://github.com/angular/angular-cli/wiki/build#--dev-vs---prod-builds) 3. Stay Up to Date 4. Import Carefully - Do NOT import 'rxjs' - Do NOT import { MaterialModule } from '@angular/material' 5. [Lazy Load](https://angular.io/guide/router#asynchronous-routing) - Lazy Load EVERYTHING! 6. Polyfill Responsibly - AOT do not use 'core-js/es6/reflect' and 'core-js/es6/reflect' (these for JIT), so we can delete from polyfill.ts [source](https://www.youtube.com/watch?v=hHNUohOPCCo)
jeffery9/vscode-odoo-snippets
1012482499
Title: m2m snippet: reverse column1 and column2 attributes Question: username_0: current: field_name_ids = fields.Many2many( string='field_name', comodel_name='model.name', relation='model.name_this_model_rel', column1='model.name_id', column2='this_model_id', ) the right way is: field_name_ids = fields.Many2many( string='field_name', comodel_name='model.name', relation='model.name_this_model_rel', column1='this_model_id', column2='model.name_id', )
GameAnalytics/GA-SDK-JAVASCRIPT
1159395999
Title: SDK error handling on initialisation process and network issues Question: username_0: I would like to handle two cases: 1. The SDK is not initialised properly for some reason, for example adblock, other tools that disables scripts, etc. 2. The SDK is not working properly because the network is blocked and it cannot sent a request for a long period of time Does the SDK provides a way that we could handle those problems mentioned above? I mean for example I would like to collect the data and push some logs to my backend if the SDK is not initialised properly but I'm afraid I don't have some explicit to do that. Correct me if I'm wrong, but I've tried to use `isSdkReady` method. Not sure if this is a correct approach but anyway I had to force my linter to use it because it's a private static method. Or maybe there's better way? Unfortunately I don't see anything in the exposed methods that could help. The second thing is with the network. I don't see any network event handlers that could trigger if the SDK cannot connect to GameAnalytics API. Any help with this would be much appreciate it :) Thanks in advance. Answers: username_1: The SDK will initialize correctly even though it didn't get to get connected with backend as it could be offline to start with. If the SDK fails to connect to GameAnalytics API, the SDK should try again every 8 seconds to send its cached events to the GA backend. I am not sure how to properly check if the SDK requests are being blocked by other plugins or scripts in the browser.
MIC-DKFZ/nnUNet
856579805
Title: Dice loss for background or not? Question: username_0: Hi, may i ask you whether you use the dice loss function for both foreground and background, or only use dice loss function on the foreground? Answers: username_1: is taken for the dice loss part. There exists an argument `do_bg` (=do background) which by default is True. Yes, therefore the average of all classes - including background and foreground - is used to calculate the dice loss. Best, Marcus username_2: Thanks for stepping in @username_1. However, the background is NOT used to compute the Dice loss: https://github.com/MIC-DKFZ/nnUNet/blob/058b695d61d34dda7f79cd36ab950a5d3e031653/nnunet/training/network_training/nnUNetTrainer.py#L108 this line is where the loss is initialized. Best, Fabian username_1: Oh yes, maybe one should adapt the default arguments to the ones that are indeed used in default training ;-) I'll delete my reply to not accidentally confuse others. Thanks for clarifying! username_2: I like to keep things consistent. People might use my functions and rely on the default values. If we change the defaults, the behavior of someone else's code might change without them noticing. That would be bad username_0: Thanks for your reply, however, I want to further ask another question, The Loss in nnUNet is Dice + CE, the CE loss contains the cacluation for class 0 (background), why Dice loss neglects the background? Is the experimental result that demonstrates without background can Dice loss w CE achieve better results? username_2: Yeah some experiments a very long time ago showed that :-) Not sure if that still holds true though. But eh - best you can get is parity when including the background so I wouldn't bother going down that route. Best, Fabian Status: Issue closed
VSCodeVim/Vim
188444909
Title: a text object works incorrectly in nested objects Question: username_0: <!-- For questions, ask us on [Slack](https://vscodevim-slackin.azurewebsites.net/) 👫. Found a bug? Delete this line and fill out the sections below. --> Please *thumbs-up* 👍 this issue if it personally affects you! You can do this by clicking on the emoji-face on the top right of this post. Issues with more thumbs-up will be prioritized. ----- ### What did you do? ```ts const a = { b: { /*cursor here*/c: "d" } } ``` Put the exact keys you pressed. any of ```ca{``` ### What did you expect to happen? ```ts const a = { b: /* cursor here for ca{ */ } ``` ### What happened instead? ```ts const a = { b: /* cursor */ } } ### Technical details: * VSCode Version: 1.7.1 * VsCodeVim Version: *0.4.1* * OS: Windows 10 same happens with ```ca[ , ca}, ca]``` and probably others Answers: username_0: seems related to #1023 username_1: @username_2 this should be fixed as well now that #1023 is fixed Status: Issue closed username_2: Thanks @username_1 !
gost/dashboard-v2
272214436
Title: fix npm warnings Question: username_0: npm WARN deprecated [email protected]: Use uuid module instead npm WARN deprecated [email protected]: Please update to minimatch 3.0.2 or higher to avoid a RegExp DoS issue npm WARN deprecated [email protected]: Please update to minimatch 3.0.2 or higher to avoid a RegExp DoS issue npm WARN deprecated [email protected]: graceful-fs v3.0.0 and before will fail on node releases >= v7.0. Please update to graceful-fs@^4.0 .0 as soon as possible. Use 'npm ls graceful-fs' to find it in the tree. npm WARN deprecated [email protected]: 🙌 Thanks for using Babel: we recommend using babel-preset-env now: please read babeljs.i o/env to update! npm WARN deprecated [email protected]: ...psst! Your project can stop working at any moment because its dependencies can change. Prevent this by migrating to Yarn: https://bower.io/blog/2017/how-to-migrate-away-from-bower/ npm WARN deprecated @types/[email protected]: See https://github.com/DefinitelyTyped/DefinitelyTyped/issues/12826 npm WARN prefer global [email protected] should be installed with -g
adazzle/react-data-grid
441883064
Title: Cell Actions not working for the last rows Question: username_0: --- ### Issue Details I'm not being able to open a cell action on the last rows. After the first time you click the scroll bar starts blinking (it opens and close the action menu automatically) Follow the evidence: ![chrome-capture](https://user-images.githubusercontent.com/11162653/57400602-3f2fff80-71aa-11e9-8196-ae7ac27713db.gif) You can also see it happening through the official docs: https://adazzle.github.io/react-data-grid/docs/examples/cell-actions ``` Please include: - What the current behavior is - What the desired behvaior is - (If Bug) Steps to reproduce the issue - (If Feature) The motivation / use case for the feature We especially love screenshots / videos of problems, and remember The Best Issue Is A Pull Request™ ``` Answers: username_1: Any thoughts on this? Facing the same problem. username_2: +1 on this username_3: Little css approach for those who facing this ``` .react-grid-Row:nth-child(n + 6) .rdg-cell-action-menu { top: 0; transform: translateY(-100%); } ``` username_4: Just a small improvement ``` .react-grid-Row:nth-last-child(-n + 6) .rdg-cell-action-menu { top: 0; transform: translateY(-100%); } ``` Status: Issue closed username_5: `cellActions` prop will be removed in v7 but it is still possible to add cell actions using a custom formatter https://github.com/adazzle/react-data-grid/blob/canary/CHANGELOG.md https://adazzle.github.io/react-data-grid/canary/?path=/story/demos--cell-actions https://github.com/adazzle/react-data-grid/blob/canary/stories/demos/CellActions.tsx
Tencent/tdesign-vue
1084043150
Title: tag Question: username_0: **Describe Framework** - [x] tdesign-vue - [x] tdesign-vue-next - [x] tdesign-react **Describe Environment** pkg version: lts **Describe the bug** [codesandbox demo ](https://codesandbox.io/s/tdesign-vue-demo-forked-cxx00) ![image](https://user-images.githubusercontent.com/35833812/146668610-8440bbcc-c5cb-411b-9731-fbad95dca126.png) no disabled style Answers: username_1: 带颜色的tag没有disbaled效果 Status: Issue closed
nodejs/abi-stable-node
365676171
Title: Old issues and PRs on the node-addon-examples Question: username_0: Hi everyone, we are starting to work on examples to the https://github.com/nodejs/node-addon-examples repo. I noted that there are very old issues and PRs on this repo. How should we proceed? Status: Issue closed Answers: username_0: Ok thanks I will work as you suggested.
Tietoarkisto/metka
112809337
Title: Datakonfiguraation ja GUI-konfiguraation muokkaamisen ohjeistus on huonoa. Pitäisi olla kunnollinen ohjeistus kuvakaappauksien ja esimerkkien kera. Question: username_0: Puuttuu edelleen kokonaan GUI-konfiguraatioiden muokkausohjeet. Answers: username_0: Puuttuu edelleen kokonaan GUI-konfiguraatioiden muokkausohjeet. username_0: Datakonfiguraatioeditorin ohjeesta puuttuu esimerkit. username_0: Molempien konfiguraatioiden käytöstä tulee olla rautalangasta väännetyt ohjeet sekä esimerkkejä. username_1: Tästä esimerkkejä mm. [Esimerkkejä](https://github.com/Tietoarkisto/metka/wiki/Esimerkkej%C3%A4)-sivulla. EN/SV -näkymiä ei sinällään voi muokata muuten, kuin merkitsemällä, mitkä kentät ovat käännettäviä ja mitkä eivät (Esimerkiksi, ero sen välillä, merkitseekö koko CONTAINER-tyyppisen kentän käännettäväksi, vai sen sisällä olevan STRING-kentän käännettäväksi. Edeltävässä GUI-konfiguraatio luo automaattisesti uuden CONTAINER:in kaikille kielille, kun taas jälkimmäisessä luodaan saman CONTAINER:in sisään useampi versio STRING-kentästä.)
tensorflow/tensorflow
207700814
Title: Tensorflow support rnn with batch normalzition cell? Http://arxiv.org/abs/1510.01378 Question: username_0: Tensorflow support rnn with batch normalzition cell? Http://arxiv.org/abs/1510.01378 Answers: username_1: These are the existing cells: https://www.tensorflow.org/api_docs/python/rnn_cell/rnn_cells_for_use_with_tensorflow_s_core_rnn_methods You can create custom cells by extending `tf.nn.rnn_cell.RNNCell` and overwriting method `__call__(self, inputs, state, scope=None) ` username_0: @username_1 @username_2 When I use the __liner function, do a linear change, how to call batch normalztion, and use clip relu activation function? Status: Issue closed username_2: This question is better asked on [StackOverflow](http://stackoverflow.com/questions/tagged/tensorflow) since it is not a bug or feature request. There is also a larger community that reads questions there. Thanks!
vuetifyjs/vuetify
597921751
Title: [Feature Request] Add more scheduling features to calendar Question: username_0: ### Problem to solve I would love it if we could add more scheduling features in the calendar. Honestly just two features would suffice. 1. Add resource view with daily appointments, something vertical like Resource1 Resource2 10am 11am 12pm 2. Then the ability to click into a cell and get the context of the resource as well as the date/time ### Proposed solution This solution isn't too far off from the week view, but instead of incrementing days, you would loop over resources and show the same day for each one. Then you would just add a click on cells and pass context. Honestly this shouldn't be a huge addon. The reason I'm asking for this is cause I've tried so many other schedulers, there's nothing really good out there for vue. There is syncfusion and bryntum but these are plain javascript libraries that wrap themselves in in popular frameworks like vue and angular, however these cause tons of issues when trying to customize them since they are not component based at core. <!-- generated by vuetify-issue-helper. DO NOT REMOVE --> Answers: username_1: <template v-if="google" #category="{ category }"> <div class="google-category"> <div class="google-category-name">{{ category }}</div> <v-btn v-if="categoryVisuals[category]" class="google-category-avatar" fab small :color="categoryVisuals[category].color"> {{ categoryVisuals[category].avatar }} </v-btn> </div> </template> </v-calendar> </v-sheet> </div> </v-container> </template> <script> export default { data: () => ({ mode: 'stack', value: '', events: [], colors: ['blue', 'indigo', 'deep-purple', 'cyan', 'green', 'orange', 'grey darken-1'], names: ['Meeting', 'Holiday', 'PTO', 'Travel', 'Event', 'Birthday', 'Conference', 'Party'], categories: ['<NAME>', '<NAME>', '<NAME>'], categoriesAlways: ['<NAME>', '<NAME>'], categoryVisuals: { '<NAME>': { avatar: 'P', color: 'primary' }, '<NAME>': { avatar: 'J', color: 'secondary' }, }, categoryShowAll: false, categoryHideDynamic: false, categoryDays: 1, google: false, start: null, }), mounted () { window.app = this }, methods: { getEvents ({ start, end }) { const events = [] const min = new Date(`${start.date}T00:00:00`) const max = new Date(`${end.date}T23:59:59`) const days = (max.getTime() - min.getTime()) / 86400000 const eventCount = this.rnd(days, days + 20) for (let i = 0; i < eventCount; i++) { const allDay = this.rnd(0, 3) === 0 const firstTimestamp = this.rnd(min.getTime(), max.getTime()) const first = new Date(firstTimestamp - (firstTimestamp % 900000)) const secondTimestamp = this.rnd(2, allDay ? 288 : 8) * 900000 const second = new Date(first.getTime() + secondTimestamp) const type = this.rnd(0, this.names.length - 1) events.push({ name: this.names[type], [Truncated] text-align: right; padding: 8px; font-size: 11px; } .google-category-avatar { background-color: rgba(0,0,0,0.3); width: 40px; height: 40px; position: absolute; border-radius: 20px; right: 8px; top: 30px; } } } </style> ``` </details> username_2: resolved in #11198 If you have any additional questions, please reach out to us in our [Discord community](https://community.vuetifyjs.com). Status: Issue closed
OfficeDev/microsoft-teams-apps-requestateam
613119589
Title: Visibility option when using the app Question: username_0: I have added the app to my team but it isn't working correctly:- 1.I don't see the Approve Requests tab 2. I click to create a team from scratch and I am asked for the privacy setting but in the visibility dropdown there are no options. With this being a mandatory field I cannot get past this stage. I did find the deployment a bit menacing Many thanks Yvonne Answers: username_1: The app is designed in a way that you can only move forward when user clicks on Check Availability and Team Name is available. Did you replaced the SharePoint site URL with your SharePoint site in 'Check Team Availability' flow. Which you can see my navigating to My flows. It needs to be replaced at two points as per deployment guide. username_0: Yes I did ... I followed the deployment guide to the letter Get Outlook for iOS<https://aka.ms/o0ukef> username_1: Can you check the Check Team Availability run history, is it running or failed. If you can see history, click on it, see at what step it failed. username_0: Thank you. I had accidentally put a 2 in the tenant name so all is working now except I dont get the approval adaptive card and get the following error in the Flow action 'Post_an_Adaptive_Card_to_a_Teams_channel_and_wait_for_a_response' failed Error The request failed. Error code: 'BotRequestFailed'. Error Message: 'Request to the Bot framework failed with error: '{"error":{"code":"ConversationNotFound","message":"Conversation not found."}}'.'. username_2: I have the very same issue...Would love to know if there is a resolution. username_1: @username_0 @username_2 In that step we are providing groupId and ChannelId of the team where we want to post the adaptive card. Can you check that the groupId(TeamId) and ChannelId values inside Team Request Settings SharePoint list. 10th to 12th point in the given link. If there is any confusion about replacing values in channelId follow the "tip" below point 10. https://github.com/OfficeDev/microsoft-teams-apps-requestateam/wiki/Deployment-guide#step-2-create-admin-group username_3: Closing this issue due to lack of response. Status: Issue closed
w3c/transitions
510592235
Title: CR Update Request for Web of Things (WoT) Architecture Question: username_0: # Link to updated CR draft https://cdn.statically.io/gh/w3c/wot-architecture/06bcab5b/publication/cr2/index.html For informational purposes, an auto-generated diff from the previous CR is at: https://cdn.statically.io/gh/w3c/wot-architecture/06bcab5b/publication/cr2/diff.html # Link to GitHub files for updated CR draft https://github.com/w3c/wot-architecture/master/publication/cr2/ # Link to group's decision to request transition https://www.w3.org/2019/10/16-wot-minutes.html#resolution03 # Link to previous Candidate Recommendation transition request https://lists.w3.org/Archives/Member/chairs/2019AprJun/0069.html # Substantive changes - Terminology section is now informative. # Any changes in normative references? - References to all other WoT documents, in particular the Web of Things (WoT) Thing Description, have been made informative. - As all terminology definitions are now informative, references only from that section are now informative, not normative. - Normative references added: RFC2046 - Normative references removed or moved to informative section: IANA-RELATIONS, MQTT, RFC4395, RFC6838, RFC7049, RFC7231, RFC7252 # Any changes in informative references? - In order to support definitions of security and privacy-related terms, additional informative references to ISO-IEC specifications 2382, 27000, and 29100 were added. However, as we also made all terminology definitions informative in this update, these references now appear in the informative section. # Any changes in requirements? No. However, we did clarify the introduction to emphasize that this is an abstract architecture document. # Wide Review of substantive changes In addition to the wide reviews done for the previous CR transition, this version has addressed several issues raised by the Privacy IG during their review. # Issues status All issues are either closed or marked for resolution by PR transition (informative changes) or have been deferred to an update to be developed under a rechartered WG. # Formal Objections None. # Any changes in implementation information? No. # Deadline for further comments We intend to request a PR transition on Dec 10, 2019. # Any changes in patent disclosures? No. Answers: username_1: Transition approved. Comment in #178 regarding possibly normative semantics remaining in the material now made informative bears review during the remaining CR period. username_2: Thanks a lot, @username_1 ! username_3: Draft transition approved. username_4: # Link to updated CR draft https://cdn.statically.io/gh/w3c/wot-architecture/06bcab5b/publication/cr2/index.html For informational purposes, an auto-generated diff from the previous CR is at: https://cdn.statically.io/gh/w3c/wot-architecture/06bcab5b/publication/cr2/diff.html # Link to GitHub files for updated CR draft https://github.com/w3c/wot-architecture/master/publication/cr2/ # Link to group's decision to request transition https://www.w3.org/2019/10/16-wot-minutes.html#resolution03 # Link to previous Candidate Recommendation transition request https://lists.w3.org/Archives/Member/chairs/2019AprJun/0069.html # Substantive changes - Terminology section is now informative. # Any changes in normative references? - References to all other WoT documents, in particular the Web of Things (WoT) Thing Description, have been made informative. - As all terminology definitions are now informative, references only from that section are now informative, not normative. - Normative references added: RFC2046 - Normative references removed or moved to informative section: IANA-RELATIONS, MQTT, RFC4395, RFC6838, RFC7049, RFC7231, RFC7252 # Any changes in informative references? - In order to support definitions of security and privacy-related terms, additional informative references to ISO-IEC specifications 2382, 27000, and 29100 were added. However, as we also made all terminology definitions informative in this update, these references now appear in the informative section. # Any changes in requirements? No. However, we did clarify the introduction to emphasize that this is an abstract architecture document. # Wide Review of substantive changes In addition to the wide reviews done for the previous CR transition, this version has addressed several issues raised by the Privacy IG during their review. # Issues status All issues are either closed or marked for resolution by PR transition (informative changes) or have been deferred to an update to be developed under a rechartered WG. # Formal Objections None. # Any changes in implementation information? No. # Deadline for further comments We intend to request a PR transition on Dec 10, 2019. # Any changes in patent disclosures? No.
Boscop/web-view
653898538
Title: Running application using web-view dumps Chinese registry keys Question: username_0: I have been testing this package over the last few weeks and today I have noticed an odd entry in my windows registry referencing an application built using this. The path generated is named as follows `\HKEY_CURRENT_USER\潓瑦慷敲䵜捩潲潳瑦䥜瑮牥敮⁴硅汰牯牥䵜楡屮敆瑡牵䍥湯牴汯䙜䅅啔䕒䉟佒南剅䕟啍䅌䥔乏` The path contains a single key `:\projects\rust\webview\multi-window-test\target\debug\multi-window-test.exe` with a DWORD value of `2af8`. Is this something to be concerned about, with the contents of this package? Answers: username_1: I was testing unicode support on Windows onetime, forgot to remove the test string and removed it in some later release. Maybe you tried some older version of webview? I saw this issue pop up a few times.
usnm-vertnet/usnm-birds
158360688
Title: Monthly VertNet data use report for 2016-5, resource nmnh_birds Question: username_0: Your monthly VertNet data use report is ready! You can see the HTML rendered version of the reports with this link: http://tools-usagestats.vertnet-portal.appspot.com/reports/5df38344-b821-49c2-8174-cf0f29f4df0d-Birds/201605/ Raw text and JSON-formatted versions of the report are also available for download from this link. In addition, a copy of the text version has been uploaded to your GitHub repository, under the "Reports" folder. Also, a full list of all reports can be accessed here: http://tools-usagestats.vertnet-portal.appspot.com/reports/5df38344-b821-49c2-8174-cf0f29f4df0d-Birds/ You can find more information on the reporting system, along with an explanation of each metric, here: http://www.vertnet.org/resources/usagereportingguide.html Please post any comments or questions to: http://www.vertnet.org/feedback/contact.html Thank you for being a part of VertNet.
jupyter-widgets/ipyleaflet
986855315
Title: map not visible Question: username_0: hi, i was looking at your work on https://ipyleaflet.readthedocs.io/en/latest/api_reference/basemaps.html# with basemaps, specifically, and noticed all your Strava maps were greyed out. There was a note box that suggested reaching out on here if that situation occured on the site. As a side note, im interested to see what it produces! Is it a way of plotting streets and drivable routes? Answers: username_1: Hi, we depend on `xyzservices`, and it seems [the Strava server is currently down](https://github.com/geopandas/xyzservices/blob/f6f5e8640596721400af5c2bdbefe2fc150dae72/CHANGELOG.md#xyzservices-2021081-august-12-2021). username_2: From what I see the problem is just a typo double slash, '//tiles/' in the Strava URLs on [basemaps.py#199 and further](https://github.com/jupyter-widgets/ipyleaflet/blob/master/ipyleaflet/basemaps.py#L199): ``` Strava=Bunch( All=dict( url='https://heatmap-external-a.strava.com//tiles/all/hot/{z}/{x}/{y}.png?v=19', max_zoom=15, attribution='Map tiles by <a href="https://labs.strava.com/heatmap">Strava 2017</a>', name='Strava.All' ), . . ``` So https://heatmap-external-a.strava.com/tiles/all/hot/10/163/393.png will render but https://heatmap-external-a.strava.com//tiles/all/hot/10/163/393.png not. username_2: And https://github.com/geopandas/xyzservices/pull/85 (the more `xyzservices` becomes relevant!). username_3: Thanks for reporting. So it might only be a matter of updating the docs. We maybe need to update the stable branch. username_3: Thanks for reporting. So it might only be a matter of updating the docs. We maybe need to update the stable branch. username_3: Rebuilding now username_2: Not too familiar how xyzservices is used, thought it was WIP via PR #857. To me it seems that only https://github.com/jupyter-widgets/ipyleaflet/blob/master/ipyleaflet/basemaps.py#L199 needs updating... username_3: Indeed. I thought the xyzservices PR was merged and released, but it is still open https://github.com/jupyter-widgets/ipyleaflet/pull/857. So yeah we need to finish this. username_2: I can make a PR to update https://github.com/jupyter-widgets/ipyleaflet/blob/master/ipyleaflet/basemaps.py#L199. username_3: thanks! Status: Issue closed
IntergatedCircuits/USBDevice
909315145
Title: HID InData callback Question: username_0: Is there any particular reason why the HID doesn't have an InData callback? Answers: username_1: It's not added because it hasn't been a use-case so far, the HID applications I've worked on triggered the IN data transmission on external events. Does your application require the use of it? username_0: Yes, I try to use your library in my CMSIS DAP project. It needs IN and OUT endpoints. username_1: OK, I'm adding this request to my work queue, ETC 1 week. If it's more urgent, you can do it yourself as well, just add an `InData` function to `hid_cbks` and redirect it to the application callback. username_0: Thank you. username_1: Hi Ali, please try this master now, and close the ticket if it satisfies your requirements. Status: Issue closed
AndroidDagashi/AndroidDagashi
443054041
Title: What's New in Play Question: username_0: https://android-developers.googleblog.com/2019/05/whats-new-in-play.html I/Oで発表されたPlayの新しい機能です。version codeを気にせず社内でテストできるinternal app sharing、アプリ内で更新を促せるin-app updates、最近のレーティングを重視するようになったレーティングアルゴリズムの変更、などが含まれています。 Answers: username_1: Internal App Sharingを見てFabric Betaの後継プロダクトの進捗が気になり確認してみたのですが、どうやらアーリーアクセスが始まっているようです。 https://get.fabric.io/roadmap
electron-userland/electron-builder
606787399
Title: can you add supports for makensis INPUTCHARSET OUTPUTCHARSET? Question: username_0: * **Version**: [email protected]@app-builder-lib * **Electron Version**: 5.0.4 * **Target**: build <!-- Enter your issue details below this comment. --> My windows charset is gbk, when i use makensis.exe to compile utf-8 nsis script, i got a "Bad text encoding" error. ![image](https://user-images.githubusercontent.com/17950041/80283187-60398980-8748-11ea-9225-b206fe97d12b.png) Makensis.exe has option to specify input charset, but electron-builder seem doesn't have this option. Can you add supports for makensis INPUTCHARSET OUTPUTCHARSET? Thanks!! Temporarily, i modified "node_modules/electron-builder/node_modules/app-builder-lib/out/targets/nsis/NsisTarget.js" and "NsisTarget.js.map" by insert " args.push("-INPUTCHARSET", "UTF8", "-OUTPUTCHARSET", "UTF8");" into line 724, to support utf-8 nsis script. ![image](https://user-images.githubusercontent.com/17950041/80283237-9d9e1700-8748-11ea-89f5-90a2150c0c5b.png) ![image](https://user-images.githubusercontent.com/17950041/80283251-b9a1b880-8748-11ea-8bfd-dce8292db289.png) Answers: username_1: same problem here.We have to use custom nsis script in electorn-builder, but got "Bad text encoding" error.We have to use INPUTCHARSET option with makensis. Any fix for that? username_2: These is useful to fix custom scripts encoding issues. Why closed. username_3: * **Version**: [email protected]@app-builder-lib * **Electron Version**: 5.0.4 * **Target**: build <!-- Enter your issue details below this comment. --> My windows charset is gbk, when i use makensis.exe to compile utf-8 nsis script, i got a "Bad text encoding" error. ![image](https://user-images.githubusercontent.com/17950041/80283187-60398980-8748-11ea-9225-b206fe97d12b.png) Makensis.exe has option to specify input charset, but electron-builder seem doesn't have this option. Can you add supports for makensis INPUTCHARSET OUTPUTCHARSET? Thanks!! Temporarily, i modified "node_modules/electron-builder/node_modules/app-builder-lib/out/targets/nsis/NsisTarget.js" and "NsisTarget.js.map" by insert " args.push("-INPUTCHARSET", "UTF8", "-OUTPUTCHARSET", "UTF8");" into line 724, to support utf-8 nsis script. ![image](https://user-images.githubusercontent.com/17950041/80283237-9d9e1700-8748-11ea-89f5-90a2150c0c5b.png) ![image](https://user-images.githubusercontent.com/17950041/80283251-b9a1b880-8748-11ea-8bfd-dce8292db289.png) username_3: Is there a particular reason why we wouldn't want the input and output charsets to be utf8? Seems like those args would be beneficial to just use by default. Status: Issue closed
quarkusio/quarkus
901286979
Title: OIDC introspection (opaque access token, not jwt) Question: username_0: I'm using quarkus 1.11.1. and oidc _service_ application type. Is there any option to disable introspection for opaque access tokens? Answers: username_1: @username_0 In 1.13.x or 2.0.0.M2 one can use `quarkus.oidc.token.allow-jwt-introspection=false` - but it only blocks the JWT token introspection. What is your requirement ? Are you working with JWT but would like to block the opaque tokens ? One option to try is to disable the auto discovery and only set the `jwks-path` - in this case no introspection will be possible for any token - not sure how it will work in `1.11.x` - but should fail with `401` in `1.13.x`/`2.0.0.M2` username_0: @username_1 Perfect that will do. Thank you ;) Status: Issue closed username_1: @username_0 Please re-open if it is not 401 for the opaque tokens when introspection path is not available in 1.13.x/2.0.0.x, thanks username_1: It will fail, but we don't have a test yet where an opaque token is rejected because no introspection path is available username_1: Perfect, thanks :-) username_0: sorry i meant 500 username_0: I'm using quarkus 1.11.1. and oidc _service_ application type. Is there any option to disable introspection for opaque access tokens? Thank you. username_1: @username_0 OK :-), which version is it ? username_0: 1.13.x username_1: I think I may need to introduce `quarkus.oidc.token.allow-opaque-token-introspection` - since if it is a valid opaque token and the server has not been configured with the introspection path it may well be `500`, especially if it is expected the opaque tokens are used. It would also work with or without auto discovery (as opposed to the suggested indirect solution) and will be easy to understand what it is for. username_0: quick question @username_1 is it possible to have multiple clients calling a quarkus api with an opaque access token? probably I'm missing something but if the lib calls the introspection, which requires a client id and secret, how do I handle multiple client ids? Status: Issue closed
nsftx/chameleon-bundle-material
510556774
Title: Image in Vertical list overlaps the defined container height Question: username_0: When placing an Image to a Vertical List and defining the height to the Vertical List, setting width and height of an image to 100% should place an image to defined container height, but it does not. It rather goes of the defined frame and takes 100% of the page. This behaviour is present on deployed app, when in builder, it looks fine. See images: _in builder:_ ![image](https://user-images.githubusercontent.com/50828021/67276113-49c62300-f4c4-11e9-9782-7b172847d394.png) _on deployed app:_ ![image](https://user-images.githubusercontent.com/50828021/67276204-7843fe00-f4c4-11e9-87b9-d39a07d72ab7.png) **Steps to reproduce:** 1. Place Panel to the page and set height to 100vh and width to 100% 2. Drag and drop Vertical List to the Panel and set a height in percentages, **lower** than 100% 3. Place an Image to the Vertical List and set width and height to 100%<issue_closed> Status: Issue closed
skrafft/react-native-jitsi-meet
845184130
Title: Why there's no index.js on this project? Question: username_0: Sorry if my question is too stupid, but I just installed it on my React Native (w/Typescript) clean project and I get this error: `Module not found: Can't resolve 'react-native-jitsi-meet'`. Answers: username_1: Hi, It's because `@types/react-native-jitsi-meet` is not available to provide TypeScript compile checks. add // @ts-ignore to suppress the error. username_0: @username_1 I've added the @ts-ignore, but nothing changes. The same error persists. username_1: Could please give us more information, if you write this, you can't have this typescript error ``` // @ts-ignore import JitsiMeet, { JitsiMeetView } from 'react-native-jitsi-meet'; ``` username_1: From #84 react-native-jitsi-meet is using WebRTC, WebRTC is not compliant with expo projects. You have to eject your project if you want to use react native jitsi meet.
botmakers-net/bot-templates
635276562
Title: Bot de prise de rendez-vous pour les entreprises d'enlèvement de déchets Question: username_0: Améliorez la notoriété de votre marque, promouvez et vendez vos services de suppression des déchets pour attirer plus de clients en utilisant le canal de marketing des médias sociaux # 1 #rendez-vous-reservation https://botmakers.net/chatbot-templates/bot-de-prise-de-rendez-vous-pour-les-entreprises-d-enlevement-de-dechets
Azure/azure-functions-powershell-worker
920231677
Title: msonline module - where statements not working Question: username_0: Hey Team, I am trying to get a script working but any time I pipe a "where" or "Where-object" into it, I get zero results. i.e. $UserNames = Get-MsolUser -TenantId $tenantid -ALL | Where {($_.licenses).AccountSkuId -ilike "*$($licenseName)"} | Sort-Object DisplayName If i run the same script locally it works perfectly fine, but just does not work in functions. If i remove the Where statement it runs fine. Please help Answers: username_1: Hello @username_0 -- Thank you for reporting this issue. Are you importing the module using the `-UseWindowsPowerShell` ? username_0: Yes, definitely. The module doesnt work unless you do that. I am wondering if its something to do with "deserialized" objects. as it says this when importing the module. Module msonline is loaded in Windows PowerShell using WinPSCompatSession remoting session; please note that all input and output of commands from this module will be deserialized objects. and some more digging I have found the licenses property returns. "Licenses": [ "Microsoft.Online.Administration.UserLicense" ], Instead of the actual license names :S . So I am not sure where to go from here. I cannot use the azuread module cause it does not work with the partner center delegated admin like it should. But I need a solution to get this into azure functions asap. username_1: Thanks for the reply @username_0. There is another potential workaround you can try which uses the Windows PowerShell version in the worker (this is the VM where your function code is executed). The workaround is described here: https://github.com/Azure/azure-functions-powershell-worker/issues/232#issuecomment-536744760. Here is a full example: https://github.com/eamonoreilly/ManageAzureActiveDirectoryWithPowerShellFunction/blob/master/ManageAzureAD/run.ps1 username_0: Great idea, and it seems to work! Just need to work out getting the right output from it! Genius! Thanks so much for sharing that, that will get me out of trouble for now. username_1: Great to hear the workaround is working. I will follow-up with the PowerShell Team on the serialization issue. Thanks. Status: Issue closed
ocornut/imgui
589883527
Title: unresolved external symbol LNK2019 Question: username_0: hello, I was interested in your development and decided to do something and as a result I get errors LNK2019. ---- **Version/Branch of Dear ImGui:** Version: last **Back-end/Renderer/Compiler/OS** Compiler: Visual Studio 2019 _(if the question is related to building or platform specific features)_ Operating System: Windows 10 **My Issue/Question:** I decided to create a regular imgui menu, and decided to check on the game (Saint's Row: The Third). As a result, it did not work to create imgui menu in any way due to LNK2019 error. I translated the translator in advance, I apologize. **Screenshots/Video** XXX _(you can drag files here)_ ![image](https://user-images.githubusercontent.com/50081707/77860437-f5b02f00-7217-11ea-996b-a72c12dcdf92.png) ``` // Here's some code anyone can copy and paste to reproduce your issue // dear imgui: standalone example application for DirectX 9 // If you are new to dear imgui, see examples/README.txt and documentation at the top of imgui.cpp. #include "imgui.h" #include "imgui_impl_dx9.h" #include "imgui_impl_win32.h" #include <d3d9.h> #define DIRECTINPUT_VERSION 0x0800 #include <dinput.h> #include <tchar.h> // Data static LPDIRECT3D9 g_pD3D = NULL; static LPDIRECT3DDEVICE9 g_pd3dDevice = NULL; static D3DPRESENT_PARAMETERS g_d3dpp = {}; // Forward declarations of helper functions bool CreateDeviceD3D(HWND hWnd); void CleanupDeviceD3D(); void ResetDevice(); LRESULT WINAPI WndProc(HWND hWnd, UINT msg, WPARAM wParam, LPARAM lParam); // Main code int main(int, char**) { // Create application window WNDCLASSEX wc = { sizeof(WNDCLASSEX), CS_CLASSDC, WndProc, 0L, 0L, GetModuleHandle(NULL), NULL, NULL, NULL, NULL, _T("ImGui Example"), NULL }; ::RegisterClassEx(&wc); HWND hwnd = ::CreateWindow(wc.lpszClassName, _T("Dear ImGui DirectX9 Example"), WS_OVERLAPPEDWINDOW, 100, 100, 1280, 800, NULL, NULL, wc.hInstance, NULL); // Initialize Direct3D if (!CreateDeviceD3D(hwnd)) { CleanupDeviceD3D(); ::UnregisterClass(wc.lpszClassName, wc.hInstance); [Truncated] case WM_SIZE: if (g_pd3dDevice != NULL && wParam != SIZE_MINIMIZED) { g_d3dpp.BackBufferWidth = LOWORD(lParam); g_d3dpp.BackBufferHeight = HIWORD(lParam); ResetDevice(); } return 0; case WM_SYSCOMMAND: if ((wParam & 0xfff0) == SC_KEYMENU) // Disable ALT application menu return 0; break; case WM_DESTROY: ::PostQuitMessage(0); return 0; } return ::DefWindowProc(hWnd, msg, wParam, lParam); } ``` Answers: username_1: Please google or ask about your error on Stack Overflow or another programming forum, this is unrelated to dear imgui. Thank you. Status: Issue closed username_1: Sorry but #1586
vinodtapmobi/havok-operations
52799583
Title: P1 issue: Page rendering is not happening properly when more than 2 banners are added in banner control at Lumata node Question: username_0: Tested UR at editor: http://p2.tapmobi.in/app/gui/editor.html?pageId=521 Tested URL at Lumata node: http://l1.tapmobi.in/r.aspx?pageid=521 Actual: When user adds more than 2 banners in Banner block control, The added banners are not reflecting in the Slots screen at editor page as well as platform node on mobile as below in the screenshot: ![image](https://cloud.githubusercontent.com/assets/8113291/5546290/5b5d5dfa-8b64-11e4-9454-c0ac3167961d.png) Banners are displaying properly as expected in the below screenshot, when there are 2 or less are available/added as below in the screenshot: ![image](https://cloud.githubusercontent.com/assets/8113291/5546302/ab0b8354-8b64-11e4-898a-129a93846eee.png) Note: Banners are displaying properly in Preview page and in browsers by using Modfify header/User Agent string at platfrom node for the same. Answers: username_0: Closed Status: Issue closed
halo-dev/halo-theme-hux
600809389
Title: 所有分类的页面都无法访问 Question: username_0: 如下图,所有分类都显示无法找到模板,请问是 halo 更新了1.3.x 版本的问题吗? ![image](https://user-images.githubusercontent.com/43313765/79426925-73d14b80-7ff6-11ea-81ae-4fa2eaed5449.png) Answers: username_1: 这主题本身就没有分类模板。 username_0: 好吧,确实只有标签页,我再想想办法,谢谢! Status: Issue closed username_0: 如下图,所有分类都显示无法找到模板,请问是 halo 更新了1.3.x 版本的问题吗? ![image](https://user-images.githubusercontent.com/43313765/79426925-73d14b80-7ff6-11ea-81ae-4fa2eaed5449.png) Status: Issue closed
telerik/kendo-ui-core
382149041
Title: Support TreeList incell editing with drag and drop Question: username_0: ### Enhancement Ticket 1349649 ### Reproduction of the problem ``` editable: { move:true, mode: "incell" } ``` Edit a cell and click to another cell. The changed value will be lost. [https://dojo.telerik.com/@bubblemaster/iFAFIVix](https://dojo.telerik.com/@bubblemaster/iFAFIVix) ### Current behavior The draggable functionality prevents the mousedown event. As a result, the change event of the editor input does not fire, which in turn prevents the MVVM from saving the updated value. ### Expected/desired behavior It would be great to have the ability to use both of these features together. Perhaps by wiring a drag event handler internally and dealing with any changes before dragging. ### Workaround Add an event handler to the `cellClose` event and trigger change programmatically to any inputs with the `data-bind` attribute as well as to any widgets: ``` cellClose: function(e) { var input = e.container.find("input[data-bind*='value']"); input.trigger("change"); var numeric = input.data("kendoNumericTextBox"); if(numeric){ numeric.trigger("change"); } } ``` ### Environment * **Kendo UI version:** 2018.3.1017 Answers: username_1: We can evaluate an ignore option to the user events similar to draggable: https://docs.telerik.com/kendo-ui/api/javascript/ui/draggable/configuration/ignore username_2: Added a note about the current limitation to the TreeList documentation here based on customer feedback: [https://docs.telerik.com/kendo-ui/controls/data-management/treelist/editing](https://docs.telerik.com/kendo-ui/controls/data-management/treelist/editing) If the problem is resolved, we should remove the note. username_3: DevExpress's TreeList widget gets around this issue by adding a special left-most column where the user can click/tap for drag-and-drop operations. All the other columns can still allow in-cell editing. I think you guys should consider a similar option. ![image](https://user-images.githubusercontent.com/10633170/93512287-1c5f4080-f8f2-11ea-90c6-8b27d12188de.png)
jayfk/statuspage
162125895
Title: Adding new systems ? Question: username_0: Once we have set up the page, how do we add new systems ? Answers: username_1: You need to add a new label and run the update process (or, if you are using the latest RC, refresh the page). Make sure to use the systems label color `171717 `. Status: Issue closed
influxdata/influxdb-client-go
1091864413
Title: Add WriteBatch to WriteAPIBlocking Question: username_0: The problem in my situation is that `WriteAPIBlocking` accepts only on individual points or line-protocol records (possibly in a slice, but still separated). In my application, if I need to persist unwritten points the natural encoding is newline-separated (or terminated) line protocol records in a byte sequence. However, when I read that back off disk and prepare to resend it the current API requires me to split it into individual records at the newline, so that WriteRecord can then add the newlines back and aggregate it again. This is cumbersome, but trivially avoided by extending the API to allow writing a batch. Answers: username_1: The already existing `WriteRecord` function can also take just a single string. And such string can be a batch. Status: Issue closed
google/blockly
650900649
Title: FieldDate is not included in the master branch Question: username_0: Hello - I am looking at the Blockly documentation and it looks like the Date field is included (https://developers.google.com/blockly/guides/create-custom-blocks/fields/built-in-fields/date?hl=en) and the following is recommended to enable Date field: __Warning: Due to its limited use and large footprint, FieldDate is not compiled into Blockly by default. To use it you will need to require it and rebuild. Add goog.require('Blockly.FieldDate') to your project, or uncomment it inside blockly.js to enable it. FieldDate also depends on the Closure Library which is no longer compiled into Blockly by default. You will need to add google-closure-library to your package.json and rebuild with the --closure-library flag: gulp build --closure-library.__ However, unfortunately, I could not see any code or reference related to FieldDate in the master branch content in Github. Am I miss something somewhere or is there anything wrong in the documentation. Many thanks, -fatih Answers: username_1: Thanks for filing @username_0. We've moved FieldDate to it's own package and is now published at https://www.npmjs.com/package/@blockly/field-date Will keep this issue to track updating the documentation to reflect that. username_2: Closing this issue in favor of #4094 that more clearly explains how to update documentation. Status: Issue closed
lovelmh13/myBlog
952633782
Title: mysql 版本号比较 Question: username_0: 取每组版本号并向前补0至N位(比方5位、10位),最后拼接好在进行比较 ```sql CONCAT( LPAD( SUBSTRING_INDEX( SUBSTRING_INDEX( cver, '.', 1 ), '.', - 1 ), 10, '0' ), LPAD( SUBSTRING_INDEX( SUBSTRING_INDEX( cver, '.', 2 ), '.', - 1 ), 10, '0' ), LPAD( SUBSTRING_INDEX( SUBSTRING_INDEX( cver, '.', 3 ), '.', - 1 ), 10, '0' ) ) > CONCAT( LPAD( SUBSTRING_INDEX( SUBSTRING_INDEX( repaired_cver, '.', 1 ), '.', - 1 ), 10, '0' ), LPAD( SUBSTRING_INDEX( SUBSTRING_INDEX( repaired_cver, '.', 2 ), '.', - 1 ), 10, '0' ), LPAD( SUBSTRING_INDEX( SUBSTRING_INDEX( repaired_cver, '.', 3 ), '.', - 1 ), 10, '0' ) ) ``` 这个 sql ,每个组成部分最多可以包含10位数字。 它将转换为:X.XX.XXX > Y.YY.YYY → '000000000X00000000XX0000000XXX' > '000000000Y00000000YY0000000YYY'
kubernetes/kubernetes
364185148
Title: Fix descriptor lock release logic for block volume unmapDevice Question: username_0: **Is this a BUG REPORT or FEATURE REQUEST?**: /kind bug **What you expected to happen**: - Current call sequence of [```unmapDeviceFunc```](https://github.com/kubernetes/kubernetes/blob/master/pkg/volume/util/operationexecutor/operation_generator.go#L1069) generated by ```GenerateUnmapDeviceFunc``` is as follows: (1) Check that symlink doesn't exist in globalMapPath to ensure that no one is accessing a device to detach (2) Call ```TearDownDevice``` for driver to detach the device from the node (3) Delete globalMapPath (4) Delete descriptor lock by removing loop device For backend storage that can't detach a device while descriptor lock is held, the order of calling (2) and (4) makes problem. As a workaround, [rbd driver release the descriptor lock inside (2) before detaching a device](https://github.com/kubernetes/kubernetes/blob/master/pkg/volume/rbd/rbd.go#L962). This makes difficult to implement ```TearDownDevice``` properly for csi driver that can work with all backend storage, because it needs to handle both backend which requires descriptor lock to be released before detaching a device and backend which doesn't requires it. - This issue should be resolved by changing the sequense as follows: (1) Check that symlink doesn't exist in globalMapPath to ensure that no one is accessing a device to detach (2) Delete descriptor lock by removing loop device (3) Call ```TearDownDevice``` for driver to detach the device from the node (4) Delete globalMapPath Note that the issue of failure in releasing descriptor lock in ```unmapDeviceFunc```, which is described in local attach drivers' ```TearDownDevice``` [comments](https://github.com/kubernetes/kubernetes/blob/master/pkg/volume/iscsi/iscsi_util.go#L673), should be resolved now. Because in current implementation of ```mapVolumeFunc```, [taking file descriptor lock](https://github.com/kubernetes/kubernetes/blob/master/pkg/volume/util/operationexecutor/operation_generator.go#L912) and [updating ```volumeToMount```](https://github.com/kubernetes/kubernetes/blob/master/pkg/volume/util/operationexecutor/operation_generator.go#L919) are done to the same ```devicePath```. Therefore, there should be no need to release descriptor lock inside ```TearDownDevice```. **How to reproduce it (as minimally and precisely as possible)**: Run below e2e tests on environment with ```BlockVolume=true``` - ```--ginkgo.focus="In-tree.*iscsi.*Pre-provision.*block.*volumeMode.*should create.*"``` - ```--ginkgo.focus="In-tree.*rbd.*Pre-provision.*block.*volumeMode.*should create.*"``` **Anything else we need to know?**: /sig storage This issue is opened as a separated issue from https://github.com/kubernetes/kubernetes/pull/68635 CSI drivers that requires the descriptor lock to be release before detaching device won't work well on ```TearDownDevice```, until this issue is fixed. **Environment**: - Kubernetes version (use `kubectl version`): # ```kubectl version Client Version: version.Info{Major:"1", Minor:"13+", GitVersion:"v1.13.0-alpha.0.1580+07e81cb8ff590d-dirty", GitCommit:"<PASSWORD>", GitTreeState:"dirty", BuildDate:"2018-09-26T19:23:23Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"13+", GitVersion:"v1.13.0-alpha.0.1580+07e81cb8ff590d-dirty", GitCommit:"<PASSWORD>", GitTreeState:"dirty", BuildDate:"2018-09-26T19:23:23Z", GoVersion:"go1.10.3", Compiler:"gc", Platform:"linux/amd64"}``` - Cloud provider or hardware configuration: kvm - OS (e.g. from /etc/os-release): ```Fedora release 27 (Twenty Seven)``` - Kernel (e.g. `uname -a`): ```Linux csi-k8s 4.13.9-300.fc27.x86_64 #1 SMP Mon Oct 23 13:41:58 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux``` - Install tools: ```local-up-cluster.sh``` - Others:
bitsongofficial/go-bitsong
894216306
Title: Mining fan token amount Question: username_0: I tried to mint some fantoken, when we `mint` a new `fantoken` it will add 6 standard decimal and the result could be out of the proposed intentions. ## Command ``` bitsongd tx fantoken mint clay --recipient bitsong1lemmdcgn087pd380p54q0f45pe7vrylcetu529 --amount 1000 --from validator --chain-id test ``` ## What I expect I expect that with this command I will mint `1000 clay`. Currently the chain mint `1000000000uclay` I think the best approach is to change the command to accept something like that ``` ..... fantoken mint uclay --amount 1000000000 ``` with that expression the account is sure to get the specified amount. Answers: username_1: fixed in https://github.com/bitsongofficial/go-bitsong/tree/module/fantoken
Smattr/rumur
500630015
Title: Murphi preprocessor Question: username_0: A common use case turns out to be transforming Murphi models into something else. While this can be accomplished by linking against librumur, it is even nicer for code generators if they only have a subset of constructs to deal with. It should be possible for us to ship a standalone executable that exposes a collection of source-to-source transformations. Example of a simple one would be to weaken `a != b` into `!(a = b)`. Example of a more complicated one would be transforming a switch into an if-then-else. It would be extremely desirable if such transformations could as much as possible preserve formatting and whitespace.<issue_closed> Status: Issue closed
studentinsights/studentinsights
147910198
Title: Automate some IE QA Question: username_0: For now, our practice is to check the demo site and production site in Internet Explorer whenever we push a change that affects layout. Good idea from @kevinrobinson via an in-person chat at Code for Boston #346: after deploy, send a script to go look at the web app in IE, take screenshots, and post in Slack. Answers: username_0: This library may or may not be useful: https://github.com/watir/watir https://watir.com/ username_1: Browserstack will cost $29.99 per month. It allows you to create a virtual machine in Mac, Linux or Windows and put IE 11 in an automated way. There is a 100 minute unlimited free trial. Information about how to use their services is as follows: https://www.browserstack.com/automate/python https://www.browserstack.com/automate/ruby https://www.browserstack.com/automate/java https://www.browserstack.com/automate/node username_1: I got an example to work with Browserstack, pushed a branch: username_1:add_browserstack Status: Issue closed
learningequality/studio
365610505
Title: Email links are non-HTTPS Question: username_0: ## Summary When for instance inviting someone to a channel, the link should be HTTPS, otherwise one is redirected to a non-HTTPS page for logging in. ![image](https://user-images.githubusercontent.com/374612/46310878-e3bce400-c5c0-11e8-998a-a5d7e9f3db23.png) Fix 1: Change the HTTP server configuration to redirect all HTTP requests to HTTPS. This might be dangerous because we have some HTTP-only traffic for channel downloads? Fix 2: Add an HTTPS setting to force the protocol of URLs in emails. We can use `request.is_secure()` while sending emails, but then we are also relying on a request object, which is only relevant in views. See: https://docs.djangoproject.com/en/2.1/topics/security/#ssl-https ## Category Security<issue_closed> Status: Issue closed
cuberite/cuberite
210459112
Title: Implement SRV records Question: username_0: I read about implementing the 'cuberite new' subcommand as a way to spawn new instances of cuberite which could be useful as a virtual hosting solution. If you are going to implement such subcommand, then atleast add support for SRV records which can direct the connecting user to the corresponding Cuberite server. Answers: username_1: I've been using an SRV record for my Cuberite server without issues. Status: Issue closed username_2: SRV records are a client side issue - closing as not applicable.
pyenv/pyenv-virtualenv
299842730
Title: switching to different virtualenv does not change pip version accordingly Question: username_0: Cross posting the issue from [#1015](https://github.com/pyenv/pyenv/issues/1015) in `pyenv` repo because this is the same bug but also affects pyenv-virtualenv. It is rather frustrating. I have the following in my `.bashrc` on Ubuntu 16.06: ```bash eval "$(pyenv init -)" eval "$(pyenv virtualenv-init -)" ``` # Problem Description the pip executable being used does not change along with the python version upon evoking `pyenv active <virtualenv>` or `deactivate` for both system installs and pyenv installs alike. Evoking ` eval "$(pyenv init -)"` directly results in the pip executable updating as appropriate. However, changing back into or out of a virtualenv with undo this. In my situation, it is defaulting to a pip for pyenv python version I installed, `Python 3.5.2`. Here is the observed behavior: ```bash 💀 ↛ python --version Python 2.7.12 💀 ↛ python3 --version Python 3.5.2 💀 ↛ pip --version pip 9.0.1 from /home/spook/.local/lib/python3.5/site-packages (python 3.5) 💀 ↛ pyenv activate demo pyenv-virtualenv: prompt changing will be removed from future release. configure `export PYENV_VIRTUALENV_DISABLE_PROMPT=1' to simulate the behavior. (demo) 💀 ↛ python --version Python 3.6.0 (demo) 💀 ↛ pip --version pip 9.0.1 from /home/spook/.local/lib/python3.5/site-packages (python 3.5) (demo) 💀 ↛ eval "$(pyenv init -)" (demo) 💀 ↛ pip --version pip 9.0.1 from /home/spook/.pyenv/versions/3.6.0/envs/demo/lib/python3.6/site-packages (python 3.6) (demo) 💀 ↛ pyenv deactivate 💀 ↛ pip --version pip 9.0.1 from /home/spook/.local/lib/python3.5/site-packages (python 3.5) ``` Answers: username_0: lmao. # solution `pyenv update` Problematic behavior has disappeared. Will update if it does come back. Thanks to everyone who maintains this package. Status: Issue closed username_0: Hmm. I still get this behavior. username_0: Cross posting the issue from [#1015](https://github.com/pyenv/pyenv/issues/1015) in `pyenv` repo because this is the same bug but also affects pyenv-virtualenv. It is rather frustrating. I have the following in my `.bashrc` on Ubuntu 16.06: ```bash eval "$(pyenv init -)" eval "$(pyenv virtualenv-init -)" ``` # Problem Description the pip executable being used does not change along with the python version upon evoking `pyenv active <virtualenv>` or `deactivate` for both system installs and pyenv installs alike. Evoking ` eval "$(pyenv init -)"` directly results in the pip executable updating as appropriate. However, changing back into or out of a virtualenv with undo this. In my situation, it is defaulting to a pip for pyenv python version I installed, `Python 3.5.2`. Here is the observed behavior: ```bash 💀 ↛ python --version Python 2.7.12 💀 ↛ python3 --version Python 3.5.2 💀 ↛ pip --version pip 9.0.1 from /home/spook/.local/lib/python3.5/site-packages (python 3.5) 💀 ↛ pyenv activate demo pyenv-virtualenv: prompt changing will be removed from future release. configure `export PYENV_VIRTUALENV_DISABLE_PROMPT=1' to simulate the behavior. (demo) 💀 ↛ python --version Python 3.6.0 (demo) 💀 ↛ pip --version pip 9.0.1 from /home/spook/.local/lib/python3.5/site-packages (python 3.5) (demo) 💀 ↛ eval "$(pyenv init -)" (demo) 💀 ↛ pip --version pip 9.0.1 from /home/spook/.pyenv/versions/3.6.0/envs/demo/lib/python3.6/site-packages (python 3.6) (demo) 💀 ↛ pyenv deactivate 💀 ↛ pip --version pip 9.0.1 from /home/spook/.local/lib/python3.5/site-packages (python 3.5) ``` username_1: It looks like you have pip installation at `/home/spook/.local/lib/python3.5/site-packages` and it has precedence than pip installation managed by pyenv/pyenv-virtualenv via shell's `$PATH` configuration. In general, this must be some shell configuration issue of your environment. BTW, both `pyenv-init` and `pyenv-virtualenv-init` will try to put its command path at the top of `$PATH`. It means that the ordering of the execution matters to manage what to be invoked by given command. Running `pyenv-init` multiple times means that it forces shell to pick up pyenv's shims regardless of rest of shell configuration. I'd not recommend to do so unless you know what you're doing. https://github.com/pyenv/pyenv/blob/v1.2.1/libexec/pyenv-init#L87-L96 https://github.com/pyenv/pyenv-virtualenv/blob/v1.1.1/bin/pyenv-virtualenv-init#L86-L99 Status: Issue closed username_0: I'm 99% sure I have my shell configured correctly and have no other commands that should affect the evaluation of my $PATH generally or those specific to pyenv/python. That said, since updating like stated above, switching pyenv-virtualenvs results in the proper pip most of the time. I just double check now with `pip --version` to make sure. Do you have any idea why I may sometimes not get a proper order evaluation order/execution of the shim on my $PATH? Could having multiple shells opened (with different pyenv-virtualenvs activated) possibly effect it? This isn't a real problem anymore but I would just like to understand more about the program! Thank you for maintaining such a great tool :smile:
MikeAxtell/ShortStack
47870289
Title: super slow in miRNA finding if genome is fragmented Question: username_0: Several users have noted that miRNA finding becomes super-slow in cases where the reference genomes are highly fragmented (like lots of small scaffolds instead of a few large chromosomes). Seems to be that samtools faidx (used internally by maple to get sequences for folding) is slow in these cases. Will try to address ... Answers: username_0: Solved as of 3.0 beta commits. Highly fragmented genomes (> 50 chromosomes and N50 length < 1Mb) are now 'stitched' together into larger contigs which drastically improves runtimes. Status: Issue closed
yingkaisha/keras-unet-collection
1047964079
Title: Patch Embedding Dimension in TransUNet Question: username_0: I think the patch embedding dimensions in your implementation might be larger than the original TransUNet paper. In the original TransUNet, it is mentioned that "For the “base” model, the **hidden size D**, **number of layers**, MLP size, and number of heads are set to be 12, 768, 3072, and 12, respectively". The hidden size D is the embedding dimension of the transformer's output which is set to 12 (unless the authors made a mistake in the order of the numbers). However, in your implementation, it is set to 768. Am I missing something? Answers: username_1: I feel that they made a mistake here. `embedding size = hidden size D = 768` is somewhat a default hyperparameter choice inherited from the BERT-base. Given the tensor size prior to embedding (height times width times channel), 12 is too low. username_0: That makes sense. It's most likely a mistake. One more question. In the paper, what do they mean by "number of layers" which is equal to 768? username_1: My guess: ``` "number of layers" = number of transformers = 12 "MLP size" = number of MLP nodes per transformer = 3072 "number of heads" = number of self-attention heads = 12 "hidden size D" = number of embedded dimensions = 768 ``` There are no guarantees, and you can contact the original authors of transunet. Status: Issue closed
cretueusebiu/vform
255344819
Title: How upload image/file? Question: username_0: i was using `FormData()` for file/image upload. but i dont know how to implement with this. please help me Answers: username_1: Take a look at [this example](https://github.com/username_1/vform/blob/master/example/upload.html). Status: Issue closed username_2: Hi, I am using vForm and would like to upload multiple files on the <input type="file" multiple /> input field. I had a look at the example that @username_1 provided but it is not a working solution for me. I am using FileReader() with the input fields.
canonical/operator
625724434
Title: Remove default target for Framework.observe Question: username_0: Our experience so far has shown that the implicit dependencies for Framework.observe push people down a path that actually makes their charms less maintainable. Especially when it comes to writing components, it is better to have clearly named functions and clearly defined targets. For components, it is usually better to have private methods, because the handlers are an implementation detail rather than something you want to advertise to people who want to interact with your component. This is a breaking change (all charms that have been using `self.framework.observe(self.on.config_changed, self)` will break. So we want to make this change before we do an official release version.<issue_closed> Status: Issue closed
pbkhrv/handshake-telegram-bot
928806928
Title: Fuzzy search thru names in /alerts Question: username_0: If i say `/alerts *easy*`, bot should show details of all my alerts for names that contain the word `easy` in them. Why: I have lots of alerts set, this would make it easier for me to see alert details for a particular name without having to type it out fully.
medeia/medeia
457874647
Title: Do we need `BsonDocumentCodec` Question: username_0: When deriving the encoder via `deriveBsonEncoder` we automatically get a `BsonDocumentEncoder` for case classes. If we use `deriveBsonCodec` however, we only get a `BsonEncoder` and we have to derive both separately to get the document encoder. Maybe we should also have a `BsonDocumentCodec`?<issue_closed> Status: Issue closed
twatter-soen341/Twatter
415312884
Title: Acceptance Test for User Story #25 Question: username_0: ## Acceptance Test | User Story | `#25` | | --: | :--| | **Test Priority** | NORMAL | | **Test Title** | User can create a post | | **Description** | User can write some content and save it as a new post on Twatter. | | **Tester Name** | <NAME> | ## Test Result | Step | Test Step | Test Data | Expected Results | Acutal Results | Status | Notes | | :--: | :-- | :-- | :-- | :-- | :--: | :-- | | 1 | Navigate to home page by clicking the home icon. | | Route to home page | | | | | 2 | Click on the textbox under "Create Twat". | | Display dialog box to create "Twat" | | | | | 3 | Write some content. | | Content is typed in textarea | | | | | 4 | Click "TWAT" to post. | | Dialog box closes and the post is added in home page and user's profile. | | | | _**Step**_ *&* _**Test Step**_ *should be filled in description. The rest should be filled and reposted as comment.* Answers: username_0: ## Test Result | Step | Test Step | Test Data | Expected Results | Acutal Results | Status | Notes | | :--: | :-- | :-- | :-- | :-- | :--: | :-- | | 1 | Navigate to home page by clicking the home icon. | | Route to home page | Route to home page | Pass | | | 2 | Click on the textbox under "Create Twat". | | Display dialog box to create "Twat" | Display dialog box to create "Twat" | Pass | | | 3 | Write some content. | "Roast me!" | Content is typed in textarea | Content is typed in textarea | Pass | | | 4 | Click "TWAT" to post. | "Roast me!" | Dialog box closes and the post is added in home page and user's profile. | Dialog box closes and the post is added in home page and user's profile. | Pass | User stor #25 Passed Accepance test #161 | Status: Issue closed username_1: Acceptance Test Completed.
unoacm/code_court
214543914
Title: Update executor api to allow for reporting of failure conditions Question: username_0: Possible failure conditions: - run timed out - run gave too much output - run gave no output - invalid run parameters - runtime error [maybe] - compilation error [maybe] This api will be used by the executor to report error conditions, and the defendant frontend to display error conditions. Answers: username_1: Here are a few more that might be useful: https://icpcarchive.ecs.baylor.edu/index.php?option=com_content&task=view&id=14&Itemid=30 Status: Issue closed
dbeaver/dbeaver
756196492
Title: Dashboard name is not displayed in Delete popup Question: username_0: #### System information: - Operating system (distribution) and version Mac - DBeaver version 7.3.0 - Additional extensions #### Steps to reproduce, if exist: 1. Create a dashboard 2. Click on it and press Delete Expected: confirmation popup appears with the Dashboard name Actual result: brackets with null are displayed in the popup message ![Screenshot 2020-12-03 at 16 19 39](https://user-images.githubusercontent.com/51405061/101024136-bdba7200-3584-11eb-9f37-63960daf2821.png) Answers: username_1: verified Status: Issue closed
momo-niu/selfintro
642856644
Title: Get Started with springboot. Question: username_0: 1. Follow these steps to build an app with springboot. [https://spring.io/guides/gs/spring-boot/](Guides) 2. Wrap one or more controller to provide a RESTful web service. [https://spring.io/guides/gs/rest-service/](Guides)
python-discord/bot
693997427
Title: !help output in muted #bot-commands Question: username_0: When someone uses `!help` in another channel than `#bot-commands`, the output is redirected to `#bot-commands`. This is still the case when send message permissions in `#bot-commands` are removed, which could lead to spamming or "drowning" a potential message about what is happening, in the channel. Answers: username_1: Us having to disable send message permissions in bot-commands specifically is a rare event, and it isn't the end of the world if this issue happens anyway. Messages with information can be pinned in the channel, and updates on an incident would usually be in `#announcements` or `#changelog` anyway. I'm not sure changing this behavior is really worth it imo. Status: Issue closed
nwootton/MMM-UKLiveBusStopInfo
302796868
Title: NextBuses toggle schedule Question: username_0: In order to get more use of the NextBuses API, would it be possible to implement start/stop time ranges to set _nextBuses=yes_ on a schedule? I expect most people have a regular commute where they'd like more accurate bus times for a short period of each day ... e.g. ``` nextBusesStart: 07.45 nextBusesStop: 08.15 ``` If that could be combined with a general start/stop time for all updates then you could stop polling the API completely overnight (our buses don't run overnight anyway). e.g. ``` start: 07.30 stop: 19.30 ``` Based on those examples I think you could have it updating every 1 minute and still keep under the 1000 requests per day ... Answers: username_1: Yeah, have a look at the StartTime branch, I think he solved this there dude username_2: It's a prototype only on that brach. I never really had the time to work on it fully. Not even sure that the code in there works, so caveat emptor! username_3: Well I'll be - seems to work for me! Consider this a +1 for merging this into the master branch when you next get some time ... username_3: In the end, I actually went with using https://github.com/ianperrin/MMM-ModuleScheduler That lets me setup more than one module entry with different updateinterval and showRealTime/showDelay settings for different times of day, and switching off the module completely overnight, giving more control than the SmartTime branch, though it's a bit trickier to configure. username_0: After a month of occasional tweaking, I've been unable to get this working reliably with ModuleScheduler. It's close, but something seems to be stopping the nextbus live data. I'm not sure if this has anything to do with ModuleScheduler as it seems to get stuck even without this. It either gets stuck showing 'Loading bus info ...' or shows no scheduled departures even when there are some using the live api documentation. Sometimes it works but not for long. I've been unable to tie it down to any particular time of day. Could you see if there's anything suspicious with my ATCO code? ``` { module: 'MMM-UKLiveBusStopInfo', position: 'bottom_left', header: 'To Leeds', classes: 'scheduler', config: { module_schedule: {from: '15 7 * * *', to: '15 8 * * *' }, atcocode: '3200YNA00515', // ATCO code for specific bus stop app_id: '###', app_key: '###', limit: 5, nextBuses: "yes", updateInterval: 60000, showRealTime: true, showDelay: true, showBearing: true, } }, ``` username_2: Immediate response is that your nextBuses value should be 'yes' not "yes" and you have a trailing comma after showBearing - but those will depend on how strict the json parser in MM is. TransportAPI has no issues with your ATCO - Marlborough Drive, Tadcaster going NE. Looks like it's route 840 going to Leeds Station. As I don't use ModuleScheduler, I'm afraid there's very little else I can suggest. Sometimes these issues are down to the bus operator not getting the info to the TransportAPI in a timely manner or even with incorrect data. You can try adding ```debug: true ``` to the bottom of the config (after showBearing) and start the mirror in dev mode. That will show you all the messages flowing around the mirror modules and into the Bus module. username_2: Have you checked the usage stats on the TransportAPI to see if the issues correspond with spikes shown there? Status: Issue closed username_0: The json formatting tweaks looked good in a quick test last night, but by morning it had stopped working again ... I'll figure out how to start MM in dev mode and get some more tests setup later this week, then update this issue when I have more solid info.
apache/submarine
1049491139
Title: [Submarine Spark Security] Different masking policies for different tables fail in the case of union Question: username_0: Hi, We ran into a problem with Submarine's Spark Security for Column Masking For example, select col_1, col_2 from tbl_1 union select col_3, col_4 from tbl_2 If masking policies for col_2 and col_4 are different, col_4 will use col_2's policy instead of its own because col_4 is not in "plan.output" ![image](https://user-images.githubusercontent.com/67689926/141068525-6bb2f58e-3b57-40ed-b91a-f14c618e05c3.png) Answers: username_1: I also encountered the same problem. username_2: What is the expected output in this case? username_3: Hi,I was also using spark security plugin to authorize SparkSQL,but when i execute command "show tables " or any sql command ,it report below errors: 22/01/11 10:23:26 WARN PolicyRefresher: cache file does not exist or not readable '/etc/ranger/sparkServer/policycache/sparkSql_sparkServer.json' I have no idea why would this happen even though i configured a sparkService in hive in ranger? ![image](https://user-images.githubusercontent.com/19393229/148870991-bf4e0ca3-9ba9-45e2-a009-d6342edaa120.png) username_4: The [Submarine Spark Security] functionality has been moved to the apache/incubator-kyuubi standalone project. Status: Issue closed
haskell/haskell-language-server
1158258269
Title: hls consistently crashes on eval code lens Question: username_0: ### Your environment OS: arch linux 5.16.10 LSP client: neovim nightly built-in LSP client HLS version: from package haskell-language-server-static 1.6.1.0-1 at the AUR GHC version: 9.0.2 ### Steps to reproduce Create a file `test.hs` with the following contents: ``` -- >>> 3 + 5 ``` Evaluate the code lens from the eval plugin. ### Expected behaviour A new line should be inserted with the result: ``` -- >>> 3 + 5 -- 8 ``` ### Actual behaviour HLS crashes. Until a few weeks ago, the eval plugin worked flawlessly. Now HLS basically always crashes when executing an eval code lens, with some rare exceptions. However, the action is always completed normally if the expression raises an exception: ``` -- >>> show (+) -- No instance for (Show (Integer -> Integer -> Integer)) -- arising from a use of ‘show’ ``` ### Include debug information After triggering the code lens: ``` 2022-03-03 11:45:51.500641495 [ThreadId 170] INFO hls: finish: codeLens.GhcSession (took 0.00s) 2022-03-03 11:45:51.50063769 [ThreadId 168] INFO hls: finish: Wingman.getIdeDynflags.GetModSummaryWithoutTimestamps (took 0.00s) 2022-03-03 11:45:51.500743127 [ThreadId 167] INFO hls: finish: eval.GetParsedModuleWithComments (took 0.00s) 2022-03-03 11:45:51.500831964 [ThreadId 169] INFO hls: finish: ModuleName.GetParsedModule (took 0.00s) 2022-03-03 11:45:51.500847129 [ThreadId 171] INFO hls: finish: (took 0.00s) 2022-03-03 11:45:51.50094092 [ThreadId 178] INFO hls: finish: codeLens.TypeCheck (took 0.00s) 2022-03-03 11:45:51.501295017 [ThreadId 182] INFO hls: finish: codeLens.GetBindings (took 0.00s) 2022-03-03 11:45:51.501339683 [ThreadId 180] INFO hls: finish: Wingman.codeLensProvider.GetAnnotatedParsedSource (took 0.00s) 2022-03-03 11:45:51.50151244 [ThreadId 177] INFO hls: finish: RefineImports (took 0.00s) 2022-03-03 11:45:51.501542606 [ThreadId 191] INFO hls: finish: codeLens.GetGlobalBindingTypeSigs (took 0.00s) 2022-03-03 11:45:51.501607892 [ThreadId 196] INFO hls: finish: Wingman.codeLensProvider.GetBindings (took 0.00s) 2022-03-03 11:45:51.502004945 [ThreadId 198] INFO hls: finish: Wingman.emptyCaseScrutinees.TypeCheck (took 0.00s) 2022-03-03 11:45:51.502176917 [ThreadId 199] INFO hls: finish: Wingman.emptyCaseScrutinees.False (took 0.00s) 2022-03-03 11:45:51.985303435 [ThreadId 219] INFO hls: finish: eval (took 0.00s) 2022-03-03 11:45:51.985334051 [ThreadId 264] INFO hls: finish: runEvalCmd.getModSummary (took 0.00s) haskell-language-server-wrapper: callProcess: /usr/bin/haskell-language-server-9.0.2 "--lsp" (exit -11): failed ``` Answers: username_1: Any idea of what has changed in the last few weeks? username_2: Same issue here, running Fedora 35, GHC 8.6.5 and HLS 1.6.1 (also happens on 1.5.1). Related: #2363 username_0: Not really. It might have stopped working after a GHC update, but I only updated HLS after the code lenses stopped working.
hdmf-dev/hdmf
750129694
Title: validating an hdmf file can no longer be done by executing the module Question: username_0: ## Description The HDMF documentation shows the following process for validating a file: https://hdmf.readthedocs.io/en/stable/validation.html In talking with @rly, this used to work prior to extracting HDFM out of PyNWB, and now only works through the `pynwb.validate` module: https://pynwb.readthedocs.io/en/stable/validation.html Two options for resolving are to: 1. remove this from the documentation 2. re-add this functionality matching the documentation @rly suggested that we implement the second. ## Steps to Reproduce Attempt to validate a hdmf file using the command provided in the documentation: ``` $ python -m hdmf.validate -p namespace.yaml test.h5 ``` ## Environment Python Executable: Conda Python Version: Python 3.7 Operating System: Linux HDMF Version: 2.2.0 ## Checklist - [x] Have you ensured the feature or change was not already [reported](https://github.com/hdmf-dev/hdmf/issues) ? - [x] Have you included a brief and descriptive title? - [x] Have you included a clear description of the problem you are trying to solve? - [x] Have you included a minimal code snippet that reproduces the issue you are encountering? - [x] Have you checked our [Contributing](https://github.com/hdmf-dev/hdmf/blob/dev/docs/CONTRIBUTING.rst) document? Answers: username_0: I tagged this as a bug because it sounds like it was intended functionality that no longer works, but I don't know what the bug criteria is for this project, so this could very easily be an enhancement. I'd be happy to submit a PR for this. @rly suggested it could be very similar to this: https://github.com/NeurodataWithoutBorders/pynwb/blob/dev/src/pynwb/validate.py username_1: That is a reasonable label. In this particular case, it would also be fine to label it as an enhancement since this is functionality that by itself has not existed in HDMF, but I think either option is fine. username_0: I made a PR to update the documentation in #482. Should we keep this open, or create a new enhancement issue for implementing the documented functionality and close this issue? It seems like implementing this in the code might not be a high priority since the validation functionality is currently available in pynwb, do you guys agree? username_1: I don't think there are any pressing use cases that need this functionality in HDMF itself right now, but I'm sure it will become necessary at some point. It is good to at least have an issue that documents that this still needs to be done.
robotframework/robotframework
703379142
Title: Include exception type in error message Question: username_0: **Why**: In our organization we are trying to automate failure analysis, it will be like based on few keys in error message will mark test case as Automation or Application Issue. Example: Let us assume test case consist keyword `Should Be True "Demo" != "Demo1"` and it got failed, we will get error message like `"Demo" != "Demo1"` In above error message **=!** is my key and I update testcase analysis comment that its application issue since validation got failed. Accuracy of this approach is around or less than 50% and we want to improve accuracy of automation analysis. **What**: If robotframework provide exception type as prefix in error message this helps us to understand more about issue and mark test case accordingly. Example: 1. Robotframework: If any `Should *` keyword got failed error message should be like `ValidationException: <actual error>` 2. Selenium: If any `Wait *` keyword got failed error message should be `ElementException: <actual selenium throwed exception & error>` We believe this approach improve our auto failure analysis accuracy upto 80% or more
alibaba/MNN
478768176
Title: 对argmax层的维度识别错误 Question: username_0: 如题,caffe中的prototxt的argmax层axis为1,转换成mnn格式后,在MNN架构下运行,在CPUArgMax.cpp添加打印,发现argmax层的前一层维度可以正确输入,但argmax层中的width变为了1,channel保持不变,尝试改变prototxt中axis,重新转换mnn,发现无论更换为0,1,2还是3,argmax的输出结果唯独都不变化,想请教下这是什么原因导致的啊 Answers: username_1: tf.argmax(logits, 3, output_type=tf.int32) have the same problem, Looking forward to this bug being resolved as soon as possible. Thanks username_2: fixing username_3: pull最新的代码,还是argmax层维度识别错误 Status: Issue closed username_5: MNN-0.2.0.7发现同样的问题,argmax层中的width变为1,请问已经修复了吗? username_3: 我是把最后一层的argmax层单独拿出来,自己写个函数实现argmax的功能 ------------------ Vehicle Engineering , South China University of Technology 余蒙 Cell:155-2106-4363 E-mail:<EMAIL> ------------------ Original ------------------
fossasia/open-event-frontend
249880024
Title: Footer: All links are not working Question: username_0: Links on footer are not working. ![screenshot from 2017-08-13 17-20-28](https://user-images.githubusercontent.com/1583873/29250926-c47a4fb2-804b-11e7-80c5-ae044d400d7d.png) Answers: username_1: @username_0 I would like to work on this issue, please assign this issue to me. username_2: sir, is this issue is still present? As i have open the website link and all the footer link are working. please reply sir thanks username_3: hello sir, sir, The contact us in footer is not taking to the page where the user can contact you, so either a page is needed or the contact us option should be removed and same as with the help button also. And the categories are just taking to the top of page and to the categories divisions. Lastly, there is no use of languages changer at the footer in the pages as no change is happening so its better to remove the option. Else no issue can be seen it the footer. username_4: @username_3 The contact us in the footer is a dummy page that the admin of the app creates. So for now, we have put dummy content in it. It depends on the admin what he wants to add. The horizontal links in the footer can be created from ```admin/users/pages```. Hence that part is okay. Also the languages part at the footer is necessary. Our aim is to create eventyay.com. Please check the behaviour on eventyay. username_3: @username_4 @username_0 **Current behavior:** In 350px width the footer does some random float of words which make a haste. ![issue 699](https://user-images.githubusercontent.com/29895469/30342134-45347196-9816-11e7-90ec-720d64307076.PNG) **Expected behavior:** terms & Condition should be not apart I would like to work on this . username_5: Hey i would like to work on this issue.Is this still present or resolved? username_6: completed via #1459 #1304 #800 Status: Issue closed