repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
PaddlePaddle/Paddle
854574494
Title: model.summary()后出现KeyError: 'output_shape'错误 Question: username_0: - 标题:自主搭建resnet模型报错 - 版本、环境信息:    1)PaddlePaddle版本:2.0.1    2)系统环境:Paddle version: 2.0.1 Paddle With CUDA: False OS: debian stretch/sid Python version: 3.7.4 CUDA version: None cuDNN version: None.None.None Nvidia driver version: None - 模型信息    1)模型名称 图一 ![image](https://user-images.githubusercontent.com/65533171/114192591-40629c00-9980-11eb-9ac7-6f17cc86299d.png) keras复现版本(图二) ![image](https://user-images.githubusercontent.com/65533171/114192794-7869df00-9980-11eb-8dc7-c5aaf7275194.png) 2)使用数据集名称 (matlab生成数字信号调制识别数据集_8class shape=(800000, 1, 2, 100)) https://aistudio.baidu.com/aistudio/datasetdetail/73835 3)模型代码 如下.zip文件 [net_resnet_stack.zip](https://github.com/PaddlePaddle/Paddle/files/6286343/net_resnet_stack.zip) - 问题描述:运行model.summary((None, 1, 2, 100))后出现如下报错信息 - 报错信息: ---------------------------------------------------------------------------AttributeError Traceback (most recent call last)/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/hapi/model_summary.py in hook(layer, input, output) 215 try: --> 216 summary[m_key]["output_shape"] = _get_output_shape(output) 217 except: /opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/hapi/model_summary.py in _get_output_shape(output) 193 else: --> 194 output_shape = list(output.shape) 195 return output_shape /opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/fluid/dygraph/layers.py in __getattr__(self, name) 1049 return _buffers[name] -> 1050 return object.__getattribute__(self, name) 1051 AttributeError: 'MaxPool2D' object has no attribute 'shape' During handling of the above exception, another exception occurred: KeyError Traceback (most recent call last)<ipython-input-6-6c680bc1eda0> in <module> 31 # model_resstack 32 model_resstack = paddle.Model(model_resstack) ---> 33 model_resstack.summary((None, 1, 2, 100)) /opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/hapi/model.py in summary(self, input_size, dtype) 1879 else: 1880 _input_size = self._inputs -> 1881 return summary(self.network, _input_size, dtype) 1882 1883 def _verify_spec(self, specs, shapes=None, dtypes=None, is_input=False): /opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/hapi/model_summary.py in summary(net, input_size, dtypes) 147 148 _input_size = _check_input(_input_size) --> 149 result, params_info = summary_string(net, _input_size, dtypes) 150 print(result) 151 </opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/decorator.py:decorator-gen-343> in summary_string(model, input_size, dtypes) /opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/fluid/dygraph/base.py in _decorate_function(func, *args, **kwargs) 313 def _decorate_function(func, *args, **kwargs): 314 with self: --> 315 return func(*args, **kwargs) [Truncated] 904 for forward_post_hook in self._forward_post_hooks.values(): /opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/fluid/dygraph/container.py in forward(self, input) 93 def forward(self, input): 94 for layer in self._sub_layers.values(): ---> 95 input = layer(input) 96 return input 97 /opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/fluid/dygraph/layers.py in __call__(self, *inputs, **kwargs) 903 904 for forward_post_hook in self._forward_post_hooks.values(): --> 905 hook_result = forward_post_hook(self, inputs, outputs) 906 if hook_result is not None: 907 outputs = hook_result /opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/hapi/model_summary.py in hook(layer, input, output) 217 except: 218 warnings.warn('Get layer {} output shape failed!') --> 219 summary[m_key]["output_shape"] 220 221 params = 0 KeyError: 'output_shape' Answers: username_1: @username_0 你好,方便把paddle版本调用summary的详细代码贴出来吗? username_0: ![image](https://user-images.githubusercontent.com/65533171/114265773-48c0e280-9a25-11eb-8e27-3a885b6f89b3.png) ![image](https://user-images.githubusercontent.com/65533171/114265783-51b1b400-9a25-11eb-8504-3afdc1085cfd.png) 您好,这样可以吗? username_1: @username_0 MaxPool2D返回的是一个可调用对象,并不是一个Tensor。 所以你代码中引用`MaxPool2D`可以改为: ```python if self.maxpool: self.maxpool = nn.MaxPool2D(kernel_size=(2,1), stride=(2,1), padding='VALID') out = self.maxpool(out) ``` 可参考MaxPool2D的官方示例。 https://www.paddlepaddle.org.cn/documentation/docs/zh/api/paddle/nn/layer/pooling/MaxPool2D_cn.html#daimashili
pointless-lang/pointless
724920264
Title: try catch results in crash Question: username_0: `try`-`catch` crashes for me. Here's a minimal failing example: ``` output= try throw SomeLabel {some= "thing"} catch is(SomeLabel) err => println("???") ``` The stack trace I get from this: ``` Unhandled exception: type '_GrowableList<ASTNode>' is not a subtype of type 'ASTNode' #0 dispatch (package:pointless/src/interpreter.dart:94) #1 eval (package:pointless/src/interpreter.dart:47) #2 evalCheck (package:pointless/src/interpreter.dart:63) #3 dispatch (package:pointless/src/interpreter.dart:294) #4 eval (package:pointless/src/interpreter.dart:47) #5 Env.addDefName.<anonymous closure> (package:pointless/src/env.dart:81) #6 Thunk.getValue (package:pointless/src/thunk.dart:33) #7 Env.lookupName (package:pointless/src/env.dart:121) #8 Env.getOutput.<anonymous closure> (package:pointless/src/env.dart:144) #9 _SyncIterator.moveNext (dart:core-patch/core_patch.dart:165) #10 runFlag (package:pointless/src/debug.dart:22) #11 runProgram (package:pointless/src/debug.dart:38) #12 main (package:pointless/pointless.dart:28) #13 _startIsolate.<anonymous closure> (dart:isolate-patch/isolate_patch.dart:299) #14 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:168) ``` It crashes both in the repl and in files. I don't know any dart, so I'm not sure what to do with this. :)
robotframework/RIDE
84454033
Title: Resolve paths to Library/Resource/Variables files when prefixed with ${EXECDIR} Question: username_0: Many test suites in our Robot project have settings like the following: Library ${EXECDIR}/execution/lib/DSVLibrary.py WITH NAME DSV Resource ${EXECDIR}/implementation/resources/WorkflowKeywords.txt If the suites are opened with the "Open Directory" button, we know ${EXECDIR} will be set to the directory you've opened. It would be great if RIDE could resolve these paths to enable keyword-source navigation. ## Why? This helps us avoid copypasta errors where Test suites at different parts of the directory tree Example: Feature/Test.robot Feature/Test2.robot Feature/Subfeature/Test.robot The above suites can all import the above libraries with the same lines of text. Answers: username_0: I am not sure what would be the best implementation for this. There are at least two options: 1) set ${EXECDIR} to ${CURDIR} 2) allow configuring ${EXECDIR} via preferences 2) is obviously more flexible. Opinions? username_0: I think it would be best to resolve ${EXECDIR} as the same location that pybot starts in when you press the Start/Run button. That way, pybot and RIDE would find resources the same way and reduce unpleasant surprise. This is similar to option #1, but I think it's slightly different when you're running a test suite with nested folders. I'll try a quick test of this later, but I think ${CURDIR} would be a bad substitute when test suites have nested folders like my example above. I *think* it would work out like this, assuming I use "File/Open Folder" on `/home/ed/dev/foo/tests`: Test file ${CURDIR} Desired ${EXECDIR} toplevel.robot /home/ed/dev/foo/tests/ /home/ed/dev/foo/tests/ sub/first.robot /home/ed/dev/foo/tests/sub/ /home/ed/dev/foo/tests/ sub/second.robot /home/ed/dev/foo/tests/sub/ /home/ed/dev/foo/tests/ sub/more/bottom.robot /home/ed/dev/foo/tests/sub/more/ /home/ed/dev/foo/tests/ As for Option #2: Currently ${EXECDIR} resolves to `ride.py`'s working directory, so that's already possible by starting ride.py in a command prompt or by editing the Working Directory on [copies of?] the "RIDE" Desktop/Start-menu shortcut. username_0: I agree on the first point. I think that would make the most obvious and least surprising implementation. And you are right that ${CURDIR} in general is not good substitution, what I actually meant was the ${CURDIR} of the root suite, which is exactly the directory opened with RIDE. username_0: Oh! That sounds perfect. I was thinking about how option #2 might work, and I wasn't able to figure out anything that made sense in a world with both: A. Multiple test suites on the same computer B. Test suites shared in source control, cloned or checked out in different places for each person Well, I suppose keeping per-suite settings indexed by location, but yuck? :) Status: Issue closed
google/grinder.dart
68690462
Title: Automatically add grinder to `dev_dependencies` on grinder:init Question: username_0: on ```bash pub global run grinder:init ``` Status: Issue closed Answers: username_1: Generally, it's intended to be used as `pub run grinder:init` (without global). That way you can specify which version of grinder you want for a given project, and initialize a grinder script which works with it. Den can help here though: ``` pub global activate stagehand pub global activate den stagehand package den install grinder --dev pub get pub run grinder:init ``` :)
MaskRay/ccls
561985904
Title: autocompletion with pointer correction goes wrong Question: username_0: ### Expected behavior When using ccls with coc.nvim in neovim, sometimes we may forget if a variable is a a pointer or not. I notice that ccls will show up auto-completions even if I use dot on a pointer type. But unfortunately, when insert -> after select a auto-completion, the previous entered dot is still there. It behaves as follows: ```c++ #include <vector> int main() { std::vector<int>* foo; foo.->push_back(0); // extra dot } ``` The expected behavior should be `foo->push_back()` without a dot. ### Steps to reproduce 1. type a dot after a pointer value 2. select a suggestion from auto-completion list 3. the dot you entered is still there ### System information * ccls version (`git describe --tags --long`): * clang version: 9.0.0 * OS: macOS Mojave * Editor: nvim * Language client (and version): coc.nvim Answers: username_0: A related issue in coc.nvim https://github.com/neoclide/coc.nvim/issues/1396 Status: Issue closed username_0: ### Expected behavior When using ccls with coc.nvim in neovim, sometimes we may forget if a variable is a a pointer or not. I notice that ccls will show up auto-completions even if I use dot on a pointer type. But unfortunately, when insert -> after select a auto-completion, the previous entered dot is still there. It behaves as follows: ```c++ #include <vector> int main() { std::vector<int>* foo; foo.->push_back(0); // extra dot } ``` The expected behavior should be `foo->push_back()` without a dot. ### Steps to reproduce 1. type a dot after a pointer value 2. select a suggestion from auto-completion list 3. the dot you entered is still there ### System information * ccls version (`git describe --tags --long`): * clang version: 9.0.0 * OS: macOS Mojave * Editor: nvim * Language client (and version): coc.nvim username_1: seem that it can be solve by set client.snippetSupport and install coc-snippets see https://github.com/MaskRay/ccls/wiki/Customization#clientsnippetsupport example configuration ```json "ccls": { "command": "ccls", "args": [ "--log-file", "/tmp/ccls.log" ], "filetypes": [ "c", "cpp", "objc", "objcpp" ], "rootPatterns": [ ".ccls-cache", ".vim/", ".git/", "compile_commands.json" ], "initializationOptions": { "cache": { "directory": "./.ccls-cache" }, "compilationDatabaseDirectory": "./build/Debug", "client":{ "snippetSupport":true } } }, ``` username_0: I've resolved it by writing a coc.nvim extensions. Close this issue for now. Status: Issue closed
isaac-rand/sister-teams-explorer
637321711
Title: Brief Error Message Question: username_0: Right when you open the [app](https://isaacrand.shinyapps.io/Sister-Teams-Explorer/) there is briefly a message which says that an error has occurred. It then goes away and everything works fine. It's not a huge problem, but it's a problem
rapid7/ruby_smb
268775800
Title: Improve SMB2 rename and delete logic Question: username_0: SMB2 rename and delete functionalities, implemented in the File class, are both using `SetInfoResponse` packet with `FileRenameInformation` and `FileDispositionInformation` FSCC File Information classes. However, the File Information classes are instantiated in the File methods and used to set the packet's attribute: - https://github.com/rapid7/ruby_smb/blob/master/lib/ruby_smb/smb2/file.rb#L223 - https://github.com/rapid7/ruby_smb/blob/master/lib/ruby_smb/smb2/file.rb#L149 This design can be improved by simply using the File Information code after instantiating `SetInfoResponse` and let the this class set the related attributes. As a reference, the SMB1 TRANS2 implementation is currently using this design to implement sub-commands (https://github.com/rapid7/ruby_smb/blob/master/lib/ruby_smb/smb1/packet/trans2/request.rb and https://github.com/rapid7/ruby_smb/blob/master/lib/ruby_smb/smb1/packet/trans2/subcommands.rb).<issue_closed> Status: Issue closed
alibaba/MNN
634087748
Title: Problem running using Vulkan and OpenCL Question: username_0: # 平台(如果交叉编译请再附上交叉编译目标平台): # Platform(Include target platform as well if cross-compiling): Linux, x86_64 # Github版本: # Github Version: 直接下载ZIP包请提供下载日期以及压缩包注释里的git版本(可通过``7z l zip包路径``命令并在输出信息中搜索``Comment`` 获得,形如``Comment = bc80b11110cd440aacdabbf59658d630527a7f2b``)。 git clone请提供 ``git commit`` 第一行的commit id Provide date (or better yet, git revision from the comment section of the zip. Obtainable using ``7z l PATH/TO/ZIP`` and search for ``Comment`` in the output) if downloading source as zip,otherwise provide the first commit id from the output of ``git commit`` f5dae040a0b60678726a5893b7b2778e15ce690c # 编译方式: # Compiling Method gcc ``` 请在这里粘贴cmake参数或使用的cmake脚本路径以及完整输出 Paste cmake arguments or path of the build script used here as well as the full log of the cmake proess here or pastebin ``` ``` option(MNN_USE_SYSTEM_LIB "For opencl and vulkan, use system lib or use dlopen" ON) option(MNN_BUILD_HARD "Build -mfloat-abi=hard or not" OFF) option(MNN_BUILD_SHARED_LIBS "MNN build shared or static lib" OFF) option(MNN_FORBID_MULTI_THREAD "Disable Multi Thread" OFF) option(MNN_OPENMP "Use OpenMP's thread pool implementation. Does not work on iOS or Mac OS" OFF) option(MNN_USE_THREAD_POOL "Use MNN's own thread pool implementation" ON) option(MNN_BUILD_TRAIN "Build MNN's training framework" OFF) option(MNN_BUILD_DEMO "Build demo/exec or not" OFF) option(MNN_BUILD_TOOLS "Build tools/cpp or not" OFF) option(MNN_BUILD_QUANTOOLS "Build Quantized Tools or not" OFF) option(MNN_EVALUATION "Build Evaluation Tools or not" OFF) option(MNN_BUILD_CONVERTER "Build Converter" OFF) option(MNN_SUPPORT_TFLITE_QUAN "Enable MNN's tflite quantized op" ON) option(MNN_DEBUG_MEMORY "MNN Debug Memory Access" OFF) option(MNN_DEBUG_TENSOR_SIZE "Enable Tensor Size" OFF) option(MNN_GPU_TRACE "Enable MNN Gpu Debug" OFF) option(MNN_PORTABLE_BUILD "Link the static version of third party libraries where possible to improve the portability of built executables" OFF) option(MNN_SEP_BUILD "Build MNN Backends and expression seperately. Only works with MNN_BUILD_SHARED_LIBS=ON" OFF) option(NATIVE_LIBRARY_OUTPUT "Native Library Path" OFF) option(NATIVE_INCLUDE_OUTPUT "Native Include Path" OFF) option(MNN_AAPL_FMWK "Build MNN.framework instead of traditional .a/.dylib" OFF) option(MNN_FMA_ENABLE "x86 routine use fma extension" OFF) option(MNN_WITH_PLUGIN "Build with plugin op support." OFF) $ cmake .. -- The C compiler identification is GNU 8.3.0 -- The CXX compiler identification is GNU 8.3.0 -- The ASM compiler identification is GNU -- Found assembler: /usr/bin/cc -- Check for working C compiler: /usr/bin/cc -- Check for working C compiler: /usr/bin/cc -- works -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Detecting C compile features -- Detecting C compile features - done -- Check for working CXX compiler: /usr/bin/c++ -- Check for working CXX compiler: /usr/bin/c++ -- works -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Detecting CXX compile features -- Detecting CXX compile features - done [Truncated] With openCL I don't get any output, with Vulkan I get something but incorrect. I tried the following combinations for backend: ``` MNN::ScheduleConfig schConfig; schConfig.type = MNN_FORWARD_VULKAN; schConfig.backupType = MNN_FORWARD_CPU; MNN::ScheduleConfig schConfig; schConfig.type = MNN_FORWARD_VULKAN; schConfig.backupType = MNN_FORWARD_VULKAN; MNN::ScheduleConfig schConfig; schConfig.type = MNN_FORWARD_OPENCL; schConfig.backupType = MNN_FORWARD_CPU; MNN::ScheduleConfig schConfig; schConfig.type = MNN_FORWARD_OPENCL; schConfig.backupType = MNN_FORWARD_OPENCL; ``` Status: Issue closed Answers: username_1: SliceTF Has been supported for vulkan and opencl
yegor256/netbout
130981100
Title: Exclamation mark icon on production after recent release Question: username_0: After releasing version 2.23 to production I see exclamation mark icon on my avatar: ![image](https://cloud.githubusercontent.com/assets/5467276/12778848/57d6150a-ca65-11e5-889a-80a67ef7c9ad.png) It shouldn't be there, as I have my email address set properly. Answers: username_1: @username_2 valid bug username_2: @username_1 I tagged this as "bug" username_2: @username_0 I set milestone here to `3.1`, let me know if it is wrong username_2: @username_0 thanks for reporting! I topped your account for 15 mins, transaction 56b329ea4f694f57700000c7 username_2: @username_3 do this task username_3: @username_0 @username_1 I just created a new user and I am not getting this issue It seems this is only related to already created users ? I do not know how I can test this locally. ![screenshot from 2016-02-15 16-14-33](https://cloud.githubusercontent.com/assets/14311743/13052457/3f708010-d3ff-11e5-81a0-1655e1d0c3c1.png) username_3: @username_0 Now I don't see this issue in my account to. It is gone now could you please close the issue. username_0: @username_3 I still see it in my profile username_3: Here is my account, but I have a very strange issue now #1015. ![screenshot from 2016-02-15 19-53-07](https://cloud.githubusercontent.com/assets/14311743/13057709/3ec36b54-d41e-11e5-90f8-c505c3653fb3.png) username_3: @username_1 I can not reproduce this with my account and cant test it ? username_1: @username_3 let's wait until fix for #1015 will be released username_3: @username_2 waiting for #1015 to be released username_2: @username_3 agreed, we'll wait for #1015 username_3: @username_1 I think it hase something to do with : ``` private static XeSource source(final Alias alias) throws IOException { final String email; final String newemail; final String[] emails = alias.email().split("!"); if (emails.length > 1) { email = emails[0]; newemail = emails[1]; } else { email = alias.email(); newemail = ""; } return new XeAppend( "alias", new XeDirectives( new Directives() .add("name").set(alias.name()).up() .add("locale").set(alias.locale().toString()).up() .add("photo").set(alias.photo().toString()).up() .add("email").set(email).up() .add("newEmail").set(newemail) ) ); } ``` in XeAlias, it does return newemail even when there is no newemail ? username_3: @username_0 Could you tell me what it says if you hover over it ? username_0: @username_3 it says that my email is not yet verified username_3: @username_0 does it specify the email because I want to know what the newEmail is returning, is it the same email you are now using or an old one ? username_0: @username_3 the same as I'm using now username_3: @username_0 Ok, I will see it is very interesting. username_2: @username_0 we were waiting for #1015 - it is closed already username_3: @username_2 Please assign someone else because I can not find the issue here. username_2: @username_3 30 points was deducted from your rating username_2: @username_3 all right, we'll find someone else for this task username_2: @username_4 this is your task username_4: @username_0 , I could reproduce this now. This appears to be an issue with the DynamoDB sync and is just a timing thing. For me this happens: * Change email, I get my verification link and experience #853 too * If I click the link I get something like "no verification needed" as a notice * Then I wait for say 15 minutes * New email appears instead of old one at last and I see the exclamation mark, the fact that I clicked the link before when I got the "not needed" notice did not validate my email. * But if I click it after I see the exclamation mark and the email address has changed and wait another 15 minutes, the icon disappears and all seems properly validated :) ... I'm on it fixing this, just in case you want to remove that exclamation mark ;), you should be able to remove it by simply visiting the validation link you got at some point, if you still have it that is ;) username_4: @username_0 PR #1066 added, reproducing and fixing(well by turning off the cache) the above explanation. username_0: @username_4 great job! username_4: @username_0 thanks :), I think we can close here right? Workaround merged, rest goes elsewhere I suppose. username_4: @username_0 ping :) Status: Issue closed username_0: @username_4 thanks! username_4: @username_2 please pay attention to this ticket as well as my other closed tickets in your project. I closed all of them in a very short period of time each, still I'm blocked by having them pile up as open tasks and receive an unrealistic speed rating by them staying open long after the PRs are closed and payment on them is handled. Thanks :) username_2: @username_5 please, review this task for compliance with our [quality rules](http://at.teamed.io/qa.html) username_5: @username_2 Looks good! username_2: @username_5 many thanks username_2: @username_4 paied 10 mins to @username_5 for QA review (payment ID is `79352562`); **38 mins** added to your account (payment number `AP-0CX14905429037629`), many thanks for your contribution! 66 hours and 1 min spent here.; there is a bonus for fast delivery (m=3961); +38 added to your rating, at the moment it is: [+730](http://www.netbout.com/b/36646?open=rating)
rometools/rome
484924321
Title: Packing Issue - Alternatives Rome.Utils / Firebase? Question: username_0: Hello. When i will create APK Signed in AndroidStudio i have a trouble. * im using this = implementation 'com.rometools:rome:1.12.1' * In debug mode its OK, not problem and RSS works fine * But when i create a signed Package i have a error: 1 º **Duplicate jar entry [com/rometools/utils/Alternatives.class]** 2 º **Caused by: java.lang.RuntimeException: java.io.IOException: Can't write [C:\Users\Lucas\Desktop\OnePlaceX\br.com.oneplace\build\intermediates\transforms\proguard\release\0.jar] (Can't read [C:\Users\Lucas\Desktop\OnePlaceX\br.com.oneplace\build\intermediates\transforms\FirebasePerformancePlugin\release\187(;;;;;;;**.class)] (Can't read [com] (Can't read [rometools] (Can't read [utils] (Can't read [Alternatives.class] (Duplicate jar entry [com/rometools/utils/Alternatives.class]))))))** 3º My proguard: **-dontwarn com.rometools.rome.** -keep class com.rometools.** { *; } -keep interface com.rometools.** { *; }** ![image](https://user-images.githubusercontent.com/28266902/63649769-56263d80-c718-11e9-81cd-1efb005f5a32.png) 4º Before use this version of rome 1.12.1, i was using this: **implementation files('libs/android-rome-feed-reader-1.0.0.jar') implementation files('libs/jdom-1.1.1-android-fork.jar')** The package build works fine..but my RSS not work properly like version 1.12.1. Somebody can help me with this problem? 3 days thinking and searching something and nothing. Other people have the same error, here the link [https://stackoverflow.com/questions/53854658/duplicate-jar-entry-after-upgrade-to-gradle-3-2-0-3-2-1](url) Answers: username_0: I make some tests here and the Signed Build APK Works with version 1.12.1 BUT 1 - ** I have to remove the firebase performance plugin** 2 - What i removed? **classpath** 'com.google.firebase:perf-plugin:1.3.1** **implementation "com.google.firebase:firebaseperf:${FIREBASE_PERFORMANCE_VERSION_SDK_VERSION}** 3º I think have some conflict, but i need use performance firebase plugin and ROME RSS 1.12.1 What can i do ? Thanks in advanced! username_1: Maybe Rome is not working on Android at all. I never tried to use it on Android but there are plenty of artickes on the web, that older versions of jdom are not supported on Android and that even the latest version of jdom has some issues on Android. username_0: So... but i dont understand why the error have some kind of problem with firebase perf plugin . Now or i use rome or firebase performance.. What you suggest to do? username_1: Sorry, I can't help. I am not an Android developer. I recommend you to ask your question on StackOverflow. Status: Issue closed
kubernetes-sigs/cluster-api-provider-azure
609362212
Title: LoadBalancer service is not accessible Question: username_0: /kind bug **What steps did you take and what happened:** Create a LoadBalancer service, public IP is allocated, but the service is not accessible through public IP. **Anything else you would like to add:** kube-controller-manager is adding correct inbound security rules to node-nsg with destination to be the allocated public IP. But capz securitygroup reconcile reset the rules to default and remove those rules. **Environment:** - cluster-api-provider-azure version: v0.4.2 - Kubernetes version: (use `kubectl version`): - OS (e.g. from `/etc/os-release`): Answers: username_0: /assign username_1: @username_0 I discussed with @username_3, we should do something similar to what the Azure cloud provider is doing: when reconciling the NSG, first get the existing NSG if it is already created and get the existing rules. Then, only add the rules we need for the CAPZ cluster to be functional if they do not already exist but preserve all the additional existing rules. We should also use ETAGs to make sure we do not update the NSG while something else (eg. Azure Cloud Provider) is updating it. If we don't, that could cause a race condition where we fetch the existing rules [a, b, c] and in the meantime something else deletes b but CAPZ updates the rules to [a, b, c, d] and thus we undo the rule deletion. This is where Azure Load Balancer updates the NSG and checks the ETAG to make sure its version of the NSG is still current: https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/legacy-cloud-providers/azure/azure_backoff.go#L149-L175 And this is where Azure Load Balancer determines what rules to apply: https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/legacy-cloud-providers/azure/azure_loadbalancer.go#L1112-L1307 Let me know if any of this doesn't make sense. username_0: Thanks @username_1 for sharing this. Make sense to me. username_1: /priority important-soon username_0: /assign username_2: I went through the changes required for this issue. As suggested, azure-cloud-provider uses its own client https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/legacy-cloud-providers/azure/clients/armclient/azure_armclient.go#L44 to communicate with Azure resources which has the provision to specify ETAG Headers: ``` WithHeader("If-Match", autorest.String(etag)) ``` azure-cluster-api-provider uses client from `azure-sdk-for-go` which does not have this provision. To add support for Headers would hardly be a 3-line change but in https://github.com/Azure/azure-sdk-for-go repository So to have ETAG provision there are three options: 1. Import armClient from legacy-cloud-provider(Is this import allowed. It’s not used anywhere in the code) 2. Make changes to azure-sdk-for-go client 3. Create a new client for the this repo(Too much of work to do. Feasibility?) @username_1 Can you please suggest which option would be better? Or is there any other way to proceed. username_0: /unassign username_0: /assign username_2 username_3: @username_2 The older versions (< 1.18) of azure-cloud-provider uses the `azure-sdk-for-go`. We have etag checks in those releases for security groups - https://github.com/kubernetes/kubernetes/blob/release-1.16/staging/src/k8s.io/legacy-cloud-providers/azure/azure_backoff.go#L184-L215. We should be able to do the same here. username_2: Thanks @username_3 I got your point
iamgabrielma/Python-for-stock-market-analysis
930215319
Title: Add validation method so tickers in tickerList are unique Question: username_0: At the moment `tickerList` may have duplicated tickers, which causes the [JSON output](https://raw.githubusercontent.com/username_0/Python-for-stock-market-analysis/main/testData/2021-06-25-rsi.json) to mess with the `id`'s as well. **Problems:** 1. Items do not have an unique - consecutive - ID 2. When we fetch the data via the iOS app, the does not load as the SwiftUI `Identifiable` protocol requires unique ID's
kaerosen/tilemaps
670563529
Title: Extra features Question: username_0: Great to see this algorithm implemented in R. I worked on the tile map paper. After developing the original algorithm, we considered a couple of extra features: 1) Repel non-neighbors 2) Allow using the concave hull of the centroids rather than the original boundary These were not added in time for the paper, but are included in the demos: [https://observablehq.com/@username_0/make-a-tile-map](https://observablehq.com/@username_0/make-a-tile-map) [https://username_0.github.io/tile-maps-eurovis-demo/](https://username_0.github.io/tile-maps-eurovis-demo/) I don't think these features are included in this package? If not, it might be nice to add them at some point? Answers: username_1: Thanks for letting me know about these extra features! They are currently not included in this package, but I will definitely look into adding them in the future. username_2: using this as R code: install.packages(c("tilemaps", "sf")) library(tilemaps) library(sf) library(dplyr) library(ggplot2) governors <- governors %>% mutate(tile_map = generate_map(geometry, square = FALSE, flat_topped = TRUE)) head(governors) ggplot(governors) + geom_sf(aes(geometry = tile_map)) + geom_sf_text(aes(geometry = tile_map, label = abbreviation), fun.geometry = function(x) st_centroid(x)) + theme_void() I am trying to build a Canadian Tile map. I got the geometry (lpr_000b16a_e.shp found https://open.canada.ca/data/en/dataset/47bd4f2e-1c77-49f8-8406-dc4dca64ee6b) from statistics Canada. when I am using the R code above, I am getting this error: Error: Problem with `mutate()` input `tile_map`. x regions are not contiguous i Input `tile_map` is `generate_map(geometry, square = FALSE, flat_topped = TRUE) How can I create a tile map of the Canada provinces? username_1: Hi, I have opened a new issue (#4) to address this question. Status: Issue closed
altugcagri/boun-swe-573
449286891
Title: Make Changes On Frontend Api Question: username_0: **Is your feature request related to a problem? Please describe.** Change the related API calls according to back-end URL's. Make sure all changes done properly. Test front-end application with back-end of yours.
eclipsesource/jsonforms
283317050
Title: npm run build doesn't work on Windows because ../ (forward slash) is not recognized Question: username_0: (node:10276) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code. Answers: username_1: Found multiple things: NPM closed this issue by a bot: https://github.com/npm/npm/issues/13789 Yarn has a solution: https://github.com/yarnpkg/yarn/issues/1729 SO suggest to use symlinks https://stackoverflow.com/questions/28078780/relative-paths-in-package-json @edgarmueller any ideas? username_2: i just ran into the same issue while trying to run build Status: Issue closed
INFURA/hackathons
547153774
Title: Infura API Optimization Prize (500 Dai) Question: username_0: ### Infura Ethereum API Optimization Prize ### Prize Bounty 500 Dai ### Challenge Description Each Ethereum network call costs $ and can be vulnerable to network clogging. We'd like to see you think about how you can lessen either the number or the gas cost of eth_calls that your dapp makes, and showcase improvements. This is a wide-ranging topic and we invite you to be creative. Additionally, you can use contracts like makerdao's Multicall (https://github.com/makerdao/multicall) either directly or for inspiration. ### Submission Requirements A valid submission will include a short project video demo as well as a text description of how the number/complexity/gas cost of eth_calls is reduced, including a before and after comparison. ### Submission Deadline January 23, 2020 ### Judging Criteria The project demo and code will be evaluated by Infura team members, and the one best optimizing eth_calls will be awarded the prize. Judging includes criteria based on execution, project goal, originality, and code quality. ### Winner Announcement Date The winner will be announced via our Twitter and/or other channels within 1 week of the hackathon ending. Answers: username_1: Had a need to look ether and token balances of several wallets and then seen this bounty so decided to make something useful for others as well. As ethcall uses multicall, number of calls made effectively goes down to 1. Status: Issue closed
PhelanBavaria/ancienttimeline
202389008
Title: Basic ideas Question: username_0: So, having looked through the ideas and playing the game for a bit, I think we should redo the basic ideas. Most don't really fit with the time, and while they could be renamed and such as some have, it'd probably be easier and better to just redo them. In addition, we could look at policies, though I'm not sure what to do for them. ET uses different ideasets for early eras, so I suppose we could do something like that and maybe have some more dynamic ones as well (available to specific cultures, religions, locations, techs, times &c.) What do you think? Answers: username_1: Sounds good username_2: I started I guess. I already changed Humanist Ideas to Altruistic ideas. username_0: So, these are some of the ones I've got so far. Haven't got all of them done yet, and the numbers aren't final. [00_basic_ideas.txt](https://github.com/PhelanBavaria/ancienttimeline/files/785848/00_basic_ideas.txt) username_0: Right, I've finished a few idea groups. I'll localise them in a bit and I'll add some more. Do we want to ad the 'early ideas' from ET and/or do we want to create an equivalent 'really early' group of ideasets for the beginning of AT?
vimalloc/flask-jwt-extended
396246096
Title: Allow JWT_DECODE_AUDIENCE to be an array Question: username_0: Hi. I have spent much time with your library and I use it a lot including for verification of outside OIDC issued tokens. I would like to propose JWT_DECODE_AUDIENCE to be an array so token verify would allow verification from multiple audiences. Scenarios. Identity Clients accessing API: - CLI tool could be one consumer (direct grant) - Web client could be second consumer (implicit flow) - Web developer running on localhost could be third consumer (client has different redirects) All issued by same provider, all valid. Only difference is each client has a different AUDience. It seems an a common practice to allow multiple audiences to be allowed to be verified in api. https://github.com/auth0/node-jsonwebtoken/issues/4 So if I could do JWT_DECODE_AUDIENCE = ['ai.mysoft.web','ai.mysoft.cli','ai.mysoft.localhost'] that would be great. Does that make sense? Answers: username_0: FROM: https://openid.net/specs/openid-connect-core-1_0.html#IDToken ``` aud REQUIRED. Audience(s) that this ID Token is intended for. It MUST contain the OAuth 2.0 client_id of the Relying Party as an audience value. It MAY also contain identifiers for other audiences. In the general case, the aud value is an array of case sensitive strings. In the common special case when there is one audience, the aud value MAY be a single case sensitive string. ``` If you allow, I'd like to contribute oidc use example including autodiscovery and assembly of RSA public key from JWK endpoint data. username_1: Yeah, `aud` should absolutely allow arrays, it was an oversight that wasn't caught when adding that feature. PyJWT already supports `aud` being an array, so all we need to do for this extension is update the config helper file, add the documentation, and write a unit test. Would you be willing to submit a PR to add these features? If not, I can add them when I have a free moment, probably this weekend. Thanks! Status: Issue closed username_1: So it looks like an array can already be passed in `JWT_DECODE_AUDIENCE`, whatever is passed in there is forwarded directly to PyJWT behind the scenes. I've updated the documentation to make this clearer, and added a unit test to make sure that this functionality is never accidentally removed in future versions of this library. Cheers :+1: username_0: @username_1 as promised https://github.com/username_1/flask-jwt-extended/pull/222
Lusito/forget-me-not
585534448
Title: Cookies not cleared at browser startup Question: username_0: Cookies are not cleared at browser startup which means: after browser startup, I am still logged in into Facebook. Answers: username_1: Hello and thanks for the report. Since this is still working correctly here, It is very likely, that you have configured it incorrectly. You need to enable cleanup on startup in the settings tab of the add-on. Also your rules need to be configured to remove facebook data. Please check out the tutorial included in the add-on. Status: Issue closed
Kanazawanaoaki/Style-Transfer-colab
593030608
Title: TODO Question: username_0: - READMEの画像をいい感じにする。 [こんな感じ](https://github.com/albumentations-team/albumentations/blob/master/README.md)に横並びにした画像を作って貼り付けてしまう。 - 改善する。 もっと綺麗にできるように[他の実装](https://github.com/keras-team/keras/blob/master/examples/neural_style_transfer.py)とかも見て改善する。 - 他の手法も試してみる。 [ここ](https://elix-tech.github.io/ja/2016/08/22/art.html#improve)に載っている改善とか [この記事](https://ai-scholar.tech/articles/treatise/style-ai-224)で紹介されている手法
kiali/kiali
669725322
Title: Create a simple kiali helm installer Question: username_0: Collecting feedback from the community, it seems that for some basic scenarios people don't want to use the kiali-operator to install Kiali. This issue is open to study the possibility to write a very simple way to have an instance of Kiali installed in Istio without using the operator. Note, this is not in any way, any change in the strategy to use the operator, that is designed to manage multiple and complex scenarios. Finally, the simple kiali helm installer should have a very limited demo scope, it shouldn't be a recommended way to install Kiali in production scenarios where operator should be the way to go. Answers: username_1: I agree we could produce and publish a basic "demo" kiali server helm chart. There are some things people will lose if they opt-out of the operator: 1. You lose the ability for multi-tenancy support like Kiali has for Maistra 2. You lose the ability to auto-detect and correct unauthorized changes to the Kiali resources (e.g. if someone deletes the Kiali Deployment or edits the Kiali ConfigMap, the operator is no longer there to auto-revert - it is on you to detect those changes and manually revert) 3. You lose the ability to change the accessible namespaces on the fly (the operator today can force Kiali to have cluster roles given to it only for the namespaces in the mesh - thus added a level of security). The basic helm chart will have accessible_namespaces=** thus allowing Kiali to see everything. That's how it is with the upstream Istio release, so this is no different than how people are used to using it, but I bring it up anyway. 4. You lose the OpenShift Console Links (operator creates those on the fly as namespaces are added to the accessible namespaces list). There are probably other things I'm forgetting, those are just off the top of my head. Status: Issue closed
MicrosoftDocs/office-docs-powershell
643643953
Title: EnableAuthAdminReadSession is missing Question: username_0: [KB4559435](https://support.microsoft.com/en-us/help/4559435/) describes that EnableAuthAdminReadSession has been added and links for more info to the [Set-OrganizationConfig](https://docs.microsoft.com/en-us/powershell/module/exchange/set-organizationconfig) page. However, this parameter isn't discussed on this page. A bit more clarification regarding this parameter would be nice as well (also on the KB article page), like for instance for what scenarios you’d want this and what the default is. Thanks! Robert --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 03895b30-2bb7-9e96-28db-def76ae59029 * Version Independent ID: 8056b3ee-4fab-3a5c-935b-a9d0e9ad82a4 * Content: [Set-OrganizationConfig (exchange)](https://docs.microsoft.com/en-us/powershell/module/exchange/set-organizationconfig?view=exchange-ps) * Content Source: [exchange/exchange-ps/exchange/Set-OrganizationConfig.md](https://github.com/MicrosoftDocs/office-docs-powershell/blob/master/exchange/exchange-ps/exchange/Set-OrganizationConfig.md) * Service: **exchange-powershell** * GitHub Login: @username_2 * Microsoft Alias: **username_2** Answers: username_1: Get-RetentionCompliancePolicy username_1: @officedocsbot assign @username_1 Status: Issue closed
beefoo/climate-lab
250097027
Title: Research into what are some warming/cooling events/regions Question: username_0: By looking at the global and regional temperature records, see if we can identify specific climate events that we can call out, e.g.: - Cooler-than-usual or warmer-than-usual set of years - Various regional heat waves - Specific historical markers of significance (e.g. various periods industrialization like WW2) Answers: username_0: Dip between 1940 to 1970 can be explained by aerosols emitted by industrial activities and volcanic eruptions: https://www.newscientist.com/article/dn11639-climate-myths-the-cooling-after-1940-shows-co2-does-not-cause-warming/ username_0: The Mt. Agung eruption in March 1963 produced a decrease of about 0.5°C in the mean temperature of the tropical troposphere http://science.sciencemag.org/content/194/4272/1413 username_0: [Clean air acts](https://en.wikipedia.org/wiki/Clean_Air_Act) reduced aerosol emissions, thus warming effects of GHGs slowly outweighed cooling effects of aerosols username_0: Three largest volcanos of the 20th century seems to align with short-term cooling [Santa María Volcano in 1902](https://en.wikipedia.org/wiki/Santa_Mar%C3%ADa_(volcano)) [Mt Novarupta eruption of 1912](https://en.wikipedia.org/wiki/Novarupta) [Mt Pinatubo of 1991](https://en.wikipedia.org/wiki/Mount_Pinatubo) username_0: 1870 is the approximate beginning of the [2nd Industrial Revolution](https://en.wikipedia.org/wiki/Second_Industrial_Revolution) username_0: [Strong El Nino contributed to 1940-1942 warming](http://onlinelibrary.wiley.com/store/10.1256/wea.248.04/asset/2005601203_ftp.pdf?v=1&t=j8ddjopz&s=9662788abf1afd959761bd4dc9c3e04593935fcb)
RohitSurwase/UCE-Handler
346821167
Title: The lib doesn't work on API 26? Question: username_0: it works fine on my Samsung s6 (API 24), but it doesn't work on my other Samsung s7 (API 26). Do you have any idea? Answers: username_1: @username_0 Sorry for the delay (I have not got notified). Could you please provide more details about the issue? Error logs would be fine. Status: Issue closed
material-components/material-components-web
275110544
Title: Remove events from persistent drawer Question: username_0: I want to add some extra features into mdc-persistent-drawer however there is events like pointerup, pointerdown, click, pointermove. These are completely useless and I cant find way to remove them. Tried to unbind, .off, override but no luck. Answers: username_1: Yep, I have to remove click listener also. I guess Pointer/Scroll events disabling added in order to stop main page content from scrolling. Have no idea, why disabling clicks are required. ``` var persistentDrawer = new drawer.MDCPersistentDrawer(...) var drawer = persistentDrawer._root_.querySelector('nav') drawer_.removeEventListener('click', persistentDrawer.foundation_.drawerClickHandler_) ``` username_2: @username_0 why do you have to remove these event handlers? username_0: I want to customize drawer with various elements and for example if I want to add fancier list then elemenets in list wont be clickable (gray background on mousedown...) Status: Issue closed username_3: We will be removing the `drawerClickHandler_` which currently stops propagation of the click event. This is being done in #1138, similar discussion on #1004.
Th3Shadowbroker/OuroborosMines
714021381
Title: Item duration isn't decreased when mining Question: username_0: **Describe the bug** Originally described by @laspi94 as `mining on the region whith the flag, item no loss durability.` **To Reproduce** Steps to reproduce the behavior: 1. Mine within a region. 2. Mine. 3. Item durability isn't decreased. **Expected behavior** The mining tool is supposed to lose durability.<issue_closed> Status: Issue closed
AppImage/AppImageKit
383988025
Title: Portable Apps and their settings: Portable Home Question: username_0: Hi, lead developer of [ORB Applications](https://www.orbital-apps.com/) here. We have developed a "standard" called **[Portable Home](https://github.com/username_0/orb-specification/issues/2)**, designed for the portability of data used by [Portable Apps](https://www.orbital-apps.com/portable). I was wondering it there are plans to have "Portable AppImages" (Portable Apps)? Is there interest in adopting something like this "standard"? Nothing is set in stone regarding this standard, so it can be changed, if necessary. Feel free to comment [on the Github page](https://github.com/username_0/orb-specification/issues/2) or here. Answers: username_1: @username_2 we think that whether data should travel alongside the AppImage or not should be determined by the user, not the author of an AppImage. username_1: @username_0 do you think all functionality is now in place and this ticket can be closed? Please see https://docs.appimage.org/user-guide/portable-mode.html Status: Issue closed username_2: @username_1 I think that's a bad idea. This works fine for rather advanced linux users, but the average user expects to be able to download a "portable edition" without possibly screwing up the portability with manually creating directories. I think like this, the functionality is simply useless for a larger part of the userbase... At the very least there would need to be a graphical dialog at launch to choose the mode. But that again would need a possible way *inside* the application to prompt for this username_1: You could put into your AppImage a custom AppRun script that would call ``` ... mkdir -p "${APPIMAGE}.home" export HOME=${APPIMAGE}.home .... ``` I think. The downside is that the user can no longer decide. username_1: @username_2 do you have a concrete application in mind? username_2: Sure, this is e.g. a common problem with IDEs: it is often useful to ship programming language-specific flavors (special set of addons & settings), but if the user already has a system-wide install all the settings might be messed up and interfering, so it is vital to be able to offer such a flavored IDE as a portable application. IDEs are also the kind of app where users will be very not amused if the settings are botched up, since they're usually highly configured to personal needs. Apart from that, any creative application which the user might want to use on the run and keep everything with them, and they would possibly end up disappointed if they discovered on the run they messed up making it portable and half the settings are gone. It's just better to be able to ship a clearly separate, clearly portable build that is always portable no matter what, so that such room for error is eliminated username_2: Is this something that is done inside the app image, or outside? Because if it can't be done inside without external scripts, it doesn't really help. Also, there is always the use case for portable IDEs, portable audio editing apps, ... what have you, and for those the use case of "just give me a portable version I can't accidentally make not portable" always remains. I was just trying to give more specific examples since you asked for it, but I don't think those are super relevant for the general use case of an always-portable-no-matter-what version username_1: Inside. Point is, you don't really need a portable "version", since every (recent) AppImage can be used in a "portable" way. username_2: You're not wrong, but depending on the target audience it's still useful to have an explicit portable version - but you made it abundantly clear that is possible to do, so that's good enough for me :+1: sorry for the comment spam username_1: Thanks @username_2 username_3: Say I have a folder next to my appimage called program.appimage.home.config Will it store both config and the home directory files there, neither or only one? Or do I simply not need it? username_4: If your AppImage is called `a.AppImage`, you can create the directories `a.AppImage.home` and/or `a.AppImage.config`. The former should include configuration files. You should probably test this, though, as these are experimental features. username_3: So technically only .home should be necessary? username_4: It highly depends on the application. In theory, as long as the application references everything from `$HOME`, I'd say yes. But you should really test this. username_3: Ok, so far, no appimage has misbehaved without the .config folder (as in, none has written outside of the .home folder unless specified). Thank you. username_4: There is the [XDG base dir spec](https://specifications.freedesktop.org/basedir-spec/basedir-spec-latest.html), which most frameworks support. But you cannot predict the environment variables on a system, and there are implementation specific details. Therefore, these directories are considered experimental, as they do not attempt to make guarantees but are a best effort solution.
ant-design/ant-design
347853922
Title: [collapse] add extra field like that in card component Question: username_0: - [x] I have searched the [issues](https://github.com/ant-design/ant-design/issues) of this repository and believe that this is not a duplicate. ### What problem does this feature solve? It would be a good place to put actions which may manipulate a collapse panel or its content. Though it can be implemented by creating a `float: right` div in `header`, I know, it's not a clean solution. ### What does the proposed API look like? Just like `card` component. ```jsx const extra = 'string' || () => <p>string</p>; return ( <Panel header="This is panel header 1" key="1" extraField={extra}> <p>{text}</p> </Panel> ) ``` <!-- generated by ant-design-issue-helper. DO NOT REMOVE --> A screenshot may be helpful :D <img width="783" alt="2018-08-06 5 38 06" src="https://user-images.githubusercontent.com/12122021/43709728-3159241e-99a0-11e8-8dd6-302dfb49cf27.png"><issue_closed> Status: Issue closed
MelvorIdle/melvoridle.github.io
1088841239
Title: [Bug]: <mobile sometimes does little to no actions with the app closed> Question: username_0: ### Describe the bug There’s been a few times where I start a task before going to sleep and I will gain nothing over night afk ### Reproduction Steps Started thieving before closing app and woke up with about 20 minutes of actions done ### Expected behaviour _No response_ ### Screenshots _No response_ ### Console output _No response_ ### Which platforms are you experiencing this bug on? - [ ] Chrome - [ ] Firefox - [ ] Edge - [ ] Safari - [X] Mobile App (iOS) - [ ] Mobile App (Android) - [ ] Steam - [ ] Other (Please Specify) ### Which version of the game are you experiencing this bug on? v1.0 ### Game Subversion ?1419 ### Are you using any scripts or extensions? No Answers: username_1: Many fixes for various offline progress issues have been released over the last few days. Please ensure you are running at least v1.0.3 (?1843). Please open a new issue if the problem persists. Status: Issue closed
DNNAssociation/DnnSummit2019.Mobile
391428548
Title: Update Location Model, ViewModel and View to pull data from the website Question: username_0: ## Description of problem Currently, the Location data is stubbed in code. This needs to be changed to pull data from the website API. ## Description of solution Update the Location Model, ViewModel and View to pull data from the website. ## Description of alternatives considered n/a ## Additional context n/a ## Screenshots n/;a ## Affected version * [x] 1.0.0 ## Affected device * [x] Android * [x] iOS<issue_closed> Status: Issue closed
MarkRedeman/ProfessorFrancken
146269294
Title: Choose an automated code review service Question: username_0: Currently I've enabled [nitpick-ci](https://nitpick-ci.com) and [styleci](https://styleci.io/). Styleci requires some additional configuration and nitpick ci can't review our tests files because they are using snake cased function names.<issue_closed> Status: Issue closed
moment/moment-timezone
151998505
Title: Bug when parsing unix timestamp and using default time zone Question: username_0: As reported in moment/moment#3158 there is an issue where parsing an 'X' token will not work correctly with a default timezone. ```js moment.tz.setDefault("America/Chicago"); moment("1461906597", "X").isValid(); // false ``` This is an issue with the way updateOffset handles default time zones. ```js if (mom._z === undefined) { if (zone && needsOffset(mom) && !mom._isUTC) { mom._d = moment.utc(mom._a)._d; mom.utc().add(zone.parse(mom), 'minutes'); } mom._z = zone; } ``` This bit of code will attempt to take the initial _a array from the moment config, and use it to construct a new UTC moment. With a timestamp, _a is an array with a bunch of undefined values, ergo this doesn't work. Answers: username_1: I added a unit test for this to my fork of the repo: https://github.com/username_1/moment-timezone/commit/dde70303f5e148362090bb508918cb40a46d8bca. username_2: Sorry, is there a solution for this issue at the moment (besides of the unit test added)? Or, is there any other way to apply kind of "default" timezone to all subsequent moment().format() calls? username_3: The workaround for now is to pass the timestamp as a number instead of a string. ```js moment.unix(1461906597) // seconds moment(1461906597123) // milliseconds ``` If you have a string, then use `+` or `parseInt` to convert to number. username_2: Thank you! Will it apply my default timezone, or just help to avoid errors? username_3: It's a better form to use anyway, and yes - it will work with the default time zone. The bug is only that the default timezone isn't working with `x` or `X` string formats.
FB-18-19-PreAP-CS/math-helper-mosgood549
363273243
Title: Don't use sleep Question: username_0: I would just think it would be nicer if either the selection section shows up right away after using a formula, or an output shows up saying something like "would you like to use another formula? (y)es or (n)o" i just thought that might look better. Answers: username_1: I use sleep to make it easier to see the math helper homepage, and it helps the user see the answer without having to scroll through the shell. I think that I'm going to keep it there, because I like the way that it organizes my code. Status: Issue closed
ExchangeCore/Concrete5-CKEditor
112214719
Title: Setup appropriate default plugins Question: username_0: I just kind of tossed a few in there as an example. We should come up with a reasonable list to attempt to mimic the functionality that redactor provides by default. See https://github.com/ExchangeCore/Concrete5-CKEditor/blob/master/controller.php#L169-L183 Answers: username_1: I whittled down what I think is a good default set of plugins. ![default_plugins](https://cloud.githubusercontent.com/assets/10898145/10602661/99dacb84-76e7-11e5-9aec-028165dc7f63.png) ```php protected function setupDefaultPlugins() { $this->getConfig()->save('plugins', array( 'a11yhelp', 'basicstyles', 'blockquote', 'clipboard', 'colorbutton', 'colordialog', 'contextmenu', 'dialogadvtab', 'elementspath', 'enterkey', 'entities', 'floatingspace', 'font', 'format', 'htmlwriter', 'image', 'indentblock', 'indentlist', 'justify', 'link', 'list', 'liststyle', 'magicline', 'removeformat', 'resize', 'showblocks', 'showborders', 'sourcearea', 'specialchar', 'stylescombo', 'tab', 'table', 'tabletools', 'toolbar', 'undo', 'wysiwygarea' ) ); } ``` username_0: This is a pretty solid list. Do you have any particular love for the blockquote / specialchar functionality? At a glance I'm not sure I see blockquote being used much by my basic users, ditto for the symbols stuff. With it being so easy to turn on I feel more inclined to strip this down as far as possible and let people build it back up to their liking. I don't feel very strongly about this either way but anything we can do to dwindle down options to eliminate confusion seems ideal to me. username_1: You are right, those two aren't essential. Last night I tried to remove the copy, paste, and cut, but didn't realize it was "Clipboard". ![default_plugins](https://cloud.githubusercontent.com/assets/10898145/10622455/b497fdd4-7755-11e5-94e8-2760af2a1dd5.png) ```php protected function setupDefaultPlugins() { $this->getConfig()->save('plugins', array( 'a11yhelp', 'basicstyles', 'colorbutton', 'colordialog', 'contextmenu', 'dialogadvtab', 'elementspath', 'enterkey', 'entities', 'find', 'floatingspace', 'font', 'format', 'htmlwriter', 'image', 'indentblock', 'indentlist', 'justify', 'link', 'list', 'liststyle', 'magicline', 'removeformat', 'resize', 'showblocks', 'showborders', 'sourcearea', 'stylescombo', 'tab', 'table', 'tabletools', 'toolbar', 'undo', 'wysiwygarea' ) ); } ``` username_0: Ok i'm closing this per https://github.com/ExchangeCore/Concrete5-CKEditor/commit/58c1621792a95073371681ed62aeac2ae943097b we can always revisit later if necessary Status: Issue closed
syndesisio/syndesis
319593869
Title: Salesforce connection keep showing configuration issue message Question: username_0: ## This is a... <!-- Check ONLY one of the following options with "x" --> <pre><code> [ ] Feature request [x ] Regression (a behavior that used to work and stopped working in a new release) [x ] Bug report <!-- Please search GitHub for a similar issue or PR before submitting --> [ ] Documentation issue or request </code></pre> ## The problem <!-- Briefly describe the issue you are experiencing (or the feature you want to see implemented on Syndesis). + For BUGS, tell us what you were trying to do and what happened instead. + For NEW FEATURES, describe the _User Persona_ demanding it and its use case. --> A valid salesforce connector configuration was set up and when the connection was created the following error message appear: ![sfdc1](https://user-images.githubusercontent.com/8625482/39531386-9765338a-4e5d-11e8-8be6-d1966eddfa8a.png) ![sfdc2](https://user-images.githubusercontent.com/8625482/39531388-9797be72-4e5d-11e8-8da5-bf4a518f7c07.png) ## Expected behavior <!-- Describe what the desired behavior would be, enlistin gthe acceptance criteria. --> There should not be a warning, as the configuration is correct.
web3j/web3j
309266755
Title: Unable to make transaction using admin.personalSendTransaction method. Question: username_0: I am using geth and I have used personalNewAccount to create an account. To make the transaction I wrote this. `HttpService httpService = new HttpService(envConfiguration.getTestNetServerUrl()); Admin admin = Admin.build(httpService); Transaction trx = Transaction.createEtherTransaction(from, nonce, gasprice, gaslimit, to, amount); EthSendTransaction ethSendTransaction = admin.personalSendTransaction(trx, existingpassword).send(); transactionHash = ethSendTransaction.getTransactionHash();` I'm getting output transactionHash = null Status: Issue closed Answers: username_0: There was some issue in sending amount value. Thanks.
postmanlabs/postman-docs
802386435
Title: image not showing on Variables page Question: username_0: https://learning.postman.com/docs/sending-requests/variables/#sharing-and-persisting-data <img width="945" alt="Screen Shot 2021-02-05 at 10 34 14 AM" src="https://user-images.githubusercontent.com/36343528/107074629-da9a3e00-679d-11eb-8930-094b7cf3d1a7.png"><issue_closed> Status: Issue closed
fishcakez/dbg
125799179
Title: Dbg crashes when tracing calls Question: username_0: Attempted to follow instructions in README.md, but getting erorr with verison 1.0.1, Erlang 18, Elixir 1.2.0. Ran iex -S mix in dbg directory after mix compile then followed steps in README.m and got the following: Erlang/OTP 18 [erts-7.2.1] [source] [64-bit] [smp:8:8] [async-threads:10] [hipe] [kernel-poll:false] [dtrace] Interactive Elixir (1.2.0) - press Ctrl+C to exit (type h() ENTER for help) iex(1)> Dbg.trace(self(), :call) %{counts: %{nonode@nohost: 1}, errors: %{}} iex(2)> Dbg.call(&Map.new/0) %{counts: %{nonode@nohost: 1}, errors: %{}} iex(3)> Map.new() %{} iex(4)> ** dbg got EXIT - terminating: {trace_handler_crashed, {function_clause, [{'Elixir.IEx.Config',default_option, [width], [{file,"lib/iex/config.ex"},{line,63}]}, {'Elixir.IEx.Config', '-default_config/0-fun-0-',1, [{file,"lib/iex/config.ex"},{line,60}]}, {'Elixir.Enum','-map/2-lists^map/1-0-',2, [{file,"lib/enum.ex"},{line,1088}]}, {'Elixir.Enum','-map/2-lists^map/1-0-',2, [{file,"lib/enum.ex"},{line,1088}]}, {'Elixir.IEx.Config',configuration,0, [{file,"lib/iex/config.ex"},{line,56}]}, {'Elixir.Dbg.Handler',handle_event,2, [{file,"lib/dbg/handler.ex"},{line,9}]}, {dbg,handle_traces,4, [{file,"dbg.erl"},{line,819}]}, {dbg,tracer_loop,2, [{file,"dbg.erl"},{line,783}]}]}} =ERROR REPORT==== 9-Jan-2016::23:05:46 === ** Generic server 'Elixir.Dbg.Watcher' terminating ** Last message in was {'DOWN',#Ref<0.0.4.689>,process,<0.121.0>, {trace_handler_crashed, {function_clause, [{'Elixir.IEx.Config',default_option, [width], [{file,"lib/iex/config.ex"},{line,63}]}, {'Elixir.IEx.Config','-default_config/0-fun-0-',1, [{file,"lib/iex/config.ex"},{line,60}]}, {'Elixir.Enum','-map/2-lists^map/1-0-',2, [{file,"lib/enum.ex"},{line,1088}]}, {'Elixir.Enum','-map/2-lists^map/1-0-',2, [{file,"lib/enum.ex"},{line,1088}]}, {'Elixir.IEx.Config',configuration,0, [{file,"lib/iex/config.ex"},{line,56}]}, {'Elixir.Dbg.Handler',handle_event,2, [{file,"lib/dbg/handler.ex"},{line,9}]}, {dbg,handle_traces,4,[{file,"dbg.erl"},{line,819}]}, {dbg,tracer_loop,2, [{file,"dbg.erl"},{line,783}]}]}}} ** When Server state == #{'__struct__' => 'Elixir.Dbg.Watcher', dbg_ref => #Ref<0.0.4.688>, tracer => nil, tracer_ref => nil} ** Reason for termination == ** {trace_handler_crashed, {function_clause, [{'Elixir.IEx.Config',default_option, [width], [{file,"lib/iex/config.ex"},{line,63}]}, {'Elixir.IEx.Config','-default_config/0-fun-0-',1, [{file,"lib/iex/config.ex"},{line,60}]}, {'Elixir.Enum','-map/2-lists^map/1-0-',2, [{file,"lib/enum.ex"},{line,1088}]}, {'Elixir.Enum','-map/2-lists^map/1-0-',2, [{file,"lib/enum.ex"},{line,1088}]}, {'Elixir.IEx.Config',configuration,0, [{file,"lib/iex/config.ex"},{line,56}]}, {'Elixir.Dbg.Handler',handle_event,2, [{file,"lib/dbg/handler.ex"},{line,9}]}, {dbg,handle_traces,4,[{file,"dbg.erl"},{line,819}]}, {dbg,tracer_loop,2,[{file,"dbg.erl"},{line,783}]}]}} Answers: username_1: Unfortunately this is due to a bug in Elixir 1.2.0. It is fixed in 1.3.0-dev and 1.2.1-dev. https://github.com/elixir-lang/elixir/commit/a6b80a19c50905ff7f0de4e689e1339fa883a367
microsoft/coe-starter-kit
1107466496
Title: [CoE Starter Kit - BUG] isCOE field on environment table Question: username_0: ### Describe the issue Need to be able to know which envt is CoE in apps. Can already tell in flows if you are in the envt. Thought about having this be an envt variable but there doesnt seem to be a need, we can just put in the table ### Expected Behavior _No response_ ### What solution are you experiencing the issue with? Core ### What solution version are you using? 3.20.6 ### What app or flow are you having the issue with? Admin | Sync Template v3 ### Steps To Reproduce _No response_ ### Anything else? _No response_ Answers: username_0: local fixed Feb
forbole/big-dipper-default-interface
743453729
Title: Setup meta tags Question: username_0: **Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] **Describe the solution you'd like** A clear and concise description of what you want to happen. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.<issue_closed> Status: Issue closed
gbif/portal-feedback
294129379
Title: Having this issue Question: username_0: fbitem-7490229a59148a44a969f4b702276af26e940516 System: Chrome 63.0.3239 / Windows 10 0.0.0 Referer: https://www.gbif.org/standards Window size: width 1185 - height 591 API log Site log System health at time of feedback: OPERATIONAL<issue_closed> Status: Issue closed
dbeaver/dbeaver
494416684
Title: Forced query cancel Question: username_0: I see many threads asking for query cancel to work in environments where that's impossible due to the lack of support on the driver/server side. What DBeaver **could** do, however, it to turn the ”cancel query” button into a ”force disconnect" button after the first click, instead of disabling it. Clicking on that button should show a dialog reminding the user that clicking that button will disconnect DBeaver from the query, but the query might continue running on the database server. Confirming that will unlock DBeaver interface allowing it to be used for other queries. This looks like an easy enough fix (although i don't know how much control you have on that button), and will give people what they care about, which is often not query cancel but the ability to run a new query. Answers: username_1: In fact it already works exactly as you described. If query cancel doesn't actually cancel it then DBeaver tries to cancel whole connection. This is controlled by following pref option: ![image](https://user-images.githubusercontent.com/6398845/65030529-62cd3880-d948-11e9-851c-9c7f13349a65.png) Meaning of this parameter: when query cancel operation takes longer than this value (ms) DBeaver will close whole connection. By default it was set to 0 (= disabled). Perhaps we should change it to some reasonable value.. username_0: my suggestion is for more manual version of that. Setting a timeout is difficult because sometimes you deal with a query which actually takes a long time to cancel (e.g. needs to rollback a large update) while sometimes you deal with a select you don't care about. If you reclick on the button you clearly state your intention instead. username_2: Wow that works perfectly. I was about to suggest something similar (Postgres "cancel" commands sometimes get lost if network is down or flakey, resulting in indefinitely running query with no ability to cancel it). After "x" seconds of a cancelled query not working, change button from greyed out "canceled" to a new button "force disconnect" or add a new button or what not. username_3: Bumping this issue. In our instance, this bug dynamic is related to: VPN connection to AWS Redshift I have not messed with the Error Handling or SQL Processing settings, but "Cancel execution" [after 10000 ms] seems like a workable solution, provided it doesn't leave the execution hanging server side. Does closing the dBeaver application locally also prompt the server side execution to end? username_3: Just updated to 6.3.5 and it seems it may be fixed?!? Hopefully not getting too excited too early, but this is a big help if resolved. Thanks to DBeaver dev team! Status: Issue closed username_5: I submitted a new issue that focus more on @username_0 's ultimate conclusion rather than what's said in his initial post: https://github.com/dbeaver/dbeaver/issues/9272 username_0: thanks for reopening, I didn't notice my issue got closed!
ros2/ros2
285571045
Title: removed note about ASIO not found on macOS Question: username_0: @0wu I removed the note you added about turning on building of dependencies within fast-rtps when asio is not found on macOS: https://github.com/ros2/ros2/wiki/OSX-Development-Setup/_compare/2af42dc2bde5bb6b700683c9435c7fadd55616d6...a8369e7f20d5bb4e361be66de51e7107c8bdcbbe That should be covered by installing asio from Homebrew. If that's not working then it's likely a bug in the asio installed by Homebrew or in Fast-RTPS's build. Can you confirm that asio is installed by homebrew and linked? Status: Issue closed Answers: username_1: @0wu I'm going to close this due to inactivity, please feel free to comment here if you still face the issue and we can reopen it
nhedlund/intrinio
339123725
Title: Authentication using Web API Key Question: username_0: The web API key does not require a username/password combo Answers: username_1: Are you sure? When I look at the current documentation from Intrinio they still use username/password combos as API keys: https://intrinio.com/tutorial/web_api username_1: Last time I checked they had not added API keys to their documentation but now that they have I will add it to the library. Username and password still works. Good suggestion, I checked their python-sdk package and from what I can see they do not use Pandas as return types and don't return the complete datasets by default. username_2: Thanks - I thought I posted another comment last night, but I found out they terminated my free subscriptions. I'm really annoyed that they dropped any free tier of their service, including for existing users, and will probably look at Python libraries that pull from IEX. I don't need anything near real-time pricing data. What's the unique value that you see in Intrinio over IEX or [other free options](https://www.reddit.com/r/Python/comments/8b5xuw/apis_that_work_for_stock_data/)? username_1: For me Intrinio gives you split- and dividend-adjusted stock prices that I need for accurate backtesting, for a reasonable monthly price. Quandl is another good provider. Most or all providers of free stock prices only offer unadjusted data or very limited historic data, at least what I have found. username_1: The web API key does not require a username/password combo username_1: It seems that they have changed the API a lot from V1 to V2 that would cause a major rewrite of this library. Even if I prefer working with Pandas datasets and libraries that returns full datasets instead of using paging it seems better to refer to the official SDK that targets the new API. Status: Issue closed username_2: Thanks. If it's helpful, it looks like the Yahoo API can still be made to work with pandas-datareader: https://towardsdatascience.com/python-for-finance-stock-portfolio-analyses-6da4c3e61054 https://pypi.org/project/fix-yahoo-finance/ I don't know what about the fix is 'temporary', but this API returns a pandas dataframe and included both actual and adjusted (I assume split/dividend) market close numbers. I've only looked far enough to see that the example walkthrough on the blog post returns data.
j3k0/cordova-plugin-purchase
492729613
Title: Ionic says plugin not installed when it is installed Question: username_0: Hi, I try to use the plugin in Ionic 5 but I'm getting warnings: ``` [ng] [console.warn]: "Native: tried calling InAppPurchase2.order, but the InAppPurchase2 plugin is not installed." [ng] [console.warn]: "Install the InAppPurchase2 plugin: 'ionic cordova plugin add cc.fovea.cordova.purchase'" ``` Of course I installed the plugin. ``` $ cordova plugin list cc.fovea.cordova.purchase 8.1.1 "Purchase" ``` I added some debug logs to node_modules/@ionic-native/core/decorators/common.js before the lines that prints warnings and I see that plugin reference is "store" and plugin instance created by calling `pluginInstance = getPlugin(pluginRef)` returns null. Any ideas how to fix it? Answers: username_0: Fixed by installing cordova-support-google-services plugin and putting google-services.json in platforms/android/app/src. Here is an instruction how to get google-services.json: https://support.google.com/firebase/answer/7015592?hl=en username_1: Installing "cordova-support-google-services" for android Plugin doesn't support this project's cordova-android version. cordova-android: 9.0.0, failed version requirement: <9.0.0 Skipping 'cordova-support-google-services' for android Adding cordova-support-google-services to package.json ``` Is there a way to fix this for android >9.0.0?
istio-ecosystem/authservice
550342052
Title: Session management Question: username_0: Once we have finished implementing server-side sessions, we should consider: 1. Issue new session IDs as often as possible to help users avoid things like [session fixation attacks](https://github.com/OWASP/CheatSheetSeries/blob/master/cheatsheets/Session_Management_Cheat_Sheet.md#renew-the-session-id-after-any-privilege-level-change) 1. Session timeouts 1. Offer a configuration option for max session idle time 1. Offer a configuration option for max absolute session lifetime 1. Also expire session info from the session storage to reduce resource usage https://github.com/OWASP/CheatSheetSeries/blob/master/cheatsheets/Session_Management_Cheat_Sheet.md Answers: username_0: Closed by https://github.com/istio-ecosystem/authservice/pull/67 Status: Issue closed
elastic/observability-docs
956745764
Title: [Request] Update APM agent - server - compatibility docs page Question: username_0: ## Description *(Describe what needs to be documented. Link to relevant issues.)* update [agent-server-compatibility](https://www.elastic.co/guide/en/apm/get-started/current/agent-server-compatibility.html#agent-server-compatibility) --> remove EOL agent versions ## Collaboration * [ ] The docs team will lead producing the content * [ ] The product team will provide the initial content and the docs team will edit / review * [x ] The docs team will define with product team the structure and location, and the product team will provide the initial content * [ ] Other (please describe) ## Suggested Target Release *(Let us know if this feature is needed for a specific release.)* ## Stakeholders @username_0
TimothyEarley/Groovy-Game-Core
157404642
Title: GL capabilities not created before test Question: username_0: Test throws a java.lang.NoClassDefFoundError [see here](https://travis-ci.org/username_0/Groovy-Game-Core/builds/133783017) Answers: username_0: This probably means the natives could not loaded. Hard to test since they are loading on my local machine.
fission/fission
375381970
Title: JVM improperly terminated breaks CI Question: username_0: ``` [ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12.4:test (default-test) on project hello-world: Execution default-test of goal org.apache.maven.plugins:maven-surefire-plugin:2.12.4:test failed: The forked VM terminated without saying properly goodbye. VM crash or System.exit called ? -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/PluginExecutionException [FAILED]: /test_environments/test_java_env.sh ``` Answers: username_1: @username_0 Can you please paste URL of one of failed builds so I can take a look at logs? username_1: Detailed stack trace: ``` ------------------------------------------------------- T E S T S ------------------------------------------------------- Forking command line: /bin/sh -c cd /usr/src/mymaven && /usr/lib/jvm/java-8-openjdk-amd64/jre/bin/java -jar /usr/src/mymaven/target/surefire/surefirebooter2572449849247098580.jar /usr/src/mymaven/target/surefire/surefire6253537605233400285tmp /usr/src/mymaven/target/surefire/surefire_02904009310726627281tmp Error: Could not find or load main class org.apache.maven.surefire.booter.ForkedBooter Results : Tests run: 0, Failures: 0, Errors: 0, Skipped: 0 [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 53.868 s [INFO] Finished at: 2018-10-31T07:13:17Z [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12.4:test (default-test) on project hello-world: Execution default-test of goal org.apache.maven.plugins:maven-surefire-plugin:2.12.4:test failed: The forked VM terminated without saying properly goodbye. VM crash or System.exit called ? -> [Help 1] org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12.4:test (default-test) on project hello-world: Execution default-test of goal org.apache.maven.plugins:maven-surefire-plugin:2.12.4:test failed: The forked VM terminated without saying properly goodbye. VM crash or System.exit called ? at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:213) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:154) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:146) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81) at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:56) at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:305) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:192) at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:105) at org.apache.maven.cli.MavenCli.execute (MavenCli.java:954) at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:288) at org.apache.maven.cli.MavenCli.main (MavenCli.java:192) at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke (Method.java:498) at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:289) at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:229) at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:415) at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:356) Caused by: org.apache.maven.plugin.PluginExecutionException: Execution default-test of goal org.apache.maven.plugins:maven-surefire-plugin:2.12.4:test failed: The forked VM terminated without saying properly goodbye. VM crash or System.exit called ? at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:148) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:208) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:154) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:146) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81) at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:56) at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:305) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:192) at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:105) at org.apache.maven.cli.MavenCli.execute (MavenCli.java:954) at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:288) at org.apache.maven.cli.MavenCli.main (MavenCli.java:192) at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke (Method.java:498) at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:289) at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:229) at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:415) at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:356) ``` username_1: Closing as commented in PR #960 Status: Issue closed
sublimelsp/LSP
948181843
Title: Add ability to hide diagnostics on hover Question: username_0: There is a quite big diagnostic window appears on hover. ![Screenshot from 2021-07-20 05-54-15](https://user-images.githubusercontent.com/6022888/126251450-c75cd5fb-c13e-4452-8ae3-4e8b8a9f5075.png) But there is no setting to disable/hide it. Could you kindly add it? Something like "show_hover_diagnostics_severity_level". Or can you point me to line of code where it called, so i can override it? Answers: username_1: But the `show_diagnostics_severity_level` setting should be affecting the popup too. If you set it to 3 so that `hint` severity is not reported then the diagnostic should not show up in the popup either. (From the color of the diagnostic I'm assuming it's a "hint" diagnostic) username_0: Yes it is hint, but `show_diagnostics_severity_level` doesn't affect it. It only removes circles near line numbers and underlines near variables. I already set it to 3, and as you see circles and underlinse are disappears, but on hover we still see hints. username_1: That's not how it's supposed to work and not how it works here. Did you try restarting ST? And also, are you on ST4? username_0: Yes, i am on ST4. And i restarted ST several times. username_1: You are right. I didn't wait long enough before checking. Then it looks like a bug to me./ Status: Issue closed
bitcoinjs/bitcoinjs-lib
246520315
Title: Generate uncompressed key Question: username_0: How can I generate uncompressed address? This method outputs compressed version only: `keyPair.getAddress()` Thanks! Answers: username_1: ``` js keyPair.compressed = true/false keyPair.getAddress() ``` Status: Issue closed username_1: Or, from the start. `ECPair.makeRandom({ compressed: false })` username_0: Awesome, many thanks!
rancher/rke2
673003329
Title: FIPS - Etcd secrets encryption Question: username_0: We should look into what our options are for encrypting sensitive etcd information. For example Secrets, ConfigMaps, etc. Further, if we implement this the fips documentation should be updated as well as a part of this effort. Answers: username_1: I think the question we now have to answer is whether or not the algorithm we use by default for encyrpting is a fips validated one. If it isnt, that might be ok and we just have to document how to switch to one that is. username_0: This is covered in https://docs.rke2.io/advanced/#secrets-encryption-config Closing. QA will be taking a holistic look at the docs so this is already covered and I glanced over and seems fine to me. Status: Issue closed
ziglang/zig
708954372
Title: Wrong size of packed struct with bit flags Question: username_0: I have defined below packed struct. There are 14 'bool' fields and u18 padding field. I was expecting @sizeOf(RESOURCE_STATES) to be 4. Unfortunately @sizeOf(RESOURCE_STATES) is 5. `pub const RESOURCE_STATES = packed struct { VERTEX_AND_CONSTANT_BUFFER: bool = false, INDEX_BUFFER: bool = false, RENDER_TARGET: bool = false, UNORDERED_ACCESS: bool = false, DEPTH_WRITE: bool = false, DEPTH_READ: bool = false, NON_PIXEL_SHADER_RESOURCE: bool = false, PIXEL_SHADER_RESOURCE: bool = false, STREAM_OUT: bool = false, INDIRECT_ARGUMENT_AND_PREDICATION: bool = false, COPY_DEST: bool = false, COPY_SOURCE: bool = false, RESOLVE_DEST: bool = false, RESOLVE_SOURCE: bool = false, padding: u18 = 0, }; ` The same technique but for smaller structs works as expected: `pub const HEAP_FLAGS = packed struct { SHARED: bool = false, DENY_BUFFERS: bool = false, ALLOW_DISPLAY: bool = false, SHARED_CROSS_ADAPTER: bool = false, DENY_RT_DS_TEXTURES: bool = false, DENY_NON_RT_DS_TEXTURES: bool = false, HARDWARE_PROTECTED: bool = false, padding: u25 = 0, }; ` @sizeOf(HEAP_FLAGS) == 4. Thanks, Michal Answers: username_1: Packed structs are currently a bit broken unfortunately. The issues to watch about this are #2627 and #3133
postlund/pyatv
528692812
Title: KeyError: '4' setting up Apple TV Question: username_0: **Describe the bug** Getting this exception when entering the pin in the configuration flow. Happens every time. 2019-11-25 19:42:13 ERROR (MainThread) [custom_components.apple_tv.config_flow] Unexpected exception Traceback (most recent call last): File "/config/custom_components/apple_tv/config_flow.py", line 175, in async_step_pair_with_pin await self._pairing.finish() File "/usr/local/lib/python3.7/site-packages/pyatv/mrp/pairing.py", line 45, in finish await self.pairing_procedure.finish_pairing(self.pin_code) File "/usr/local/lib/python3.7/site-packages/pyatv/mrp/auth.py", line 64, in finish_pairing atv_proof = pairing_data[tlv8.TLV_PROOF] KeyError: '4' **To Reproduce** Add the integration for the Apple TV, and enter the PIN shown on the Apple TV. **Expected behavior** **System Setup (please complete the following information):** - OS: [e.g. Debian]: Raspbian 4.19.75-v7+ #1270 SMP - Python: [e.g. 3.6.2] 3.7.3 - pyatv: [e.g. 0.3.10] pyatv-0.4.0a4 **Additional context** Here's the log file: [pyatv.log](https://github.com/username_1/pyatv/files/3891811/pyatv.log) Answers: username_1: Hmm, this is strange. First of all, the snippet you posted above is not from the attached log (timstamp differs). So this must be from two different points in time, seems to be about 10min apart. In the log, `pyatv` seems to connect and set everything up as expected (but I see that I have to make some adjustments to the connection flow, but that's for later). So the pairing must have succeeded at some point, otherwise there would be no credentials. There wouldn't even be any entities in Home Assistant as no config entry will be created before pairing has succeeded for both MRP/DMAP and AirPlay. So I have a hard time seeing that it fails *every* time? Can you elaborate on exactly what you are doing? Would be good if you removed the Apple TV from the Integrations page and restart Home Assistant. You could even bump to the latest update of the component, using `pyatv` 0.4.0a5 that I pushed earlier this morning. username_0: Just tried it with your latest Apple TV integration release (4944a51), and pyatv 0.4.0a5. Got a different error. [pyatv-2.log](https://github.com/username_1/pyatv/files/3892088/pyatv-2.log) username_1: Great, that seemed to do the trick. Now it's the same problem as in #279, a fix will be available soon. username_1: New update is pushed, please update component and try again! username_1: @username_0 Can you verify if you still see this issue and close it if you don't? username_0: This bug is fixed. I’ll close the issue when I get to a real computer Status: Issue closed username_1: Awesome! 👍
docker/for-mac
249824998
Title: Docker Error Question: username_0: ### Expected behavior ### Actual behavior ### Information - Full output of the diagnostics from "Diagnose & Feedback" in the menu - A reproducible case if this is a bug, Dockerfiles FTW - Page URL if this is a docs issue or the name of a man page ### Steps to reproduce the behavior Docker for Mac: version: 1.13.0 (0c6d765c5) macOS: version 10.12.5 (build: 16F73) logs: /tmp/0CFAA88E-6E92-4918-B7D4-999E46062052/20170812-134012.tar.gz failure: sysctl -a reports no hypervisor support [ERROR] vmnetd /var/tmp/com.docker.vmnetd.socket does not exist Unexpected error (No such file or directory) connecting to /var/tmp/com.docker.vmnetd.socket /private/tmp/vmnetd/com.docker.vmnetd is not running [OK] dns [ERROR] driver.amd64-linux com.docker.driver.amd64-linux -db is not running [OK] virtualization VT-X [OK] app [ERROR] moby /Users/alvaro/Library/Containers/com.docker.docker/Data/com.docker.driver.amd64-linux/console-ring does not exist [OK] system [OK] moby-syslog [ERROR] db /Users/alvaro/Library/Containers/com.docker.docker/Data/s40 does not exist Unexpected error (No such file or directory) connecting to /Users/alvaro/Library/Containers/com.docker.docker/Data/s40 com.docker.db is not running [OK] env [ERROR] virtualization kern.hv_support sysctl -a reports no hypervisor support [ERROR] slirp Unexpected error ((Failure "Error connecting socket to 9p endpoint unix:/Users/alvaro/Library/Containers/com.docker.docker/Data/s51: Unix.Unix_error(Unix.ENOENT, \"connect\", \"\")")) connecting to /Users/alvaro/Library/Containers/com.docker.docker/Data/s51 com.docker.slirp is not running slirp check failed with: (Failure "Error connecting socket to 9p endpoint unix:/Users/alvaro/Library/Containers/com.docker.docker/Data/s52: Unix.Unix_error(Unix.ENOENT, \"connect\", \"\")") [ERROR] osxfs com.docker.osxfs is not running [OK] moby-console [OK] logs [ERROR] docker-cli cannot find docker docker-cli check failed with: (Failure "docker -v: timeout after 10.00s") [OK] menubar [ERROR] disk Docker.qcow2 missing: the VM has never been started
clearlydefined/service
812401688
Title: Update historic maven package definitions Question: username_0: Definitions for Maven components in ClearlyDefined used to have incorrect source and other urls (see #755). With this pull request #794, all new Maven definitions will have the correct urls. This has been merged and deployed. However, there is still the matter of updating all of the existing Maven definitions (800,000+) to have correct URLs. I initially took the approach of [creating a script to update the "definitions-paged" Mongo DB](https://github.com/clearlydefined/correct-cd-maven-urls). This worked in my local development environment. However, it did not work in our [development environment](https://dev.clearlydefined.io/?sort=releaseDate&sortDesc=true) This is because, in our development and production environments, we use Azure blob storage (along with Mongo) to store definition information. As @jeffmcaffer put it in our internal Slack: "the definitions come from blob storage. You could completely delete the cosmos db and ClearlyDefined would continue to work (though search would be affected). 2:12 The key takeaway here is that blob is the golden store. Everything else should be derived from that." A better approach to update the historic data may be to: * Modify [this API call](https://github.com/clearlydefined/service/blob/324017ef0ccfe73eb9d1593faf777d19b621a216/routes/definitions.js#L65) to take a parameter to force recomputation (and storage) of the definition. This is currently done automatically for calls to this API from a localhost, but there is not currently a way to do this outside of localhost. This should be a straightforward change. * Download info for all 800,000 maven components into a big json file * Write a script to parse through the 800,000 components and call POST /definitions for each of them (sleeping for at least a second or two between each one) * Likely call the script on chunks of 100,000, rather than trying to do all 800,000 at once. Status: Issue closed Answers: username_0: Definitions for Maven components in ClearlyDefined used to have incorrect source and other urls (see #755). With this pull request #794, all new Maven definitions will have the correct urls. This has been merged and deployed. However, there is still the matter of updating all of the existing Maven definitions (800,000+) to have correct URLs. I initially took the approach of [creating a script to update the "definitions-paged" Mongo DB](https://github.com/clearlydefined/correct-cd-maven-urls). This worked in my local development environment. However, it did not work in our [development environment](https://dev.clearlydefined.io/?sort=releaseDate&sortDesc=true) This is because, in our development and production environments, we use Azure blob storage (along with Mongo) to store definition information. As @jeffmcaffer put it in our internal Slack: "the definitions come from blob storage. You could completely delete the cosmos db and ClearlyDefined would continue to work (though search would be affected). 2:12 The key takeaway here is that blob is the golden store. Everything else should be derived from that." A better approach to update the historic data may be to: * Modify [this API call](https://github.com/clearlydefined/service/blob/324017ef0ccfe73eb9d1593faf777d19b621a216/routes/definitions.js#L65) to take a parameter to force recomputation (and storage) of the definition. This is currently done automatically for calls to this API from a localhost, but there is not currently a way to do this outside of localhost. This should be a straightforward change. * Download info for all 800,000 maven components into a big json file * Write a script to parse through the 800,000 components and call POST /definitions for each of them (sleeping for at least a second or two between each one) * Likely call the script on chunks of 100,000, rather than trying to do all 800,000 at once. username_0: I've updated this repo https://github.com/clearlydefined/correct-cd-maven-urls, which contains a script to correct the records through GET /definitions requests. I've tested this in my local environment and in the azure dev environment and it seems to work for the vast majority of records. There are a few that time out that we will likely need to handle individually, but updating the vast majority through an automated scripts will meet our needs at this time. username_0: At this point I believe the script has updated the majority of maven components in ClearlyDefined. If anyone sees ones that are incorrect, please either re-open this issue or open a new issues. Status: Issue closed
typora/typora-issues
252192272
Title: Install on Windows without installer Question: username_0: As I don't have admin rights on my work pc, it would be nice if I could "install" Typora without the use of the windows installer, i.e. maybe as a zip. Answers: username_1: maybe same as https://github.com/typora/typora-issues/issues/371 Current concern is: we need admin rights if we want to let "Typora" show up in the "Open With" Context menu automatically. It is also needed if we want to put recent opened files in list, when user right click on the icon in taskbar. Status: Issue closed username_2: For what it's worth, I wanted to use typora at work, but cannot. I am not allowed admin rights. I'd **happily** give up the context and open with for the ability to actually use it. Like other installers, you should just have one installer, but say "you don't have admin, therefore, we can install without x & y, or you can cancel." username_2: In the meantime, I'm going to have to continue using Atom, as it does install without admin rights.
steam-test1/VPlusHUD
594581793
Title: Healthcircle negative value Question: username_0: This has nothing to do with the latest update to it because I have seen it for a long time just never bothered to report it since how rare it is. Sometimes enemies seem to have negative health which is a little odd. ![Untitled](https://user-images.githubusercontent.com/30779314/78506304-e16cb480-7778-11ea-8b67-dd88ea644e0b.png) Answers: username_1: Just testing a fix as the enemy health bar already includes a fix for that issue username_0: It happened using the latest version, but so far that's the only time I have seen it happen since updating Status: Issue closed username_0: will reopen if issue persists
mongoid/mongoid-locker
41039145
Title: Allow customizing locked_at field (clash with Devise). Question: username_0: mongoid-locker uses hardcoded field name `:locked_at`. This is also the name that is hardcoded into `Devise::Lockable` module. In our app that led to a nasty bug where users would be spontaneously logged out. Very annoying. I tried to make the field configurable but I hated the result (code becomes quite uglier). So I won't be posting a PR now, I'm afraid. In our app I fixed it by forking and hardcode-renaming the field: https://github.com/username_0/mongoid-locker/tree/alternate_field A clash with such popular gem as Devise should be a serious problem. I'm actually surprised it didn't surface earlier. Status: Issue closed Answers: username_1: Resolved in https://github.com/mongoid/mongoid-locker/pull/55. Please try HEAD! username_0: Awesome news! I felt kinda bad about not following up on this :)
MicrosoftDocs/windowsserverdocs
579374153
Title: Azure VM Question: username_0: Note Azure information is incorrect, you will probably need to install the following package before extended support patches will be received. Extended Security Updates (ESU) Licensing Preparation Package (KB4538483) More information here: https://techcommunity.microsoft.com/t5/windows-it-pro-blog/obtaining-extended-security-updates-for-eligible-windows-devices/ba-p/1167091 --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 628f3cbe-8ba7-a242-5305-c29a688f31c6 * Version Independent ID: cbacbc46-3e12-6434-e247-e5b7ed18dd5e * Content: [Windows Server 2008 and 2008 R2 extended security updates](https://docs.microsoft.com/en-us/windows-server/get-started/extended-security-updates#feedback) * Content Source: [WindowsServerDocs/get-started/extended-security-updates.md](https://github.com/MicrosoftDocs/windowsserverdocs/blob/master/WindowsServerDocs/get-started/extended-security-updates.md) * Product: **windows-server** * Technology: **server-general** * GitHub Login: @iainfoulds * Microsoft Alias: **iainfou** Status: Issue closed Answers: username_1: Note Azure information is incorrect, you will probably need to install the following package before extended support patches will be received. Extended Security Updates (ESU) Licensing Preparation Package (KB4538483) More information here: https://techcommunity.microsoft.com/t5/windows-it-pro-blog/obtaining-extended-security-updates-for-eligible-windows-devices/ba-p/1167091 --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 628f3cbe-8ba7-a242-5305-c29a688f31c6 * Version Independent ID: cbacbc46-3e12-6434-e247-e5b7ed18dd5e * Content: [Windows Server 2008 and 2008 R2 extended security updates](https://docs.microsoft.com/en-us/windows-server/get-started/extended-security-updates#feedback) * Content Source: [WindowsServerDocs/get-started/extended-security-updates.md](https://github.com/MicrosoftDocs/windowsserverdocs/blob/master/WindowsServerDocs/get-started/extended-security-updates.md) * Product: **windows-server** * Technology: **server-general** * GitHub Login: @iainfoulds * Microsoft Alias: **iainfou** username_2: #assign:dknappettmsft
SR-Sunny-Raj/Hacktoberfest2021-DSA
1021733884
Title: Defense of a Kingdom SPOJ Question: username_0: Theodore implements a new strategy game “Defense of a Kingdom”. On each level a player defends the Kingdom that is represented by a rectangular grid of cells. The player builds crossbow towers in some cells of the grid. The tower defends all the cells in the same row and the same column. No two towers share a row or a column. The penalty of the position is the number of cells in the largest undefended rectangle. Help Theodore write a program that calculates the penalty of the given position. Example Input: 1 15 8 3 3 8 11 2 8 6 Output: 12 Answers: username_0: please assign this to me
decred/dcrd
146480905
Title: dcrdctl clear history Question: username_0: The current option menu shows, [h]elp print this message [l]ist list all available commands [p]rotect toggle protected mode (for passwords) [q]uit/ctrl+d exit Id like to have this option: [c]lear clear history of commands<issue_closed> Status: Issue closed
redsunservers/LoadoutBugTracker
298111525
Title: Badge Collector and Mod Panel overlap Question: username_0: ![image](https://user-images.githubusercontent.com/33851923/36355986-57c27ef0-14a0-11e8-8de3-a6bd82f6b178.png) Answers: username_1: What is the size of the window you're using? Works fine at 1920x1080. username_2: similar thing happens on mobile (2560x1440) https://i.imgur.com/zXVn3kK.png username_3: cdn related, try to force a refresh with your mobile. Status: Issue closed
dotnet/csharplang
336454692
Title: Proposal: Approximately-equal operator ~= Question: username_0: Motivation: The existing Equals `==` operator is overridden for types for common comparison scenarios. However the nature of "equality" in .NET (and software in general) is often more nuanced. For example I often find myself needing to perform ordinal case-insensitive string comparisons, but the `==(String,String)` operator invokes `String.Equals(String,String)` which performs an ordinal, case-sensitive, string comparison - this means I can't use the nice `==` shorthand syntax but need to replace them with the verbose `String.Equals( x, y, StringComparison.OrdinalIgnoreCase )`. I'd like to propose that the C# language adds a new operator `~=` (or `~~` or `=~`) which has the semantics of an "approximately equal" comparison. 1. The primary use-case being to invoke an ordinal, case-insensitive comparison when used with `String` operands - however it has applications for other types too: 1. For IEEE-754 types, where `==` often doesn't work as-intended, the `~=` operator could be used to check for a near-enough comparison without needing to repeat the equality-within-tolerance code in the `Double.Equals` [documentation page](https://msdn.microsoft.com/en-us/library/ya2zha7s(v=vs.110).aspx). An obvious shortcoming is the necessity to parameterize the operation: while ordinal case-insensitive string equality is straightforward, it's an anglocentric presumption of mine that it has greater utility than an ordinal, accent-insensitive string equality (for many other languages) or even more general: [a Soundex string comparison](https://en.wikipedia.org/wiki/Soundex). This parameterization could be done by treating the `~=` operator as a ternary operator (with 3 operands), the parameter being specified with some kind of separator after the second operand. For example: bool floatApproxEquals = ( 0.3333f ~= 1f/3f ); // binary operator with default tolerance of 0.01 bool floatApproxEquals = ( 0.3333f ~= 1f/3f : 0.001f ); // ternary, explicit tolerance of 0.001 bool stringApproxEquals = ( "abc" ~= "ABC" ); // binary operator with default argument CompareOptions.IgnoreCase bool stringApproxEquals = ( "abc" ~= "ABC" : CompareOptions.IgnoreNonSpace ); // ternary, explicit CompareOptions value The operator overload signature would look like: public class Foo { // Binary public static Boolean operator~=( Foo first, Foo second ) { // ... } // Ternary public static Boolean operator~=( Foo first, Foo second, FooComparisonOptions mode ) { // ... } } As a bonus, this works well with Linq when used against a SQL data-source where `==(String,String)` is actually a case-insensitive comparison (according to most DBMS collations), so by adopting `~=` the Linq implementation could specify the precisely required `COLLATE` argument to textual comparison operations (and perhaps restore `==` to stricter ordinal, case-sensitive, semantics shared with C#): dbContext.People.Where( person => person.Name == "<NAME>" ).ToList(); // SELECT ... WHERE Name = "<NAME>" COLLATE Latin1_General_CS_AS dbContext.People.Where( person => person.Name ~= "<NAME>" ).ToList(); // SELECT ... WHERE Name = "<NAME>" COLLATE Latin1_General_CI_AI dbContext.People.Where( person => person.Name ~= "<NAME>" : CompareOptions.IgnoreNonSpace ).ToList(); // SELECT ... WHERE Name = "<NAME>" COLLATE Latin1_General_CS_AI Answers: username_1: The moment you need to parameterize an operator... well.. then an operator probably isn't the right tool for the job. username_2: Well written proposal. Really. But I disagree that we need an operator for this. username_0: I'm sober now, closing :) Status: Issue closed
gaearon/react-hot-loader
306020900
Title: How do I unit test hot loader behavior in my app? Question: username_0: I'd like to make test assertions for my app's RHL behavior, e.g. "If `ComponentA` is replaced by `ComponentB`, then I expect `ComponentB` to have this state." However, I don't see any API for doing so. RHL's internal tests [use a `register()` method](https://github.com/gaearon/react-hot-loader/blob/9ef4c35aa62695c46fdbca6929de1db77ad22ba2/test/AppContainer.dev.test.js#L26) for this purpose, which doesn't appear to be exported [in the dev build](https://github.com/gaearon/react-hot-loader/blob/9ef4c35aa62695c46fdbca6929de1db77ad22ba2/src/index.dev.js). Is there an approach that I'm overlooking? Answers: username_1: ```js export default ReactHotLoader ``` and then ```js reactHotLoader.register(type, NAME, FILENAME); ``` But why do you need to test it? I will not recommend to have RHL enabled in testing environment due 2 points: - it does not make sence - it actually breaking Jest(#872) and Enzyme(#897) username_0: So, I have a use case that looks like this: ```jsx class MyComponent extends React.Component { constructor(props) { super(props); this.state = {thing: props.defaultThing}; } componentWillMount() { this._transformedThing = transform(this.state.thing); } componentWillUpdate(nextProps, nextState) { this._transformedThing = transform(nextState.thing); } } ``` I want a test case in my app to confirm that, if I make changes to `MyComponent` in a dev environment where RHL is in use, then the instance variable will reflect the preserved state. Something along the lines of: ```jsx import { shallow } from 'enzyme'; const wrapper = shallow(<MyComponent defaultThing="foo" />); expect(wrapper.instance()._transformedThing).toEqual(transform("foo")); wrapper.setState({thing: "bar"}); expect(wrapper.instance()._transformedThing).toEqual(transform("bar")); // Now simulate MyComponent being replaced by RHL with a new component // with the same implementation... expect(hotReplacedWrapper.state("thing")).toBe("bar"); expect(hotReplacedWrapper.instance()._transformedThing).toEqual(transform("bar")); ``` Unfortunately, I can't think of any way to simulate a component instance's state being altered between the execution of its `constructor()` and `componentWillMount()`, as RHL does. username_1: why not to add your case to RHL's own tests? username_1: When ComponentA is replaced by ComponentB, the following occurs: - All ComponentA instances are __NOT__ unmounted (componentWillUnmount()) - A ComponentB instance is created, and ComponentA instnce is created (constructor()) - All variables unchanged between a new ComponentA instance and existing ComponentA instances, are replaced by ComponentB variables. - As result the state got unchanged. - All arrow function or _bound_ members, as long they are variables, is replaced and updated. - ComponentA prototype is replaced by ComponentB prototype. As result ComponentA does not exists anymore. - All ComponentB instances are NOT mounted (componentWillMount()), as long they were not unmounted. As result - hot update went seamlessly - constructor of ComponentA was called once - constructor of ComponentB was called once - render method of ComponentA instance might be called at least twice. - no lifecycle method was called. That was actually a thing to discuss - we are detecting changes in lifecycle methods, and displaying a warning about it. We are not executing them due to "no time travel" rule. Aka component already exists, we and wont change its past. It is possible to not to update these components, but unmount them in real. Changes in lifecycle methods are rare, but applying that changes - it might be something user could expect. username_0: Ah, I see. Thanks, that's a useful clarification. Somehow I'd been under the impression that react-hot-loader was unmounting the changed components and mounting their replacements. Status: Issue closed
ankitshaw/athena-PersonalAssistant
296321985
Title: Apk Version Question: username_0: can you give me the link of the apk version? Answers: username_1: apk version? @username_0 this is the link to sdk version https://github.com/username_1/athena-PersonalAssistant/blob/de51189cc39310fe0cd2d038981e2d80fe1c4be6/app/build.gradle Status: Issue closed
MassTransit/MassTransit
675495800
Title: Test Harness: long-running consumer shouldn't trigger the inactivity timeout Question: username_0: ### Is this a bug report? Yes ### Can you also reproduce the problem with the latest version? Yes 7.0.2 .NET Core 3.1 ### Steps to Reproduce Query harness.Comsumed.Any() after sending a message to a long-running consumer. ### Expected Behavior It should wait until the consumer completes. ### Actual Behavior It returns after the inactivity timeout and does not wait for the consumer. ### Reproducible Demo Solution attached. [MTTestTimeout.zip](https://github.com/MassTransit/MassTransit/files/5045033/MTTestTimeout.zip) Answers: username_1: I missed a key aspect of the bus inactivity monitor when I ported it into the test harness. That's been resolved now. Status: Issue closed
ebaumstarck/NewYorkTimes
172006813
Title: NYTimes Assignment Feedback Question: username_0: Hey Emma, :+1: nice work. Here are some notes after checking out the code: - Be sure to [use styles.xml whenever possible](http://guides.codepath.com/android/Styles-and-Themes#defining-and-using-styles) to keep your layout files clear of duplicate and unnecessary view presentation code. - Good use of a `Toolbar` within your activities. Be sure to [check out how to collapse the Toolbar on scroll](http://guides.codepath.com/android/Using-the-App-ToolBar#reacting-to-scroll) using the `CoordinatorLayout`. - You might want to improve the look and feel of the article item results. You should play with the padding, spacing, font sizes and look into [the CardView](http://guides.codepath.com/android/Using-the-CardView). - Consider adding [progress bars](http://guides.codepath.com/android/Handling-ProgressBars#progress-within-actionbar) to your app whenever a network request goes out. - Glad to see you used a date-time picker. To make sure the picker looks better on older devices (pre-Lollipop), you can check out [third-party libraries](https://github.com/wdullaer/MaterialDateTimePicker) for improved backwards compatibility. - Be sure to try out the [RecyclerView](http://guides.codepath.com/android/Using-the-RecyclerView), the successor to the ListView and GridView. - In the future, you might want to take a closer look at [Retrofit](http://guides.codepath.com/android/Consuming-APIs-with-Retrofit) for networking as this library is quite popular and used in many apps. - Good to see you got sharing working using the `ShareActionProvider` system. Keep in mind that this is simply leveraging the implicit intent system behind the scenes. Refer to [this guide on implicit intents](http://guides.codepath.com/android/Common-Implicit-Intents) to see other common usages of them. Let me know if you have any other thoughts or questions about this assignment. Looking forward to seeing your Twitter submission!
zio/zio-intellij
932408301
Title: Run all tests in package not supported Question: username_0: The following exceptions(s) occurred in the ZIO for IntelliJ plugin: ``` com.intellij.openapi.util.InvalidDataException: Test configuration kind 'All in package' is not supported. at zio.intellij.testsupport.ZTestRunConfiguration.fromTestConfiguration(ZTestRunConfiguration.scala:44) at zio.intellij.testsupport.ZTestRunConfiguration.$anonfun$runnerInfo$2(ZTestRunConfiguration.scala:52) at scala.Option.getOrElse(Option.scala:201) at zio.intellij.testsupport.ZTestRunConfiguration.zio$intellij$testsupport$ZTestRunConfiguration$$runnerInfo(ZTestRunConfiguration.scala:52) at zio.intellij.testsupport.ZTestRunConfiguration$ZioTestCommandLineState.<init>(ZTestRunConfiguration.scala:82) at zio.intellij.testsupport.ZTestRunConfiguration.$anonfun$runStateProvider$1(ZTestRunConfiguration.scala:59) at org.jetbrains.plugins.scala.testingSupport.test.AbstractTestRunConfiguration.getState(AbstractTestRunConfiguration.scala:190) at com.intellij.execution.runners.ExecutionEnvironment.getState(ExecutionEnvironment.java:222) at com.intellij.execution.impl.DefaultJavaProgramRunner.execute(DefaultJavaProgramRunner.java:96) at com.intellij.execution.impl.ExecutionManagerImpl.executeConfiguration(ExecutionManagerImpl.kt:731) at com.intellij.execution.impl.ExecutionManagerImpl.access$executeConfiguration(ExecutionManagerImpl.kt:69) at com.intellij.execution.impl.ExecutionManagerImpl$executeConfiguration$2.accept(ExecutionManagerImpl.kt:669) at com.intellij.execution.impl.ExecutionManagerImpl$executeConfiguration$2.accept(ExecutionManagerImpl.kt:69) at com.intellij.openapi.application.impl.NonBlockingReadActionImpl$Submission.lambda$safeTransferToEdt$6(NonBlockingReadActionImpl.java:577) at com.intellij.openapi.application.TransactionGuardImpl.runWithWritingAllowed(TransactionGuardImpl.java:218) at com.intellij.openapi.application.TransactionGuardImpl.access$200(TransactionGuardImpl.java:21) at com.intellij.openapi.application.TransactionGuardImpl$2.run(TransactionGuardImpl.java:200) at com.intellij.openapi.application.impl.ApplicationImpl.runIntendedWriteActionOnCurrentThread(ApplicationImpl.java:781) at com.intellij.openapi.application.impl.ApplicationImpl.lambda$invokeLater$4(ApplicationImpl.java:319) at com.intellij.openapi.application.impl.FlushQueue.doRun(FlushQueue.java:84) at com.intellij.openapi.application.impl.FlushQueue.runNextEvent(FlushQueue.java:133) at com.intellij.openapi.application.impl.FlushQueue.flushNow(FlushQueue.java:46) at com.intellij.openapi.application.impl.FlushQueue$FlushNow.run(FlushQueue.java:189) at java.desktop/java.awt.event.InvocationEvent.dispatch(InvocationEvent.java:313) at java.desktop/java.awt.EventQueue.dispatchEventImpl(EventQueue.java:776) at java.desktop/java.awt.EventQueue$4.run(EventQueue.java:727) at java.desktop/java.awt.EventQueue$4.run(EventQueue.java:721) at java.base/java.security.AccessController.doPrivileged(Native Method) at java.base/java.security.ProtectionDomain$JavaSecurityAccessImpl.doIntersectionPrivilege(ProtectionDomain.java:85) at java.desktop/java.awt.EventQueue.dispatchEvent(EventQueue.java:746) at com.intellij.ide.IdeEventQueue.defaultDispatchEvent(IdeEventQueue.java:969) at com.intellij.ide.IdeEventQueue._dispatchEvent(IdeEventQueue.java:839) at com.intellij.ide.IdeEventQueue.lambda$dispatchEvent$8(IdeEventQueue.java:449) at com.intellij.openapi.progress.impl.CoreProgressManager.computePrioritized(CoreProgressManager.java:808) at com.intellij.ide.IdeEventQueue.lambda$dispatchEvent$9(IdeEventQueue.java:448) at com.intellij.openapi.application.impl.ApplicationImpl.runIntendedWriteActionOnCurrentThread(ApplicationImpl.java:781) at com.intellij.ide.IdeEventQueue.dispatchEvent(IdeEventQueue.java:496) at java.desktop/java.awt.EventDispatchThread.pumpOneEventForFilters(EventDispatchThread.java:203) at java.desktop/java.awt.EventDispatchThread.pumpEventsForFilter(EventDispatchThread.java:124) at java.desktop/java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThread.java:113) at java.desktop/java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:109) at java.desktop/java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:101) at java.desktop/java.awt.EventDispatchThread.run(EventDispatchThread.java:90) ``` ### Extra details: ZIO Test right click test package and 'Run ZIO Tests in scala' ### Additional information: <details> ZIO plugin version: 2021.1.10.1 Scala plugin version: 2021.1.384 ZIO version(s): 1.0.9 Scala version(s): 2.13.1 sbt version: 1.3.8 IntelliJ version: IntelliJ IDEA 2021.1.2 Build #IC-211.7442.40 Java version: 11.0.11+9-b1341.57 Operating System: amd64 Linux (5.4.0-77-generic, amd64) JVM version: Dynamic Code Evolution 64-Bit Server VM JetBrains s.r.o. </details> Answers: username_1: Looks like the error reporter is working :)) Thanks, unfortunately this feature is not implemented yet, but I'll try adding it... username_1: Relates to #26 username_2: Would be very useful to me too as I have a multimodule sbt project.
skywinder/ActionSheetPicker-3.0
105526829
Title: Error while using CoreActionSheetPicker Question: username_0: ![screen shot 2015-09-09 at 11 40 27 am](https://cloud.githubusercontent.com/assets/13223874/9754533/c086a80e-56e7-11e5-86bf-15ace5141c90.png) Answers: username_1: what object is `sender` in your case? username_0: I am using UITableViewController. username_1: You have to specify `UIView` as origin in this case. Status: Issue closed
google/guava
148376359
Title: Exception handling is not very elegant Question: username_0: public class ExceptionTest { @Test public void testExceptionToString(){ try{ throw new RuntimeException("RuntimeException-test"); }catch (Exception e){ String ss = Throwables.getStackTraceAsString(e); System.out.println("exception:"+ss); } } @Test public void testExceptionCausalChain(){ try{ throw new RuntimeException("RuntimeException-test"); }catch (Exception e){ List<Throwable> chain = Throwables.getCausalChain(e); for(Throwable t :chain) { System.out.println("exception:" + String.valueOf(t)); } e.printStackTrace(); } } } -------------- I just want to focus on special package callstack! Answers: username_1: Can you clarify what you're requesting? username_0: I just want to get the call chains under the specific package as ```java String ss = Throwables.getStackTraceAsString(e,specialPackageName); ``` username_2: Can you please elaborate it ? what exactly you are asking about ? Sorry but could not get clear picture of your request . Status: Issue closed
jlippold/tweakCompatible
416456587
Title: `Notch'd` working on iOS 12.1 Question: username_0: ``` { "packageId": "cydia.kiiimo.org.notchsimulator", "action": "working", "userInfo": { "arch32": false, "packageId": "cydia.kiiimo.org.notchsimulator", "deviceId": "iPhone10,1", "url": "http://cydia.saurik.com/package/cydia.kiiimo.org.notchsimulator/", "iOSVersion": "12.1", "packageVersionIndexed": false, "packageName": "Notch'd", "category": "iOS 11.3x", "repository": "A kiiimo Respositoty ", "name": "Notch'd", "installed": "0.1", "packageIndexed": false, "packageStatusExplaination": "This tweak has not been reviewed. Please submit a review if you choose to install.", "id": "cydia.kiiimo.org.notchsimulator", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.0.9", "shortDescription": "Simulate that iPhone X notch!", "latest": "0.1", "author": "Sniper_GER", "packageStatus": "Unknown" }, "base64": "<KEY> "chosenStatus": "working", "notes": "" } ```
r-lib/devtools
494179496
Title: DT should not be an imported package Question: username_0: Because DT is an imported package, when you load devtools, it also loads htmltools, which makes it hard to do development on htmltools, especially on on Windows. There are also some packages that are listed as Imports in crosstalk's DESCRIPTION but don't load immediately because they're not imported in the NAMESPACE. Even though they're not loaded immediately, these packages do need to be installed when DT (and thus crosstalk) is installed. Two of these packages have a pretty heavy installation footprint: shiny and ggplot2. So there really two annoying issues: one is that loading devtools causes DT and htmltools to load. That one can be solved by removing `importFrom(DT,datatable)` from NAMESPACE. The other one is the heavy installation footprint. To solve that, DT would have to be removed from the Imports in the DESCRIPTION file. I see in #2085 that it was moved there specifically to cause it to be automatically installed. Maybe it could be changed to a Suggested package again and prompt for installation when it's needed? Status: Issue closed Answers: username_1: devtools is a meta package intended to be installed on rarely on developers' machines, where dependency load is not really a concern, but making setup easy for new developers is. The code previously did prompt for installation when it was needed, but I don't think these concerns warrant changing it back to the previous behavior. 67a44f1d9265fcad177ef8 removes DT from an explicit import, which should resolve the htmltools concern. username_0: Installing package into ‘/Users/winston/R/3.7’ (as ‘lib’ is unspecified) also installing the dependencies ‘zeallot’, ‘colorspace’, ‘utf8’, ‘vctrs’, ‘plyr’, ‘labeling’, ‘munsell’, ‘RColorBrewer’, ‘fansi’, ‘pillar’, ‘pkgconfig’, ‘httpuv’, ‘xtable’, ‘sourcetools’, ‘gtable’, ‘reshape2’, ‘scales’, ‘tibble’, ‘viridisLite’, ‘BH’, ‘sys’, ‘ini’, ‘backports’, ‘ps’, ‘lazyeval’, ‘shiny’, ‘ggplot2’, ‘later’, ‘askpass’, ‘clipr’, ‘clisymbols’, ‘curl’, ‘fs’, ‘gh’, ‘purrr’, ‘rlang’, ‘rprojroot’, ‘whisker’, ‘yaml’, ‘processx’, ‘R6’, ‘assertthat’, ‘rex’, ‘htmltools’, ‘htmlwidgets’, ‘magrittr’, ‘crosstalk’, ‘promises’, ‘mime’, ‘openssl’, ‘prettyunits’, ‘xopen’, ‘brew’, ‘commonmark’, ‘Rcpp’, ‘stringi’, ‘stringr’, ‘xml2’, ‘evaluate’, ‘praise’, ‘usethis’, ‘callr’, ‘cli’, ‘covr’, ‘crayon’, ‘desc’, ‘digest’, ‘DT’, ‘ellipsis’, ‘glue’, ‘git2r’, ‘httr’, ‘jsonlite’, ‘memoise’, ‘pkgbuild’, ‘pkgload’, ‘rcmdcheck’, ‘roxygen2’, ‘rstudioapi’, ‘rversions’, ‘sessioninfo’, ‘testthat’, ‘withr’ Warning in install.packages : unable to access index for repository https://cloud.r-project.org/bin/macosx/el-capitan/contrib/3.7: cannot open URL 'https://cloud.r-project.org/bin/macosx/el-capitan/contrib/3.7/PACKAGES' Packages which are only available in source form, and may need compilation of C/C++/Fortran: ‘colorspace’ ‘utf8’ ‘vctrs’ ‘plyr’ ‘fansi’ ‘httpuv’ ‘sourcetools’ ‘reshape2’ ‘scales’ ‘tibble’ ‘sys’ ‘backports’ ‘ps’ ‘lazyeval’ ‘later’ ‘askpass’ ‘curl’ ‘fs’ ‘purrr’ ‘rlang’ ‘yaml’ ‘processx’ ‘htmltools’ ‘promises’ ‘mime’ ‘openssl’ ‘commonmark’ ‘Rcpp’ ‘stringi’ ‘xml2’ ‘covr’ ‘digest’ ‘ellipsis’ ‘glue’ ‘git2r’ ‘jsonlite’ ‘pkgload’ ‘roxygen2’ ‘testthat’ Do you want to attempt to install these from sources? (Yes/no/cancel) ``` username_1: Maybe you know someone on the shiny team who can work on reducing these dependencies in DT and crosstalk ;) FWIW covr does not need shiny to generate or serve the reports, it simply uses DT to create the tables for the static HTML, which I guess I could generate without the DT package, but it doesn't seem like a great use of my time. username_0: I have to admit that you have a good point about the crosstalk dependencies. I filed an issue at rstudio/crosstalk#72. username_2: Dependency of devtools against a web framework is a major design flaw and should be addressed as a bug. I don't want shiny in my environment that has nothing to do or test with shiny.
gatsbyjs/gatsby
860082278
Title: Reference style links in MDX broken in static builds Question: username_0: <!-- Please fill out each section below, otherwise, your issue will be closed. This info allows Gatsby maintainers to diagnose (and fix!) your issue as quickly as possible. Useful Links: - Documentation: https://www.gatsbyjs.com/docs/ - How to File an Issue: https://www.gatsbyjs.com/contributing/how-to-file-an-issue/ Before opening a new issue, please search existing issues: https://github.com/gatsbyjs/gatsby/issues --> ## Description When building the static assets for a site with MDX content that uses [reference-style links][] and a site prefix, the site prefix does not get appended to the link destination for internal links. [reference-style links]: https://www.markdownguide.org/basic-syntax/#reference-style-links ### Steps to reproduce 1. `npm init gatsby` 2. Select **No CMS** and **No styling system** 3. Select **Add markdown and MDX support** feature 4. Edit `gatsby-config.js` and [add a path prefix][] ```diff 1 module.exports = { + 2 pathPrefix: `/my-site/`, 3 siteMetadata: { 4 title: "My Gatsby Site", 5 }, 6 plugins: [ 7 "gatsby-plugin-mdx", 8 { 9 resolve: "gatsby-source-filesystem", 10 options: { 11 name: "pages", 12 path: "./src/pages/", 13 }, 14 __key: "pages", 15 }, 16 ], 17 }; ``` 5. Create a new mdx file (`src/pages/test.mdx`) with the following content: ```mdx This is an [inline link](/path/to/page/). This is a [reference style link][]. [reference style link]: /path/to/page/ ``` 6. Build static assets with the prefix flag: `npm run build --prefix-paths` 7. Serve the static HTML: `npm run serve` 8. Navigate to: http://localhost:9000/test/ [add a path prefix]: https://www.gatsbyjs.com/docs/how-to/previews-deploys-hosting/path-prefix/ ### Expected result [Truncated] ```sh System: OS: macOS 10.15.7 CPU: (8) x64 Intel(R) Core(TM) i7-7820HQ CPU @ 2.90GHz Shell: 5.7.1 - /bin/zsh Binaries: Node: 12.19.0 - ~/.nvm/versions/node/v12.19.0/bin/node Yarn: 1.22.10 - /usr/local/bin/yarn npm: 6.14.8 - ~/.nvm/versions/node/v12.19.0/bin/npm Languages: Python: 2.7.16 - /usr/local/bin/python Browsers: Chrome: 89.0.4389.128 Safari: 14.0.3 npmPackages: gatsby: ^3.2.1 => 3.3.0 gatsby-plugin-mdx: ^2.3.0 => 2.3.0 gatsby-source-filesystem: ^3.3.0 => 3.3.0 ``` Answers: username_1: Thanks for the issue and reproduction! I'm also seeing this following your steps. Unfortunately, at this time I'm not sure why that might happen. Maybe it's a bug in the underlying packages such as `remark`? username_0: It looks like the path prefix is added in [`gatsby-plugin-mdx/utils/get-source-plugins-as-remark-plugins.js`][]. However, it only visits `link` nodes and not `definition` nodes. The following change should fix this: ```diff async function transformer(markdownAST) { // Ensure relative links include `pathPrefix` - visit(markdownAST, `link`, node => { + visit(markdownAST, [`link`, `definition`], node => { if ( node.url && node.url.startsWith(`/`) && !node.url.startsWith(`//`) ) { ``` I can create a PR for this change if you want. [`gatsby-plugin-mdx/utils/get-source-plugins-as-remark-plugins.js`]: https://github.com/gatsbyjs/gatsby/blob/master/packages/gatsby-plugin-mdx/utils/get-source-plugins-as-remark-plugins.js#L23 username_1: Yeah, looking at https://astexplorer.net/#/gist/e6f30d6ee1d24d361ad6baa6e8cd8c62/6e86918d2c1fcb1c964c9c5694e598c1f63ed447 adding `definition` there makes sense! A PR would be highly appreciated! username_0: PR created: #30967 Status: Issue closed
MicrosoftDocs/OfficeDocs-SkypeForBusiness
567148555
Title: Intune enrollment Question: username_0: "If you have Microsoft Intune, devices are automatically enrolled in Intune. After a device is enrolled, device compliance is confirmed and conditional access policies are applied to the device." Need more information around this because my device did not enroll in Intune automatically like it says. The device only appears in Teams. It probably has something to do with existing Intune Conditional Access rules but how do we resolve? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 8a627b5d-3edc-1add-b921-1dcd49c38653 * Version Independent ID: a2e2bd3a-31c0-471a-05fb-fd64bd942c4a * Content: [Manage your devices in Microsoft Teams - Microsoft Teams](https://docs.microsoft.com/en-us/microsoftteams/device-management#feedback) * Content Source: [Teams/device-management.md](https://github.com/MicrosoftDocs/OfficeDocs-SkypeForBusiness/blob/live/Teams/device-management.md) * Service: **msteams** * GitHub Login: @LanaChin * Microsoft Alias: **v-lanac** Answers: username_1: @username_0 Thank you for submitting feedback and contributing to the docs. We are currently investigating this. username_2: @username_0 thank you for reaching out and helping improve MS Docs. Please have a look at this link and let me know if it helps at all - https://techcommunity.microsoft.com/t5/microsoft-teams-blog/microsoft-teams-ip-phones-and-intune-enrollment/ba-p/330432 username_0: It was my enrollment restrictions causing the problem. I basically need to allow android device admin enrollment but first add the corporate identifier so I can continue to keep personal devices blocked. The article should at least be updated to say ....devices are automatically enrolled in Intune depending on your enrollment restrictions and conditional access policies. Also, if enrollment fails, the device seems to work perfectly fine anyway and can still be managed in Teams so it would be good to list the advantages of having desk phones enrolled in intune. username_1: @username_0 Thank you for submitting feedback. We understand that this issue has been resolved. Please feel free to re-open this issue if there is a specific area of the docs that we can improve or make better. Thank you. Status: Issue closed
mdn/browser-compat-data
926913844
Title: Sec-CH-UA User Client Hints - Time to document? Question: username_0: *User Agent Client Hints* went into [Chrome 89](https://www.chromestatus.com/feature/5995832180473856). - The [spec](https://wicg.github.io/ua-client-hints/) is a W3C Draft Community Group Report - which presumably counts as "on a standard track". - Mozilla's [position](https://mozilla.github.io/standards-positions/) is "not harmful" - There is no signal from Safari according to [that Chrome 89 info](https://www.chromestatus.com/feature/5995832180473856) The MDN convention on [When to document new technologies](https://developer.mozilla.org/en-US/docs/MDN/Guidelines/Conventions_definitions#when_to_document_new_technologies) doesn't give me clear direction on whether we document this on MDN and in BCD. I feel it probably is time, but perhaps I'm overly influenced by chromium's effective market share. Is it time? FYI @Elchi3 @username_1 Answers: username_1: Ordinarily, I'd say that if a browser is shipping a feature, then it should be represented in BCD. That feels like an obvious threshold, so I'm concerned that I'm missing some key detail that would suggest otherwise in this case. username_0: You're not missing any key detail. The question arises because [When to document new technologies ](https://developer.mozilla.org/en-US/docs/MDN/Guidelines/Conventions_definitions#when_to_document_new_technologies) for MDN includes the level of interest from other browser vendors as part of the calculation. If BCD has slightly different standard of "is in a browser" that's all I need to know. Will look at this on Friday. Status: Issue closed
ossrs/srs
740392822
Title: SRS从集群服务器拉流转发播放应该如何配置 Question: username_0: 您好,我的源服务器(master节点)是多个SRS的服务集群,服务集群中只有一个节点有用户需要的流。如下图所示: ![image](https://user-images.githubusercontent.com/16345166/98756796-3c007a00-2406-11eb-9cae-b272e82db366.png) 据我所知,使用SRS回源拉流只能从配置文件中指定的主节点进行回源拉流,我想请问一下当主节点有多个时能否支持边缘节点的回源拉流? Answers: username_1: 暂不支持,RTC级联比较复杂。 Status: Issue closed username_1: 先说明下词汇: * SRS Edge,边缘服务器,就是这里的SRS(边缘节点)。 * SRS Origin,援站服务器,就是这里的SRS(主节点)。 支持集群,需要如下配置: 1. Origin Cluster,配置成源站集群,具体参考文档。 2. Edge Cluster,回源指定多个Origin,具体参考文档。
caplin/FlexLayout
672999759
Title: How to run tests? Question: username_0: I tried `yarn test` and it gives me ``` $ jasmine Started No specs found Finished in 0.003 seconds Incomplete: No specs found ``` What is the proper command to run tests? Answers: username_1: The tests appear to be broken. There was only one test (for the model) and it looks like it has not been run since the update to typescript. username_0: Any plans to set up tests (and e2e tests maybe)? I can help with it probably. But tests should be run on PRs or they don't make a lot of sense. username_0: Hey @username_1. Do you want to follow up on this? I can investigate what is required to run tests on commits.
Project-OSRM/osrm-backend
313738952
Title: Exception thrown when osrm-datastore --list called with no shared memory regions exists Question: username_0: In 5.17 `osrm-datastore` has a new option `--list` that will display shared memory regions and metadata about them. However when run when no shared memory regions have been allocated, we throw an exception. ``` /osrm-datastore --list=1 terminate called after throwing an instance of 'osrm::util::exception' what(): No shared memory block 'osrm-region' found, have you forgotten to run osrm-datastore?include/storage/shared_monitor.hpp:83 ``` This should be handled more nicely - running list without shmem regions allocated is a reasonable use case.
filiprak/path-tracer
874232740
Title: Compiling errors when using the command Question: username_0: HI dear author, It's truly a honor to write a letter to you, I'm building your project nowadays and found the error when building as following, I wonder if there is a chance that you know the reason? :) ![image](https://user-images.githubusercontent.com/51076736/116843073-5e36c000-ac11-11eb-9b68-61a61e0958ae.png) thank you best regards to you William
jlippold/tweakCompatible
418127509
Title: `LocalIAPStore` working on iOS 12.0.1 Question: username_0: ``` { "packageId": "com.repo.xarold.com.localiapstore", "action": "working", "userInfo": { "arch32": false, "packageId": "com.repo.xarold.com.localiapstore", "deviceId": "iPhone8,1", "url": "http://cydia.saurik.com/package/com.repo.xarold.com.localiapstore/", "iOSVersion": "12.0.1", "packageVersionIndexed": false, "packageName": "LocalIAPStore", "category": "Tweaks", "repository": "Xarold Repo", "name": "LocalIAPStore", "installed": "1.4-2", "packageIndexed": true, "packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.", "id": "com.repo.xarold.com.localiapstore", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.1.3", "shortDescription": "Pwns In-App purchases on iOS 5.1.1 - 7.1.X. Compatible with ARMv8 (64 Bit). Enable in settings. PRESS 'DISMISS' ON POPUP!", "latest": "1.4-2", "author": "anondev", "packageStatus": "Unknown" }, "base64": "<KEY>4ifQ==", "chosenStatus": "working", "notes": "" } ```
jameslnewell/xhr-mock
254067579
Title: Global modifier (\g) in regex parameter brings the match to a fail Question: username_0: Hello, team. ## **What's Up** I want to match the set of URLs, which are different with query parameters only. To achieve this goal, I pass regex instead of string to _mock.get_ as the first parameter according to your specification `.get(url | regex, fn)` This approach misses the passed URL one time in a set of two requests if the regex value contains the global modifier (\g). ## **Example** ``` var mock = require('xhr-mock'); mock.setup(); mock.get(/(api\/users).*/g, function(req, res) { return res.status(200).body('<h1>Google</h1>'); }); // --------- var axios = require('axios'); setInterval(function() { axios({url: 'api/users'}).then( function(res) { console.log('loaded', res.data); }, function(error) { console.log('ERROR', error); } ); }, 2000); ``` As the result, we will get the error at every second request. However, if we remove the global modifier from the regex, mock will start to work as expected. Status: Issue closed Answers: username_1: "team" 🤣 I wish. Try `1.9.1`.
JuliaPackaging/BinaryBuilder.jl
720844785
Title: Infinite cycle of library linking Question: username_0: I'm getting VERY close to building PulseAudio, which I'm excited about. Everything runs smoothly, but at the end of the wizard I get infinitely repeating cycles of the following messages: ``` [ Info: Linked library libpulsecore-13.0.so has been auto-mapped to libpulsecore-13.0.so [ Info: Linked library libpulsecommon-13.0.so has been auto-mapped to libpulsecommon-13.0.so [ Info: Linked library libpulse.so.0 has been auto-mapped to libpulse.so.0 [ Info: Linked library libgdbm.so.6 has been auto-mapped to libgdbm.so.6 [ Info: Linked library libcap.so.2 has been auto-mapped to libcap.so.2 [ Info: Ignored system libraries libm.so.6, libpthread.so.0, librt.so.1, libc.so.6, libdl.so.2 [ Info: Checking lib/pulse-13.0/modules/libprotocol-http.so with RPath list SubString{String}["/workspace/destdir/lib/pulseaudio", "/workspace/destdir/lib"] ``` For reference, info to build PulseAudio is below ``` ## URL https://freedesktop.org/software/pulseaudio/releases/pulseaudio-13.0.tar.gz ## Script cd pulseaudio-13.0 sed -i 's/ -ffast-math//' src/Makefile.am aclocal automake ./configure --prefix=${prefix} --build=${MACHTYPE} --host=${target} --with-database=gdbm make make install ## Dependencies - libsndfile_jll - libatomic_ops_jll - SpeexDSP_jll - Libtool_jll - JSON_C_jll - Gettext_jll - libcap_jll - alsa_jll - Dbus_jll - Gdbm_jll - XML2_jll ``` Also, sidenote, PulseAudio gives very loud warnings about being built without udev. It looks like udev comes with systemd, and systemd is...impossible to build I think (apk doesn't even have a systemd package) so I've given up on that front. Answers: username_0: I think I might have been confused here; this works eventually. Maybe the repeated messages were just a process taking a long time? Status: Issue closed
haoxins/gulp-file-include
958363900
Title: Build only the files that will be changed Question: username_0: Some of my partials are only included on certain files and not all. So when I make changes on these partial files I only want to build files that will be changed. Is this possible to achieve this somehow?
xarbit/PrusaSlicer-flatpak
638404284
Title: USB drive/SD Card not visible from within the application when trying to export Question: username_0: Version 2.2.0 When exporting sliced gcode, the plugged in USB drive or SD card isn't accessible or visible to the app. When doing the same from Prusa supplied AppImages, this is not an issue. In the screenshot, left is from the AppImage, right is from the FlatPak ![Screenshot_20200614_131957](https://user-images.githubusercontent.com/1296705/84600912-61cb2800-ae42-11ea-8a24-17eea295ea7d.png) Answers: username_1: Confirmed. This affects me as well on Ubuntu / PopOS 20.04 as well. I think it is a permissions issue. https://github.com/prusa3d/PrusaSlicer/issues/4772
realm/realm-object-server
289709856
Title: Can't connect to ROS running on Ubuntu 16.04 (WSL) Question: username_0: Hi! I'm new to Realm and I am trying to run a first server using the Getting Started tutorial. I've a Windows 10 Pro machine, so I've installed ROS in the WSL (Windows Subsystem for Linux) running Ubuntu 16.04. Installation process went perfectly and server launches when I run 'npm start'. Sadly, in Realm Studio I'm not able to connect to the server, it hangs on the main screen with three dots animation. On server I can see no messages or errors. ## Goals Run Realm for first time! ## Expected Results Be able to connect to the realm using Realm Studio ## Actual Results Screen hangs on 'Connecting to http://localhost:9080/ in grey with animation of three dots ## Steps to Reproduce - Install Ubuntu though Microsoft Store - Install ROS as stated in docs - Create an app using 'ros init TestServer' and run it with 'npm start' - Install Realm Studio - Launch Realm Studio and try to connect ## Version of Realm and Tooling - Realm Object Server Version: 2.6.2 - Flavor: - [X] Developer - [ ] Professional - [ ] Enterprise - Server OS & Version: Ubuntu 16.04 running in Windows Subsystem for Linux - Client SDK Version: ? - Client OS & Version: ? ## Logs <!--- Please provide the output of the following commands on the (Linux) server: --> <details> <summary>systemctl status realm-object-server.service</summary> <code> No such file or directory </code> </details> <details> <summary>sudo journalctl -u realm-object-server.service</summary> <code> No journal files were found </code> </details> <details> <summary>cat /var/log/realm-object-server.log</summary> <code> No such file or directory </code> </details> Answers: username_1: Do you have a firewall that could be blocking connections? username_0: Hi! - I tried `telnet localhost 9080` and connects - i switched off my AV firewall (and checked that Windows firewall is off as well) and no Realm Studio still does not connect. username_2: Should note we have never tested ROS in Windows Subsystem for Linux so it might not be working due to that. username_0: I understand perfectly. Is there something I could do to help testing this scenario? username_3: @username_0 We are right now very focused on alleviating those getting started issues by providing a cloud service - that should help you get started in a minute :-) You can signup here: https://cloud.realm.io/ username_4: Is WSL support in the pipeline? Would love to see it as we're doing some of our development on Windows, and would like an environment that's fully local in those cases.
evanw/esbuild
869521336
Title: Improve error message on parsing await expression Question: username_0: 2 │ await f(); ╵ ^ 1 error ``` This is pretty hard to understand especially when this function definition is non-trivial. Not sure how much this factor into speed consideration since improving the error message here might involve changing how the parser works? But it would be nice to get it fixed. Thanks! Answers: username_1: Sure. This seems simple enough. Here's [the implementation from V8](https://github.com/v8/v8/blob/dc712da548c7fb433caed56af9a021d964952728/src/parsing/parser-base.h#L897-L919): ```c++ void ExpectSemicolon() { // Check for automatic semicolon insertion according to // the rules given in ECMA-262, section 7.9, page 21. Token::Value tok = peek(); if (V8_LIKELY(tok == Token::SEMICOLON)) { Next(); return; } if (V8_LIKELY(scanner()->HasLineTerminatorBeforeNext() || Token::IsAutoSemicolon(tok))) { return; } if (scanner()->current_token() == Token::AWAIT && !is_async_function()) { ReportMessageAt(scanner()->location(), flags().allow_harmony_top_level_await() ? MessageTemplate::kAwaitNotInAsyncContext : MessageTemplate::kAwaitNotInAsyncFunction); return; } ReportUnexpectedToken(Next()); } ``` So basically: if you hit a syntax error and the previous token was an `await`, complain about the `await` instead. Status: Issue closed username_1: 2 │ await f(); ╵ ~~~~~ example.js:1:0: note: Consider adding the "async" keyword here 1 │ function f() { // This should really be an async function │ ^ ╵ async 1 error ``` username_0: Thanks for such speedy turnaround 🙌🏽 ! I have so much more faith in esbuild :D
hcayless/TEI-Guidelines
70166855
Title: @require on app. crit. elements? Question: username_0: We can represent the fact that different readings exclude each other, even at different locations, using @exclude. Is there a use case for a linking element with semantics opposite to @exclude, i.e. @require? This would allow us to know that if a particular reading is applied, its dependencies should also. Might be useful in cases of syntactic agreement separated by several words.
expo/expo
380267950
Title: Local URIs for Assets are being URL-Encoded Twice Question: username_0: ### Environment ``` Expo CLI 2.3.8 environment info: System: OS: macOS 10.14 Shell: 3.2.57 - /bin/bash Binaries: Node: 8.12.0 - /usr/local/opt/node@8/bin/node Yarn: 1.12.3 - /usr/local/bin/yarn npm: 6.4.1 - /usr/local/opt/node@8/bin/npm Watchman: 4.9.0 - /usr/local/bin/watchman IDEs: Xcode: 10.1/10B61 - /usr/bin/xcodebuild npmPackages: expo: 31.0.5 => 31.0.5 react: 16.5.0 => 16.5.0 react-native: https://github.com/expo/react-native/archive/sdk-31.0.1.tar.gz => 0.57.1 react-navigation: 2.18.2 => 2.18.2 npmGlobalPackages: expo-cli: 2.3.8 Diagnostics report: https://exp-xde-diagnostics.s3.amazonaws.com/react-epfl-c994e2d3-5131-4821-b4db-df6e6dd3284f.tar.gz ``` Running on the iOS simulator. ### Steps to Reproduce I recently updated to Expo SDK 31 and it looks to me like `Asset` local URIs are being url-encoded twice. Therefore I see paths like the following: ``` .../Library/Caches/ExponentExperienceData/%2540react-epfl%252Fgraasp/ExponentAsset-00643c1d9e497681ceda68f702582183.html ``` Where `%2540` is double-encoding of `%40` and `%252F` of `%2F`. When saving the `Asset` the path with `%40` and `%2F` does exist locally, but not the one that has been double-encoded. My code, which was working with SDK 30, does the following: 1. Create an `html`-type `Asset` from a web resource. 2. Load the `Asset` in a `WebView` using its `localUri`. ### Expected Behaviour When I pass the `localUri` of an `Asset` to a `WebView` for rendering, the `WebView` should render fine. ### Actual Behaviour The `localUri` of the `Asset` seems to be encoded twice, resulting in the file not being found when the `WebView` tries to retrieve it. I get a warning like the following: ``` [15:18:50] [Unhandled promise rejection: Error: Unable to open URL: file:///.../Library/Caches/ExponentExperienceData/%2540react-epfl%252Fgraasp/ExponentAsset-6c8f176f808f51a7e8596966172f5d1d.html] - node_modules/react-native/Libraries/BatchedBridge/NativeModules.js:146:41 in createErrorFromErrorData - node_modules/react-native/Libraries/BatchedBridge/NativeModules.js:95:55 in <unknown> - ... 5 more stack frames from framework internals ``` ### Reproducible Demo Here is a reproducible demo: https://snack.expo.io/@username_0/issue-with-asset-encoding You can see in the logs how the `localUri` is being url-encoded twice. Note that although in both the `html`-type `Asset` case and the `jpg`-type `Asset` case the `localUri` is being encoded twice, it does not affect an `Image` component from rendering, but it does affect a `WebView` component from rendering. Maybe because in some cases there is double decoding going on. Also note that if you change the `WebView` to use the `uri` as a `source` instead of the `localUri`, the `WebView` renders immediately, but otherwise in the example as is, the `WebView` keeps on trying to load. Locally I get a big yellow warning, but that does not happen on the Snack. Answers: username_0: I just dug deeper and it turns out that the issue with the `WebView` is due to this issue in ReactNative: https://github.com/facebook/react-native/issues/20835 The `WebView` is now loading fine. The double-encoding is clearly still happening, so I'm not sure if I should keep this issue open or go ahead and close it. username_1: This also happens in SDK 30
llvm/circt
850570374
Title: error: unexpected error: 'rtl.module' op Found port without a name. Port names are required for Verilog synthesis. Question: username_0: A bunch of the integration tests are broken with error messages like this: ``` /home/jodemme/circt/integration_test/EmitVerilog/lint.mlir:10:1: error: unexpected error: 'rtl.module' op Found port without a name. Port names are required for Verilog synthesis. rtl.module @B(%a: i1 { rtl.inout }) -> (i1 {rtl.name = "b"}, i1 {rtl.name = "c"}) { ^ /home/jodemme/circt/integration_test/EmitVerilog/lint.mlir:10:1: error: unexpected error: 'rtl.module' op Found port without a name. Port names are required for Verilog synthesis. rtl.module @B(%a: i1 { rtl.inout }) -> (i1 {rtl.name = "b"}, i1 {rtl.name = "c"}) { ^ /home/jodemme/circt/integration_test/EmitVerilog/lint.mlir:16:1: error: unexpected error: 'rtl.module' op Found port without a name. Port names are required for Verilog synthesis. rtl.module @A(%d: i1, %e: i1) -> (i1 {rtl.name = "f"}) { ^ /home/jodemme/circt/integration_test/EmitVerilog/lint.mlir:21:1: error: unexpected error: 'rtl.module' op Found port without a name. Port names are required for Verilog synthesis. rtl.module @AAA(%d: i1, %e: i1) -> (i1 {rtl.name = "f"}) { ^ /home/jodemme/circt/integration_test/EmitVerilog/lint.mlir:26:1: error: unexpected error: 'rtl.module' op Found port without a name. Port names are required for Verilog synthesis. rtl.module @AB(%w: i1, %x: i1) -> (i1 {rtl.name = "y"}, i1 {rtl.name = "z"}) { ^ /home/jodemme/circt/integration_test/EmitVerilog/lint.mlir:26:1: error: unexpected error: 'rtl.module' op Found port without a name. Port names are required for Verilog synthesis. rtl.module @AB(%w: i1, %x: i1) -> (i1 {rtl.name = "y"}, i1 {rtl.name = "z"}) { ^ /home/jodemme/circt/integration_test/EmitVerilog/lint.mlir:32:1: error: unexpected error: 'rtl.module' op Found port without a name. Port names are required for Verilog synthesis. rtl.module @shl(%a: i1) -> (i1 {rtl.name = "b"}) { ^ -- ``` Answers: username_1: This is because I changed the representation of port names in rtl.module. I'll take a look later today when I have a chance. Sorry for the breakage! Status: Issue closed username_1: actually, this was quick. 1429a9dd
tidymodels/parsnip
431296286
Title: Mega dump of issues Question: username_0: I'm sorry this is awful but I don't have time right now to turn this into atomic and nice issues, so I'm just dumping this here because there are so many bugs and gah it was frustrating. ``` r library(tidymodels) #> -- Attaching packages --------- tidymodels 0.0.2 -- #> v broom 0.5.2.9001 v purrr 0.3.2 #> v dials 0.0.2 v recipes 0.1.5 #> v dplyr 0.8.0.1 v rsample 0.0.4 #> v ggplot2 3.1.0 v tibble 2.1.1 #> v infer 0.4.0 v yardstick 0.0.3 #> v parsnip 0.0.2 #> -- Conflicts ------------ tidymodels_conflicts() -- #> x purrr::discard() masks scales::discard() #> x dplyr::filter() masks stats::filter() #> x dplyr::lag() masks stats::lag() #> x recipes::step() masks stats::step() # you need the dev version of tidyr for the following # devtools::install_github("tidyverse/tidyr") library(tidyr) library(furrr) #> Loading required package: future set.seed(27) # the one true seed fit_on_fold <- function(spec, prepped) { x <- juice(prepped, all_predictors()) y <- juice(prepped, all_outcomes()) fit_xy(spec, x, y) } predict_helper <- function(fit, new_data, recipe) { if (inherits(new_data, "rsplit")) { obs <- as.integer(new_data, data = "assessment") new_data <- bake(recipe, assessment(new_data)) } else { obs <- 1:nrow(new_data) new_data <- bake(recipe, new_data) } predict(fit, new_data, type = "prob") %>% mutate(obs = obs) } spread_nested_predictions <- function(data) { data %>% unnest(preds) %>% pivot_wider( id_cols = obs, names_from = model_id, values_from = contains(".pred") ) } super_learner <- function(library, recipe, meta_spec, data) { folds <- vfold_cv(data, v = 10) [Truncated] sl9 <- super_learner(lib9, rec, meta, train) #> Error: Both weight decay and dropout should not be specified. predict(sl9, test) #> Error in predict(sl9, test): object 'sl9' not found lib10 <- create_library( boost_tree(mode = "classification", mtry = 3, trees = 100, min_n = 3, learn_rate = 0.1, loss_reduction = 0.1, sample_size = 1), grid_random(tree_depth, size = 5) ) sl10 <- super_learner(lib10, rec, meta, train) #> Error in xgboost::xgb.DMatrix(data = newdata, missing = NA): 'data' has class 'character' and length 1725. #> 'data' accepts either a numeric matrix or a single filename. predict(sl10, test) #> Error in predict(sl10, test): object 'sl10' not found ``` <sup>Created on 2019-04-09 by the [reprex package](https://reprex.tidyverse.org) (v0.2.1)</sup> Answers: username_1: The second-to-last error (from `xgboost::DMatrix`) is something I just hit as well. I believe it happens because the data include character columns and `as.matrix` gets called: https://github.com/tidymodels/parsnip/blob/b05d93c930a4850d571bea8c8a27292d9d4a691f/R/boost_tree.R#L340. However, the same code is used at training time so I'm not sure why the error doesn't occur until prediction. As is noted elsewhere in the code, `as.matrix` should probably be `model.matrix()`. username_0: There should be no character data in my case though since there's a `step_dummy(all_nominal())` in the recipe. Status: Issue closed username_2: A lot of these have been addressed but any that haven't, please open new issues! 🙏
codesmithtools/Templates
98779637
Title: PLINQO L2S - Runtime Configurable Table Names Question: username_0: ``` Here is my use case: I am developing a module for DotNetNuke using PLINQO. The problem is that DNN has config options to set a database table prefix (ex: "DNN_MyTable") and a database owner ("user1" instead of "dbo"). I need a way to dynamically change the table name and/or schema for my LINQ entities at runtime. Because PLINQO doesn't support this, it's impossible to use PLINQO to build commercial modules for DNN. This feature would also solve problems for other commercial applications, and scenarios where the schema name is different in each environment (dev, staging, production). I found a potential solution, but it would need to be part of the PLINQO code generation. http://weblogs.asp.net/guybarrette/archive/2008/07/23/linq-to-sql-dynamic-mappin g.aspx Thanks for your consideration! ``` Original issue reported on code.google.com by `<EMAIL>` on 13 May 2011 at 6:01<issue_closed> Status: Issue closed
kubernetes/cloud-provider-openstack
806666590
Title: [cinder-csi-plugin] Test enabling / disabling CSI migration with a running application Question: username_0: To make sure that CSI migration works as expected, please test (manually): 1. Install Kubernetes on OpenStack with the CSI driver installed + CSI migration disabled (e.g. Kubernetes 1.20, where the migration is disabled by default). 2. Run some workloads (StatefulSets?) that uses some volume(s). 3. Enable the CSI migration. 4. Check that the workloads still work with the volumes. 5. Disable the CSI migration. 6. Check that the workloads still work with the volumes. Note that enabling / disabling migration is not straightforward! There is a specific order in which the alpha feature gates must be enabled/disabled in Kubernetes components, see https://github.com/kubernetes/community/blob/master/contributors/design-proposals/storage/csi-migration.md#upgradedowngrade-migrateunmigrate-scenarios for details! CI job for this is nice to have. Answers: username_0: @username_0 will test with OpenShift / Kubernetes 1.20, keeping unassigned to get more tests with other Kubernetes vendors or even vanilla Kubernetes. username_0: /kind feature username_0: With OpenStack 4.7 (= Kubernetes 1.20) on a 3 node cluster I ran 10 StatefulSets, each with 3 replicas (=30 Pods) and enabled / disabled the CSI migration of Cinder. Apart from obvious downtime during draining nodes (to enable/disable the migration), I did not find anything obviously wrong. username_1: With Kubernetes 1.20, I ran a 5 node cluster spread across three availability zones (where compute and storage AZ are the same). I enabled Topology support in `csi-cinder-controllerplugin`, with a custom build of `csi-provisioner` (because we wait for a release). ``` kubectl get nodes --show-labels NAME STATUS ROLES AGE VERSION LABELS control-plane Ready control-plane,master 18d v1.20.2 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=b7c4fa0b-7960-4311-a86b-507dbf58e8ac,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=fra,failure-domain.beta.kubernetes.io/zone=ix1,kubernetes.io/arch=amd64,kubernetes.io/hostname=control-plane,kubernetes.io/os=linux,node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.kubernetes.io/instance-type=b7c4fa0b-7960-4311-a86b-507dbf58e8ac,topology.cinder.csi.openstack.org/zone=ix1,topology.kubernetes.io/region=fra,topology.kubernetes.io/zone=ix1 worker-es1-1 Ready <none> 18d v1.20.2 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=b7c4fa0b-7960-4311-a86b-507dbf58e8ac,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=fra,failure-domain.beta.kubernetes.io/zone=es1,kubernetes.io/arch=amd64,kubernetes.io/hostname=worker-es1-1,kubernetes.io/os=linux,node.kubernetes.io/instance-type=b7c4fa0b-7960-4311-a86b-507dbf58e8ac,topology.cinder.csi.openstack.org/zone=es1,topology.kubernetes.io/region=fra,topology.kubernetes.io/zone=es1 worker-es1-2 Ready <none> 18d v1.20.2 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=b7c4fa0b-7960-4311-a86b-507dbf58e8ac,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=fra,failure-domain.beta.kubernetes.io/zone=es1,kubernetes.io/arch=amd64,kubernetes.io/hostname=worker-es1-2,kubernetes.io/os=linux,node.kubernetes.io/instance-type=b7c4fa0b-7960-4311-a86b-507dbf58e8ac,topology.cinder.csi.openstack.org/zone=es1,topology.kubernetes.io/region=fra,topology.kubernetes.io/zone=es1 worker-ix2-1 Ready <none> 18d v1.20.2 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=b7c4fa0b-7960-4311-a86b-507dbf58e8ac,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=fra,failure-domain.beta.kubernetes.io/zone=ix2,kubernetes.io/arch=amd64,kubernetes.io/hostname=worker-ix2-1,kubernetes.io/os=linux,node.kubernetes.io/instance-type=b7c4fa0b-7960-4311-a86b-507dbf58e8ac,topology.cinder.csi.openstack.org/zone=ix2,topology.kubernetes.io/region=fra,topology.kubernetes.io/zone=ix2 worker-ix2-2 Ready <none> 16d v1.20.2 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/instance-type=b7c4fa0b-7960-4311-a86b-507dbf58e8ac,beta.kubernetes.io/os=linux,failure-domain.beta.kubernetes.io/region=fra,failure-domain.beta.kubernetes.io/zone=ix2,kubernetes.io/arch=amd64,kubernetes.io/hostname=worker-ix2-2,kubernetes.io/os=linux,node.kubernetes.io/instance-type=b7c4fa0b-7960-4311-a86b-507dbf58e8ac,topology.cinder.csi.openstack.org/zone=ix2,topology.kubernetes.io/region=fra,topology.kubernetes.io/zone=ix2 ``` I created 5 StatefulSets of 2 Pods upfront without CSIMigration, then I added 2 more with CSIMigration. In total, I enabled and disabled CSIMigration 50 times, without spotting issues. Topology worked like a charm (thanks @username_0 ). username_2: /remove-lifecycle stale username_2: @username_0 @username_1 could we close this now? Any more tests pending? username_0: I think we're fine, two independent vendors tested it /close
djozsef/openpgp-docs
739709433
Title: Creating CSR for a key stored on card Question: username_0: Please mention that during `gpgsm` execution we need to answer N to the offer to create self-signed certificate. I answered y and ended up with a certificate instead of certificate signing request :) Status: Issue closed Answers: username_1: @username_0 thank you, included your suggestion
jonreid/iOSAlertViewActionSheetUnitTesting
38355715
Title: Make CocoaPod? Question: username_0: Would it be possible to turn this into a CocoaPod to make it easier to include in projects? Thanks. Status: Issue closed Answers: username_1: I've made a pod for a new project: [MockUIAlertViewActionSheet](https://github.com/username_1/MockUIAlertViewActionSheet)
Hackebein/docker-garrysmod
624690101
Title: Workshop Collection Question: username_0: The server works without workshop collection, but the opposite isn't working. Answers: username_1: I'm using it every day with a workshop collection and it's working nice for me. (As written down for 'Example for TTT') Can you please provide more information? username_0: These are the screenshots for my Docker Compose configuration file, Digital Ocean droplet specification my UFW firewall setup. ![compose-gmod](https://user-images.githubusercontent.com/51810664/82978085-e0843080-a00d-11ea-9207-142de1822dbe.PNG) ![droplet-specification](https://user-images.githubusercontent.com/51810664/82978087-e11cc700-a00d-11ea-91ac-d38b999c6ff5.PNG) ![Capture](https://user-images.githubusercontent.com/51810664/82978138-04e00d00-a00e-11ea-912a-02a81b23815f.PNG) username_1: That log is incomplete because you missed to add "-it" in your compose file. You need to add this: ``` stdin_open: true tty: true ``` I see your TTT Collection (https://steamcommunity.com/sharedfiles/filedetails/?id=2073568192) is huge. I think the missing part in the log will show us that the server is working as expected and downloading all the workshop content. Status: Issue closed username_0: Hello, I see the logs now, it is downloading the addons from the collection as you mentioned, so I presume it is working, thank you and have a good day.
datalad/datalad
666436520
Title: Idea (failing so far): Do all necessary for ssh tests hostnames setup within our `setup_package` fixture Question: username_0: continuation from https://github.com/datalad/datalad/pull/4762#issuecomment-664504230 #### What is the problem? To run ssh tests we need to tune `/etc/hosts` to setup the `datalad-test` hostname... With #4762 it would become required for all ssh tests. In some cases I will not have such ability even if ssh is installed and I can `ssh localhost`. With this apsect (local boxes and all CIs with /etc/hosts and/or `.ssh/config` tune ups) in mind it hit me: did we ever try to just establish `~/.ssh/config` within our custom `$HOME` we use for testing? If we didn't and there is no show stopper (although looks like there is - see below), then an ultimate solution, which would avoid any box/CI "manual" configuration (of /etc/hosts and ssh) would be: - add datalad configuration items `datalad.tests.ssh_hosts.datalad_test` and `datalad.tests.ssh_hosts.datalad_test2` which would both default to `localhost:22` - in https://github.com/datalad/datalad/blob/master/datalad/__init__.py#L128 establish `~/.ssh/config` with `Host` entries for each of those `datalad-test`s (there is some `if` where we reuse GIT_HOME but I think it is purely for debian pkg building where we do not run any ssh tests anyways) - in those few tests were exact knowledge of the port is desired, it could be parsed out from config (e.g. via SSHRI) if use of `ssh -G` seems like a heavier thing to do. Then it should work as is for any local testing (even if we add more named hosts there) and for CI we would need to just provide that CIs custom configuration through the exports of two environment variables without additional helpers. But then I decided to give it a quick try to fail. <details> <summary>Q&D patch for the fixture</summary> ```shell $> git diff diff --git a/datalad/__init__.py b/datalad/__init__.py index 4504aace0..da15dc41d 100644 --- a/datalad/__init__.py +++ b/datalad/__init__.py @@ -120,11 +120,19 @@ def setup_package(): os.environ['HOME'] = new_home if not os.path.exists(new_home): os.makedirs(new_home) + os.makedirs(os.path.join(new_home, '.ssh')) with open(os.path.join(new_home, '.gitconfig'), 'w') as f: f.write("""\ [user] name = <NAME> email = <EMAIL> +""") + with open(os.path.join(new_home, '.ssh', 'config'), 'w') as f: + f.write("""\ +Host datalad-test + Port 22 + Hostname localhost + StrictHostKeyChecking no """) _TEMP_PATHS_GENERATED.append(new_home) ``` </details> failure `#1`: for some reason `ssh` simply ignores `HOME` and still tries to go for `/home/yoh/.ssh/`: ```ssh $> SSH_TTY= SSH_AUTH_SOCK= SSH_CLIENT= SSH_CONNECTION= SSH_AGENT_PID= USER=blah HOME=/home/yoh/.tmp/datalad_temp_zj4pfl0z ssh -v -v datalad-test OpenSSH_8.1p1 Debian-5, OpenSSL 1.1.1g 21 Apr 2020 debug1: Reading configuration data /home/yoh/.ssh/config debug1: /home/yoh/.ssh/config line 337: Applying options for * debug1: Reading configuration data /etc/ssh/ssh_config debug1: /etc/ssh/ssh_config line 19: Applying options for * debug1: Control socket "/home/yoh/.ssh/controls/yoh@datalad-test:22" does not exist debug2: resolving "datalad-test" port 22 ssh: Could not resolve hostname datalad-test: Name or service not known $> cat /home/yoh/.tmp/datalad_temp_zj4pfl0z/.ssh/config Host datalad-test Port 222 Hostname localhost StrictHostKeyChecking no ``` so it seems that internally `ssh` does not care about `$HOME` at all! (note: `$HOME` is *nix specific for ssh, according to https://github.com/PowerShell/Win32-OpenSSH/issues/1146 on windows there is `%userprofile%` but that one might be readonly, heh heh -- failure `#2`) So may be this idea is no go, unless someone sees some way Answers: username_1: Thank you for reporting this issue! However, at this point I am closing it at a time of ~600 open issues, in an attempt to regain control over the issue tracker. This particular issue was posted more than 6 months ago, has not received a single response, and does not describe an immediate software defect. *Closing this issue does not imply that it is not relevant*. If any reader objects the closing of this issue, please feel more than free to reopen it with an update on how it is relevant for the current state of the project, and which concrete actions to address it are ongoing or planned. Status: Issue closed
Codepoints/awesome-codepoints
274527020
Title: U+10FFFF is the last defined character Question: username_0: Private Use characters in Planes 15 and 16 are limited to all the codepoints **except** the last two of each plane, which are *noncharacters*. (U+F0000–U+FFFFD, U+100000–U+10FFFD) More here: https://en.wikipedia.org/wiki/Private_Use_Areas Answers: username_1: Agreed, thanks for the correction! I'll update that part. Status: Issue closed
jbradberry/universe
957545067
Title: Inhabited planets can build using their production and minerals Question: username_0: Inhabited planets need a queuing system to build things, particularly planetary facilities (mines and factories). - [ ] add a new system that executes planetary order queues - [ ] decrease the relevant resources for planetary orders that have been worked on - [ ] stop work when all resources are exhausted for the year - [ ] leaving partially completed items in the queue to be worked on in subsequent turns - [ ] partially completed items need to keep track of how many action resources and minerals have been consumed so far - [ ] mines may be queued and built using nothing more than action resources - [ ] factories may be queued and built using germanium plus action resources
pdm-project/pdm
1090075983
Title: distutils hack Question: username_0: Related to #623 Not a bug nor a feature request. Just a question and possible enhancement. Since a few days, I'm getting this `distutils-precedence.pth` file installed, which triggers error messages just like in #623. ```console % cat __pypackages__/3.9/lib/distutils-precedence.pth import os; var = 'SETUPTOOLS_USE_DISTUTILS'; enabled = os.environ.get(var, 'local') == 'local'; enabled and __import__('_distutils_hack').ensure_shim(); ``` ``` Error processing line 1 of /media/data/dev/mkdocstrings/__pypackages__/3.9/lib/distutils-precedence.pth: Traceback (most recent call last): File "/home/username_0/.basher-packages/pyenv/pyenv/versions/3.9.9/lib/python3.9/site.py", line 169, in addpackage exec(line) File "<string>", line 1, in <module> ModuleNotFoundError: No module named '_distutils_hack' Remainder of file ignored ``` I guess it comes from setuptools: - https://stackoverflow.com/questions/69038398/python-module-distutilis-hack - https://github.com/pypa/setuptools/blob/main/_distutils_hack/__init__.py Not sure why the `_distutils_hack` module cannot be found. Not sure if it's completely useless when using PDM, though everything works fine even when I delete this `distutils-precedence.pth` file. Answers: username_1: Are you using a `.pth` based install cache? username_0: Yes, symlinks didnt work yet so I stayed on version 1.11.3 and pth based cache. username_1: That is simply because `distutils-precedence.pth` is executed before `setuptools.pth`, which loads the path of `_distutils_hack` username_0: Hmmm. Should I raise an issue on setuptools tracker then? Not sure if this actionable at all in PDM. Feel free to close. username_1: No, it is not setuptools' fault. It is the nature of .pth file loading, they are looked one by one. Prefixing the pdm generated pth file with `aaaaa_` can fix this, but that looks ugly. Status: Issue closed username_2: I landed here when I ran into the same problem both locally and in [my CI](https://gitlab.com/username_2/flufl.lock/-/jobs/1950193258). I can see that @username_1 had some suggested workaround in #623 but that does not seem to help when testing locally because: * The error is comming from setuptools * `pdm config --local feature.install_cache` already returns `False` * I'm running from a clean git checkout (i.e. after `git clean -dxffi`) I'm running under tox with multiple versions of Python, from 3.7 to 3.10. ``` py37 run-test-pre: PYTHONHASHSEED='2171121435' py37 run-test: commands[0] | pdm config --local use_venv True py37 run-test: commands[1] | pdm config --local feature.install_cache False py37 run-test: commands[2] | pdm install -G testing ⠋ Resolving packages from lockfile... ⠙ Resolving packages from lockfile... ⠹ Resolving packages from lockfile... ⠸ Resolving packages from lockfile... Synchronizing working set with lock file: 14 to add, 2 to update, 0 to remove ⠋ Updating setuptools 60.1.1 -> 60.3.1... ⠙ Updating setuptools 60.1.1 -> 60.3.1... ⠹ Updating setuptools 60.1.1 -> 60.3.1... ⠸ Updating setuptools 60.1.1 -> 60.3.1... Error processing line 1 of /Users/barry/projects/flufl/lock/.tox/py37/lib/python3.7/site-packages/distutils-precedence.pth: Traceback (most recent call last): File "/usr/local/lib/python3.7/site.py", line 168, in addpackage exec(line) File "<string>", line 1, in <module> ModuleNotFoundError: No module named '_distutils_hack' Remainder of file ignored Error processing line 1 of /Users/barry/projects/flufl/lock/.tox/py37/lib/python3.7/site-packages/distutils-precedence.pth: Traceback (most recent call last): File "/usr/local/lib/python3.7/site.py", line 168, in addpackage exec(line) File "<string>", line 1, in <module> ModuleNotFoundError: No module named '_distutils_hack' Remainder of file ignored ⠼ Updating setuptools 60.1.1 -> 60.3.1... Error processing line 1 of /Users/barry/projects/flufl/lock/.tox/py37/lib/python3.7/site-packages/distutils-precedence.pth: Traceback (most recent call last): File "/usr/local/lib/python3.7/site.py", line 168, in addpackage exec(line) File "<string>", line 1, in <module> ModuleNotFoundError: No module named '_distutils_hack' Remainder of file ignored Error processing line 1 of /Users/barry/projects/flufl/lock/.tox/py37/lib/python3.7/site-packages/distutils-precedence.pth: Traceback (most recent call last): File "/usr/local/lib/python3.7/site.py", line 168, in addpackage exec(line) File "<string>", line 1, in <module> ModuleNotFoundError: No module named '_distutils_hack' Remainder of file ignored [Truncated] processing line 1 of /Users/barry/projects/flufl/lock/.tox/py37/lib/python3.7/site-packages/distutils-precedence.pth: Traceback (most recent call last): File "/usr/local/lib/python3.7/site.py", line 168, in addpackage exec(line) File "<string>", line 1, in <module> ModuleNotFoundError: No module named '_distutils_hack' Remainder of file ignored ⠧ Updating setuptools 60.1.1 -> 60.3.1... ⠇ Updating setuptools 60.1.1 -> 60.3.1... ⠏ Updating setuptools 60.1.1 -> 60.3.1... ⠋ Updating setuptools 60.1.1 -> 60.3.1... ⠙ Updating setuptools 60.1.1 -> 60.3.1... ⠹ Updating setuptools 60.1.1 -> 60.3.1... ⠸ Updating setuptools 60.1.1 -> 60.3.1... ✔ Update setuptools 60.1.1 -> 60.3.1 successful ``` I see this issue was closed, but I don't know if that's available in pdm 1.12.2, which is I think the latest version on PyPI and in brew. username_1: How can I reproduce it locally? Does it fail on all python versions from 3.7 to 3.10? I am using PDM 1.12.3 and Python 3.10 but can't reproduce.