repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
cyberway/cyberway
483177787
Title: Replaying of mainnet fails on nodeosd version 2.0.0 Question: username_0: **STR** 1. Configure to connect to mainnet 2. Start resync **Actual result** Resync fails on block 121492 with error: ``` error 2019-08-21T02:58:11.318 nodeos producer_plugin.cpp:319 on_incoming_block ] 3030000 block_validate_exception: Block exception receipt does not match {"producer_receipt":{"status":"executed","cpu_usage_us":10186,"net_usage_words":31,"ram_kbytes":1,"storage_kbytes":1,"trx":[1,{"signatures":["<KEY>","<KEY>"],"compression":"none","packed_context_free_data":"","packed_trx":"1bf95a5d90da2a0819b700000000000002001a76f1e80a706400000000a84c77d501304b4279f82a98b3000000804d9731ad57304b4279f82a98b37076e0b8179476fc4472652d63796265727761792d73696c6e79692d692d6e6164657a686e79692d626c6f6b636865696e2d313536353938343639323334372d3135363632343237353834393210270000000080ab8e470000e04725b7e9ad0100000000000070640000e04725b7e9ad100000000000007064304b4279f82a98b300"}]},"validator_receipt":{"status":"executed","cpu_usage_us":10186,"net_usage_words":31,"ram_kbytes":1,"storage_kbytes":2,"trx":[1,{"signatures":["SIG_K1_KgdX195w<KEY>","SIG_K1_<KEY>"],"compression":"none","packed_context_free_data":"","packed_trx":"1bf95a5d90da2a0819b700000000000002001a76f1e80a706400000000a84c77d501304b4279f82a98b3000000804d9731ad57304b4279f82a98b37076e0b8179476fc4472652d63796265727761792d73696c6e79692d692d6e6164657a686e79692d626c6f6b636865696e2d313536353938343639323334372d3135363632343237353834393210270000000080ab8e470000e04725b7e9ad0100000000000070640000e04725b7e9ad100000000000007064304b4279f82a98b300"}]}} nodeos controller.cpp:1284 apply_block {} nodeos controller.cpp:1314 apply_block rethrow {} nodeos controller.cpp:1357 push_block stacktrace: 0# 0x0000000000E6300C in /opt/cyberway/bin/nodeos 1# 0x00000000005D0EDE in /opt/cyberway/bin/nodeos 2# 0x0000000000AA5AE7 in /opt/cyberway/bin/nodeos 3# 0x0000000000ABC835 in /opt/cyberway/bin/nodeos 4# 0x0000000000A9B393 in /opt/cyberway/bin/nodeos 5# 0x0000000000A1A43B in /opt/cyberway/bin/nodeos 6# 0x00000000005D430E in /opt/cyberway/bin/nodeos 7# 0x0000000000882D3B in /opt/cyberway/bin/nodeos 8# 0x00000000006FDEEB in /opt/cyberway/bin/nodeos 9# 0x00000000007287F5 in /opt/cyberway/bin/nodeos 10# 0x00000000006EF68B in /opt/cyberway/bin/nodeos 11# 0x000000000071D6D1 in /opt/cyberway/bin/nodeos 12# 0x000000000047FE4D in /opt/cyberway/bin/nodeos 13# 0x0000000000473D7F in /opt/cyberway/bin/nodeos 14# __libc_start_main in /lib/x86_64-linux-gnu/libc.so.6 15# 0x000000000047104A in /opt/cyberway/bin/nodeos ``` **Expected result** No errors<issue_closed> Status: Issue closed
CartoDB/mobile-sdk
1074318995
Title: handling cross tile polygons labels Question: username_0: In my custom version of openmaptiles i have landuse polygons which can have a name. I have an issue while rendering those labels when polygon is split on 2 tiles. The label render twice. What i "would like" is to get one label in the middle of the polygon. I can try use min text distance but it will not always work depending on the polygon size. You can see the tiles limits in openstreetmap (thin white lines) <img width="1284" alt="Screenshot 2021-12-08 at 11 57 36" src="https://user-images.githubusercontent.com/655344/145197377-7bd0a32c-ce2f-4194-955f-011106cbcbe9.png"> Answers: username_1: This is problematic with tiles, unfortunately. That is also the main reason why there are layers like 'water_name', 'transportation_name' in vector tiles in addition to 'water' and 'transportation'. In case the plygon feature includes id (or perhaps you can use 'name' field as id), the label should be rendered once but it may jump around when zooming or panning the map. username_0: @username_1 ok you confirm what i was thinking. Did not really want to add another layer just for those names as it would make the tiles bigger but i get it. Now about your point about id indeed landcover and landuse dont have id (for size reasons i would guess). All you say makes sense and i understand you cant do much more. I ll stick with this for now. Thanks for the explanation Status: Issue closed
dotnet/efcore
1003949121
Title: Can't generate a migration using HiLo Question: username_0: When I generate the migrations, I have this error for all the primary keys that are configured to use HiLo ![image](https://user-images.githubusercontent.com/4983925/134301252-85724a7f-26a2-48a3-a8a5-1163b5e78ba4.png) ### Include provider and version information EF Core version: 6.0.0-rtm.21471.18 Database provider: Microsoft.EntityFrameworkCore.SqlServer Target framework: .NET 6.0 Operating system: Windows 10 Pro IDE: Visual Studio 2019 16.3 Answers: username_1: Thanks, I can see the bug and have submitted #26136 to fix this for 6.0. username_0: Thank you @username_1 Status: Issue closed username_1: When I generate the migrations, I have this error for all the primary keys that are configured to use HiLo ![image](https://user-images.githubusercontent.com/4983925/134301252-85724a7f-26a2-48a3-a8a5-1163b5e78ba4.png) ### Include provider and version information EF Core version: 6.0.0-rtm.21471.18 Database provider: Microsoft.EntityFrameworkCore.SqlServer Target framework: .NET 6.0 Operating system: Windows 10 Pro IDE: Visual Studio 2019 16.3 username_1: As a workaround, you can replace SqlServerModelBuilderExtensions with SqlServerPropertyBuilderExtensions on all property configurations. Status: Issue closed
neillturner/kitchen-puppet
97048356
Title: Kitchen: Message: Failed to complete #converge action: [undefined local variable or method `modules_path' for Question: username_0: Seems like 0.0.29 release is broken: ``` E, [2015-07-24T12:50:03.996708 #25439] ERROR -- Kitchen: Class: Kitchen::ActionFailed E, [2015-07-24T12:50:03.996737 #25439] ERROR -- Kitchen: Message: Failed to complete #converge action: [undefined local variable or method `modules_path' for #<Kitchen::Provisioner::PuppetApply:0x007f913e1752d8>] E, [2015-07-24T12:50:03.996765 #25439] ERROR -- Kitchen: ---Nested Exception--- E, [2015-07-24T12:50:03.996839 #25439] ERROR -- Kitchen: Class: NameError E, [2015-07-24T12:50:03.996889 #25439] ERROR -- Kitchen: Message: undefined local variable or method `modules_path' for #<Kitchen::Provisioner::PuppetApply:0x007f913e1752d8> E, [2015-07-24T12:50:03.996921 #25439] ERROR -- Kitchen: ------Backtrace------- E, [2015-07-24T12:50:03.996949 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/kitchen-puppet-0.0.29/lib/kitchen/provisioner/puppet_apply.rb:535:in `modules' E, [2015-07-24T12:50:03.996977 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/kitchen-puppet-0.0.29/lib/kitchen/provisioner/puppet_apply.rb:814:in `prepare_modules' E, [2015-07-24T12:50:03.997006 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/kitchen-puppet-0.0.29/lib/kitchen/provisioner/puppet_apply.rb:361:in `create_sandbox' E, [2015-07-24T12:50:03.997032 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/provisioner/base.rb:61:in `call' E, [2015-07-24T12:50:03.997061 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/instance.rb:366:in `block in converge_action' E, [2015-07-24T12:50:03.997089 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/instance.rb:488:in `call' E, [2015-07-24T12:50:03.997117 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/instance.rb:488:in `synchronize_or_call' E, [2015-07-24T12:50:03.997144 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/instance.rb:453:in `block in action' E, [2015-07-24T12:50:03.997169 #25439] ERROR -- Kitchen: /opt/rbenv/versions/2.1.6/lib/ruby/2.1.0/benchmark.rb:279:in `measure' E, [2015-07-24T12:50:03.997196 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/instance.rb:452:in `action' E, [2015-07-24T12:50:03.997224 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/instance.rb:362:in `converge_action' E, [2015-07-24T12:50:03.997254 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/instance.rb:341:in `block in transition_to' E, [2015-07-24T12:50:03.997330 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/instance.rb:340:in `each' E, [2015-07-24T12:50:03.997363 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/instance.rb:340:in `transition_to' E, [2015-07-24T12:50:03.997390 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/instance.rb:160:in `verify' E, [2015-07-24T12:50:03.997416 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/instance.rb:189:in `block in test' E, [2015-07-24T12:50:03.997442 #25439] ERROR -- Kitchen: /opt/rbenv/versions/2.1.6/lib/ruby/2.1.0/benchmark.rb:279:in `measure' E, [2015-07-24T12:50:03.997470 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/instance.rb:185:in `test' E, [2015-07-24T12:50:03.997504 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/command.rb:176:in `public_send' E, [2015-07-24T12:50:03.997531 #25439] ERROR -- Kitchen: /my/user/home/.bundle/gems/test-kitchen-1.4.1/lib/kitchen/command.rb:176:in `block (2 levels) in run_action' ``` Answers: username_1: I guess we should not merge pull requests ( https://github.com/username_3/kitchen-puppet/pull/58 ) without rspec tests anymore username_2: I completely missed that there weren't any tests. Sorry about this regression, I should have reviewed that better. I'll revert that PR, so master is clean and ask @username_3 if we can pull the 0.0.29 release. Apologies for any inconvenience this caused username_3: i didn't have time to test. just reviewed code. i will yank the gem. -) username_2: Thanks @username_3. Could you also revert the PR? #58 Turns out I don't have revert perms :) username_3: reverted. username_0: thanks guys! username_3: close as codd reverted. not sure what i'd do about the contribution Status: Issue closed
fluttercommunity/flutter_sms
623971475
Title: Documentation improvements Question: username_0: Gentlefolk. Can I ask for some documentation improvements. 1) what is the difference between sendSMS and launchSMS? 2) what actions are required to support dual sim cards. Answers: username_1: 1. There should be a doc comment that explains it for each method 2. No support at this time but PRs are welcome 3. General log messages Status: Issue closed username_1: sendSMS is for using the url scheme, and launchSMS is for opening the native dialog
SecurityInnovation/Smart-Contract-CTF
314363552
Title: Top-left menu link to SI homepage is broken Question: username_0: The top-left corner link to the SI home page is a messed up anchor link as it starts with a ``#``, needs to be absolute. Currently links to: ``http://blockchain-ctf.securityinnovation.com/#/https://securityinnovation.com`` Rendered code: ```<a target="_blank" class="pure-menu-link" href="#/https://securityinnovation.com">Security Innovation</a>``` Answers: username_1: Fixed in commit 70d441221f2a81f16cad790ef5685e82d25beaf0 Status: Issue closed
kubernetes/kubernetes
920409863
Title: Container Lifecycle Hooks exit on non-0 exit code Question: username_0: Hi, i don't understand why Container Lifecycle Hooks are instantly killed by Kubernetes (Exit Code 137) when some command is used that returns a non-0 exit code. Sometimes this is expected behavior and i don't understand why this is not possible with Kubernetes. Answers: username_1: /kind support /sig node username_2: https://kubernetes.io/docs/tasks/configure-pod-container/attach-handler-lifecycle-event/#discussion. The Container's state will only change to Running when the postStart hook is successfully executed, so I don't understand what your question is? username_2: /triage needs-information username_3: Kubernetes does not use issues on this repo for support requests. If you have a question on how to use Kubernetes or to debug a specific issue, please visit our [forums](https://discuss.kubernetes.io/). It looks like we have some documentation for this already as linked by @username_2 but I will also mark this with a docs tag. /kind documentation /close
snapframework/snap-server
52068779
Title: Allow custom formats for the Access logs; like JSON. Question: username_0: I have been reading through the code and I am pretty sure that I need to override the logA method in order to get custom logging support. I [asked a question on SO about it just in case][1]. But I don't think that I can currently override this logA functionality without changing the code inside of Snap Server. If this is correct then, for this issue, it would be great if I could inject a method into the snap server configuration that would intercept the access log generation and let me write my own access log lines in a custom format. [1]: http://stackoverflow.com/questions/27496925/haskell-snap-framework-configuring-custom-logging Answers: username_0: @username_1 I'm currently working on this code in a branch because I really really want and need this functionality because it would allow me to output all of my logs in JSON format. If I were to write this code such that it could be merged into both 0.9-stable and master would you accept both PR's? You can see my initial attempts here: https://github.com/snapframework/snap-server/compare/0.9-stable...username_0:issue/62-custom-access-and-error-log-handlers username_0: Please have a look at my initial attempts and let me know if I am going in the right direction. I'd like feedback now before I spin my wheels too fast in one direction. :D username_0: I have tested this branch against my Snap application My Reminders and it is working: ![My Reminders JSON log](https://s3.amazonaws.com/uploads.hipchat.com/10804/83693/PXjauYS0wpoxurf/Screen%20Shot%202015-05-20%20at%206.21.24%20pm.png) So my branch seems to be workable to accomplish the goal of logging your access and error logs in a custom format. You can see an example of me using this branch in the My Reminders code here: https://bitbucket.org/atlassianlabs/my-reminders/branch/issue/MR-7-json-logs-for-my-reminders#diff username_0: Ping. These PR's have been open for a week now without comment. Sorry that I'm pinging so soon but: 1. I suspect that there may still be more work to do on these PR's; especially if you want me to use the OutputStream code. 1. The sooner this code gets released the sooner that I can hookup better Logstash / Kibana logging to my services. Which I think they desperately need. Cheers and sorry if I'm commenting too frequently! Just tell me to stop. :smile: I won't mind! username_1: I was on vacation all last week and didn't get to it on the weekend because I have a houseguest. I'll take a look today. username_0: Wow, I am such a pest. I am really sorry. Thank you so much for looking at this so soon. In my defence I guess I'm just excited about submitting changes back to snap. username_1: Believe me, I'm quite happy you sent the patches :) username_0: control your log file output format. The two functions are setAccessLogHandler and setErrorLogHandler. These methods respectively will let you control how the access and error log lines are rendered respectively. This is useful if you wist to modify the log line output; for example, if you wish to log in a custom data format (like JSON) then now you can. For more information please read the docs. What do you think? I was trying for short and to the point. username_0: @username_1 Also, I have bumped the snap-server version to 0.9.6 because I have only made additive changes. Not sure if anybody else has made breaking changes that would require a bump to 0.10.0? At any rate, you can see that my additions to 0.9-stable still build against snap 0.14-stable: https://bitbucket.org/snippets/username_0/rkpK7 Cheers! username_1: See my comment on the changeset, we will need a major bump for this. Your initial wording sounds OK to me, but the best way to handle this will be to open a pull request on snap-website, we can discuss the minutae there. username_0: I have responded: https://github.com/snapframework/snap-server/pull/71/files#r31794779 I hope that changing the Config instance is not enough to prevent a backport. After all, it implements Monoid and even the docs recommend to create it with the monoid instance.
coleifer/huey
525804822
Title: Cancel a long running task? Question: username_0: Hi @username_1 Please advice how to correctly cancel a long running task. There is `CancelExecution` exception but it works only in pre-execute hooks. I could kill the worker, but if the task has a retries, it will start again. Answers: username_1: There's no way to do that, unfortunately. You can't easily interrupt a running thread, and it adds a lot of complexity -- should an interrupted task be marked as an error? Should it be retried? What about subsequent executions? The solution is to modify your task to track its time and cancel itself if it has the potential to run for a long time. You can do this with socket timeouts if it's making a slow network call, or if there is a loop to perform a check during the loop body. Status: Issue closed
CRAlpha/react-native-wkwebview
195732031
Title: clearCacheForSingleFile is not work .and How set cache mode??? Question: username_0: usr play three website'ex: http://xxx.com/gamex.html', i want remove one cache .so ,i use clearCacheForSingleFile, like this: clearCacheForSingleFile('http://xxx.com/gamex.html'); ??? it's not work fine~ how i do ,pls help ~ _ ~<issue_closed> Status: Issue closed
greenplum-db/gpdb
264481240
Title: Incorrect result with CUBE Question: username_0: On GPDB: ``` postgres=# select x,y,count(*), grouping(x), grouping(y),grouping(x,y) from generate_series(1,1) x, generate_series(1,1) y group by cube(x,y); x | y | count | grouping | grouping | grouping ---+---+-------+----------+----------+---------- 1 | 1 | 1 | 0 | 0 | 0 | 1 | 1 | 1 | 0 | 2 | | 1 | 1 | 0 | 2 1 | | 1 | 0 | 1 | 1 (4 rows) ``` PostgreSQL produces a different result: ``` postgres=# select x,y,count(*), grouping(x), grouping(y),grouping(x,y) from generate_series(1,1) x, generate_series(1,1) y group by cube(x,y); x | y | count | grouping | grouping | grouping ---+---+-------+----------+----------+---------- | | 1 | 1 | 1 | 3 1 | 1 | 1 | 0 | 0 | 0 1 | | 1 | 0 | 1 | 1 | 1 | 1 | 1 | 0 | 2 (4 rows) ``` Note the different values in the last column. I believe PostgreSQL got this right, and the GPDB result is incorrect. Answers: username_1: @dhanashreek89 @hsyuan Can we please get eyes on this - it's a fix for a wrong results. @schubert for awareness username_2: Still exist on latest master username_0: This was fixed on master with the PostgreSQL merge. Status: Issue closed
ScottIsAFool/Bex
106560736
Title: Location info not retrived on activity Question: username_0: Hi, i have a problem with your library, i have added the ReadActivityLocation scope to login and MapPoints to request but the location in mapPoint object have all values (latitude,longitude ecc) to 0 Thank in advance! Answers: username_1: Are you able to provide a quick repro? Or some sample code that you're using so I can give it a try? username_0: Thank for your response. for try you login with with all scope App.BexClient.CreateAuthenticationUrl(new List<Scope> { Scope.ActivityHistory, Scope.ActivityLocation, Scope.Devices, Scope.Profile,Scope.offline_access }); then you can retrive activities with ActivitiesRequest ar1 = new ActivitiesRequest(); ar1.MaxItemsReturned = 25; ar1.ActivityTypes = new string[] { "Run" }; ar1.ActivityFieldsToInclude = new ActivityFields[] { ActivityFields.Details, ActivityFields.MapPoints }; ActivitiesResponse ar = await App.BexClient.GetActivitiesAsync(ar1); than you can see MapPoints in ar.RunActivities have all Latitude and Longitude to 0 Thanks again! username_0: Hi, have you reproduced the problem, i can do anything for help you? username_1: Hi, sorry, not yet. Will try and look at it tonight or tomorrow (been a bit busy this week) username_0: no problem it's only for help you :;D username_0: Some news? username_1: I've had a look, and the run data I'm getting back isn't even including any of the location data. Looking around, it looks like this is a fault of the MS Health API. I think when they fix it on their end, it will just start working for you.
egoist/ideas
552767782
Title: My goal for open source in 2020 Question: username_0: ## My goal for 2020: Build an application framework - Support both React and Vue (perhaps Svelte too) - File-system based routing, like many PHP frameworks and Next.js / Nuxt.js, there will be two kinds of pages: - Component pages: `pages/**/*.{vue,js,jsx,ts,tsx}` - Server pages: `pages/**/*.{json,html,xml,etc}.{js,ts}`, used as http handlers to output other types of content, like API response or RSS feed. - Support multiple rendering modes: - SSR: by default your app is server-rendered and rehydrated on the client-side - SSG: it can also be statically generated if you can't deploy it as a Node.js app - SPA: like SSG but server-side rendering is disabled so you don't need to write SSR-compatible code [__You can support me to make this idea become reality__](https://github.com/sponsors/username_0), I'm confident that this will benefit a lot of developers (including me myself) and your support will definitely accelerate it!
ironjan/klausurtool-ror
158652839
Title: Hide exams that are not in a folder Question: username_0: ## How does this problem happen? 1. First edit an exam so that it is not shown in the TOC, i.e. it's not yet in the folder 2. Then go to the exam search and search for that exam ## What does happen? The exam is listed as part of that folder. ## What do you expect to happen instead? The exam should not be listed as part of that folder.
Tarskin/HappyTools
323642507
Title: Scientific notation numbers in processed data files Question: username_0: HappyTools attempts to cast numbers into floats, with a (power)-user defined precision, however it is possible that the intensity contains a number using scientific notation that causes HappyTools to crash. Answers: username_0: The latest build (b180516a) has fixed this bug. Status: Issue closed
DIPSAS/DIPS.Xamarin.UI
1096079200
Title: [Enhancement] Sheet could intercept gesture when scrollable content reaches the top Question: username_0: ## Summary If the content of the sheet is scrollable and the user scrolls to the top the sheet view could intercept the gesture and start moving downwards. Then the user can get the feel of scrolling the whole sheet away. This would be similar to the sheet views in Apple Maps. ## API Changes Should be able to toggle this behaviour with a bool property. Could use the existing `InterceptDragGesture`. ## Intended Use Case Sheet views are often used to show scrollable content. Allowing the user to smoothly scroll away the sheet when finished would be beneficial
sensu/sensu-go
380296286
Title: Determine how to handle top-level endpoints in RBAC Question: username_0: We need to figure how how to handle permissions for top-level endpoints (e.g. /health). I'm not quite sure how Kubernetes does it. Child of #2243 Answers: username_0: We currently have a single "top-level" endpoint (/health), which does not require authentication, so I'm closing this issue! Status: Issue closed
tomwojcik/starlette-context
664620110
Title: PluginUUIDBase's force_new_uuid option seems broken Question: username_0: When using `PluginUUIDBase`'s `force_new_uuid` option and setting it to `True` I'm always getting the same `uuid`. Which is the **opposite** of what I expect. It looks like when `self.value is None` a new `uuid` is generated, and this works fine when for the first request. But subsequent request seem to be using the same id. I suspect the code that needs to be fixed is missing a `self.value = None` before trying anything else here: https://github.com/username_1/starlette-context/blob/2f80262bbf4c00fb501c22ac04a5c1c408187fe6/starlette_context/plugins/plugin_uuid.py#L34 Answers: username_0: Thinking about this, why is `self.value` being used anyway? I don't understand why this is being stored on the plugin instance. It seems like a normal variable would be enough and match the intend better, and it would have avoided this bug. username_1: Hey, thanks for opening this ticket. Indeed that's a bug. I'll fix this within a few days. username_1: 0.2.3 is published. Please try again and let me know if it's all good now. username_0: Looks good from here! Thank you for the quick fix. FTR my workaround was: ```python # Fix https://github.com/username_1/starlette-context/issues/15 __original_extract_value_from_header_by_key = ( PluginUUIDBase.extract_value_from_header_by_key) async def __patched_extract_value_from_header_by_key(self, request): self.value = None return await __original_extract_value_from_header_by_key(self, request) PluginUUIDBase.extract_value_from_header_by_key = ( __patched_extract_value_from_header_by_key ) ``` **Removing** this from my code, and using `0.2.3` all my tests still pass. :fireworks: For me it looks good! Status: Issue closed username_1: Awesome! Thanks for checking that out.
apache/airflow
596276385
Title: Create guide for Dataproc operators Question: username_0: **Description** Hello, A guide that describes how to use Dataproc service operators would be useful. We have an example DAG for this service, so the guide should not be a big challenge. If anyone is interested in this task, I am willing to provide all the necessary tips and information. Other guides are available: https://airflow.readthedocs.io/en/latest/howto/operator/index.html Best regards, Kamil **Use case / motivation** N/A **Related Issues** N/A Answers: username_1: Happy to pick this up. Worked with the service a bit before. username_0: @username_1 I assigned you to this task. Status: Issue closed username_2: solved in https://github.com/apache/airflow/pull/9037
2amigos/yii2-usuario
243181971
Title: Provide the ability for a user to delete its very own account Question: username_0: Specs: - Should be configurable at the Module level - The action should fire events so developers could programmatically develop their very own action log systems. Example actions to be taken on before or after delete events: a) on before delete the user instance that is about to remove himself from the system will be passed to the event object, so the developer can easily stop its plans, clone its data to a different tables (for example a CRM system, so team could follow up on the reasons why he wishes to leave the system). b) another scenario would be to report the sys admin about the action of the user<issue_closed> Status: Issue closed
jcabi/jcabi-heroku-maven-plugin
53771716
Title: Copyright section is outdated Question: username_0: Copyright sections still feature 2014 as a year. Answers: username_1: we'll find someone to do this task, soon username_1: @username_2 it's in your hands now, please proceed username_1: @username_0 thanks for reporting! I topped your account for 15 mins, transaction 49997556 username_2: @username_1 @username_0 I created PR #5 but haven't seen activity to get it reviewed. Anything I should do? username_3: @username_2 we'll find a reviewer soon, thanks for the PR! username_2: @username_0 PR #5 has been merged. Please close the issue. Status: Issue closed username_1: @username_2 thanks, paid, **30 mins** to your account, payment ID is `AP-5VJ27404GD162805S` +30 added to your rating, current score is: [+130](http://www.netbout.com/b/35107?open=rating)
withspectrum/spectrum
311687088
Title: Handling spam communities Question: username_0: <!-- FILL OUT THE FORM BELOW OR THE ISSUE WILL BE AUTO-CLOSED **Issue Type (check one)** - [ ] Bug Report - [ ] Feature Idea - [x] Technical Discussion - [ ] Question (these will be auto-closed, please ask them on Spectrum instead https://spectrum.chat/spectrum/open) **Description (type any text below)** --> Our SEO is getting good. Which means people will start spinning up things like this: https://spectrum.chat/cbd-oil?thread=2d09eb88-434a-4c0d-8181-a47667ee5ef3 to try and get Google karma for the links. How do we want to handle spam communities + threads like this? Answers: username_1: Good question, I wonder what other user-generated content sites are doing about that username_0: Probably nothing, as a lot of other ugc sites are ad-driven, so pageviews === good. I'm not sure there's anything here to do immediately, except watch for patterns here. If it starts to scale people will trust our content/platform less in general. username_2: [These](https://webmasters.googleblog.com/2017/01/protect-your-site-from-user-generated.html) might be few best practices. but there should be an efficient way to handle spam at scale. username_0: Quick wins from that post (thanks @username_2!) - rate limit # of threads that can be posted in a given time period - rate limit # of messages that can be posted in a given time period - add nofollow to links posted in chat messages Bigger tasks that we need anyways: - community option to approve all incoming threads before they are published - implement additional spam detection systems like akismet (https://github.com/cedx/akismet.js) - we'll just need to handle false positive cases so people don't lose content they post if it is legitimate - add ability to block user from spectrum at a system level Biggest task that would be nice to have: - internal admin page that collects all inbound toxicity + spam reports, with quick actions for deleting + blocking content username_1: I think we already do this? username_1: This ties into #2451, right? username_0: do we? unsure username_0: Good call on #2451 - will build on top of that username_0: First draft to work on this going out in #2758 Status: Issue closed
jhaals/yopass
356886212
Title: HTTPS Question: username_0: Hi, I'm trying to run yopass on port 443 but setting it like below doesn't work. It does on 1337, Could you advise please? docker run --name memcached_yopass -d memcached docker run -p 443:443 -v /local/certs/:/certs \ --link memcached_yopass:memcache -d username_1/yopass -memcached=memcache:11211 -tls.key=/certs/my.key -tls.cert=/certs/my.crt docker run -p 443:443 --link memcached_yopass:memcache -d username_1/yopass -memcached=memcache:11211 Status: Issue closed Answers: username_1: Hi @username_0 You can either map the correct port using docker with `-p 443:1337`, see https://docs.docker.com/config/containers/container-networking/ Or tell yopass to listen on port 443 with `-port 443` username_0: Thank you username_1! That is perfect!
FFCK/compet-ffck
278504935
Title: Modification des inscriptions Cloud en niveau 1 Question: username_0: pour une connexion de niveau 1 (gestionnaire sportif régional) on peut modifier les inscriptions cloud de toutes les manifestations grâce au clic droit - ce qui ne devrait pas être possible. Les boutons d'accès direct sont bien vérouillés ![image](https://user-images.githubusercontent.com/33895732/33490047-13a2f5d8-d6b6-11e7-8b82-1febc0f749b8.png)
pytorch/pytorch
402366366
Title: Unexpected result with tensor::Tensor::eq_ in C++ Question: username_0: tensor([0, 0, 0], dtype=torch.int32) ``` ## Environment This happens on MacOs and Linux testes with latest file from pytorch website. Answers: username_1: This has been fixed in https://github.com/pytorch/pytorch/pull/15479, and is available in the nightly version of LibTorch. Below is the output I get with the nightly version: ``` x: 1 2 3 [ Variable[CPUIntType]{3} ] y: 0 0 0 [ Variable[CPUIntType]{3} ] 0 0 0 [ Variable[CPUIntType]{3} ] ``` Status: Issue closed
zkSNACKs/WalletWasabi
802773013
Title: Ledger Nano X - Unknown Device/Hardware Question: username_0: When trying to load the ledger nano x via Wasabi it shows as Unknown. Click to load it anyway brings up an unknown hardware error. I've attached images below: ![Error 2](https://user-images.githubusercontent.com/78661127/107127967-d101fa80-68b1-11eb-985a-cb7d9d6f3304.png) ![Error](https://user-images.githubusercontent.com/78661127/107127968-d19a9100-68b1-11eb-8103-84fc5de41611.png) Currently running Windows 10 but I've tried on several devices and a Macbook and it's the same issue. Using the latest Wasabi Wallet, version 1.1.12.3 Answers: username_1: Unfortunately, like you said, Nano X is not officially supported. I'll close this issue. Status: Issue closed
fig02/Better-OoT
438899186
Title: It would be nice to have a Toggle for all the Default changes Question: username_0: A lot of these features are always applied without the option to toggle them off. Most people who use the app now will always want them. But giving the option to disable them can open up some interesting uses for BOoT. For example, OoT Online will only support version 1.0. So in other words, Master Quest is not supported in OoT Online. However, using BOoT, you can patch 1.0 to have the MQ dungeons. Circumventing this problem entirely! But a lot of casual players might want to play mostly vanilla, and still have the option to skip owls, or have D-Pad Boots, or a few cosmetic changes, etc. A single toggle for all of the default features would open up the app to many more people. Of course, ideally, every tiny little feature would have a toggle. But that gets a bit tedious. It's a quick fix to do them all with a single toggle. Although, If you did want to do that, you could make some sort of "Advanced" tab for disabling default features one at a time. People love customizing their experience. Answers: username_1: Hi Thank you for your suggestion. This tool was designed specifically for speedrunning, and the settings options are perfect for what it was designed for. I definitely see how it could be opened up to serve more purposes, but I will not be the one to implement that since this fits me and my community's needs already. Sorry and thank you for understanding Status: Issue closed
elastic/elasticsearch-php
286693124
Title: Getting blank hits on Search Question: username_0: ### Hi I am getting empty result when search a document. Using browser search it's working fine but using elasticsearch-php client not working. ##Request: $params = [ 'index' => 'response_packets_index_v2', 'type' => 'documents', 'body' => [ 'query' => [ 'match' => [ 'list_name' => 'number' ] ] ] ]; $results = $client->search($params); ##Response: {"took":2,"timed_out":false,"_shards":{"total":5,"successful":5,"skipped":0,"failed":0},"hits":{"total":0,"max_score":null,"hits":[]}} ### System details - Operating System Ubuntu 16.04 - PHP Version 7.1 - ES-PHP client version 5.0 - Elasticsearch version 5.0 Answers: username_1: Are you connecting to the same host? Without specifying the parameters for `ClientBuilder` it connects to localhost. username_0: @username_1 i find a difference between bulk indexing and single indexing. so single index search working smoothly, but bulk index searching not working. username_2: @username_0 The URL you posted is searching the `sms_api_response_packets_index_v2` index, but is not specifying a doc type. Your PHP search is specifying the `response_packets_index_v2` index and the `documents` doc type. Can you try the PHP search against the same index as your URL query string (`sms_api_response_packets_index_v2`) and without a doctype? username_0: @username_2 i was searching with index and doc type, but not worked. I found a solution, while doing bulk mapping and bulk creating, now it's working fine. Actually, may be i am doing wrong, while create indexing. ##Wrong way indexing example ``` for ($i = 0; $i <= $count; $i++) { $params['body'][] = [ 'index' => [ '_index' => 'response_packets_index_v5', '_type' => 'response_packets_v5', 'routing' => 'company', ] ]; $params['body'][] = [ 'my_field' => $documentData[$i] ]; } return $client->bulk($params); ``` ## Right way indexing example, which is working ( Removing "my_field" from body param.) ``` for ($i = 0; $i <= $count; $i++) { $params['body'][] = [ 'index' => [ '_index' => 'sms_api_response_packets_index_v5', '_type' => 'sms_api_response_packets_v5', //'id' => 'my_id', 'routing' => 'company', ] ]; $params['body'][] = $documentData[$i]; } return $client->bulk($params); ``` So i just want to know, what is the difference between, above these two type bulk indexing. username_2: Depends on what's inside of `$documentData`. The first example will nest all of `$documentData[$i]` inside of the `'my_field'` field inside the JSON. So you probably created documents that looked like this: ```json { "my_field" : { "foo" : "bar", "title" : "the title here", "price" : 123 } } ``` whereas the second example would put the `$documentData` directly at the top level, like this: ```json { "foo" : "bar", "title" : "the title here", "price" : 123 } ``` Which is probably why your search wasn't working... the document field structure didn't match what the query was searching :) username_0: @username_2 so i want to search below data using php syntax, can you give me some example. ``` { "_index": "esponse_packets_index_v2", "_type": "documents", "_id": "0", "_score": 1, "_source": { "my_field": { "id": 1, "workflow_id": "943", "workflow_name": "Diwali", "list_name": "number", "list_id": "798", "operator": "-", "circle": "-", "country": "-", "sender_id": "MANISH", "submit_date": "2017-11-06 14:09:56", "dlrdatetime": "2017-11-06 14:10:06", "split_count": 1, "error_code": "Waiting", "error_text": "-", "currency_used": "0.2000", "text_type": "text", "error_code_status": null, "origin_type": "1", "api_response_id": 1, "response": null, "parent_id": null, "is_test": 0, "link": null, "type": 2, "message_text": "Hi This is text message", "status": null, "is_link_api": 0, "winner_branch": null, "instance_id": "724e540394481746", "created_at": "2017-11-06 14:10:06", "updated_at": "2017-11-06 14:10:06", "branch_id": 0 } } ``` Sample code, which not working. ``` $searchParams = [ 'index' => 'response_packets_index_v2', 'type' => 'api_response_packets_v2', 'body' => [ 'query' => [ 'bool' => [ 'filter' => [ 'range' => [ 'created_at' => [ 'gte' => '2017-11-06 14:09:00', 'lte' => '2017-11-06 14:12:56' ] ] ], 'must' => [ 'match' => ['error_code' => 'Waiting'] ] ] ] ] ]; return $client->search($searchParams); ``` username_3: My search has a timeout error. Status: Issue closed username_4: Outdated issue. @username_0 if this is still an issue I'll reopen it. username_5: yes its still an issue for me , I am trying to search through about 100,000 records in an index I created, when i tried the same method but only 18,000 data it worked fine, note : the same the data set on both ends
gboeing/osmnx
665194279
Title: osmnx.utils_geo.bbox_from_point(point, dist=1000, project_utm=False, return_crs=False) too slow for "big" datasets Question: username_0: Hi! 1st of all I have to say this package is amazing and super complete, really useful :) **"Issue"?** So I am using it to find Pois in 1km radius from a 55k rows dataset I am working with. The point is that I wanted to obtain the bounding box from lat/long of all my points (55k). And it took 1h27min with the `bbox_from_point(dist=1000, project_utm=False, return_crs=False)` function. This was my implementation: `trash_df["bbox"] = trash_df.swifter.apply(lambda column: ox.utils_geo.bbox_from_point((column.latitude, column.longitude), dist=1000), axis=1)` **Describe the solution you'd like to propose** So I made this fn: ``` def getBbox(lat, lon, radius=1): import math R = 6371 w = lon - math.degrees(radius/R/math.cos(math.radians(lat))) e = lon + math.degrees(radius/R/math.cos(math.radians(lat))) s = lat - math.degrees(radius/R) n = lat + math.degrees(radius/R) return (n, s, e, w) ``` And it takes 2seconds and the resulst are equal until 3rd-4th decimal Answers: username_1: Thanks @username_0. The current implementation of `utils_geo.bbox_from_point` is only really designed for one-off use and, as you noted, isn't very efficient because it: 1) projects the passed-in point to the local UTM projection, 2) buffers it, 3) projects back to lat-lng, and 4) gets the bounds of the buffer. A streamlined solution like you propose might be useful. A few questions/comments first: 1. If `R=6371`, it looks like you're working in units of kilometers right? Why not meters? 2. If `R` represents the radius of the earth, what does the variable `radius` mean? Is that the distance for the bounding box? If so, a more obvious variable name would be good. 3. The 3rd decimal place in decimal degrees would be something like ~100 meters. If creating a bbox 1000 meters in each direction from the center point, this is a pretty significant discrepancy in the results. 4. I'm not sure how you derived the formulae for the bounding box, but it doesn't look quite right to me, and I think may account for the previous point. I might instead implement something like this: ```python def get_bbox(point, dist=500): earth_radius = 6371000 #meters lat, lng = point delta_lat = (dist / earth_radius) * (180 / math.pi) delta_lng = (dist / earth_radius) * (180 / math.pi) / math.cos(lat * math.pi / 180) north = lat + delta_lat south = lat - delta_lat east = lng + delta_lng west = lng - delta_lng return north, south, east, west ``` I suspect this will yield nearly identical results to the current implementation of `utils_geo.bbox_from_point`, but be much faster. username_1: I just did some quick tests and I believe our two solutions are actually mathematically equivalent. They seem to yield results within approximately a 1 meter of the current `utils_geo.bbox_from_point` implementation. username_1: See proposed enhancement in #541. Status: Issue closed
HaxeFoundation/intellij-haxe
500556702
Title: NPE in HaxeExpressionEvaluator. Question: username_0: Randomly ran into this while editing. Not sure how to repro. ``` In file: file:///home/sni/sandbox/uwongke/poptropica-openfl/src/game/creators/scene/RaceSegmentCreator.hx java.lang.NullPointerException at com.intellij.plugins.haxe.model.type.HaxeExpressionEvaluator.handle(HaxeExpressionEvaluator.java:81) at com.intellij.plugins.haxe.model.type.HaxeExpressionEvaluator._handle(HaxeExpressionEvaluator.java:207) at com.intellij.plugins.haxe.model.type.HaxeExpressionEvaluator.handle(HaxeExpressionEvaluator.java:68) at com.intellij.plugins.haxe.model.type.HaxeExpressionEvaluator.evaluate(HaxeExpressionEvaluator.java:59) at com.intellij.plugins.haxe.model.type.HaxeTypeResolver.evaluateFunction(HaxeTypeResolver.java:515) at com.intellij.plugins.haxe.model.type.HaxeTypeResolver.getPsiElementType(HaxeTypeResolver.java:492) at com.intellij.plugins.haxe.model.type.HaxeTypeResolver.getPsiElementType(HaxeTypeResolver.java:443) at com.intellij.plugins.haxe.ide.annotator.AssignExpressionChecker.check(HaxeSemanticAnnotator.java:923) at com.intellij.plugins.haxe.ide.annotator.HaxeSemanticAnnotator.analyzeSingle(HaxeSemanticAnnotator.java:82) at com.intellij.plugins.haxe.ide.annotator.HaxeSemanticAnnotator.annotate(HaxeSemanticAnnotator.java:61) at com.intellij.codeInsight.daemon.impl.DefaultHighlightVisitor.runAnnotators(DefaultHighlightVisitor.java:121) at com.intellij.codeInsight.daemon.impl.DefaultHighlightVisitor.visit(DefaultHighlightVisitor.java:86) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.runVisitors(GeneralHighlightingPass.java:351) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.lambda$collectHighlights$5(GeneralHighlightingPass.java:284) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.analyzeByVisitors(GeneralHighlightingPass.java:311) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.lambda$analyzeByVisitors$6(GeneralHighlightingPass.java:314) at com.intellij.codeInsight.daemon.impl.DefaultHighlightVisitor.analyze(DefaultHighlightVisitor.java:70) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.analyzeByVisitors(GeneralHighlightingPass.java:314) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.collectHighlights(GeneralHighlightingPass.java:281) at com.intellij.codeInsight.daemon.impl.GeneralHighlightingPass.collectInformationWithProgress(GeneralHighlightingPass.java:225) at com.intellij.codeInsight.daemon.impl.ProgressableTextEditorHighlightingPass.doCollectInformation(ProgressableTextEditorHighlightingPass.java:84) at com.intellij.codeHighlighting.TextEditorHighlightingPass.collectInformation(TextEditorHighlightingPass.java:55) at com.intellij.codeInsight.daemon.impl.PassExecutorService$ScheduledPass.lambda$null$1(PassExecutorService.java:429) at com.intellij.openapi.application.impl.ApplicationImpl.tryRunReadAction(ApplicationImpl.java:1106) at com.intellij.codeInsight.daemon.impl.PassExecutorService$ScheduledPass.lambda$doRun$2(PassExecutorService.java:422) at com.intellij.openapi.progress.impl.CoreProgressManager.registerIndicatorAndRun(CoreProgressManager.java:591) at com.intellij.openapi.progress.impl.CoreProgressManager.executeProcessUnderProgress(CoreProgressManager.java:537) at com.intellij.openapi.progress.impl.ProgressManagerImpl.executeProcessUnderProgress(ProgressManagerImpl.java:59) at com.intellij.codeInsight.daemon.impl.PassExecutorService$ScheduledPass.doRun(PassExecutorService.java:421) at com.intellij.codeInsight.daemon.impl.PassExecutorService$ScheduledPass.lambda$run$0(PassExecutorService.java:397) at com.intellij.openapi.application.impl.ReadMostlyRWLock.executeByImpatientReader(ReadMostlyRWLock.java:164) at com.intellij.openapi.application.impl.ApplicationImpl.executeByImpatientReader(ApplicationImpl.java:204) at com.intellij.codeInsight.daemon.impl.PassExecutorService$ScheduledPass.run(PassExecutorService.java:395) at com.intellij.concurrency.JobLauncherImpl$VoidForkJoinTask$1.exec(JobLauncherImpl.java:161) at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290) at java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020) at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656) at java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594) at java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:177) ``` Status: Issue closed Answers: username_0: It's actually throwing the NPE while trying to log an error message. bfc47e79d0e36ff562cafad2ee7d8aad7b3244ac should fix it. username_0: Fixed in #959
mpc-msri/EzPC
783873520
Title: [squeeze] Generic squeeze not supported Question: username_0: We need to add general support for [tf.squeeze](https://www.tensorflow.org/api_docs/python/tf/squeeze) A squeeze of a tensor of shape [1, 2, 1, 3, 1, 1] should result in a tensor of shape [2,3]. Squeeze also has an axis dim where we can specify to collapse only specified dims. Add support for that to. Currently we only support squeeze of 4D->2D, 4D->3D, 3D->2D in Library_common.ezpc. We should generate squeeze code instead of using library functions.
ubc-vision/image-matching-benchmark
907065087
Title: version of joblib? Question: username_0: I have installed joblib=1.0.1 in pyhon3.6, but when I run run.py in python3.6, 'ImportError' is still there. Validating method 1/1: "sp_sg" ['phototourism', 'pragueparks', 'googleurban'] Running: phototourism, stereo track Running: phototourism, multiview track Running: pragueparks, stereo track Running: pragueparks, multiview track Running: googleurban, stereo track Running: googleurban, multiview track Validating key "config_phototourism_stereo" Validating key "config_phototourism_multiview" Validating key "config_pragueparks_stereo" Validating key "config_pragueparks_multiview" Validating key "config_googleurban_stereo" Validating key "config_googleurban_multiview" Working on sp_sg: phototourism/reichstag -- File feature already exists -- Computing match WARNING: ./jobs/9efa086bd2840045bf66679f8d8ad4a21a0d23201b45937cb4c741cf8aa6e00d already exists! Traceback (most recent call last): File "compute_match.py", line 19, in <module> from joblib import Parallel, delayed ImportError: No module named joblib Traceback (most recent call last): File "run.py", line 280, in <module> main(cfg) File "run.py", line 219, in main job_dict) File "run.py", line 47, in create_eval_jobs job = create_and_queue_jobs(cmd_list, cfg, dep_str) File "/cephfs/person/dihehuang/imc-2021-submit/image-matching-benchmark/utils/queue_helper.py", line 310, in create_and_queue_jobs cpu=cpu) File "/cephfs/person/dihehuang/imc-2021-submit/image-matching-benchmark/utils/queue_helper.py", line 161, in queue_job raise RuntimeError('Subprocess error!') RuntimeError: Subprocess error! Answers: username_1: That's probably an error in the job itself, not joblib. Check logs/9efa086bd2840045bf66679f8d8ad4a21a0d23201b45937cb4c741cf8aa6e00d.log or, easier, just re-run the pipeline with the `--run_mode=interactive` flag to figure out why it's crashing. username_0: I have set '--run_mode=interactive', but there is no log file in ./logs when I set sys.path.append('/usr/local/lib/python3.6/site-packages') it occurs: Traceback (most recent call last): File "compute_match.py", line 20, in <module> from joblib import Parallel, delayed File "/usr/local/lib/python3.6/site-packages/joblib/__init__.py", line 113, in <module> from .memory import Memory, MemorizedResult, register_store_backend File "/usr/local/lib/python3.6/site-packages/joblib/memory.py", line 274 raise new_exc from exc ^ SyntaxError: invalid syntax Traceback (most recent call last): File "run.py", line 281, in <module> main(cfg) File "run.py", line 220, in main job_dict) File "run.py", line 48, in create_eval_jobs job = create_and_queue_jobs(cmd_list, cfg, dep_str) File "/cephfs/person/dihehuang/imc-2021-submit/image-matching-benchmark/utils/queue_helper.py", line 310, in create_and_queue_jobs cpu=cpu) File "/cephfs/person/dihehuang/imc-2021-submit/image-matching-benchmark/utils/queue_helper.py", line 161, in queue_job raise RuntimeError('Subprocess error!') RuntimeError: Subprocess error! username_1: Oh, I see your error message was indeed related to joblib, my bad. I presume this also happens when you import the library from the REPL? It's not a benchmark issue, and it's impossible for us to diagnose it. I would try installing joblib in a new conda environment and if that works, install the rest of the dependencies while ignoring library versions. (Also: the logs only show up if you use `--run_mode=batch`, if you set it to interactive, you see error messages in the terminal.) Status: Issue closed
grycap/scar
366503977
Title: Your website is 404 Question: username_0: You link to https://grycap.github.io/scar/ but it's currently 404. Answers: username_1: Hi @username_0. Thank you for noticing this. We want to rely on the readthedocs documentation page (https://scar.readthedocs.io/en/latest/) instead of the old web in order to better keep up with the changes in the framework. This is why we decommissioned last week the web page. We want to introduce the change on the CNCF entry (https://landscape.cncf.io/selected=scar) so that the website attribute now links to the readthedocs site. Should we do another PR to introduce this change? Thanks. Status: Issue closed username_0: Updated here, thanks: https://github.com/cncf/landscape/commit/b6aff71abaf0c594e9df602d01809a352fa211a2
githubschool/github-games-tnayak16
629235386
Title: Game broken Question: username_0: When attempting to access this at https://githubschool.github.io/github-games-tnayak16/, I am getting a 404. This could be caused by a couple things: - GitHub pages needs to be enabled on master. You can fix this in the repository settings. - the index.html file is incorrectly named inde.html. We will fix this together in class. Can you please fix the first bullet @tnayak16?
kubernetes/kubernetes
145047000
Title: kubernetes-e2e-gce-autoscaling has been timing out for two weeks Question: username_0: [Example failure](https://storage.cloud.google.com/kubernetes-jenkins/logs/kubernetes-e2e-gce-autoscaling/3337) Should this be on the critical builds page? Answers: username_0: The job is disabled. Someone can feel free to pick it up anytime. Status: Issue closed
Sage/sageone_br_nfe_documentacao_api
174820634
Title: Como faço para pegar a autorização? Question: username_0: Estou fazendo a chamada do url para pegar a autorização e está me redirecionando para o site da https://app.br.sageone.com/login Estou seguindo a documentação em https://developers.sageone.com/docs/de/v2#authentication-overview: Chamando via get https://www.sageone.com/oauth2/auth?response_type=code&client_id=xxxxx&redirect_uri=https://app.sisdanca.com.br/auth/callback Onde xxxxx é o meu código de cliente gerado. Answers: username_0: E como fazer Status: Issue closed username_1: Consegui pegar o code para geração do token para trabalhar, porém tive que colocar na internet uma página para pegar o código. Como fazer isso no ambiente de desenvolvimento? No localhost. username_0: desculpe. Pode fechar. Status: Issue closed username_1: Consegui pegar o code para geração do token para trabalhar, porém tive que colocar na internet uma página para pegar o código. Como fazer isso no ambiente de desenvolvimento? No localhost. username_1: sem problemas ;) Status: Issue closed
mortezakz/DocumentWrangling
107002752
Title: Unintentional Author Name Deletion Question: username_0: When author names appear within the the context of the abstract (as opposed to coming before the abstract) they are unintentionally deleted. Need a way to detect when these are being used in sentences to avoid deletion.
swagger-api/swagger-codegen
218553392
Title: Basic Authorization and OAuth is not separated in template Question: username_0: I have define in my RestService swagger annotation: ```java @ApiOperation( value = "Get a UserConfiguration resource." ,authorizations = @Authorization(value = "basic") ) ``` And I have expect that in client I will have code generated for Basic Authorization only. There is no reason that I have Apache Oltu included like dependency. I have view the template and there is no switch like: if(isBasic).. else if(isOAuth)
icsharpcode/ILSpy
474181134
Title: foreach expected but while used instead Question: username_0: ILSpy version 5.0.0.4970-preview3 method decompiled with **while** but **foreach** expected ``` protected CompositeDevices(SerializationInfo info, StreamingContext context) : base(info, context) { SerializationInfoEnumerator enumerator = info.GetEnumerator(); while (enumerator.MoveNext()) { string name = enumerator.Current.Name; if (name != null && name == "AdditionalParameters") { AdditionalParameters = (Dictionary<string, string>)info.GetValue("AdditionalParameters", typeof(Dictionary<string, string>)); } } base.CollectionChanged += OnCollectionChanged; } ``` please also ref to very similar method CompositeDevices(IEnumerable<CompositeDevice> col) that decompiled to foreach! target: #1601 Answers: username_1: `SerializationInfoEnumerator` implements `IEnumerator` but not `IDisposable`, so the compiler doesn't generate a using statement / try-finally block. I think currently our foreach pattern only tries to match using statements. username_0: Did a little investigation - how competitors are handling this case so, .NET Reflector as well as Telerik are using while (enumerator.MoveNext()) but JetBrains and dnSpy are using foreach (SerializationEntry serializationEntry in info) so, to summarize 1) it's possilbe 2) it's "nice to have & see" feature, not an error Feel free to close this issue. username_2: See the relevant test case at https://github.com/icsharpcode/ILSpy/blob/ddf4053a45a52609023047be326419b1ec8980df/ICSharpCode.Decompiler.Tests/TestCases/Pretty/Loops.cs#L358-L366
quic/aimet
937859536
Title: AttributeError: 'Parameter' object has no attribute 'modules' Question: username_0: Hello, I am trying to compress a Faster-RCNN detection model with channel pruning. However, in the winnowing stage, called with the function `aimet_torch.winnow.winnow.winnow_model` I get the following exception: ```` File "/usr/local/lib/python3.6/dist-packages/aimet_torch/winnow/winnow.py", line 70, in winnow_model in_place, verbose) File "/usr/local/lib/python3.6/dist-packages/aimet_torch/winnow/mask_propagation_winnower.py", line 98, in __init__ self._graph = ConnectedGraph(self._model, (dummy_input,)) File "/usr/local/lib/python3.6/dist-packages/aimet_torch/meta/connectedgraph.py", line 202, in __init__ self._construct_graph(model, model_input) File "/usr/local/lib/python3.6/dist-packages/aimet_torch/meta/connectedgraph.py", line 312, in _construct_graph self._parse_trace_graph(model, model_input, module_tensor_tuples_map) File "/usr/local/lib/python3.6/dist-packages/aimet_torch/meta/connectedgraph.py", line 361, in _parse_trace_graph if not is_leaf_module(subgraph_model): File "/usr/local/lib/python3.6/dist-packages/aimet_torch/utils.py", line 199, in is_leaf_module module_list = list(module.modules()) AttributeError: 'Parameter' object has no attribute 'modules' ```` Apparently, one of the parameters of the models (the weight of the first convolution in the ResNet stem) is interpreted as a module and enters the `is_leaf_module` in line 361 in `aimet_torch/meta/connectedgraph.py`. Here are my versions of Aimet and torch: ```` Aimet==1.10.0.0.100.0.486 AimetCommon==1.10.0.0.100.0.486 AimetTorch==1.10.0.0.100.0.486 torch==1.3.1+cu100 ```` How can I solve this? Thank you, D Answers: username_1: @username_0 Thank you for reporting this. @username_2 @quic-sundarr Could you please take a look at this. username_2: Hi @username_0 , For context, in this section of code, we are attempting to build our own internal representation of the Pytorch graph which is what we refer to as Connected Graph. As a part of this, we make use of Pytorch's jit trace to provide us a graph to parse. In the graph, we typically see various types of nodes giving us module and connectivity information, including GetAttr nodes which define modules (which could in turn contain submodules), and Callmethod nodes which show how the GetAttr nodes are being called. It looks like in your case, the subgraph_model in the stack trace came from a GetAttr node, but holds a parameter instead of a torch.nn.Module. It is unclear what would lead to this, it would help if you could provide the model definition of the module that the parameter is in, so we can attempt to reproduce the issue. Also, if you wanted to try getting past the issue, I would first try adding a check after subgraph_model has been defined a few lines above, to continue on to the next node if subgraph_model is not of type torch.nn.Module. username_0: Hi, and thank you for your answer. The model I am winnowing is a Resnet18 from `torchvision.models.resnet18()`. It should be a pretty neat model definition, that is why I was thinking the problem is in incompatibilities between Aimet version and torch version. Could you please confirm the versions I reported are good to use in conjunction? Also, the input to that winnow_model function is as follows: ```` model=ResNet (from torchvision) input_shape=[1,3,600,1000] list_of_modules_to_winnow=[(Conv2d(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1), bias=False), [0, 2, 3, 4, 5, 6, 11, 12, 15, 17, 19, 21, 23, 24, 25, 26, 27, 31, 33, 34, 35, 38, 39, 41, 43, 45, 53, 55, 56, 59, 60, 63])] reshape=False ```` The parameter mistakenly interpreted as module is an `nn.Parameter`, shaped 64, 3, 7, 7, so it seems to me the first ResNet convolution (stem). Thanks for help, D username_0: Hello, closing this as upgrading torch version fixed it. ```` Aimet==1.10.0.0.100.0.486 AimetCommon==1.10.0.0.100.0.486 AimetTorch==1.10.0.0.100.0.486 torch==1.4.0+cu100 torchvision==0.5.0+cu100 ```` Best, D Status: Issue closed username_2: Thanks for trying out the pytorch upgrade @username_0, and apologies for not responding sooner regarding the versions. It looks like you are also using an old version of AIMET, and so I would suggest upgrading to our latest 1.16.2 release. That way if you run into any more issues, it will be easier for us to debug.
koppor/jabref
188920402
Title: Improve shared database Question: username_0: - [ ] Manage connections: http://discourse.jabref.org/t/managing-shared-database-connections/289 - [ ] Command line access: http://discourse.jabref.org/t/command-line-access-to-sql-databases/288 Answers: username_1: Good questions. Wouldn't be to bad for testing issues, if we had a shared database we can test on. We'll talk about that in the next devcall. Meanwhile, set up your own remote database for testing. @username_4: You should be able to move the synchronizing stuff to a background task (jabref already offers infrastructure to follow these thanks to @username_2). If there are parts of the task that would block the user from working on the main table, JabRef should display some kind of waiting animation. If there is a huge load of entries (10.000 or so, yes, we really have users with this number of entries). If there is an infinite loop it would probably be a bug, that needs fixing. Maybe @username_0 can specify further. @username_3 I believe the task would be to expand the existing cli of jabref to work with shared databases. username_2: I think the best thing to do here is to go for the same approach as for the pdf indexer. Have one background task with a queue of database queries. You can checkout the [IndexingTaskManager](https://github.com/JabRef/jabref/blob/7eb7b86203d7c149d3b48df8e0845d35aff1597c/src/main/java/org/jabref/logic/pdf/search/indexing/IndexingTaskManager.java) to get an idea of how to do that. username_3: Hi Thank you for your answer. According to the instructions, we should automate the process of entries retrieval, is the task to import a Latex Document with all the entries in it or is a simple commandline output sufficient? Regarding the export, what exactly does the user want to export? a Latex-Document? Or a simple input in the commandline username_1: I don't think that you understood the instructions right... JabRef works on `.bib`, not with latex documents. It already can sync via the gui entries with a remote database. Locally those entries are stored in a .bib` file, remotely in a database. Syncing/importing/exporting already works in the gui and does not have to be improved (well, actually it could be improved, but thats not the concern of this issue). This issue applies to improving * the management of multiple databases (to switch between multiple database, to store credentials, to switch between different profiles etc.) * making the already existing functionality of the gui accessible via the cli (syncing a `.bib` file with a remote database, running cleanup-formatters, etc.) * improving the responsiveness of the gui while synchronizing a remote database with your local `.bib` file username_0: Did you try to connect to to https://www.elephantsql.com/? This should work instantly. Please try it and report back. username_0: Yez - we do not know about the root cause. We should observe that interaction with a shared database is slow. username_4: Yes, we did and it works. (We've already sent the question 1., right before we've figured out what the elephantsql does. Sorry for that.) Thanks for the answers. username_5: It seems we are unable to run JabRef from the .bat file. When typing `JabRef.bat` (or without the `.bat` in Ubuntu) in a terminal, we get the following error: ``` Error occurred during initialization of boot layer java.lang.module.FindException: Module org.jabref not found ``` It is possible that we missed a step, since we tried that directly in the `build/scripts/` folder in IntelliJ. Could we get some help on how to run JabRef in CLI mode ? username_6: @username_5 generally speaking I'd suggest asking these kinds of questions in the Gitter, you are likelier to get a quick response there. I don't know the goals of your course, but if you want a quick fix that might allow you to proceed until someone gets back to you with a better method, I'd say use `gradlew run` with `--args` instead (e.g., ` ./gradlew run --args="-v"`). I think @username_2 did some CLI-related improvements, do you have a better workflow? username_2: Hi! You mention you are using intellij. If you also use the Intellij build system you can create a run configuration to test the CLI. I never used self generated binaries explicitly, but you could try if the pre-built binaries work for you. If they do, there may be a problem with your build.
BurntSushi/rust-csv
123931180
Title: Doc or impl bug "If `quote` is `None`, then no quoting will be used." Question: username_0: https://github.com/username_1/rust-csv/blob/e1552706c162d594f7d454b69bece07fd544ff49/src/reader.rs#L418 says ``` /// If `quote` is `None`, then no quoting will be used. ``` but `quote` is a `u8`, not an `Option<u8>`, so it doesn't look like I can turn off quoting support.<issue_closed> Status: Issue closed
BSData/robotech-rpg-tactics
162590779
Title: BattleScribe 2.0 Question: username_0: Hi guys. BattleScribe 2.0 is coming and it contains some major changes to Data Editor and the data format. While it should happily upgrade 1.15.x format data files to the new format, you never know what sneaky bugs might be lurking... It would help a lot if you could run your data through it and make sure nothing gets broken so we have no nasty surprises on release day (whenever that might be...). **Alpha Downloads:** Desktop: https://github.com/BattleScribe/Desktop-Alphas/issues/1 Android: https://github.com/BattleScribe/Android-Alphas/issues/1 iOS: https://github.com/BattleScribe/iOS-Alphas/issues/1 If you find any issues, please do let me know: https://github.com/BattleScribe/Desktop-Alphas/issues
uber/motif
592817712
Title: ScopeFactory with single unused dependency ignores the given creatable instance Question: username_0: Generated code seems to ignore given instance for the dependency source while using ScopeFactory with a single unused dependency and tries to create the instance directly. **Library version**: 0.3.3-SNAPSHOT **Repro steps or stacktrace**: Scope ```java @Scope public interface TestScope extends Creatable<TestScope.Dependencies> { interface Dependencies { String blah(); } } ``` Creating instance of scope ```java TestScope.Dependencies dependencySource = new TestScope.Dependencies() { @Override public String blah() { return "something"; } }; ScopeFactory.create(TestScope.class, dependencySource); ``` Generated code ```java @ScopeImpl( children = {}, scope = TestScope.class, dependencies = TestScope.Dependencies.class ) public class TestScopeImpl implements TestScope { private final TestScope.Dependencies dependencies; public TestScopeImpl(TestScope.Dependencies dependencies) { this.dependencies = dependencies; } public TestScopeImpl() { this(new TestScope.Dependencies() {}); } TestScope testScope() { return this; } } ``` Notice `this(new TestScope.Dependencies() {});` in constructor. Answers: username_1: Hi! I was wondering if there's any updates on this issue?
angular-hispano/angular
863289342
Title: Traducir: guide/interpolation.md Question: username_0: 📚Traducir: <!-- ✍️ editar: --> interpolation.md <!--🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 Traducción de la documentación oficial de Angular a español 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅--> ## Nombre del archivo: <!-- ✍️ editar: --> interpolation.md ## Ruta donde se encuentra el archivo dentro del proyecto de Angular <!-- ✍️ editar: --> https://github.com/angular-hispano/angular/blob/master/aio/content/guide/interpolation.md Answers: username_1: Closed by duplicated https://github.com/angular-hispano/angular/issues/124 @username_0 please work on that issue Status: Issue closed
szjug/szjug.github.io
816247498
Title: Is this JUG still active? Question: username_0: Hi there, just wanna ask if szjug is still operating. Any upcoming events? Answers: username_1: Yes, there is WeChat group But because of COVID there were no in person meetings in 2020 and 2021 so far. username_0: May I ask how to join the WeChat group? username_1: ![image](https://user-images.githubusercontent.com/1614482/117952809-15d17d80-b348-11eb-90e3-3e700f7f1c6d.png) username_1: scan code above, the code is changing, so there is no use to have it on the site Any ideas are welcomed.
openshifttips/web
508559179
Title: Basic release server script to view the different channles and the releases. Question: username_0: Not sure where or if you would like to use this or add it to your tips. I named the script `ocr`. * requires `jq` ``` #!/bin/bash PS3='Please enter your choice: ' options=("prerelease-4.1" "stable-4.1" "candidate-4.2" "fast-4.2" "stable-4.2") _Command () { echo "Showing upgrade channel: ${channel}" curl -sH 'Accept: application/json' https://api.openshift.com/api/upgrades_info/v1/graph?channel=${channel} | jq -S '.nodes | sort_by(.version | sub ("-rc";"") | split(".") | map(tonumber)) | .[]' } select opt in "${options[@]}" do channel="${opt}" _Command break done ``` Status: Issue closed Answers: username_1: I've merged the PR, thanks again!
sbraz/pymediainfo
148879850
Title: ValueError in MediaInfo.parse Question: username_0: It breaks on line `lib.MediaInfoA_Open(handle, filename.encode("utf8"), 0)` and says "Procedure probably called with too many arguments (4 bytes in excess)". I have just downloaded last version (0.7.84) of MediaInfo. Quick look at sources of MI points me that API of dll is changed - function MediaInfoA_Open now takes only two arguments, so removing last argument in the call makes it work. Answers: username_1: I can not reproduce with 0.8.4 but it's true that the function seems to only take 2 args: https://github.com/MediaArea/MediaInfoLib/blob/master/Source/MediaInfoDLL/MediaInfoDLL.cpp#L379 I cannot remember where the third argument came from. I've pushed the change, thanks for the report. Status: Issue closed
InfamousBanana/my-repository
425912590
Title: New screenshot added to gallery Question: username_0: <img src="https://locker.ifttt.com/v2/17179621/1553687130856-c88666e11aa88311/e8197fbf8ddcb95db22e1b13649354f12ebf9b1ba885f1e93265fbab87b5d7ea/a5439477-35d7-48ab-b9f6-196bd191d47c.png?sharing_key=87f4d199d5ad320e8932dbd0409e6e4b"><br> <br> via Android https://ift.tt/2uvdOrV<br> <br> March 27, 2019 at 12:45PM
Azure/azure-event-hubs-for-kafka
439399517
Title: V1 SaslHandshake+SaslAuthenticate not working Question: username_0: Description =========== [V1 SaslHandshake](https://kafka.apache.org/protocol.html#The_Messages_SaslHandshake) appears to be broken--as a result, no clients that attempt to use V1 (most native Go clients and a few other clients) can authenticate. [Confluent Kafka Go](https://github.com/confluentinc/confluent-kafka-go) always seems to use V0. As does the Java versions. I've got a [working auth flow](https://github.com/username_0/kafka) and a [fix for Sarama](https://github.com/username_0/sarama/commit/6509b6f9a616196089617e28f70ea2e21e8406ae): - https://github.com/username_0/kafka - https://github.com/username_0/sarama/commit/6509b6f9a616196089617e28f70ea2e21e8406ae How to reproduce ================ Attempt to connect using [Sarama Kafka client](https://github.com/Shopify/sarama). Has it worked previously? ========================= Has not worked previously afaik. Checklist ========= Please provide the following information: - [x] Verified that port 9093 is not blocked by firewall - [x] Verified the namespace is either Standard or Dedicated tier (i.e. it is not Basic tier, which isn't supported) - [x] Sample you're having trouble with: [https://github.com/username_0/kafka](https://github.com/username_0/kafka) - [x] Apache Kafka version: `Azure Event Hubs for Kafka` - [x] Kafka client configuration: N/A - [x] Namespace and EventHub/topic name: `would prefer to keep this private but can provide if required` - [x] Consumer or producer failure `both--auth failure` - [x] If consumer, partition ID, group ID `N/A` - [x] Timestamp in UTC `attempts made around 2019-05-02 02:00 to 03:00-ish` - [x] Client ID `testClient` - [x] Provide all logs N/A - [x] Standalone repro `willing to send` - [x] Operating system: `Windows 10 and Ubuntu WSL (up-to-date)` - [x] Critical issue (critical with workaround) Answers: username_1: Thanks for reporting the issue, @username_0. SaslHandshake version 1 is currently not supported. We will update the service so that API version response doesn't include version 1 for the SaslHandshake API until v1 is fully supported. username_0: Thank you @username_1! FWIW there are a few other more minor issues with the ApiRequest response. For example it doesn’t report being compatible with SaslAuthenticate—just handshake. Doesn’t seem to trip up clients though. I assume that means I’ll have to submit a PR for sarama to perform a more reliable check for ApiVersion (right now they do it based on the user-specified Kafka version) but that’s not on your team! username_1: Hi @username_0, the service doesn't support SaslAuthenticate API yet and hence the response of the ApiRequest call doesn't include the version info of the SaslAuthenticate API. We will add it once the service supports SaslHandshake v1. username_0: Hi @username_1 there is a SaslAuthenticate V0 request in the spec that matches SaslHandshake V0. That represents the “opaque” (no Kafka headers) sasl authentication request. Since the kafka endpoint requires SaslAuthentication it should (and does) support V0 of SaslAuthenticate. Thanks! username_1: @username_0, for SaslHandshakeRequest v0 case, a security context corresponding to the mechanism is sent separately and it is not part of SaslAuthenticate request. And, SaslAuthenticate request is used if SaslHandshakeRequest version is v1. username_0: @username_1 ah you’re very right. I misread it! Apologies for the doubt. It’s been a long day staring at the docs figuring things out—words blurred together :) Thanks for tour help!! Status: Issue closed username_3: Hello @username_1, have the issues been fixed on Azure Event hub? Does the Event Hub now support SaslHandshake V1? If not, is there are plan till when it will? thx for help
custom-cards/button-card
469633100
Title: Ellipsis adds 3px margin left Question: username_0: **Checklist** - [ x ] I updated the card to the latest version available - [ x ] I cleared the cache of my browser **Describe the bug** When text overflows some mysterious margin is added to the left side. **Version of the card** Version: 1.11.1 **Screenshots** <img width="268" alt="Skärmavbild 2019-07-18 kl 10 19 24" src="https://user-images.githubusercontent.com/2734444/61441338-b372e600-a945-11e9-94d3-e9306d646506.png"> Answers: username_1: Hey, the solution to this problem is here: https://community.home-assistant.io/t/lovelace-button-card/65981/1678?u=romrider Mainly this part: You should add: `- text-align: start`to all the fields and it will fix alignment issues. Looks like when `ellipsis` is used, it doesn’t take into account the `justify-self`. But you still need to keep `justify-self`, because when it’s actually not “ellipsed”, it will use `justify-self`! CSS nightmare :) username_0: Awesome! Thanks. Status: Issue closed
conda/conda-build
955781565
Title: git shallow clone for speeding up large repos Question: username_0: I'm trying to speed up build times for pytorch in hopes of gaining a few minutes on CIs. I'm hoping that I can save time cloning a gigantic repository. However, git clone and checkout fail for shallow clones. I feel like git should be able to clone a single branch or tag with the command ``` git clone --branch GIT_REV --depth 1 URL ``` however, it seems that the checkout happens after the clone if I specify even as deep as 5000 commits for the shallow clone. Is there a way to skip making the mirror. Cloning a specific branch: 18MB ``` git clone --branch v1.9.0 --depth 1 [email protected]:pytorch/pytorch.git Cloning into 'pytorch'... remote: Enumerating objects: 9310, done. remote: Counting objects: 100% (9310/9310), done. remote: Compressing objects: 100% (8275/8275), done. Receiving objects: 100% (9310/9310), 18.08 MiB | 9.03 MiB/s, done. remote: Total 9310 (delta 1322), reused 2845 (delta 828), pack-reused 0 Resolving deltas: 100% (1322/1322), done. Note: switching to 'd69c22dd61a2f006dcfe1e3ea8468a3ecaf931aa'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by switching back to a branch. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -c with the switch command. Example: git switch -c <new-branch-name> Or undo this operation with: git switch - Turn off this advice by setting config variable advice.detachedHead to false ``` Cloning full tree: 500MB ``` $ git clone <EMAIL>:pytorch/pytorch.git Cloning into 'pytorch'... remote: Enumerating objects: 618076, done. remote: Counting objects: 100% (6125/6125), done. remote: Compressing objects: 100% (2169/2169), done. remote: Total 618076 (delta 4619), reused 5224 (delta 3947), pack-reused 611951 Receiving objects: 100% (618076/618076), 507.99 MiB | 41.92 MiB/s, done. Resolving deltas: 100% (500231/500231), done. ``` Would you consider having an opt-out of the "mirror" + "clone the mirror" strategy that is currently implemented for git repos? xref: https://github.com/mamba-org/boa/issues/172#issuecomment-888815403 xref: https://github.com/conda/conda-build/blob/master/conda_build/source.py#L236 cc: @username_1 cc: @wolfv Answers: username_1: +1 I've seen this issue a few times. We do this in SciPy CI as well, creating a full clone is just too expensive for large repos, and unnecessary if you are just building a tag.
angular/clang-format
156087341
Title: Inconsistent handling of Decorators Question: username_0: Decorators in a Typescript file are inconsistently handled. Please see the Typescript example below. This is my .clang-format file: ```` Language: JavaScript BasedOnStyle: Google IndentWidth: 4 ColumnLimit: 80 BreakBeforeBinaryOperators: NonAssignment ```` The Typescript example (please look at the methods findById and findByQuery): ```` JavaScript import {Request, Response, NextFunction} from 'express'; // Document kann nicht direkt importiert werden, weil es in ES2015 auch eine // Klasse Document gibt import {Document as MDocument} from 'mongoose'; import {inspect} from 'util'; import {getBaseUri, CONTENT_TYPE, APPLICATION_JSON, isBlank, log, logger} from '../../shared/index'; import {Buch, validateBuch} from '../model/buch'; import {BuecherService, TitelExistsError} from '../service/buecher.service'; class BuecherRequestHandler { private buecherService: BuecherService = new BuecherService(); // async und await: // https://blogs.msdn.microsoft.com/typescript/2015/11/30/announcing-typescript-1-7 // http://tc39.github.io/ecmascript-asyncawait vor allem von Microsoft @log async getById(req: Request, res: Response, next: NextFunction): Promise<void> { const id: string = req.params.id; let buch: MDocument; try { buch = await this.buecherService.findById(id); } catch (err) { // Exception einer async. Ausfuehrung fangen: // https://strongloop.com/strongblog/comparing-node-js-promises-trycatch-zone-js-angular logger.error(`Error: ${inspect(err)}`); res.sendStatus(500); return; } if (isBlank(buch)) { logger.debug('status = 404'); res.sendStatus(404); return; } const baseUri: string = getBaseUri(req); // Link Header res.links({ self: `${baseUri}/${id}`, list: `${baseUri}`, add: `${baseUri}`, update: `${baseUri}`, remove: `${baseUri}/${id}` }); logger.debug(`getById(): buch = ${JSON.stringify(buch)}`); [Truncated] 'use strict'; buecherRequestHandler.getByQuery(req, res, next); } export function post(req: Request, res: Response, next: NextFunction): void { 'use strict'; buecherRequestHandler.post(req, res, next); } export function put(req: Request, res: Response, next: NextFunction): void { 'use strict'; buecherRequestHandler.put(req, res, next); } export function deleteFn( req: Request, res: Response, next: NextFunction): void { 'use strict'; buecherRequestHandler.deleteMeth(req, res, next); } ```` Answers: username_1: Thanks for the report! I'll take a look, though it might take a while. Just for future reference: it's much simpler for me to understand what's going on if you create a minimal snippet of TypeScript code that reproduces the issue. Status: Issue closed username_1: The problem was actually the `async` after the `@log`. This is fixed now. username_0: @username_1 The issue is gone. Thank you for your efforts.
openwrt/mt76
532930581
Title: mt76: mt7612 5 GHz channels width when using as mesh backhaul Question: username_0: Using 17.01. I've been using the 5GHz on the mt7612 for a mesh backhaul (802.11s) and noticed when I change channels and set the vht that the width does not change and is stuck at 80MHz, I have used the lower channels 36-48 and the upper 149-161. I have also set channel 165 that is suppose to be 20 MHz (Does not matter if I set it to 20/40 or 80 VHT in the /etc/config/wireless on any channels), when I capture on charts from our spectrum analyser it shows that both lower and upper channels appear to be stuck at 80 MHz width and when I use 165 it appears it is 40MHz and not 20MHz width. Has anyone else captured charts to verify that the channel width change accordingly.
wodby/php
760648517
Title: Incompatible PHP 8 pecl modules Question: username_0: Currently the following modules not installed for php 8 due to incompatibility: - amqp - imagick https://github.com/Imagick/imagick/issues/358 - pdo_slqsrv - rdkafka https://github.com/arnaud-lb/php-rdkafka/issues/328 - sqlsrv - uploadprogress - xmlrpc - blackfire - tideways_xhprof https://github.com/tideways/php-xhprof-extension/issues/101 - newrelic https://blog.remirepo.net/post/2020/09/21/PHP-extensions-status-with-upcoming-PHP-8.0 Answers: username_1: https://github.com/tideways/php-xhprof-extension/pull/96 merged and new release is out username_2: Blackfire also has been 8.0 compatible for a while now, FYI. username_3: rdkafka:5.0 was just released, which adds compatibility for php:8 username_1: Uploadprogress errors could be fixed with https://github.com/php/pecl-php-uploadprogress/pull/8 username_1: Uploadprogress got 2.0 releases username_4: I'm seeing errors like `Warning: Missing arginfo for uploadprogress_get_contents() in Unknown on line 0` so it would be great if we could upgrade to 2.0 for that one. Thanks.
NervJS/taro
736736670
Title: taro 构建是否支持node api Question: username_0: ### 这个特性解决了什么问题? 想自己写node脚本启动taro的编译 ### 这个 API 长什么样? 类似于webpack的脚本 <!-- generated by taro-issues. 请勿修改或删除此行注释 --><!--labels=enhancement--> Answers: username_1: 试试: ```js const build = require('@tarojs/cli/commands/build') build(options) ``` 使用方法详见 `@tarojs/cli/src/cli.ts` username_0: @username_1 username_0: @username_1 目前是通过这种方式自己实现的,如果可以的话,出个标准的文档吧 谢谢啦 username_1: 人力暂时不够去完善这块文档,欢迎提交你的使用经验作为文档哈~ Status: Issue closed username_2: 同问,能够提供node api吗
reprohack/reprohack_site
1176991678
Title: Incorrect username shown in navbar when inspecting another user's profile Question: username_0: When I view the site logged in as myself, the correct username appears in the navbar: <img width="1049" alt="image" src="https://user-images.githubusercontent.com/5583057/159524513-d79f6c63-85be-47b9-b02a-8d77774c97f4.png"> However if I click on another users profile, their username appears as logged in in the navbar and the edit button even appears on the profile: <img width="1040" alt="image" src="https://user-images.githubusercontent.com/5583057/159524611-eb294ebb-2ec7-488c-a4a9-ba121a3094dd.png"> If I click the edit button, the slug remains as that of the other users profile but the data in the form and navbar changes back to my own profile. Any changes made are indeed associated with my own profile. Pretty dodgy behaviour though. Any ideas @twinkarma off the top of your head how to correct it? Answers: username_0: <img width="1148" alt="image" src="https://user-images.githubusercontent.com/5583057/159524961-59b1dc9b-2cdd-474b-ada4-3a7d36661926.png">
bwindsor/typed-config
1051244428
Title: New conflicts with mypy in v1.0.3 Question: username_0: When updating to v1.0.3, mypy starts having conflict issues with setting keyword arguments for Callables/functions, for example: ``` filter = key( required=False, default=[], cast=lambda x: set(example_function(x)) ) ``` Depending on what types you provide, the error from mypy will look something like: ``` error: No overload variant of "key" matches argument types "bool", "List[<nothing>]","Callable[[Any], Set[str]]" note: Possible overload variants: note: def [T] key(*, section_name: Optional[str] = ..., key_name: Optional[str] = ..., required: Literal[False], cast: Callable[[Union[T, str]], T], default: T) -> T note: def [T] key(*, section_name: Optional[str] = ..., key_name: Optional[str] = ..., required: Literal[False], cast: Callable[[Union[T, str]], T], default: None = ...) -> Optional[T] note: <3 more similar overloads not shown, out of 5 total overloads> ``` I assume it has something to do with the type-checker implemented with version 1.x.x. Any insight is welcome 👍 Answers: username_1: From your `cast` function, it looks like you want `filter` is a `Set[str]` type. However, your default value is `[]` which is a `list`. So the first change required is to have ```python filter = key( required=False, default=set(), cast=lambda x: set(example_function(x)) ) ``` However, I believe mypy will still give you an error at this point, as the `default` value should match the return value of `cast`, and it can't tell which type is stored inside an empty set, so it will complain that `Set[<nothing>]` and `Set[str]` are not the same. So I think this should work for you as a full solution: ```python from typing import Set default_value: Set[str] = set() filter = key( required=False, default=default_value cast=lambda x: set(example_function(x)) ) ``` Let me know if this fixes the problem. username_0: Same notes as before but produces the error: ``` error: No overload variant of "key" matches argument types "bool", "Set[str]", "Callable[[Any], Set[str]]" ``` Strange that this hasn't been an issue that Mypy has picked up before this version update username_1: Would you be able to post what you're using as `example_function` in this case? username_0: The "example_function" is the `lower_comma_separated_str` ``` def comma_separated_str(text: str) -> list[str]: if isinstance(text, list): return text return text.split(",") def lower_comma_separated_str(text: str) -> list[str]: return [i.lower() for i in comma_separated_str(text)] ``` username_1: Thanks for that, I now understand the reason for the error. The `cast` function is actually called with the `default` value, so if you have ```python key(cast=int, default='3') ``` You would still end up with `3` as an `int`, since it would cast your default. So to work with `1.0.3` you cast function would need to be ``` def lower_comma_separated_str(text: Union[set[str], str]) -> list[str]: if isinstance(text, set): return list(text) return [i.lower() for i in comma_separated_str(text)] ``` which isn't partcularly logical. Now that `key` has proper typing on it (since `1.0.0`), the above `key(cast=int, default='3')` won't pass any type checkers, as the type of `default` must match the return type of `cast`. Therefore I have removed this behaviour (which was also undocumented - the README clearly says that `cast` functions must accept a string, which I think is the sensible behaviour to have). This should be available very shortly in `1.1.0`. Thanks very much for your examples highlighting this confusing behaviour! Please let me know if updating to `1.1.0` fixes it for you, or if `mypy` is continuing to complain.
wongjiahau/TTAP-Bug-Report
384147380
Title: Bug report #1063793140 Question: username_0: Object reference not set to an instance of an object. ==================== at Time_Table_Arranging_Program.Pages.Page_Login.<<Browser_OnLoadCompleted>g__ExtractData14_3>d.MoveNext() --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.AsyncMethodBuilderCore.<>c.<ThrowAsync>b__6_0(Object state) at System.Windows.Threading.ExceptionWrapper.InternalRealCall(Delegate callback, Object args, Int32 numArgs) at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(Object source, Delegate callback, Object args, Int32 numArgs, Delegate catchHandler) ==================== <HEAD><TITLE>myUTAR - The Universiti Tunku Abdul Rahman Web Portal</TITLE> <SCRIPT language=javascript> function MM_openBrWindow(theURL,winName,features) { window.open(theURL,winName,features); } function mypopup(url, sbar, resize, width, height, top, left){ tit='' reWin=window.open(url, tit, 'toolbar=no,location=no,directories=no,status=no,menubar=no,scrollbars=' + sbar + ',resizable=' + resize + ',width=' + width + ',height=' + height + ',top=' + top + ',left=' + left) } function checkPhone(evt){ evt = (evt) ? evt : window.event var charCode = (evt.which) ? evt.which : evt.keyCode if ((charCode > 46 && charCode < 58) || charCode==45 || charCode==13){ return true } else{ alert("You can only key in numeric number") return false } } function checkNumeric(evt){ evt = (evt) ? evt : window.event var charCode = (evt.which) ? evt.which : evt.keyCode if((charCode > 47 && charCode < 58) || charCode==13){ return true } else{ alert("You can only key in numeric number") return false } } function IsNumeric(strString) { var strValidChars = "0123456789.-/"; var strChar; var blnResult = true; if (strString.length == 0) return false; for (i = 0; i < strString.length && blnResult == true; i++) { strChar = strString.charAt(i); if (strValidChars.indexOf(strChar) == -1) { blnResult = false; } } return blnResult; } function logout(myPath,logoutURL){ //alert(myPath+logoutURL) [Truncated] <TD>KBS02</TD></TR> <TR align=center> <TD rowSpan=2>147</TD> <TD rowSpan=2>P</TD> <TD rowSpan=2>6</TD> <TD rowSpan=2 align=right>20</TD> <TD>Tue</TD> <TD>09:00 AM - 12:00 PM</TD> <TD>3.0</TD> <TD>5</TD> <TD>KB729</TD> <TD rowSpan=2></TD></TR> <TR id=subRow_263525 align=center> <TD>Tue</TD> <TD>09:00 AM - 12:00 PM</TD> <TD>3.0</TD> <TD>9</TD> <TD>KBS02</TD></TR> <TR align=center> <TD rowSpan=2>148</TD></TR></TBODY></FORM></DIV></TR></TBODY></TABLE></TR></TBODY></TABLE></TR></TBODY></TABLE></BODY>
HPInc/jipp
1040915895
Title: responseCode 500 Question: username_0: When I used default print Settings on ubuntu system, added printer with IPP, PDD file with IPP everywhere, test Page printed successfully. When I use jipp, jprint returns responseCode 500.What's the difference between the two ways, Jipp and cups? Answers: username_1: There are lots of differences but they implement parts of the same specifications. You got a 500 because you sent data to the printer it didn't expect. If you post the code or the exchange, we may be able to help. username_0: thanks for your help Status: Issue closed
rust-lang/libc
518508628
Title: SysV IPC APIs missing on OpenBSD Question: username_0: See this issue: https://github.com/username_0/slice_deque/issues/77#issue-518380526 Answers: username_1: cf username_1/libc@8c688c22bc25757f945f2acefc2d5038cb72229f for a quick wip. @username_2 ? username_2: @username_1 thanks for the wip. I will use it as base for the PR. I will have few fixes (but I will take care of that): - `sys/shm.h` header is need in `libc-test/build.rs` for proper testsuite run - `shm_segsz` is `int` and not a `size_t` on OpenBSD (hey, not posix compliant. I hope others crates doesn't rely on that) - `struct shmid_ds` is incomplete - few constants mismatches Status: Issue closed
binary-com/binary-bot
483209815
Title: [bug] unable to stop showing buy successed and contract never close Question: username_0: ![Screenshot_2019-08-21-13-02-27](https://user-images.githubusercontent.com/31925501/63405186-815b1500-c418-11e9-829f-862a67d2c640.jpeg) Answers: username_1: Hey @username_0, Could you describe what's happening in the screenshot? Regards, Aaron username_0: @username_1 Time on GMT Wed, Aug 21, 2019 5:02 am as captured My VPS time zone is -7 22:02 Bot try to buy contract More than 1 hour @ 20:58 ~ GMT 03:58 in the capture ![image](https://user-images.githubusercontent.com/31925501/63493277-d3ba3580-c4ed-11e9-851d-29a7a1b0051b.png) showing by success on ID *380888 But contract never expire and unable to stop BOT show "stoping". it wont stop need to refresh page. username_0: Again pending....... contract not close. ![Screenshot_2019-08-22-21-58-55](https://user-images.githubusercontent.com/31925501/63521349-d9813c80-c528-11e9-9ad3-817255572848.jpeg) When try to stop. Stoping............but not stop. ![Screenshot_2019-08-22-21-59-37](https://user-images.githubusercontent.com/31925501/63521352-d9813c80-c528-11e9-8e90-12b724adf726.jpeg) username_2: yep I've had that happen occasionally. not for a while, though. Status: Issue closed
teamcapybara/capybara
409798356
Title: epeated hover does't work Question: username_0: **Note:** This is for issues with Capybara. If you have a howto type question, please ask on the mailing list as requested in the README: http://groups.google.com/group/ruby-capybara ## Meta Capybara Version: <!-- 2.8.1? --> Driver Information (and browser if relevant): <!-- selenium-webdriver 2.53.4 with Firefox 47.0.1? capybara-webkit? Poltergeist? --> ## Expected Behavior ## Actual Behavior <!-- include full stacktrace of any error --> ## Steps to reproduce <!-- Please be sure to include the code that is creating the issue along with HTML the code is being run against --> Answers: username_1: Do you have an example that reproduces this? Are you moving to somewhere else before trying to hover again? This report really doesn't provide me any way to reproduce your issue. I've created a gist - https://gist.github.com/username_1/14041cfa75a3445e6e155a9990c5d19f- that hovers multiple times on 2 elements and works fine. If you can produce an example that shows your issue then I will take a look at it, otherwise there's not much for me to go on here. username_0: This bug reproduced with firefox 65.0 and later, firefox 64.0.2 works as expected. username_1: The code I linked above works fine for me with FF 65.0.1 - Unless you can provide code that reproduces the issue there is nothing I can do here. username_0: ```ruby visit article_path find_link('Settings').hover click_link 'Edit' # opens edit form click_link 'Back' # back to show find_link('Settings').hover click_link 'Edit' # raises Capybara::ElementNotFound ``` username_1: @Insight What HTML am I supposed to run that against? I need to be able to reproduce this in order to do anything. Take the gist I linked to and modify it to show the issue if your app isn’t publicly available. username_1: Closing - will reopen if/when a way to reproduce this is provided Status: Issue closed username_0: I am reproduced bug in https://gist.github.com/username_0/1440b63dd3fbbaf4c612a0714342f66b. username_1: ## Meta Capybara Version: 3.13.2 <!-- 2.8.1? --> Driver Information (and browser if relevant): selenium-webdriver 3.141.0 geckodriver-v0.24.0-linux64 firefox-65.0 and later <!-- selenium-webdriver 2.53.4 with Firefox 47.0.1? capybara-webkit? Poltergeist? --> ## Expected Behavior ```ruby find_link('Label').hover # successful <going to other page> <going back> find_link('Label').hover # successful yet ``` ## Actual Behavior <!-- include full stacktrace of any error --> ```ruby find_link('Label').hover # successful <going to other page> <going back> find_link('Label').hover # does nothing (without any errors) ``` ## Steps to reproduce <!-- Please be sure to include the code that is creating the issue along with HTML the code is being run against --> https://gist.github.com/username_0/1440b63dd3fbbaf4c612a0714342f66b username_1: @username_0 Thanks - that code does reproduce it for me. As you stated it does appear to be a change in behavior for FF and should probably be reported to the geckodriver project. Short term It's possible to work around it in your project by forcing the mouse to move somewhere else first (call hover on any other element on the page). Medium term think Capybara can work around the issue by moving to an offset on the element before moving to its center for the FF driver. Longer term the best solution would be if geckodriver/FF fixed the behavior username_1: @username_0 Please try the issue_2156 branch to see if it fixes your issue (without breaking anything else) username_0: @username_1 Thank you for your help. issue_2156 branch fixes the issue. But i've found a changing of behavior of #drag_to in FF 65.0.1. We have ordered list of lines: ```rb verify_numbers('NUM1', 'NUM2', 'NUM3') source = find('.number tr', text: 'NUM1').find('.sort-handle') target = find('.number tr', text: 'NUM3') source.drag_to(target) verify_numbers('NUM2', 'NUM3', 'NUM1') ``` Result order is 'NUM2', 'NUM1', 'NUM3'. username_0: I've reported about this issue to geckodriver https://github.com/mozilla/geckodriver/issues/1507. username_1: @Insight If your drag/drop code is using HTML5 drag & drop the change may be because of the change in master Capybara that adds an extra move so JS can determine direction of move. If that's the case then it would be good to know whether or not the current behavior is now correct for the code you're using (and was wrong before), or whether it is now incorrect. It may also depend on exactly where the drop occurs and sub pixel accuracy determine whether the element should be inserted or appended. If the change is caused by the Capybara change (check master vs release) then please create a new issue and provide a way to replicate. If it's purely a browser version/pixel accuracy issue there may not be anything Capybara can do about it since it may be correct behavior as far as the browser is concerned. username_1: Closed via b7620b9851b5c6bb1830344545f2f1c0455102cb Status: Issue closed
C2FO/patio
162023538
Title: Add docs for "magic" methods Question: username_0: Patio adds a lot of power via its model layer and the associations. Currently I do not think there are any docs for the automatically added methods. The * model.addRelated * model.removeRelated As well as other magic methods are not documented. It'd be great to add it to the documentation so people browsing the docs knew they were there. They are tested, which is awesome, so it should be a fairly easy add.
SAP/olingo-jpa-processor-v4
394935187
Title: NoSuchMethodError Question: username_0: java.lang.NoSuchMethodError: com.google.common.collect.Sets$SetView.iterator()Lcom/google/common/collect/UnmodifiableIterator; at org.reflections.Reflections.expandSuperTypes(Reflections.java:380) at org.reflections.Reflections.<init>(Reflections.java:126) at com.sap.olingo.jpa.metadata.core.edm.mapper.impl.IntermediateServiceDocument.createReflections(IntermediateServiceDocument.java:263) <hr> java.lang.NoSuchMethodError: com.google.common.collect.Sets$SetView.iterator()Lcom/google/common/collect/UnmodifiableIterator; at org.reflections.Reflections.expandSuperTypes(Reflections.java:380) at org.reflections.Reflections.<init>(Reflections.java:126) at com.sap.olingo.jpa.metadata.core.edm.mapper.impl.IntermediateServiceDocument.createReflections(IntermediateServiceDocument.java:263) at com.sap.olingo.jpa.metadata.core.edm.mapper.impl.IntermediateServiceDocument.<init>(IntermediateServiceDocument.java:61) at com.sap.olingo.jpa.metadata.core.edm.mapper.impl.JPAServiceDocumentFactory.getServiceDocument(JPAServiceDocumentFactory.java:26) at com.sap.olingo.jpa.metadata.api.JPAEdmProvider.<init>(JPAEdmProvider.java:45) at tv.samim.iptv.yourPackage.common.OdataServlet.service(OdataServlet.java:49) <hr> ```java ServletContext context = getServletContext(); WebApplicationContext ctx = WebApplicationContextUtils .getWebApplicationContext(context); EntityManagerFactory emf = ctx.getBean("entityManagerFactory", EntityManagerFactory.class); // EntityManagerFactory myBean = (EntityManagerFactory) req.getSession().getAttribute(""); // EntityManagerFactory emf = JPAEntityManagerFactory.getEntityManagerFactory(PUNIT_NAME, new HashMap<String, Object>()); JPAEdmProvider metadataProvider = new JPAEdmProvider(PUNIT_NAME, emf, null, new String[]{"test"}); OData odata = OData.newInstance(); ``` Answers: username_1: It turned out that this is very likely a problem with `Guava` dependency and does not belong to jpa processor directly. JPA Processor uses [Reflections API](https://github.com/ronmamo/reflections), which uses `Guava` v20.0. Please check if you have an older `Guava` version in you classpath/dependencies. See also: [reflections issue #219](https://github.com/ronmamo/reflections/issues/219) [reflections issue #194](https://github.com/ronmamo/reflections/issues/194) username_1: I assume the problem has been solved. Otherwise reopen the issue. Status: Issue closed
GuntharDeNiro/BTCT
287661240
Title: [Feature Request] Bind to specific IP Question: username_0: Currently the bot binds its websocket port to all ips on a system, and makes all outgoing requests from whatever the systems first IP address is. Please consider honoring the bot.hostname value in the config; Bind bot/gui websockets to {$bot.hostname}:{$bot.[client]port} Make all outgoing requests bind to {$bot.hostname} This will facilitate running bots on the same host from multiple IP's, for example for hosting bots for multiple different users, each with their own dedicated IP, such as a group of friends sharing a dedicated server. Answers: username_1: Closing due to inactivity. Open a new issue if necessary. Status: Issue closed
Bear127/hello-world-
427065532
Title: Mastering Issues Question: username_0: Issues are a great way to keep track of tasks, enhancements, and bugs for your projects. They’re kind of like email—except they can be shared and discussed with the rest of your team. Most software projects have a bug tracker of some kind. GitHub’s tracker is called Issues, and has its own section in every repository. Answers: username_0: GitHub’s issue tracking is special because of our focus on collaboration, references, and excellent text formatting.
Xceptance/XCMailr
706289073
Title: NPE when viewing the My Emails page Question: username_0: When viewing the "My Emails" page, the following NPE might occur: ``` java.lang.NullPointerException: null at controllers.BoxHandler.queryAllMailboxes(BoxHandler.java:1029) ~[classes/:na] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_162] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_162] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_162] at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_162] at ninja.params.ControllerMethodInvoker.invoke(ControllerMethodInvoker.java:76) ~[ninja-core-6.3.0.jar:na] at ninja.FilterChainEnd.next(FilterChainEnd.java:45) ~[ninja-core-6.3.0.jar:na] at filters.SecureFilter.filter(SecureFilter.java:72) ~[classes/:na] at ninja.FilterChainImpl.next(FilterChainImpl.java:36) ~[ninja-core-6.3.0.jar:na] at ninja.NinjaDefault.onRouteRequest(NinjaDefault.java:102) ~[ninja-core-6.3.0.jar:na] at ninja.servlet.NinjaServletDispatcher.service(NinjaServletDispatcher.java:86) [ninja-servlet-6.3.0.jar:na] at com.google.inject.servlet.ServletDefinition.doServiceImpl(ServletDefinition.java:286) [guice-servlet-4.1.0.jar:na] at com.google.inject.servlet.ServletDefinition.doService(ServletDefinition.java:276) [guice-servlet-4.1.0.jar:na] at com.google.inject.servlet.ServletDefinition.service(ServletDefinition.java:181) [guice-servlet-4.1.0.jar:na] at com.google.inject.servlet.ManagedServletPipeline.service(ManagedServletPipeline.java:91) [guice-servlet-4.1.0.jar:na] at com.google.inject.servlet.FilterChainInvocation.doFilter(FilterChainInvocation.java:85) [guice-servlet-4.1.0.jar:na] at com.google.inject.servlet.ManagedFilterPipeline.dispatch(ManagedFilterPipeline.java:120) [guice-servlet-4.1.0.jar:na] at com.google.inject.servlet.GuiceFilter.doFilter(GuiceFilter.java:135) [guice-servlet-4.1.0.jar:na] at ninja.servlet.NinjaServletFilter.doFilter(NinjaServletFilter.java:106) [ninja-servlet-6.3.0.jar:na] at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1759) [jetty-servlet-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:582) [jetty-servlet-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1180) [jetty-server-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:512) [jetty-servlet-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1112) [jetty-server-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) [jetty-server-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134) [jetty-server-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.server.Server.handle(Server.java:534) [jetty-server-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:320) [jetty-server-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:251) [jetty-server-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:283) [jetty-io-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:110) [jetty-io-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.io.SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93) [jetty-io-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.executeProduceConsume(ExecuteProduceConsume.java:303) [jetty-util-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.produceConsume(ExecuteProduceConsume.java:148) [jetty-util-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.run(ExecuteProduceConsume.java:136) [jetty-util-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:671) [jetty-util-9.3.15.v20161220.jar:9.3.15.v20161220] at org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:589) [jetty-util-9.3.15.v20161220.jar:9.3.15.v20161220] at java.lang.Thread.run(Thread.java:748) [na:1.8.0_162] ``` This can happen if one of the user's mailboxes has been deleted concurrently.<issue_closed> Status: Issue closed
nunit/nunit
177038216
Title: TestCaseSource: "Too many arguments provided" for IEnumerable<> parameter type Question: username_0: @username_2 commented on [Wed Sep 14 2016](https://github.com/nunit/docs/issues/115) It appears that NUnit has started treating Array and Enumerable TestCaseSource parameters differently. I noticed this behaviour when upgrading from 3.0 to 3.4.1.0 (sorry that covers a lot of ground). Say I declare some test data: ``` protected static IEnumerable<IEnumerable<string>> EnumerableOfStrings => new[] { new[] { "one", "two" } }; ``` The following test will pass: ``` [Test, TestCaseSource(nameof(EnumerableOfStrings))] public void AcceptsArrayOfStrings(string[] array) { } ``` However this test will fail with the message **"Too many arguments provided"**: ``` [Test, TestCaseSource(nameof(EnumerableOfStrings))] public void AcceptsEnumerableOfStrings(IEnumerable<string> enumerable) { } ``` It seems with the second test that NUnit is attempting to unbox the enumerable and pass each element as an argument rather than passing the enumerable as a single parameter. --- @username_0 commented on [Wed Sep 14 2016](https://github.com/nunit/docs/issues/115#issuecomment-247175801) I can reproduce this - agreed it seems to be a bug. (Great bug report, by the way! :+1:) I'll also move this issue over to the NUnit framework repo - this ones just for doc issues. :smile: Answers: username_1: Not sure I agree this is a bug - or at least not the same bug you see. It seems to me that the source has to be __either__ for a method that takes a string or for one that takes an IEnumerable<string>. Therefore, both test methods can't be good. One has to be invalid, while the other is valid. I recall that we made some changes to the code to make it work as it does. Are you suggesting that the first one should give an error of "Not enough arguments" or that both should work? This has always been an ambiguous situation for NUnit: only one argument presented as a single array. Should we expect methods to take a string or an array of strings? Is there some way we can make this clearer? It would be interesting to see what are the minimal syntax changes to the TestCaseSource that would cause NUnit to switch over to the alternate interpretation. What happens if we specify either string[] or object[]? username_0: I understand this problem - but with the source here being an ienumerable of ienumerables, doesn't that remove the ambiguity? username_1: @username_0 You're right - this is not the situation I imagined. It could still be useful for whomever takes this one to explore the bounds of the problem by trying alternative formulations of the source. username_0: @username_2 - Are you working on a pull request to fix this? Would be great if so, not sure when I'd get round to it at the moment! username_2: Yeah I've forked the repo and have tests to reproduce the issue and expand upon the problem. Progress is slow at the moment and I'm not sure there's an easy elegant fix but I'll press on when I can. username_0: Great - will leave it with you! username_3: @username_2 did you make any progress on this over the weekend? We are planning on doing the 3.5 release this week and need to know what will be included. username_2: @username_3 nothing that doesn't undo other previous fixes username_2: Pull request #1812 created username_3: I am going to move this to 3.6. I am preparing the release. We might even be able to use your code from #1812 soon after the release. We are thinking about splitting CF out to it's own repository, so your fixes might be fine as is. username_1: We've recently been having some (meta-)discussions about discussing issues. This one should probably have been tossed into the discussion pipeline so we could decide exactly what we want to have work as a test case source. The old definition was that we supported any ienumerable returning a number of possible Types, indicated in the comments as follows: ```C# // 1. Source is null. This is really an error but if we // throw an exception we simply get an invalid fixture // without good info as to what caused it. Passing a // single null argument will cause an error to be // reported at the test level, in most cases. // 2. User provided an ITestCaseData and we just use it. // 3. An array was passed, it may be an object[] // or possibly some other kind of array, which // TestCaseSource can accept. ``` The comments are not too clear about what types of array can be accepted, so the code has to be consulted. Essentiallly, if all the args are of the same type, then we accept an array of that type as the argument. In the past, when an __outer__ array was used to provide the enumeration of cases, it was normally an array of objects because that's how .NET represents argument lists. However, since we accept any enumeration, it could also be a typed array. I don't think this ability was used much in the past but the introduction of `new[]` has probably led to its being used more frequently now. Personally, I don't think we want to extend what the attribute will accept more than necessary. In particular, I don't think a non-array should be used to represent the argument list for a test case. We already stretched a point in accepting, for example, an int[3] array to represent three integer args. The format that uses TestCaseData to hold arguments can always express any ambiguous cases clearly. I think it would be helpful to understand exactly what new use cases are being accepted by the PR under discussion - aside from just fixing the bug - before we merge it. @username_2 Is that something you could articulate? In parallel, I'm going to run your tests against the existing code. @username_2 Thanks for all the work you did on this so far. If it turns out that you went too far, that's our fault for not having discussed what we wanted ahead of time. 😢 username_2: @username_1 Thanks for the feedback, your comments about using an outer array of object make sense and I think that's a crucial piece of the puzzle in my understanding the behaviour of test case data. From the feedback so far and re-evaluating my PR I'm more inclined to think that extending the use-cases may be undesirable - as it puts more burden on NUnit to guess / figure out intent at the expense of explicitness. On reflection the changes I have introduced do not align with the original use cases for outer arrays (hence the changes to UniqueItemsTests that I was unsure about) and so I think the associated PR is not the right way to go. The original reported issue could potentially be addressed through documentation and more specialised exception messages; especially the steps required to fix a failing test are trivial in this instance (ie changing a type from an IEnumerable to an array in your TestFixutre). username_3: Just a heads up. The framework release is likely going to be at the end of the week. I am moving this issue out of the milestone for the release. username_4: Hello, This also happens for me with (**NUnit ~> 3.6.1**): ``` f# [<TestCase ([| null ; null |], ExpectedResult=2)>] static member ``Check Length`` (lines:seq<string>) = lines |> Seq.length ``` it errors stating it was trying to call the function with `(null, null)` as the argument. Can there be a property *`AsIs=true`* or [*`Splat=false`*](https://4loc.wordpress.com/2009/01/16/the-splat-operator-in-ruby/) that lets us retain control of intent. Thank you. Good day.
Macaulay2/M2
790877938
Title: Test errors with --check on clang10 Question: username_0: [This build](https://github.com/Macaulay2/M2/runs/1740207976?check_suite_focus=true) for #1847, I tried to use `--check` for running the tests on `cmake-ubuntu-latest-clang10`, and got two errors: ```m2 -- running check(8, "Core") ulimit -c unlimited; ulimit -t 700; ulimit -m 850000; ulimit -s 8192; ulimit -n 512; cd /tmp/M2-29112-0/3-rundir/; GC_MAXIMUM_HEAP_SIZE=400M "/home/runner/work/M2/M2/M2/BUILD/build/usr-dist/x86_64-Linux-Ubuntu-18.04/bin/M2-binary" -q --int --no-randomize --no-readline --silent --stop --print-width 77 <"/tmp/M2-29112-0/2.m2" >>"/tmp/M2-29112-0/2.tmp" 2>&1 /tmp/M2-29112-0/2.tmp:0:1: (output file) error: Macaulay2 killed by signal 14 /tmp/M2-29112-0/2.m2:0:1: (input file) M2: *** Error 14 ... -- running check(36, "Core") ulimit -c unlimited; ulimit -t 700; ulimit -m 850000; ulimit -s 8192; ulimit -n 512; cd /tmp/M2-29112-0/11-rundir/; GC_MAXIMUM_HEAP_SIZE=400M "/home/runner/work/M2/M2/M2/BUILD/build/usr-dist/x86_64-Linux-Ubuntu-18.04/bin/M2-binary" -q --int --no-randomize --no-readline --silent --stop --print-width 77 <"/tmp/M2-29112-0/10.m2" >>"/tmp/M2-29112-0/10.tmp" 2>&1 /tmp/M2-29112-0/10.tmp:0:1: (output file) error: Macaulay2 killed by signal 14 /tmp/M2-29112-0/10.m2:0:1: (input file) --status: somebody should fix this! [dan] --status: this is a strange one --status: it's a gb computation that seems to run out of memory far too soon --status: Mike? M2: *** Error 14 ``` I'm mostly perplexed about the latter failure. How are commented lines being printed but nothing else? Answers: username_1: I've seen these, too -- see #1392. username_0: Oh I didn't realize both were because of that. Still perplexed about the printed `--status:` lines Status: Issue closed
strimzi/strimzi-kafka-operator
1027781984
Title: Quick start instructions for 0.26 fail to start operator Question: username_0: **Describe the bug** When following the quick start instructions, the operator fails due to a 403 when attempting to watch for CRD changes in project namespaces. **To Reproduce** Steps to reproduce the behavior: 1. Follow the [quick start](https://strimzi.io/docs/operators/latest/quickstart.html) instructions up to step 6 in section 2.3 2. See the error in the K8s dashboard of the strimzi cluster operator pod **Expected behavior** It is expected that the operator is able to start **Environment (please complete the following information):** - Strimzi version: 0.26.0 - Installation method: YAML files from the GitHub releases page - Kubernetes cluster: Kubernetes 1.22.2 - Infrastructure: Minikube 1.23.2 **YAML files and logs** - [logs-from-strimzi-cluster-operator-in-strimzi-cluster-operator-7fc5b5b4b8-4dm7k.log](https://github.com/strimzi/strimzi-kafka-operator/files/7355846/logs-from-strimzi-cluster-operator-in-strimzi-cluster-operator-7fc5b5b4b8-4dm7k.log) - [strimzi-0.26.0-modified.zip](https://github.com/strimzi/strimzi-kafka-operator/files/7355863/strimzi-0.26.0-modified.zip) **Additional context** Running the other [quick start guide](https://strimzi.io/quickstarts/) works, but they differ in that everything is deployed in a single namespace. Answers: username_1: It seems to work fine for me. Are you sure you didn't missed any of the steps? In particular the step 7 from the chapter 2.3 would cause an error like this. username_0: I must have fat fingered something the first time and then gotten hung up on step 6. Thanks for the prompt reply. Everything works as intended. username_1: Great, glad it works. Thanks.
sindresorhus/got
623800224
Title: Tunneling broken by 11.0.0 rewrite Question: username_0: #### Expected behavior Tunneling to be accepted as options.agent parameter, or a working, non-global tunnel integration to be suggested by documentation. #### Checklist - [x] I have read the documentation. - [x] I have tried my code with the latest version of Node.js and Got. Answers: username_1: This is clearly covered in the release notes: https://github.com/username_1/got/releases/tag/v11.0.0 Status: Issue closed username_0: Yep, you're right, thank you!
calpoly-csai/swanton
747299251
Title: Uploading Code Changes Question: username_0: ## Objective We need to have a script to stop the assistant, unzip the new code, rasa model, stories, etc, and update the cron jobs. This will be triggered by the phone app after uploading new code. It should delete the update file after completing the unzip/update/restart of the rasa service.
ikedaosushi/tech-news
619479417
Title: Google Keep の OCR機能が超便利画像の文字を一瞬でテキストに変えてくれる機能 - ふなさんブログ Question: username_0: Google Keep &#12398; OCR&#27231;&#33021;&#12364;&#36229;&#20415;&#21033;&#65290;&#30011;&#20687;&#12398;&#25991;&#23383;&#12434;&#19968;&#30636;&#12391;&#12486;&#12461;&#12473;&#12488;&#12395;&#22793;&#12360;&#12390;&#12367;&#12428;&#12427;&#27231;&#33021; - &#12405;&#12394;&#12373;&#12435;&#12502;&#12525;&#12464;<br> <br> https://ift.tt/2y8i0mV
naptha/tesseract.js
467955076
Title: Possibility to have test result in array by section? Question: username_0: Is it possible or would it be possible to have the resulting text returned in an array or similar ? When scanning invoices where I live there is some specific info in the top left, some other in the top right and so on. Answers: username_1: Many you can check our latest version with new parameters to recognize only part of the image, check here: https://github.com/naptha/tesseract.js/blob/master/docs/examples.md#with-only-part-of-the-image-200-alpha12 username_0: Great, if I want several sections to be analyzed how do I do that? I tried adding another rectangle after the first one but that did not work. username_1: Hi @username_0, Below is a sample code for recognizing multiple rectangles at a time: ```javascript const Tesseract = require('tesseract.js'); const { TesseractWorker } = Tesseract; const worker = new TesseractWorker(); const params = [ { tessjs_image_rectangle_left: 0, tessjs_image_rectangle_top: 0, tessjs_image_rectangle_width: 500, tessjs_image_rectangle_height: 300, }, { tessjs_image_rectangle_left: 500, tessjs_image_rectangle_top: 0, tessjs_image_rectangle_width: 500, tessjs_image_rectangle_height: 300, }, ]; Promise.all( params.map(param => ( worker .recognize( 'https://tesseract.projectnaptha.com/img/eng_bw.png', 'eng', param ) .then(({ text }) => text) )) ).then((texts) => { console.log(texts); worker.terminate(); }); ``` username_0: Thank you for all the useful help. Last question that I hope you/someone can answer. Is there some other "simple" alternative to be able to detect logos in a document? I want to be able to store pictures of 50-500 company logos that are used in invoices and when I get a match I also know how the locations of all fields that I need to use tesseract on username_1: As the last question you asked is out of the scope of tesseract.js, I will close the issue for now. Status: Issue closed username_2: @username_1 Hi, may I know if this is still valid for v2.0.0-beta.2. Thanks :) username_1: In v2.0.0-beta.2, we have updated the API, please check below for example: https://github.com/naptha/tesseract.js/blob/master/docs/examples.md#with-only-part-of-the-image-200-beta1 username_2: I meant the code for multiple rectangles. Current doc doesn’t have an example for multiple rectangles. username_2: Sorry to hijack this thread. Should I create a new issue @username_1? username_1: Yes, please create a new issue to track this issue, thanks.
openml/openml-python
382494413
Title: Cannot run example code for adding dataset Question: username_0: #### Description API returns this error to python package: #### Expected Results Should upload data-set #### Actual Results Throws exception from server side: openml.exceptions.OpenMLServerException: https://www.openml.org/api/v1/xml/data/ returned code 131: Problem validating uploaded description file #### Versions NumPy 1.15.2 SciPy 1.1.0 Scikit-Learn 0.20.dev0 Backend TkAgg is interactive backend. Turning interactive mode on. OpenML 0.8.0dev <!-- Thanks for contributing! --> Answers: username_1: Thanks for reporting this. Can you please copy the code snippet that creates this error? username_0: `citation = ( """ <NAME>, <NAME>, <NAME>, <NAME>, <NAME>, <NAME>, <NAME>. KEEL Data-Mining Software Tool: Data Set Repository, Integration of Algorithms and Experimental Analysis Framework. Journal of Multiple-Valued Logic and Soft Computing 17:2-3 (2011) 255-287. """ ) description = ( 'The weather problem is a tiny dataset that we will use repeatedly' ' to illustrate machine learning methods. Entirely fictitious, it ' 'supposedly concerns the conditions that are suitable for playing ' 'some unspecified game. In general, instances in a dataset are ' 'characterized by the values of features, or attributes, that measure ' 'different aspects of the instance. In this case there are four ' 'attributes: outlook, temperature, humidity, and windy. ' 'The outcome is whether to play or not.' ) for metadata, data in query: print(metadata) for X,y in data: X['class'] = y print(X.info()) print(metadata['Description']) dataset = create_dataset( name = metadata['name'], description='test', creator='KEEL', contributor='Andriy Mulyar', collection_date='09-10-2018', language='English', licence=None, default_target_attribute='class', row_id_attribute=None, ignore_attribute=None, citation=citation, attributes='auto', data=X, version_label='v1_unprocessed', ) upload_did = dataset.publish() print('URL for dataset: %s/data/%d' % (openml.config.server, upload_did))` username_2: Hi Andriy, I bet this happens because the citation has to be basic latin: https://www.openml.org/api_docs#!/data/post_data <xs:element name="citation" minOccurs="0" type="oml:basic_latin1024"/> Possible solutions: - Update the server XSD to allow accents in the citation - Auto-convert such strings in the python API - Return a better error message For now you can just remove the accents, but I agree that we need to handle this better. username_3: Closing this as this is a problem with the main server error message. Status: Issue closed
Fraunhofer-AISEC/cpg
796855405
Title: Implementation of Templates in the CPG Question: username_0: The purpose of this issue is to document the decisions @konradweiss and I made on how to implement CPP Templates. The implementation of the CPP Templates will be split up into two parts: A first representation of the Template in the Graph and a Pass that refines the Template usage, with the same approach used by the CPP Compiler. The main reason for choosing this approach is the fact, that an extensive resolution of the templates leads to a more refined graph, but it might also increase the size of the graph considerably. Therefore, the refinement is implemented as a pass, in order to make the refinement optional if the graph size is too big. Java Generics can also be modelled as a subset of CPP Templates and therefore the current Java Generics representation will be modified to be consistent with the CPP Templates approach. The main differences are: 1. Java Generics can only be applied on Classes, whereas CPP Templates can be applied on Classes and Functions 2. In Java the parameters of the classes can only be Types, compared to CPP Templates, where values can also be provided to the Template. ## Representation To explain how Templates will be represented in the graph, we will be using the following snippet as an example. This snippet represent a template applied on a function, but this is sufficient as applying templates to classes is very similar. Code: ``` #include <iostream> using namespace std; template <class T, int N> T fixed_multiply (T val) { return val * N; } int main() { std::cout << fixed_multiply<int,2>(10) << '\n'; std::cout << fixed_multiply<float,3.0>(10.0) << '\n'; } ``` Graph: ![initial](https://user-images.githubusercontent.com/44771450/106276787-5d6c4780-6238-11eb-8bf2-4a6cdfa132d3.png) As we can see, the graph now includes a new node called Template (exact naming is open for discussion) containing three outgoing edges. The first one is the `realization` edge to the function, which is implemented by the template. The second and third edges are `parameters`-Edges pointing towards the two parameters of the template. The first parameter is a ParameterizedType as introduced by #317 and defined in #293. This represents the abstract Type T and contains two edges two both types it has been initialized in the code (int and float). The second argument is a `ConstantDeclaration`, as the values passed to Templates must be known at compile time and therefore they must be constant. This node also contains `possibleInitialization`-Edges to the Literals that have been found in the initialization of the Template. ## Refinement Pass As explained before, a more refined graph has a bigger size, since every usage of the Template leads to a custom function/class. However, a more refined graph is useful for the analysis of the code, since it may enable us to have a more exact resolution of the code (e.g. exact resolution of a call is possible since we now have type information). Therefore, this pass modifies the graph and 'realizes' every usage of the Template following the same approach as the CPP Compiler. Graph: ![pass](https://user-images.githubusercontent.com/44771450/106276833-6d842700-6238-11eb-956e-4b4a6bc45f8e.png) As we can see in the graph above, now the template contains two `FunctionDeclaration`s (of the same function) with the different realizations, and each CallExpression invokes the correct realized `FunctionDeclaration`. Additionally, the `FunctionDeclaration`s now contain pointer to the exact values that were used to realize the template. Answers: username_1: Just for reference, this is the generated clang AST for the example: ``` |-FunctionTemplateDecl 0x7f904c049038 <template.cpp:4:1, line:8:1> line:5:3 fixed_multiply | |-TemplateTypeParmDecl 0x7f904c048d50 <line:4:11, col:17> col:17 referenced class depth 0 index 0 T | |-NonTypeTemplateParmDecl 0x7f904c048e10 <col:20, col:24> col:24 referenced 'int' depth 0 index 1 N | |-FunctionDecl 0x7f904c048f98 <line:5:1, line:8:1> line:5:3 fixed_multiply 'T (T)' | | |-ParmVarDecl 0x7f904c048ea0 <col:19, col:21> col:21 referenced val 'T' | | `-CompoundStmt 0x7f904c049150 <line:6:1, line:8:1> | | `-ReturnStmt 0x7f904c049140 <line:7:3, col:16> | | `-BinaryOperator 0x7f904c049120 <col:10, col:16> '<dependent type>' '*' | | |-DeclRefExpr 0x7f904c0490e0 <col:10> 'T' lvalue ParmVar 0x7f904c048ea0 'val' 'T' | | `-DeclRefExpr 0x7f904c049100 <col:16> 'int' NonTypeTemplateParm 0x7f904c048e10 'N' 'int' | |-FunctionDecl 0x7f904c0496c8 <line:5:1, line:8:1> line:5:3 used fixed_multiply 'int (int)' | | |-TemplateArgument type 'int' | | |-TemplateArgument integral 2 | | |-ParmVarDecl 0x7f904c0495d0 <col:19, col:21> col:21 used val 'int':'int' | | `-CompoundStmt 0x7f904c82fe30 <line:6:1, line:8:1> | | `-ReturnStmt 0x7f904c82fe20 <line:7:3, col:16> | | `-BinaryOperator 0x7f904c82fe00 <col:10, col:16> 'int' '*' | | |-ImplicitCastExpr 0x7f904c82fde8 <col:10> 'int':'int' <LValueToRValue> | | | `-DeclRefExpr 0x7f904c82fd88 <col:10> 'int':'int' lvalue ParmVar 0x7f904c0495d0 'val' 'int':'int' | | `-SubstNonTypeTemplateParmExpr 0x7f904c82fdc8 <col:16> 'int' | | `-IntegerLiteral 0x7f904c82fda8 <col:16> 'int' 2 | `-FunctionDecl 0x7f904c82faf8 <line:5:1, line:8:1> line:5:3 used fixed_multiply 'float (float)' | |-TemplateArgument type 'float' | |-TemplateArgument integral 3 | |-ParmVarDecl 0x7f904c82fa00 <col:19, col:21> col:21 used val 'float':'float' | `-CompoundStmt 0x7f904c82ff08 <line:6:1, line:8:1> | `-ReturnStmt 0x7f904c82fef8 <line:7:3, col:16> | `-BinaryOperator 0x7f904c82fed8 <col:10, col:16> 'float' '*' | |-ImplicitCastExpr 0x7f904c82fea8 <col:10> 'float':'float' <LValueToRValue> | | `-DeclRefExpr 0x7f904c82fe48 <col:10> 'float':'float' lvalue ParmVar 0x7f904c82fa00 'val' 'float':'float' | `-ImplicitCastExpr 0x7f904c82fec0 <col:16> 'float' <IntegralToFloating> | `-SubstNonTypeTemplateParmExpr 0x7f904c82fe88 <col:16> 'int' | `-IntegerLiteral 0x7f904c82fe68 <col:16> 'int' 3 `-FunctionDecl 0x7f904c0491c0 <line:10:1, line:13:1> line:10:5 main 'int ()' `-CompoundStmt 0x7f904c82fd78 <col:12, line:13:1> ``` Status: Issue closed
gradle/gradle
360699797
Title: Gradle can't resolve dependencies classpath Question: username_0: ``` ### Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> Trying to run the project and modify the frontend ### Steps to Reproduce (for bugs) <!--- Provide a self-contained example project (as an attached archive or a Github project). --> <!--- In the rare cases where this is infeasible, we will also accept a detailed set of instructions. --> ``./gradlew tomcatRun`` should reproduce it ### Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> <!--- A build scan `https://scans.gradle.com/get-started` is ideal --> * Build scan URL: I'm on mac OS 10.12.6 (16G1510) I can't scan it, it gives me below error ``` Stanleys-MacBook-Pro:epc-photos-cloud-web stanley$ gradle build --scan Starting a Gradle Daemon (subsequent builds will be faster) FAILURE: Build failed with an exception. * Where: Auto-applied by using --scan * What went wrong: Plugin [id: 'com.gradle.build-scan', version: '1.16', artifact: 'com.gradle:build-scan-plugin:1.16'] was not found in any of the following sources: - Gradle Core Plugins (plugin is not in 'org.gradle' namespace) - Plugin Repositories (could not resolve plugin artifact 'com.gradle:build-scan-plugin:1.16') Searched in the following repositories: Gradle Central Plugin Repository * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org BUILD FAILED in 36s ``` Answers: username_1: Could you share a sample project, please? username_2: Closing given the lack of reaction and the description most likely indicates a networking / setup issue. Would need at least a stacktrace or a reproducer as requested before re-opening. Status: Issue closed
onitake/daliserver
128191213
Title: Updated Mas OS kext Question: username_0: The dummy kext wasn't allowing me to claim the device on Mac OS El Capitan 10.11.2 . I updated the plist with kpi libraries but -- in the end -- it was essential to declare IOProviderClass as IOUSBDevice and not IOUSBInterface. Please note the libkern library could likely be sent back a few versions but I haven't the means to test it. Also, the below is unsigned so SIP must be disabled to use it. <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd"> <plist version="1.0"> <dict> <key>OSBundleLibraries</key> <dict> <key>com.apple.iokit.IOUSBFamily</key> <string>6.0</string> <key>com.apple.kpi.libkern</key> <string>10.4.2</string> </dict> <key>CFBundleDevelopmentRegion</key> <string>English</string> <key>CFBundleGetInfoString</key> <string>DALI USB Interface Dummy Driver</string> <key>CFBundleIdentifier</key> <string>com.tridonic.DALIUSB</string> <key>CFBundleInfoDictionaryVersion</key> <string>6.0</string> <key>CFBundleName</key> <string>DALI USB Interface</string> <key>CFBundlePackageType</key> <string>KEXT</string> <key>CFBundleSignature</key> <string>????</string> <key>CFBundleVersion</key> <string>1.0.0</string> <key>IOKitPersonalities</key> <dict> <key>Device Driver</key> <dict> <key>CFBundleIdentifier</key> <string>com.apple.kpi.iokit</string> <key>IOClass</key> <string>IOService</string> <key>IOProviderClass</key> <string>IOUSBDevice</string> <key>idProduct</key> <integer>32</integer> <key>idVendor</key> <integer>6069</integer> <key>bConfigurationValue</key> <integer>1</integer> <key>bInterfaceNumber</key> <integer>0</integer> </dict> </dict> <key>OSBundleRequired</key> <string>Local-Root</string> </dict> </plist> Answers: username_1: Thanks a lot. I haven't tested on OS X in a long time. I know someone who has a kernel extension code signing certificate, so I was planning to ask them to sign the kext for me. I'll do that once I've had a chance to test your changes. Looks like github messed with the formatting, can you create a gist instead? username_0: You're welcome. Here's the gist, hope it's better. https://gist.github.com/username_0/3332c930c04e4892bef2 username_0: By the way, I went through the Perl scripts and have some comments. Lampoff.pl crashes daliserver... for what reason, I don't know? I'm happy to help find out why. We can discuss offline sometime if you would like. username_1: Uh, what? I know that `usbdali.pm` is a bit buggy and doesn't return the response correctly. I fixed that problem locally, but haven't pushed it to github yet. Are you positive it crashes daliserver? username_1: Ok, fixed `usbdali.pm` in bae41b12. username_0: I pulled in the changed file but no change. Lampoff.pl successfully turns off a lamp but the server crashes after. Two commands issued below: Braque:perl benknuth$ perl allset.pl 255 Writing 4 bytes to socket, address=254 command=254 Received status:success response:0 Braque:perl benknuth$ perl lampoff.pl 12 Writing 4 bytes to socket, address=25 command=0 Braque:perl benknuth$ [2016-00-22 20:28:13] INFO Starting daliserver [2016-00-22 20:28:13] INFO Server ready, waiting for events [2016-00-22 20:28:32] INFO Got connection from 127.0.0.1:27089 [2016-00-22 20:28:32] INFO Got frame: 0x02 0x00 0xfe 0xfe [2016-00-22 20:28:32] INFO Enqueued transfer (0x7fb7abc06cc0,0x7fb7abc06a90) [2016-00-22 20:28:32] INFO Sending data to device: 0x00000000 12 01 00 03 00 00 fe fe 00 00 00 00 00 00 00 00 ................ 0x00000010 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ 0x00000020 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ 0x00000030 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ Direction: USB<->DALI Sequence number: 01 Type: 16bit DALI Address: fe Command: fe [2016-00-22 20:28:32] INFO Received data from device: 0x00000000 12 73 00 00 fe fe ff ff 01 00 00 00 00 00 00 00 .s.............. 0x00000010 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ 0x00000020 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ 0x00000030 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ Direction: USB<->DALI Type: Send complete Address: fe Command: fe Status: ffff [2016-00-22 20:28:32] INFO Received data from device: 0x00000000 12 71 00 00 00 00 00 8a 01 00 00 00 00 00 00 00 .q.............. 0x00000010 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ 0x00000020 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ 0x00000030 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ Direction: USB<->DALI Type: Send response Response: ff Status: 008a Sequence number: 01 [2016-00-22 20:28:32] INFO Response to (0x00 0x00): 0xff [0x008a] [2016-00-22 20:28:32] INFO Connection 9 was disconnected [2016-00-22 20:28:32] INFO Closing connection 9 [2016-00-22 20:28:45] INFO Got connection from 127.0.0.1:27345 [2016-00-22 20:28:45] INFO Got frame: 0x02 0x00 0x19 0x00 [2016-00-22 20:28:45] INFO Enqueued transfer (0x7fb7abe01520,0x7fb7abe03090) [2016-00-22 20:28:45] INFO Connection 9 was disconnected [2016-00-22 20:28:45] INFO Closing connection 9 [2016-00-22 20:28:45] INFO Sending data to device: 0x00000000 12 02 00 03 00 00 19 00 00 00 00 00 00 00 00 00 ................ 0x00000010 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ 0x00000020 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ 0x00000030 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ Direction: USB<->DALI Sequence number: 02 Type: 16bit DALI Address: 19 Command: 00 [2016-00-22 20:28:45] INFO Received data from device: 0x00000000 12 73 00 00 19 00 ff ff 02 00 00 00 00 00 00 00 .s.............. 0x00000010 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ 0x00000020 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ 0x00000030 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ Direction: USB<->DALI Type: Send complete Address: 19 Command: 00 Status: ffff [2016-00-22 20:28:45] INFO Received data from device: 0x00000000 12 71 00 00 00 00 00 8a 02 00 00 00 00 00 00 00 .q.............. 0x00000010 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ 0x00000020 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ 0x00000030 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ................ Direction: USB<->DALI Type: Send response Response: ff Status: 008a Sequence number: 02 [2016-00-22 20:28:45] INFO Response to (0x00 0x00): 0xff [0x008a] [2016-00-22 20:28:45] **ERROR** Error writing 4 bytes to connection 2059272945: Bad file descriptor Segmentation fault: 11 Braque:~ benknuth$ username_1: Ouch! I'll look into this right away. It's clear that the connection is killed before the response can be sent, as some of the Perl scripts don't read the response at all. But the crash is in no way ok! Can you run daliserver in gdb and produce a backtrace? username_0: Here's the backtrace. Hope this helps! <script src="https://gist.github.com/anonymous/bced4b4bc594748b0d3f.js"></script> These two commands sent were: perl lampset.pl 12 255 Writing 4 bytes to socket, address=24 command=254 Received status:success response:0 perl lampoff.pl 12 Writing 4 bytes to socket, address=25 command=0 username_1: Umm... I didn't get your backtrace. Did you forget the link? But no matter, I found the bug. It's just a bit hard to fix. Need to think about it some more. As a temporary workaround, I pushed a modified lampoff.pl that checks the server response. This should prevent the crash. username_1: I moved the crash report into a new issue #12, please post there. username_1: To get back to the issue, can you change com.apple.kernel.libkern back to 1.8 and see if it still works on your system? username_0: Changed it back and it works with no apparent problems. username_1: Sorry, I completely missed your message. I'll commit the updated Info.plist then. Status: Issue closed
ros2/launch
450064923
Title: remove conceptual dependency on ROS_DOMAIN_ID in launch_testing Question: username_0: Currently `launch_testing` has options/features related to `ROS_DOMAIN_ID`: https://github.com/ros2/launch/blob/066009b097e06dad7f981de55b9ee05c83cf6e0b/launch_testing/launch_testing/launch_test.py#L51 This should not be the case, it is inappropriate for anything in this repository to depend on, reference, or really even mention anything from ROS. There's a TODO, which I linked to, but I opened this issue to raise the severity. This needs to be changed. @username_2 @username_1 please don't let things like this leak into `launch` or `launch_testing` in the future. Answers: username_1: Do you have any suggestions about how to fix it? The cure (removing this) seems worse than the disease right now username_1: Could we add an extension point to launch_test so that when launch_testing_ros is installed you get the isolation behavior? username_1: I guess we could also add a different runner with a different name to run tests that need ROS username_2: I fully agree. As it stands, this option could be moved to the [`ros2test`](https://github.com/ros2/ros_testing/blob/master/ros2test/ros2test/command/test.py#L24) CLI command with little to no effort (owing to the fact it's just an envvar). username_1: @username_2 will we add a different cmake function that calls the ros2cli command? The dependency graph for that cmake function will be quite long, but maybe not that much longer than launch_ros username_2: @username_1 there is one already, [`add_ros_test()`](https://github.com/ros2/ros_testing/blob/master/ros_testing/cmake/add_ros_test.cmake). username_1: @username_2 I want to refactor the domain coordinator into its own package so I can do this https://github.com/ament/ament_cmake/issues/172 so lets's make sure we coordinate on this. username_3: Sounds good. Where should the domain coordinator live? If it doesn't fit well in an existing repo, I will need a new one (I can't create one). username_1: @username_3 In the WIP I've got now, it lives in ament_cmake_ros along with a new ament_cmake_pytest_isolated and an ament_cmake_gtest_isolated package username_3: I think `ament_cmake_pytest_isolated` and `ament_cmake_gtest_isolated` should live in `ament_cmake_ros`. But the code of the domain coordinator could be in a separated repo, as it is really a python package (and not cmake stuff). It's ok for me having it there though. username_1: @username_3 Let's talk about it in the MR once I open it, but there's one cmake thing that I jammed into the domain_coordinator package - it might make sense to move it, or it might not. Status: Issue closed
mintingle/appinventor-sources
201954361
Title: Persist data when device is not connected to internet Question: username_0: See FirebaseDB's persist property. When the device is not connected to the internet, if the persist option is checked, then data will be pushed to the Cloud when a connection is re-established. This feature would be helpful for data collection apps in the field.
sebastianbergmann/phpunit
782132708
Title: Parse @covers of parent-classes too Question: username_0: PHPUnit has the feature to be strict about covers annotation. Test subjects may use multiple very commonly used methods. which would force you to write every single commonly used method in a `@covers` doc-comment. _(Note: We only talk about methods that can not be mocked for the test)_ Example: * An abstract class (e.g. `Constraint`) is template for multiple other classes (test subject). * Several other classes (e.g. test subjects like `IsAnything`, `IsFalse` and 20 more) reuse this abstract class. * The abstract itself uses tons of commonly used methods (e.g. the `Constraint::__construct` is one). * With strict coverage you would need to repeat tons of methods in each single test as `@covers` doc-comment. ```php <?php /** @covers Constraint::__construct */ class IsAnythingTest extends ConstraintTestCase {} /** @covers Constraint::__construct */ class IsFooTest extends ConstraintTestCase {} /** @covers Constraint::__construct */ class IsBarTest extends ConstraintTestCase {} /** @covers Constraint::__construct */ class IsBazTest extends ConstraintTestCase {} ``` _Note: This example does not ignore the fact that `@covers` can mark a complete class for coverage. But it does not use this to create a scenario where `@covers` need to be repeated way to often in each single test when using the "beStrictAboutCoversAnnotation" feature._ Possible solution: * `\PHPUnit\Util\Test::getLinesToBeCoveredOrUsed` should not only parse the current test-class but also all it's parents * By doing this it would also parse the traits of the parents (see `\PHPUnit\Util\Test::parseTestMethodAnnotations` ) Advantage: ```php <?php /** @covers Constraint::__construct */ class ConstraintTestCase {} class IsAnythingTest extends ConstraintTestCase {} class IsFooTest extends ConstraintTestCase {} class IsBarTest extends ConstraintTestCase {} class IsBazTest extends ConstraintTestCase {} ``` The often used methods can be centralized instead of repeating it every time. It is way easier to make use of the "beStrictAboutCoversAnnotation"-feature with this enhancement.
ethers-io/ethers.js
333634863
Title: String to bytes32 ? padZeros adds padding to front instead of back. Question: username_0: So this kind of resembles the last issue I posted. How can I go from a string to solidity bytes32 value? I'm trying `utils.hexlify(utils.padZeros(utils.toUtf8Bytes('Test'), 32))` , but it seems to add the in front rather than at the end. ` AssertionError: Names should be equal: expected '0x5465737400000000000000000000000000000000000000000000000000000000' to equal '0x0000000000000000000000000000000000000000000000000000000054657374'` Should I use utils.keccak256() instead to encode my parameters on both read and write contract calls? Answers: username_1: You should be sure to check out this issue: https://github.com/ethers-io/ethers.js/issues/66 Basically, the ability for Web3 to treat strings as bytes32 and vice versa is at best a bug, at worst, irresponsible. :) Stings and bytes32 are completely incompatible, however, Web3 supported special cases to make them sort of behave together, but leads to ambiguous output. The above issue includes sample code on how to do this, if this is _really_ what you want to do, or have to be backwards compatible with an antiquated Web3 application, but I strongly recommend against it, as it is really relying unsafe operations. With this method, you string must be less than 32 bytes long (not characters... For example, it must be less than 10 Japanese characters long, or depending on the words and composed form between 10 and 32 characters in German). This method also requires the application to understand implicit padding has occurred. The library cannot do this, since it cannot know whether it was a string or binary data. Consider the data "0x303132330000", passed into a library, it won't know if it represents a) a 14 character long literal string, b) the padded string "1234" or c) 4 bytes of data. The best thing to do, if you truly need strings is to use the Solidity string type, which is a dynamic type, with a length and all that jazz. If you can get by with hashed strings, that is also a great option, but if you only have a small set of strings to choose from, perhaps a simple enum makes sense? Hopefully that makes sense. I'll leave the ticket open though, so we can discuss it further if there are more questions. :) username_0: Hey thanks for the answer. One thing I don't understand is why is it a bug when solidity has the same behaviour? When you enter a string literal in a contract that is supposed to be a bytes32 value , it will automatically convert the value. Web3 doing this as well thus doesn't really seem like a bug but intended behaviour? Example: A proxy contract that gets the address of the library to delegate to from a registry through a string. The string however is a bytes32 value in reality. ``` pragma solidity ^0.4.23; /** * @title Proxy * @dev Gives the possibility to delegate any call to a foreign implementation. */ import '../registry/contract-registry.sol'; import '../registry/enabled.sol'; contract PRProxy is Enabled { /** * @dev Fallback function allowing to perform a delegatecall to the given implementation. * This function will return whatever the implementation call returns */ function () payable public isEnabled("controller") { ContractRegistry contractRegistry = ContractRegistry(CMC); address _impl = contractRegistry.getLibrary("pullrequests"); //THIS LINE <<<<<<<<<<<< require(_impl != address(0)); assembly { let ptr := mload(0x40) calldatacopy(ptr, 0, calldatasize) let result := delegatecall(gas, _impl, ptr, calldatasize, 0, 0) let size := returndatasize returndatacopy(ptr, 0, size) switch result case 0 { revert(ptr, size) } default { return(ptr, size) } } } } ``` This is the relevant code in the registry contract: mapping(bytes32 => address) public libraries; function getLibrary(bytes32 _name) external view returns (address) { return libraries[_name]; } username_0: I took the things you mentioned in the referenced topic into consideration. There are currently only two types of strings I store: short strings and ipfs hashes For the short strings I will use this function on the client side: ``` function stringToBytes32(text) { let result = utils.toUtf8Bytes(text) if (result.length > 32) { throw new Error('String too long') } result = utils.hexlify(result); while (result.length < 66) { result += '0'; } if (result.length !== 66) { throw new Error("invalid web3 implicit bytes32"); } return result; } For IPFS Hash conversion I will be using base58 // Return bytes32 hex string from base58 encoded ipfs hash, // stripping leading 2 bytes from 34 byte IPFS hash // Assume IPFS defaults: function:0x12=sha2, size:0x20=256 bits // E.g. "QmNSUYVKDSvPUnRLKmuxk9diJ6yS96r1TrAXzjTiBcCLAL" --> // "0x017dfd85d4f6cb4dcd715a88101f7b1f06cd1e009b2327a0809d01eb9c91f231" getBytes32FromIpfsHash(ipfsListing) { return ( "0x" + bs58 .decode(ipfsListing) .slice(2) .toString("hex") ) } // Return base58 encoded ipfs hash from bytes32 hex string, // E.g. "0x017dfd85d4f6cb4dcd715a88101f7b1f06cd1e009b2327a0809d01eb9c91f231" // --> "QmNSUYVKDSvPUnRLKmuxk9diJ6yS96r1TrAXzjTiBcCLAL" getIpfsHashFromBytes32(bytes32Hex) { // Add our default ipfs values for first 2 bytes: // function:0x12=sha2, size:0x20=256 bits // and cut off leading "0x" const hashHex = "1220" + bytes32Hex.slice(2) const hashBytes = Buffer.from(hashHex, "hex") const hashStr = bs58.encode(hashBytes) return hashStr } ``` username_1: Ahoy hoy! :) It's not the EVM that adds the padding, it is the ABI coder, which processes the data before sending it to the blockchain. The final data sent will always be word aligned (32 byte words) with a 4-byte signature. The EVM doesn't actually care about this, but all the current tools enforce this, since the compiled Solidity makes code that runs on the EVM that expects this. The EVM never pads, Solidity (or whatever language you use) builds any runtime code that pads (or masks); the exception being when you read past memory, calldata, et cetera. So, regarding Solidity performing the same behaviour (which I still argue occasionally over ;)) is not exactly what it does. There are two things to note: a) Solidity can only go in one direction (pass a string into bytes32 and it will "comply", but try passing a bytes32 into a string and the world will end) and b) Solidity is a statically typed language. a) More on the one-direction of this in Solidity has to do with the fact the types aren't compatible, but it can sort of fake it, by calling the function passing the `add(data, 32)` (the pointer to the string data, after the string length) and the receiving function will mask the value as needed. Passing a bytes32 into a string though, cannot be done easily and would almost certainly fail to be a valid UTF-8 string, most binary data isn't. b) This is the biggest difference between a language like JavaScript and Solidity. By being a dynamic language, and one in which strings are often used to represent hexadecimal encoded strings (the alternative makes using the library nearly impossible; using a Binary wrapper object), means that allowing strings in places where binary data is allowed is unsafe. Web3 does this "magic", and during audits I've seen many potential security and usability problems as a result. ``` // In Solidity, this is exactly 2 bytes, always 0x1234 // In Solidity, this is exactly 6 bytes, always "0x1234" // In JavaScript? 2 bytes of data or 6 bytes of string? "0x1234" ``` ---- So, for your example, if you are hard-coding strings into the call, "pullrequests", it is not uncommon to hardcode the hash sha3("pullrequests") as a constant. This save gas costs when the key is know (for example at compile time), but still allows for libraries to dynamically lookup values if necessary. I believe a bytes32 would also be cheaper on gas than a short string, since masking is required in the latter case. Also, if there are only a handful of fixed values, enums might also help. ``` bytes32 constant pullrequests = 0x5a240ef2e352806693493f8b2318e232cb084a2f64dd84f24d69d7b249731fe6; function getLibrary(string name) returns (address) { return getLibrary(keccak256(name)); } function getLibrary(bytes32 nameHash) (address) { return _libs[nameHash]; } ``` That said, I don't think that is even necessary, as I think all storage lookups execute a keccak256 anyways, in which case it is not (meaningly) more expensive to use a string or a bytes32. But I'm not sure what the current compiler does. For IPFS multihash, I absolutely agree, storing the bottom 32 bytes of raw hash is the way to go. I think it is ridiculous they made a 34 byte hash size... I think it would have made far more sense to keep the top byte as a version, and select just the last 31-bytes of the sha2 for the rest... They could always bump it up to a 2-byte + 30-byte in the future (just make the first by outside the range) and everyone would be happy... I mean, 20 bytes is enough for crypto-currency addresses... Le sigh. Hope this helps. :) username_1: Closing this, but if there are still further discussion points, please feel free to re-open. :) Status: Issue closed username_2: I'd like this issue reopened as I don't think proper consideration has been given to the extent by which `string` to `bytes32` coercion is in use. It has always been a feature of Web3 as a reflection of Solidity's value assignment and parameterisation conversions. It has thereby been the accepted practice for the life of Solidity itself so it makes no sense to diverge from it. The concerns raised against the coercion amount to little more than requiring additional bounds checking in the library which doesn't seem like such a hard task. Practically, it has made Remix near impossible to use for such parameters (which are fundamental to my project). username_1: I have added safe operations to v4 for this, `ethers.utils.formatBytes32String` and `ethers.utils.parseBytes32String`. There are a lot of nuances required to make the coercion safe, but I’ve updated the UTF8 library in the v4 branch to correctly detect (and fail on) overlong sequences, invalid UTF16 surrogate pairs and provide bounds checking on character ranges. You can install v4 using `npm install ethers@next`. The documentation is almost ready, at which time v4 will be the primary release. username_3: Created a little GUI for this here: https://blockchangers.github.io/solidity-converter-online/ I see there is some questions around this. username_4: When I input a 16byte hexcode string into `ethers.utils.formatBytes32String` I get an error trying to use it as a bytes32 value. When I use `ethers.utils.toUtf8Bytes` it works as expected. Anyone know if this is an incorrect approach? username_1: @username_4 What are you trying to do? A 16 byte value will be 34 bytes long and a bytes32 strong may only be 31 or fewer bytes. You could either store in as a `bytes16` or as a string, if you are trying to pass it into a contract. Make sense? username_4: I was generating a 256 bit hash which I thought would be 32 bytes long and 16 bytes in hexcode format. But putting that hexcode string into `ethers.utils.formatBytes32String` produced a value longer than 32 bytes. But based on what you are saying 16 bytes hexcode is 34 bytes long. So can I use `bytes16` solidity param for a 256 bit hash in hexcode format? Sorry this is a bit off-topic from the issue but appreciate the help anyway. username_5: Now, result is different with actual accounts. By researching, I came to know that console.log(typeof(signature.r)) is string but not bytes32. console.log(typeof(sigR)) is object but not array. As signature.r or signature.s have length with 66 which is greater than 32. How can we convert 66 characters length of string to bytes32 and pass to execute function to achieve same result i.e, recovered accounts are same to actual accounts where we sign. username_1: @username_5 In ethers, a hex data string that is 66 bytes long (i.e. `0x` prefixed with 64 *nibbles*, so the 66 **characters** represents 32 **bytes**) can be used as a bytes32 in most cases. My guess is that the `sign` method you are using in Web3 may not be using [EIP-191](https://eips.ethereum.org/EIPS/eip-191) prefixing? But I'm not sure. You can experiment with the [signDigest method](https://docs.ethers.io/v5/api/utils/signing-key/#SigningKey) to see if that answer corresponds to the address you are getting? Keep in mind, this function is not generally available in Etheruem as its use in signing messages can result in security vulnerabilities. Check out [these examples](https://docs.ethers.io/v5/api/utils/hashing/#utils-hashMessage) to see if that helps get you on track. :) username_5: @username_1 The sign method is working perfectly alright. I have tested using remix. In remix - When I give sigR and sigS in ["0xfds...","0xgfhghk..."] this format, I am able to recover deserved addresses. In remix - When I give sigR and sigS in ['0xfds...','0xgfhghk...'] this format, I am unable to recover deserved addresses. so, I am converting ['0xkjhjkvhjl...','0xkjhgdff...'] to ["0xfds...","0xgfhghk..."] format to call solidityContract.methods.execute function, but here I am getting error as {TypeError: param.map is not a function at ABICoder.formatParam (node_modules/web3-eth-abi/lib/index.js:218:22) } probably, when I checked it is taking my conversion as '["0xfds...","0xgfhghk..."]' format. So, when I tried to give ['0xkjhjkvhjl...','0xkjhgdff...'] this format to solidityContract.methods.execute function - i am unable to recover desired accounts. I think the array is always considered as object but not array. that too, bytes32 is considered as string. So, here how do I convert object - [ '0xceb1aaf5c07b59a1fa4fbc1477e0f7a8f38f5<KEY>', '<KEY>' ] format to [ "<KEY>", "<KEY>"] - this format as this format got executed in remix. username_5: @username_1, please re-open the issue as above mentioned problem is not solved. username_1: @username_5 I don’t think this is the issue you mean though. A Bytes32String is quite unrelated to bytes32 in general. The methods involved here are strictly for converting between legacy contracts (written pre-2015-ish; some new age contracts do this too, but it is heavily discouraged) which encoded UTF-8 strings as null-terminated and zero-padded bytes32 `BytesLike` objects and the equivalent UTF-8 code points converted to a string. I don’t think these are the methods you are looking for…
rjwang/llvvAnalysis
144496759
Title: Missing PDF files: llvvAnalysis/DMAnalysis/data/*.csv Question: username_0: CSV files containing PDF information are required in mainAnalyzer, but are not present in the repo. E.g. "llvvAnalysis/DMAnalysis/data/NNPDF30_lo_as_0130_hessian_60.csv". Answers: username_1: it is solved now Status: Issue closed username_0: Great, thanks for fixing. I tested it successfully.
fatedier/frp
817305729
Title: login to server failed: dial tcp 192.168.127.12:7000: connectex: No connection could be made because the target machine actively refused it. dial tcp 192.168.127.12:7000: connectex: No connection could be made because the target machine actively refused it. Question: username_0: login to server failed: dial tcp 192.168.127.12:7000: connectex: No connection could be made because the target machine actively refused it. dial tcp 121.61.63.**:7000: connectex: No connection could be made because the target machine actively refused it. Answers: username_1: 请关闭系统防火墙 Status: Issue closed
zzboncak/the-beer-cellar-client
567076328
Title: A user should be automatically logged in when a account is created Question: username_0: Notes: - Need to hide "login" and "register" buttons if the user has an Auth token in local storage (Est. Time: 20 Minutes) - Need the POST /api/user to issue a JWT with the newly created credentials so as to log the new user in (Est. Time: 20 minutes) - Need to redirect the user to the /cellar route once JWT is issued and stored (Est. Time: 20 minutes). Total time: 1 hour
nodejs/help
570800791
Title: Read-only file with fs.createWriteStream Question: username_0: Is there a way to create a read-only file with `fs.createWriteStream` after the last bit is written or do I have to call `chmod` after the fact? Answers: username_1: @username_0 - what is the use case for it? while the stream is flowing, the destination has to be writable right? if the destination has to be read-only, it has to be after the last bit is written (as you already said), in which case the stream's life cycle has ended. So I would see this has detached from stream's scope, and so yes, doing `chmod` after the write looks like the most reasonable thing to me! Status: Issue closed username_1: closing, pls reopen if you still have issue
FlyBase/drosophila-anatomy-developmental-ontology
701962066
Title: Lineage bundle Question: username_0: Proposed definition: A primary neurite tract or primary axon tract consisting of neurites of all of the neurons in a single lineage or hemilineage. See https://www.cell.com/cms/10.1016/j.neuron.2020.08.005/attachment/f7602644-8c73-45ad-a36d-b2d36af4e758/mmc1.pdf for examples. Problem case: there is bundle in the VNC that has two lineages. Bodge this for now? Note - for PNTs in the brain, which are of a single hemilineag but don't have all the cells, we can make part_of relationships to lineage/hemilineage clone terms. Answers: username_1: see also #783 #779 Looking back over the (Hartenstein) brain papers, I think it seems reasonable to define this as consisting of a single lineage/hemilineage - this seems to be how they are defined for the brain (PATs and SATs), even if two are found together. We may also want a term for a 'multi-lineage bundle' where a small number of lineage bundles form a larger bundle e.g. DPLp1/2 of brain or 20/22 of VNC. Status: Issue closed
department-of-veterans-affairs/va.gov-team
1030939140
Title: [Quality Assurance Audit] Post Production Audit of VAMC VA Southeast Louisiana health care for links, standards, and design standards.[BA] Question: username_0: ## User Story As a VHA Digital Media web manager, the system websites need to be audited for standardization and mistakes, so websites can be updated an corrected to meet design standards. ## Goal Remove deviations from standards. ## Resources - Tools - Documentation [Audit checklist](https://governmentcio.sharepoint.com/sites/VSATeam/Shared%20Documents/Forms/AllItems.aspx?newTargetListUrl=%2Fsites%2FVSATeam%2FShared%20Documents&viewpath=%2Fsites%2FVSATeam%2FShared%20Documents%2FForms%2FAllItems%2Easpx&viewid=0df60f54%2Ddd97%2D4321%2Da9e5%2D9c92d607ae91&id=%2Fsites%2FVSATeam%2FShared%20Documents%2FVAMC%2DWeb%5FModernization%2FChecklists%5Fand%5Ftools) ## Tasks Using the audit check list, review all areas of Fully published websites (Pass or Fail) web pages and provide notes: - [x] Review all pages using checklist. ## Acceptance Criteria - [x] Rename the file as [system name] audit final [date] and save in: [QA_Documents folder](https://governmentcio.sharepoint.com/sites/VSATeam/Shared%20Documents/Forms/AllItems.aspx?newTargetListUrl=%2Fsites%2FVSATeam%2FShared%20Documents&viewpath=%2Fsites%2FVSATeam%2FShared%20Documents%2FForms%2FAllItems%2Easpx&viewid=0df60f54%2Ddd97%2D4321%2Da9e5%2D9c92d607ae91&id=%2Fsites%2FVSATeam%2FShared%20Documents%2FVAMC%2DWeb%5FModernization%2FQA%5FDocuments) - [x] Document with corrections/proposed changes is saved in QA Documents folder Answers: username_1: Task #31741 - Post Production Audit of VAMC VA Southeast Louisiana health care for links, standards, and design standards.[BA] - is Complete and moved to Validate. Status: Issue closed
chesdenis/DataGenies
611166309
Title: Redo the reflection that works with property Question: username_0: We should expand the ability to use the Property attribute. ```CSharp public class PropertiesAttribute : Attribute { public string Name { get; set; } public PropertiesAttribute(string name) { this.Name = name; } } ``` We need to add a name to this attribute and make sure that values are extracted from clusterConfig by this name. ```CSharp [Application] public class HelloWorldApp { [Properties("GreetingMessage")] <- in Json "GreetingMessage" : "Hello World!" public string GreetingString { get; set; } <-name of the variable can be any :) [Properties("GoodbyeMessage")] <- in Json "GreetingMessage" : "Bye-bye World!" public string GreetingString { get; set; } [Start] public void OnStart() { Console.WriteLine(GreetingString); } [Stop] public void OnStop() { Console.WriteLine(GoodbyeMessage); } } ```<issue_closed> Status: Issue closed
nuoye-blog/nuoye-blog.github.io
636695176
Title: Android-day1 | nuoye Question: username_0: https://username_0.github.io/2020/05/25/3b4bd189/ 前言因linux下的前景较窄,所以想了解下android下的一些内容,准备了好久,今天就正式吧。目前第一本书:《第一行代码:Android》。 正文Android系统架构Linux内核层 系统运行库层 应用框架层 应用层 Android应用开发特色四大组件活动(Activity)、服务(Service)、广播接收器(Broadcast Receiver)、内容提供器(Content Provide
flutter/flutter
696696682
Title: [cloud_firestore] Update to 0.14.0+2 breaks app Question: username_0: ## Problem I just updated my firebase libs to the newest version and my `cloud_firestore` functions are broken. I updated them according to the changelog and the breaking changes. Here is an example:<br> ## Code ```dart UserService().removeList(UserService().currentUserNoListen(context), listId); return sharedCollection().doc(listId).delete(); ``` ## Error Output ``` Error: Can't use 'doc' because it is declared more than once. [ ] return sharedCollection().doc(listId).delete(); ``` ## Flutter doctor -v ``` [✓] Flutter (Channel beta, 1.21.0-9.2.pre, on Mac OS X 10.15.6 19G2021, locale en-US) • Flutter version 1.21.0-9.2.pre at /Users/username_0/bin/flutter • Framework revision 81a45ec2e5 (13 days ago), 2020-08-27 14:14:33 -0700 • Engine revision 20a9531835 • Dart version 2.10.0 (build 2.10.0-7.3.beta) [✓] Android toolchain - develop for Android devices (Android SDK version 30.0.0) • Android SDK at /Users/username_0/Library/Android/sdk • Platform android-30, build-tools 30.0.0 • Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java • Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593) • All Android licenses accepted. [✓] Xcode - develop for iOS and macOS (Xcode 11.7) • Xcode at /Applications/Xcode.app/Contents/Developer • Xcode 11.7, Build version 11E801a • CocoaPods version 1.9.3 [✓] Chrome - develop for the web • Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome [✓] Android Studio (version 4.0) • Android Studio at /Applications/Android Studio.app/Contents • Flutter plugin version 49.0.2 • Dart plugin version 193.7547 • Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593) [✓] VS Code (version 1.48.2) • VS Code at /Applications/Visual Studio Code.app/Contents • Flutter extension version 3.14.1 [✓] Connected device (3 available) • iPhone SE (2nd generation) (mobile) • C1A35F4C-3C77-4201-A280-FD773A1E3E70 • ios • com.apple.CoreSimulator.SimRuntime.iOS-13-7 (simulator) • Web Server (web) • web-server • web-javascript • Flutter Tools • Chrome (web) • chrome • web-javascript • Google Chrome 85.0.4183.102 • No issues found![✓] Flutter (Channel beta, 1.21.0-9.2.pre, on Mac OS X 10.15.6 19G2021, locale en-US) • Flutter version 1.21.0-9.2.pre at /Users/username_0/bin/flutter • Framework revision 81a45ec2e5 (13 days ago), 2020-08-27 14:14:33 -0700 • Engine revision 20a9531835 • Dart version 2.10.0 (build 2.10.0-7.3.beta) [✓] Android toolchain - develop for Android devices (Android SDK version 30.0.0) [Truncated] [✓] Android Studio (version 4.0) • Android Studio at /Applications/Android Studio.app/Contents • Flutter plugin version 49.0.2 • Dart plugin version 193.7547 • Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6222593) [✓] VS Code (version 1.48.2) • VS Code at /Applications/Visual Studio Code.app/Contents • Flutter extension version 3.14.1 [✓] Connected device (3 available) • iPhone SE (2nd generation) (mobile) • C1A35F4C-3C77-4201-A280-FD773A1E3E70 • ios • com.apple.CoreSimulator.SimRuntime.iOS-13-7 (simulator) • Web Server (web) • web-server • web-javascript • Flutter Tools • Chrome (web) • chrome • web-javascript • Google Chrome 85.0.4183.102 • No issues found! ``` Answers: username_1: Hi @username_0 From what I can see, the issue is related to a FlutterFire plugin rather than to Flutter itself. Please open the issue in the dedicated [repository](https://github.com/FirebaseExtended/flutterfire/issues). Closing, as this isn't an issue with Flutter itself. If you disagree, please write in the comments, providing your `flutter doctor -v`, your `flutter run -v`, your `pubspec.yaml` a minimal reproducible code sample that does not use 3rd party plugins, and I will reopen it. Thank you Status: Issue closed
hackiftekhar/IQKeyboardManager
375513318
Title: Prevent from hiding keyboard when button inside textField pressed Question: username_0: I've subclassed UITextField to add a secure text entry toggle button in password textField, but every time i click, my keyboard hides, what is not expected behaviour in my case. Is there any way to keep First Responder for this textField on click? Tried to override shouldResignOnTouchOutsideMode and resignFirstResponder() in button action func. Thanks forehead! Answers: username_1: I think it might be your code logic which is hiding keyboard. Please check again. If not then I can say you can change it on the fly without resigning first responder. Let me know if you still face problems with this issue. Status: Issue closed
ngwalton/wbba_tools
1064728693
Title: Get "old-style crs object detected" error from USAboundaries, related to crs update Question: username_0: I think I fixed this, it needed some work in USAboundaries, which has been done by that author. Updated to new version of USAboundaries, and also replaced the depredated us_boundaries with us_counties. https://issueexplorer.com/issue/r-spatial/sf/1758 Status: Issue closed Answers: username_0: I think I fixed this, it needed some work in USAboundaries, which has been done by that author. Updated to new version of USAboundaries, and also replaced the depredated us_boundaries with us_counties. https://issueexplorer.com/issue/r-spatial/sf/1758 Status: Issue closed
awslabs/fhir-works-on-aws-deployment
1029388711
Title: Machine to Machine Authentification Question: username_0: **What's on your mind?** How to enable machine to machine authentication. Basically Another api will call to fhir server. My another api only accept client id and client secrate Which we can use to get token from Cognito. How we can enable by default to generate client secret during deployment of fhir on aws. also I don't want to enable api key. **Versions (please complete the following information):** - Release Version installed [e.g. v1.0.3] Answers: username_1: Hi @username_0, Can you add more details on what you are trying to achieve here? If your intention is to call FHIR APIs from your own API - you should be able to use a Cognito token belonging to the configured user pool which was used to bring up FHIR server. It is completely up to you if you want to use the same user pool for your own APIs too. If you don't want to enable api keys, you would need to remove `private` parameter for all the paths: 1. [path '/'](https://github.com/awslabs/fhir-works-on-aws-deployment/blob/mainline/serverless.yaml#L95) 2. [path 'proxy'](https://github.com/awslabs/fhir-works-on-aws-deployment/blob/mainline/serverless.yaml#L101) Let me know if you have any follow up questions. Thanks username_0: @username_1 I am exposing FHIR to Commonwell. Commonwell only take fhir server url, client id, client secret and Oauth url. In my case , Commonwell only can only generate token using Client Credential Oauth flow right? In Project read me it is mentioned that "To access the API, you must use the ID token. This ID token must include scopes of either openid, profile or aws.cognito.signin.user.admin." We need to provide Scope from those three option. But in case of Client Credential oauth flow we need to generate custome scope. I am bit confused how to make it work using Client id and client screat given that commonwell only support that option. username_2: Hi, Hopefully the smart branch was useful for your use case. I'm closing this issue now, but feel free to reopen if you have further questions Status: Issue closed
MicrosoftDocs/azure-docs
460922557
Title: Please add guidance for unit testing Question: username_0: How can I inject my mocks? Answers: username_1: Thanks for comment. Can you confirm if your feedback is related to Azure/Cloud or On-premise scenario? From the comment above, this seems unrelated to a cloud solution. Can you share the link to the documentation you are referencing above? username_1: any update please? username_1: We will now close this issue. If there are further questions regarding this matter, please reply and we will gladly continue the discussion. Status: Issue closed
JeanRev/TeamcityDockerCloudPlugin
222475747
Title: Existing Docker container are killing when i modified docker cloud profile Question: username_0: Hi, We have one docker cloud profile, where it has one linux host and with multiple docker images. Lets assume, i modified anyone of docker image(not running) and saved, then running docker container is killing even build is running on that. Answers: username_1: Hi @username_0 ! If I understand the issue correctly, then it probably relates to #16. The plugin is tracking containers using both labels and environment variables. When a managed container (started by the plugin) is committed to a new image, those meta-data will be persisted as well. When a container is started from such image, the plugin will detect it, and since no internal state will be associated with it, destroy it. This clean-up mechanism rightfully aims to dispose containers that may have survived a server shutdown, but it should not target containers that are derived from a plugin managed container. This will probably not be part of the next release (to be released in the coming weeks), but will be addressed shortly afterward. Status: Issue closed
Azure/app-service-announcements-discussions
373702469
Title: [Discussion] Azure Functions Runtime 2.0.12161 release Question: username_0: Discussion issue for the Azure Functions [runtime release 2.0.12161](https://github.com/Azure/azure-functions-host/releases/tag/v2.0.12161). Answers: username_1: Hi, just wanted to check to see if there is a force cut-over date setup for this function app version release? username_2: Hi, is this version being rolled out? Status: Issue closed
DataDog/dd-agent
108570926
Title: Test status is "okay" with command errors Question: username_0: From ```rake ci:run```: ``` .... tests.core.test_check_status.test_check_status_fail ... ok tests.core.test_check_status.test_check_status_pass ... ok tests.core.test_check_status.test_persistence ... ok tests.core.test_check_status.test_persistence_fail ... ok test_dupe_tags (tests.core.test_common.TestAggregator) ... ok testEdgeCases (tests.core.test_common.TestCore) ... ok Tests that the app tags are sent if specified so ... Usage: nosetests [options] nosetests: error: no such option: -s ok test_collector (tests.core.test_common.TestCore) ... Usage: nosetests [options] nosetests: error: no such option: -s ok test_counter (tests.core.test_common.TestCore) ... ok test_gauge (tests.core.test_common.TestCore) ... ok test_get_proxy (tests.core.test_common.TestCore) ... ok test_min_collection_interval (tests.core.test_common.TestCore) ... ok test_name (tests.core.test_common.TestCore) ... ok Starting with Agent 5.0.0, there should always be a local forwarder ... ok ... ``` I doubt "no such option: -s" qualifies as the tests running "ok" This is using nose==1.3.4 Status: Issue closed Answers: username_1: @username_0 Thanks for the feedback! the `-s` parameter is used to tell nose to get into pdb when a breakpoint is hit. This is not required for the test suite to be executed. So it looks like the tests are passing properly. username_2: This isn't supposed to show up in the build logs, so I would be for investigating it a bit, maybe we are skipping/missing tests because of that? example: https://travis-ci.org/DataDog/dd-agent/jobs/82912178#L263 The `-s` flag is passed here https://github.com/DataDog/dd-agent/blob/master/ci/common.rb#L202 username_2: From ```rake ci:run```: ``` ... tests.core.test_check_status.test_check_status_fail ... ok tests.core.test_check_status.test_check_status_pass ... ok tests.core.test_check_status.test_persistence ... ok tests.core.test_check_status.test_persistence_fail ... ok test_dupe_tags (tests.core.test_common.TestAggregator) ... ok testEdgeCases (tests.core.test_common.TestCore) ... ok Tests that the app tags are sent if specified so ... Usage: nosetests [options] nosetests: error: no such option: -s ok test_collector (tests.core.test_common.TestCore) ... Usage: nosetests [options] nosetests: error: no such option: -s ok test_counter (tests.core.test_common.TestCore) ... ok test_gauge (tests.core.test_common.TestCore) ... ok test_get_proxy (tests.core.test_common.TestCore) ... ok test_min_collection_interval (tests.core.test_common.TestCore) ... ok test_name (tests.core.test_common.TestCore) ... ok Starting with Agent 5.0.0, there should always be a local forwarder ... ok ... ``` I doubt "no such option: -s" qualifies as the tests running "ok" This is using nose==1.3.4 Status: Issue closed
gRally/dev
326345998
Title: menu bug. some settings at config menu are missing. especially pacenotes settings has issue Question: username_0: # User: oyunbozan --- #### What happened? #### Steps to reproduce: #### Considerations... ### [log](https://www.grally.net/u/uploads/e5fe3e79fcdc4bc96b792db681ac2134d280e8f2.zip) `v1.0.3.1` Status: Issue closed Answers: username_1: duplicate of #537