repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
dsuch/pymqi
703296257
Title: Unexplainable error when using the Filter class on Debian 10 with python 2.7.16 Question: username_0: I got some strange "attribute not found" errors on self.pub_filter.operator when using the Filter class with an integer filter on the number of messages in a queue. After changing the FilterOperator and Filter classes to basically move the operator name into the FilterOperator class and get rid of the Filter object in the FilterOperator, everything worked fine. The attached file contains the diff against 1.12.0 code as available on pypi.org [pymqi-1.12.0a.patch.tar.gz](https://github.com/dsuch/pymqi/files/5236674/pymqi-1.12.0a.patch.tar.gz) Answers: username_1: Hi @username_0, Can you show code that raise the error and stack trace? Does the IDE (which one) get an error in debug mode or only when running the .py file? username_0: Hi @username_1, Well... I just installed the original code of PyMQI and checked with Eclipse IDE 2020-06 with PyDev 7.7.0 the following code: ``` import sys import argparse from pymqi import QueueManager,CMQC,CMQCFC,MQMIError,PCFExecute,Filter def getListOfNonEmptyLocalQueues(qmgrName): try: pcf = None queueName = '*'.encode() queueType = CMQC.MQQT_LOCAL queueDepthFilter = Filter(CMQC.MQIA_CURRENT_Q_DEPTH).greater(0) pcfArgs = {CMQC.MQCA_Q_NAME: queueName, CMQC.MQIA_Q_TYPE: queueType, CMQCFC.MQIACF_Q_ATTRS : CMQC.MQCA_Q_NAME} qmgrHndl = QueueManager(qmgrName) # running in bind mode! pcf = PCFExecute(qmgrHndl,dynamic_queue_name=b'PYMQI.*') response = pcf.MQCMD_INQUIRE_Q(pcfArgs,[ queueDepthFilter ]) except MQMIError as e: if pcf is not None: pcf.disconnect() if e.comp == CMQC.MQCC_FAILED and e.reason == CMQC.MQRC_UNKNOWN_OBJECT_NAME: return [] else: raise else: pcf.disconnect() return (queueInfo[CMQC.MQCA_Q_NAME].strip() for queueInfo in response) if __name__ == '__main__': parser = argparse.ArgumentParser() parser.add_argument('qmgr', metavar='<qmgr>', type=str, help='name of queue manager to run against') arguments = vars(parser.parse_args()) for qName in getListOfNonEmptyLocalQueues(arguments['qmgr']): sys.stdout.write(qName+'\n') ``` Running it inside Eclipse and on command line everything is fine. But debugging it step-by-step in Eclipse, the following exception is thrown: ``` Traceback (most recent call last): File "/opt/eclipse-2020-03/plugins/org.python.pydev.core_7.7.0.202008021154/pysrc/pydevd.py", line 3206, in <module> main() File "/opt/eclipse-2020-03/plugins/org.python.pydev.core_7.7.0.202008021154/pysrc/pydevd.py", line 3199, in main globals = debugger.run(setup['file'], None, None, is_module) File "/opt/eclipse-2020-03/plugins/org.python.pydev.core_7.7.0.202008021154/pysrc/pydevd.py", line 2273, in run return self._exec(is_module, entry_point_fn, module_name, file, globals, locals) File "/opt/eclipse-2020-03/plugins/org.python.pydev.core_7.7.0.202008021154/pysrc/pydevd.py", line 2280, in _exec pydev_imports.execfile(file, globals, locals) # execute the script File "/home/user/Development/DZ-Bank/hg/mqjournal_api_exit/mqjournal_python/src/testGetListOfNonEmptyQueues.py", line 43, in <module> for qName in getListOfNonEmptyLocalQueues(arguments['qmgr']): File "/home/user/Development/DZ-Bank/hg/mqjournal_api_exit/mqjournal_python/src/testGetListOfNonEmptyQueues.py", line 15, in getListOfNonEmptyLocalQueues queueDepthFilter = Filter(CMQC.MQIA_CURRENT_Q_DEPTH).greater(0) File "build/bdist.linux-x86_64/egg/pymqi/__init__.py", line 2622, in __call__ pymqi.Error: Operator [__class__] is not supported. ``` My changed code allows stepping through it as well - without exception. So I guess the problem isn't with PyMQI but with the Eclipse debugger... Regards, Frank username_0: Just updated Eclipse to 2020-09 and PyDev to 8.0.0, but the error when debugging with the original code still happens. username_1: Yes :) I know about this issue, I got it in VS Code too. Thank you for patch! I merged it to master. I you have no other questions, please close the issue. Status: Issue closed
mpscholten/typesafe-enum
503886903
Title: Let PHPUnit setting file be correct XML data format Question: username_0: As title, lettin the `phpunit.xml.dist` file be correct XML data format and it should do following works: - Adding the XML declaration for this `phpunit.xml.dist` setting file. - Enhance PHPUnit setting attributes. Status: Issue closed Answers: username_1: Fixed iva #10
rakudo/rakudo
655676306
Title: Support for auth in --to flag Question: username_0: <!-- The template below contains optional suggestions. You can omit some information. --> ## The Problem With the `--doc=xxx` flag, a Pod::To::XXX.render is used to render the POD embedded in a page. That leaves space for a single Pod::To::XXX which is system-wide ## Expected Behavior If possible, there should be a way to select one of several Pod::To::HTML, for instance. This has been highlighted by the introduction of an alternative Pod::To::HTML in Raku/ecosystem#516. Possibly auth could allow different authored renderers to coexist, but other selectors (by version, let's say) might be possible. ## Actual Behavior It selects the default provider of Pod::To::XXX Answers: username_1: Honestly, I think the syntax to this flag is wrong! It assumes the 'Pod::To::' prefix, which makes it very ungeneric. After all, why should a render not be named 'HTML-Renderer", rather than "Pod::To::HTML-Rendered"? username_2: This is rather interesting. What if you try to pass auth as well? I think the way here is to make `--doc` flag understand more complete module signatures, including `ver`, `auth` etc. username_1: Maybe I wrote too quickly. Currently, `--doc=xxxx` looks for the `Pod::To:xxxx` module. The compiler is specifically reducing the namespace. Why not some other namespace, such as `--doc=Render::In::HTML` ? I am objecting to the fairly arbitrary `Pod::To::` prefix. It is not a question of 'pleasing everyone', but of being generic. In addition, it would easily be possible to DWIM, first by allowing any Module name as xxx after `--doc=xxx` and if a simple `xxxx` is not found, then look for `Pod::To::xxxx`. It would also bring this part of the compiler up to date (before the Precomp Modules) if any of the qualifiers ver and auth were to be allowed, more specific candidates winning a tie. For clarity, by tie, I mean that if HTML:auth<xxx> is specified and HTML and HTML:auth<xxxx> are installed, then the :auth<xxxx> is used. username_2: The code in question is at https://github.com/rakudo/rakudo/blob/master/src/Perl6/Actions.nqp#L1433 To close this ticket, one should write a patch where the parameter passes is parsed and additional %opts are passed to this call instead of `{}` as it is now. username_1: I've submitted a PR to close this ticket as suggest by @username_2 username_0: Additionally, it's not clear, if two Pod::To::XXX are installed, which one will be chosen for rendering.
App-vNext/Polly
628833744
Title: Does polly handle all HTTPClient multithreading and concurrency issues? Question: username_0: **Summary: What are you wanting to achieve?** This blog describes how to use Httpclient in a multithreaded app http://www.michaeltaylorp3.net/httpclient-is-it-really-thread-safe/ <br/> <br/> **What code or approach do you have so far?** Exploring what implementations are necessary to use between a C# core app and many different HTTP urls, in concurrency. <br/> <br/> Answers: username_1: @SwimlaneBuilder polly doesnt take any opinion on how to handle Httpclient. so it is up to you to use Httpclient correctly when plugging into Polly. Status: Issue closed
NLog/NLog.Extensions.Logging
550753965
Title: Unable to manage loglevel from appSettings.json Question: username_0: Hi, I'm trying to combine the log configuration from code/appSettings.json/NLog.config in a .net core console application (.NET Core 3.1). I started from the online example adding the load of the appSettings.json. So I have In the code ``` // Inspecting config at runtime I see the value: [0] = {[Logging:LogLevel:Default, Warning]} //... .AddLogging(loggingBuilder => { loggingBuilder.ClearProviders(); loggingBuilder.SetMinimumLevel(Microsoft.Extensions.Logging.LogLevel.Trace); loggingBuilder.AddNLog(config); }) ``` In appSettings.json ``` { "Logging": { "LogLevel": { "Default": "Warning" } } } ``` In Nlog.config ``` <logger name="*" minlevel="Trace" writeTo="f" /> ``` I expected to have an output filtered on Warning level, but I have all the levels contained in the code; 2020-01-16 11:54:26.9520 TRACE Doing hard work! Action1 2020-01-16 11:54:26.9881 DEBUG Doing hard work! Action1 2020-01-16 11:54:26.9881 INFO Doing hard work! Action1 2020-01-16 11:54:26.9881 WARN Doing hard work! Action1 2020-01-16 11:54:26.9881 ERROR Doing hard work! Action1 2020-01-16 11:54:26.9881 FATAL Doing hard work! Action1 It works when I change the min level in code or nlog.config from trace to warning: 2020-01-16 12:27:47.1587 WARN Doing hard work! Action1 2020-01-16 12:27:47.1927 ERROR Doing hard work! Action1 2020-01-16 12:27:47.1927 FATAL Doing hard work! Action1 What am I doing wrong? This is the simple project: [ConsoleAppLog.zip](https://github.com/NLog/NLog.Extensions.Logging/files/4070962/ConsoleAppLog.zip) Answers: username_1: People that decide to use NLog usually also want to disable all MEL-filtering to avoid the confusion with two filtering systems. So the wiki-tutorial is targeted those users. I guess people who are MEL-users first will probably just use `new HostBuilder().CreateDefaultBuilder().Build()` (Will setup everything with all guns enabled). But if staying with the simple example, then you need to remove: `loggingBuilder.SetMinimumLevel(Microsoft.Extensions.Logging.LogLevel.Trace);` And add: `loggingBuilder.AddConfiguration(config.GetSection("Logging"));` So it looks like this: ```c# serviceCollection.AddLogging(loggingBuilder => { loggingBuilder.ClearProviders(); loggingBuilder.AddConfiguration(config.GetSection("Logging")); loggingBuilder.AddNLog(config); }) ``` `ILoggingBuilder.AddConfiguration` can be found at Nuget: [Microsoft.Extensions.Logging.Configuration](https://www.nuget.org/packages/Microsoft.Extensions.Logging.Configuration/) username_1: Btw. don't use Nuget-package [NLog.config](https://www.nuget.org/packages/NLog.Config) for Net Core projects. I have updated the NLog tutorial about this: https://github.com/nlog/nlog/wiki/NLog-Install username_0: Perfect, it works with the AddConfiguration() I agree, it is probably not convenient to have multiple configurations (MEL + NLog). I did it just to understand how the chain works. Status: Issue closed
goldsmithwill/BoggleBuddy
210254722
Title: Add Timer Question: username_0: There should be a timer that gives the user a set amount of time to enter as many words as possible. When the timer hits 0, it should allow the user to go to the next level or tell them it is game over.<issue_closed> Status: Issue closed
mne-tools/mne-python
749455377
Title: set_3d_view doesn't set view in docs Question: username_0: #### Describe the bug I think that `set_3d_view` does not work any more, all the examples in https://mne.tools/stable/generated/mne.viz.set_3d_view.html#mne.viz.set_3d_view look incorrect to me. I don't know what some looked like before but the fNIRS one is definitely now not correct. #### Steps to reproduce https://mne.tools/stable/auto_examples/visualization/plot_3d_to_2d.html#sphx-glr-auto-examples-visualization-plot-3d-to-2d-py https://mne.tools/stable/auto_tutorials/preprocessing/plot_70_fnirs_processing.html#id1 #### Expected results I expect the function to set the 3d view, but it docent seem to #### Actual results See doc links #### Additional information Answers: username_1: @username_2 can you have a look? username_2: Thanks for reporting this, I'm on it! username_0: Thanks Status: Issue closed
react-native-google-signin/google-signin
962235417
Title: Error 403: restricted_client Question: username_0: Hi, I am getting 403 restricted client when I try to launch the google sign in web page, it says: "This app is not yer configured to make OAuth request. To do that, set up app's OAuth consent screen in the Google Cloud Console". I am using Firebase and followed all the instructions in the documentation to use this library but I haven't been able to do it. Please any help would be appreciated. Answers: username_1: Same username_0: Well I am not sure if that solves it, I went and configure a OAUTH google consent configuration but maybe that setup in firebase just did the work. Anyways thanks username_2: hello and thanks for reporting, this is not an issue with this library but with the setup; please read more here: https://github.com/firebase/quickstart-js/issues/324 Hope this helps :) Status: Issue closed
BCDevOps/devops-requests
879874029
Title: Request for BCGOV GitHub Organization Membership - chadhosk Question: username_0: ## Step 0 **Are you looking for access to the Openshift Console?** No **Are you the Technical Lead?** Yes **Do you actually need the team member to be a member of the organization?** Yes, so they can access ZenHub ## Step 1 Visit DevHub for detailed instruction on creating a request: Done ## Step 2 Make sure no duplicated request exists, search here: Done ## Step 4 Are you requesting to: - [x] Invite a user ## Step 5 Fill out the following fields * Project Name: FOI Modernization * GitHub Org: bcgov * Full Name: <NAME> * GitHub ID: @chadhosk * GitHub Account 2FA Confirmed: yes * Email address: <EMAIL> * Organization: Information Access Operations * Project Role: Sponsor * GitHub Team to Join (if exists): https://github.com/orgs/bcgov/teams/foi-modernization * Existing GitHub Repo: https://github.com/bcgov/foi-flow, https://github.com/bcgov/foi-requests **Note** that the platform services team will be removing GitHub Org access for users that are not active for six months. Once the access has been removed, a new `Access Request` has to be made by the product owner.<issue_closed> Status: Issue closed
WooYellowCube/wooyellowcube
174299009
Title: <?= could cause problems with PHP < 5.4 Question: username_0: According to http://php.net/manual/en/language.basic-syntax.phptags.php the <?= tag is not activated by default with PHP < 5.4 The echo shortcut tag is used in the views. WordPress provides the _e() method which echoes the string immediately. I would recommend to use <?php _e(); ?> instead of <?= __() ?><issue_closed> Status: Issue closed
cloudcreativity/laravel-json-api
431612200
Title: Class is not a resource adapter. Question: username_0: I'm using the package with Laravel 5.8 I'm probably missing something simple, but i can't for the life of me figure out what is happening here. I have followed the docs, created an api `php artisan make:json-api v1` with the following config: ``` return [ 'resolver' => \CloudCreativity\LaravelJsonApi\Resolver\ResolverFactory::class, 'namespace' => 'Api', 'by-resource' => true, 'resources' => [ 'countries' => \App\Models\Country::class, ], 'use-eloquent' => true, 'url' => [ 'host' => null, 'namespace' => '/api/v1', 'name' => 'api:v1:', ], 'jobs' => [ 'resource' => 'queue-jobs', 'model' => \CloudCreativity\LaravelJsonApi\Queue\ClientJob::class, ], 'supported-ext' => null, 'encoding' => [ 'application/vnd.api+json', ], 'decoding' => [ 'application/vnd.api+json', ], 'providers' => [], ]; ``` Created a resource called "countries" `php artisan make:json-api:resource countries v1` Then added route: ``` Route::group(['middleware' => 'auth:api'], function() { JsonApi::register('v1') ->withNamespace('Api') ->routes(function ($api) { $api->resource('countries')->readOnly(); }); }); ``` In `RouteServiceProvider.php` changed `mapApiRoutes`: ``` protected function mapApiRoutes() { Route::middleware('api') ->namespace($this->namespace) ->group(base_path('routes/api.php')); } ``` And `AppServiceProvider.php`: ``` namespace App\Providers; use CloudCreativity\LaravelJsonApi\LaravelJsonApi; use Illuminate\Support\ServiceProvider; class AppServiceProvider extends ServiceProvider [Truncated] * * @return void */ public function register() { } /** * Bootstrap any application services. * * @return void */ public function boot() { LaravelJsonApi::defaultApi('v1'); } } ``` but when i request `http://127.0.0.1:8000/api/v1/countries` i get the following error: `Class [Api\Countries\Adapter] is not a resource adapter.` Answers: username_1: Presumably the `namespace` config setting should be `App\Api`? username_0: You my friend is an absolute life saver, did not catch that! Changed the namespace and changed namespace in resource files. Now it works! Status: Issue closed username_1: Great, glad to hear you got it sorted!
skuhl/RobotRun
237998662
Title: register expression Question: username_0: If you create an expression PR[2,2] = Pr[2,**2**] + 50 the point index on the right expression cannot be edited directly. If you press edit you have to enter the register index first and then point index even if the cursor is on the point index. but for expression on the left it works fine. so try to implement the same function on the right. Status: Issue closed Answers: username_1: Vincent has added this functionality in the most recent release.
quintel/etmodel
780615578
Title: Green gas should be clearly shown in the energy flow sankeys Question: username_0: A client pointed this out to me and I agree! At the moment green gas falls into the category 'biomass & waste' in the sankey and this category does not clearly state that it includes green gas: ![Screenshot 2021-01-06 at 15 48 07](https://user-images.githubusercontent.com/32833996/103781557-d102a680-5036-11eb-9591-5d73624f1161.png) I think there are several ways to fix this: 1. rename the category 'green gas, biomass & waste' 2. include green gas in the 'natural gas' category and rename this category to 'network gas' or 'methane' 3. add the 3 green gas production methods as one block to the sankey. then the route would be 'biomass & waste' -> green gas production -> 'green gas' I am in favour of option 2, as the Sankey is already quite full for option 3 and option 1 is a bit messy IMO. Additionally, I would like to propose to replace the term 'network gas' everywhere in the ETM with 'methane' as the term network gas is not often used (maybe only in the ETM?!). Let me know what you think! Answers: username_1: With the recent update of the CO2 sankey we decided to include green gas (and bio oil) in the biomass category rather than the network gas (and oil) category. I think the main reason was that it is not clear to many people what 'network gas' is. I would prefer to go with option 1 as that takes 5 seconds to implement. In any case, whatever we do we should update the CO2 sankey as well. username_2: My preference would be to keep green gas separate from network gas (or natural gas), so that users can quickly recognise which energy sources are renewable. Therefore I would not go for option 2. I think that option 3 is the most accurate, as the chart would then show the conversion losses that occur when biogas is made from wet biomass. The argument against option 3 is that it might make the chart 'messy' and that it is more difficult to implement. If this argument weighs strongly enough, then my preference would be to go for option 1. Status: Issue closed
aaronwolen/bingd
67198662
Title: Calling `AnnotationHub()` throws error Question: username_0: ``` Error in unlist(.metadata(snapshotUrl, columns = "RDataPath")) : error in evaluating the argument 'x' in selecting a method for function 'unlist': Error in (function (l, isMulti) : could not find function "splitAsList" ``` `splitAsList()` is an IRanges function.<issue_closed> Status: Issue closed
metinkale38/prayer-times-android
309710859
Title: Foreground Service Error in android 8.0 Question: username_0: 03-29 15:57:51.239 22145-22145/? E/AndroidRuntime: FATAL EXCEPTION: main Process: com.qibla.direction.finder, PID: 22145 android.app.RemoteServiceException: Context.startForegroundService() did not then call Service.startForeground() at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1870) at android.os.Handler.dispatchMessage(Handler.java:105) at android.os.Looper.loop(Looper.java:164) at android.app.ActivityThread.main(ActivityThread.java:6809) at java.lang.reflect.Method.invoke(Native Method) at com.android.internal.os.Zygote$MethodAndArgsCaller.run(Zygote.java:240) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:767) Answers: username_1: seems like you changed the targetApi to Oreo. please revert it back to 25, because it needs some changes... username_0: are you working on it ?? I have to test it on Oreo please update it . username_1: i didnt test with Oreo yet, but normally it should work with target api 25. you should not need startForegroundService with lower targetApi. Status: Issue closed
serilog/serilog-sinks-email
891861036
Title: Email Sink not sending Email (gmail smtp) Question: username_0: Hello, I am using the below code, but it is not working, not even giving me any error. I am also using MSSql Sink for DB logging (for that, the config is coming from `appSettings.json`), and it is working fine. I have tried to use the email sink by loading config from `appSettings.json`, but it doesn't work either. Any help is appreciated. ```cs Serilog.Debugging.SelfLog.Enable(msg => Console.WriteLine(msg)); Log.Logger = new LoggerConfiguration() .MinimumLevel.Debug() .MinimumLevel.Override("Microsoft", LogEventLevel.Warning) .Enrich.FromLogContext() .ReadFrom.Configuration(configBuilder) .WriteTo.Email(new EmailConnectionInfo { FromEmail = "*****<EMAIL>", ToEmail = "******<EMAIL>", MailServer = "smtp.gmail.com", NetworkCredentials = new NetworkCredential("*****<EMAIL>", "******"), EnableSsl = true, Port = 465, EmailSubject = "Serilog Log Event" }, batchPostingLimit: 1) .CreateLogger(); try { Log.Information("Application Starting Up"); CreateHostBuilder(args).Build().Run(); } catch (Exception ex) { Log.Fatal(ex, "Application Start-Up Failed"); } finally { Log.CloseAndFlush(); } ``` Status: Issue closed Answers: username_0: Hello, I am using the below code, but it is not working, not even giving me any error. I am also using MSSql Sink for DB logging (for that, the config is coming from `appSettings.json`), and it is working fine. I have tried to use the email sink by loading config from `appSettings.json`, but it doesn't work either. Any help is appreciated. ```cs Serilog.Debugging.SelfLog.Enable(msg => Console.WriteLine(msg)); Log.Logger = new LoggerConfiguration() .MinimumLevel.Debug() .MinimumLevel.Override("Microsoft", LogEventLevel.Warning) .Enrich.FromLogContext() .ReadFrom.Configuration(configBuilder) .WriteTo.Email(new EmailConnectionInfo { FromEmail = "*****<EMAIL>", ToEmail = "******<EMAIL>", MailServer = "smtp.gmail.com", NetworkCredentials = new NetworkCredential("*****<EMAIL>", "******"), EnableSsl = true, Port = 465, EmailSubject = "Serilog Log Event" }, batchPostingLimit: 1) .CreateLogger(); try { Log.Information("Application Starting Up"); CreateHostBuilder(args).Build().Run(); } catch (Exception ex) { Log.Fatal(ex, "Application Start-Up Failed"); } finally { Log.CloseAndFlush(); } ``` username_1: did you find out how to do it? can you help me too? username_2: In my case, it was just a little wait. The problem arises when we set a breakpoint on the next method, or sending an email is the last step of the application. For this purpose, I created a very advanced waiting method, which I recommend to copy after calling the logger :) for (int i = 0; i < 1000000; i++) { Console.WriteLine(i); }
andreasremdt/simple-translator
837772129
Title: Question: Is it SEO friendly? Question: username_0: How does google robot see the site? In english or german or spanish? Answers: username_1: That depends more or less on your configuration (e.g. when the page is translated). First, it's a best practice to use a fallback language, even before the translator kicks in. Looking at this code example ```html <p data-i18n="text">some fallback text</p> ``` you'd have _some fallback text_ as the default text, which is displayed when the page is initially rendered. This is also what Google's Bot would see. If you'd leave it empty, like ```html <p data-i18n="text"></p> ``` then I would expect some SEO issues, yes. What happens next depends on whether or not you translate the page automatically on page load. If so, I am not entirely sure what Google's Bot would see, but I assume it's the text that got rendered out initially, not the applied translation. Generally speaking, there are lots of SEO pitfalls when implementing translations. As an example, it's recommended to use different URLs for different languages, like `website.com/en/` or `en.website.com`. Maybe you'll find [this thread](https://stackoverflow.com/questions/1828317/internationalization-and-search-engine-optimization) helpful, it gave me some insights on how these things should be handled. Otherwise, if you have some more questions, feel free to drop them here! Status: Issue closed
tensorflow/tpu
363964916
Title: faster rcnn on cloud TPU Question: username_0: I'm trying to run faster rcnn resnet inception v2 on a cloud TPU instance, I know that there is no config in the object detection API that is compatible with TPUs, but I am getting always the following error: Compilation failure: Detected unsupported operations when trying to compile graph _functionalize_body_20[] on XLA_TPU_JIT:Where, RandomShuffle, Where, RandomShuffle, Where, RandomShuffle, Where, RandomShuffle, Where, RandomShuffle, Where, RandomShuffle, Where, RandomShuffle, Where, RandomShuffle, Where, RandomShuffle, Where, RandomShuffle, Where, Where, RandomShuffle, Where, RandomShuffle, Where, Where, RandomShuffle, Where, RandomShuffle, Where, Where, RandomShuffle, Where, RandomShuffle, Where, Where, CropAndResize, CropAndResizeGradImage, CropAndResizeGradBoxes so I was wondering if there is a way to change these operations to be compatible with TPU Answers: username_1: Could you please share the portion of the code where these operators are being used? There must be corresponding TPU compatible operations, but I need to look into the code around these operations first. username_2: @username_0 https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tpu_compatibility.md username_3: Any help her with TPU compatible config for faster RCNN?
pingcap/tidb-operator
494443523
Title: Avoid rollover in stability test to workaround #901 Question: username_0: In stability test, we may encounter #901 in the configmap rollout case. While #901 cannot be addressed in a short time, we can avoid rollover to workaround this issue, which will help stability test run more smoothly.<issue_closed> Status: Issue closed
svenhjol/Charm
738886543
Title: Redstone dust hit by glow ball disappear. Question: username_0: Well, I don't think items should disappear. This doesn't make any sense, besides a snowball doesn't have such a behavior. Glow ball should just be destroyed in contact with items such as redstone dust or flowers, just like a snowball. What about the name? Currently it is "glow ball" with a space between words. It's ok but maybe consider "glowball". Status: Issue closed Answers: username_1: Like #354 what should be the behavior here username_0: Well, I don't think items should disappear. This doesn't make any sense, besides a snowball doesn't have such a behavior. Glow ball should just be destroyed in contact with items such as redstone dust or flowers, just like a snowball. What about the name? Currently it is "glow ball" with a space between words. It's ok but maybe consider "glowball". Status: Issue closed username_1: Now drops glowstone dust if it can't be placed. It's also been renamed to "Glowball".
akinomyoga/ble.sh
865666662
Title: ble.sh multiline command is not working Question: username_0: **ble version**: <!-- The result of `echo $BLE_VERSION` (version, commit-hash) --> `0.4.0-devel3+0506df2` **Bash version**: <!-- The result of `echo $BASH_VERSION, $MACHTYPE` --> `5.0.17(1)-release, x86_64-pc-linux-gnu` <!-- I recently discovered ble.sh and wanted to give it a try and was truly impressed about the auto complete. But, I am facing a issue when I try to do a multiline command. See the demo below: ![Peek 2021-04-23 10-53](https://user-images.githubusercontent.com/35771765/115806894-8e68ac80-a422-11eb-936c-19f0dbe3c68c.gif) --> Answers: username_1: Have you read [README §3.1](https://github.com/username_1/ble.sh#31-use-multiline-mode)? In the multiline mode, <kbd>RET</kbd>/<kbd>C-m</kbd> is intended for newlines. To run the command, you can use <kbd>C-j</kbd>. Status: Issue closed username_0: Thank you!! username_1: OK! Thanks for trying `ble.sh`!
jupyter-incubator/sparkmagic
211560210
Title: Refactor SQLQuery, SparkStorecommand and Command Question: username_0: The SparkStoreCommand class introduced in #333 looks a lot like SQLQuery class. The straight-forward action is to refactor so that the new class and SQLQuery inherit from some hybrid. However. if we include Command, they are all 'commands' in the general sense, and all (directly or indirectly) inherit from ObjectWithGuid. Due to that, perhaps we should have SQLQuery, SparkStoreCommand and future types of commands to inherit from Command.
st235/ExpandableBottomBar
652011853
Title: avoid recreation of fragment Question: username_0: hello @username_1 sins the MVVM support it is awesome, but we need one more enhancement in the library, so Fragment not create every time when any bottom menu item clicked. Thanks Answers: username_1: Hi @username_0, sorry for a bit late replay. Did you check the default BottomBarNavigation view? Does it behave in an expected way? Also, did I understand you correctly - double-clicking on the same item causes Fragment recreation? If you don't mind, it would be really cool, if you can record a small video, which reflects wrong behaviour. Looking forward to hearing from you username_0: Hello @username_1 I ma very glade to see you again, i know the default behavior navigation bottom, but this is not use only for show activity or fragment, but in this library we move forward already, it have MVVM support, but default doesn't, we need to avoid the recreation of fragment, now please check what is happen right now. double-clicking on the same item causes Fragment recreation, no it's not, it recreate again when click on other menu item, it will recreate again and again i make one little video,it will clear your query ![ezgif com-video-to-gif](https://user-images.githubusercontent.com/16591300/87222181-5f031a80-c38f-11ea-85e0-e7e7eb9206e5.gif) Please check once. username_0: Hello @username_1 Please reply, it is bit late. i am waiting, Looking forward to hearing from you Thanks username_1: Hi @username_0, sorry, it was a really crazy week. I'll be able to take a look at the problem only at the end of this week. PS: However, it's always welcomed any kind of contribution here. So, if you have any ideas on how to solve this problem, do not hesitate, please, to create a pull request. Kind regards, <NAME> username_0: Hello @username_1 not an issue, please see on the issue end of the week, but i have one solution for that. `public class MainActivity extends AppCompatActivity { final Fragment fragment1 = new HomeFragment(); final Fragment fragment2 = new DashboardFragment(); final Fragment fragment3 = new NotificationsFragment(); final FragmentManager fm = getSupportFragmentManager(); Fragment active = fragment1; @Override protected void onCreate(Bundle savedInstanceState) { super.onCreate(savedInstanceState); setContentView(R.layout.activity_main); Toolbar toolbar = (Toolbar) findViewById(R.id.toolbar); setSupportActionBar(toolbar); BottomNavigationView navigation = (BottomNavigationView) findViewById(R.id.navigation); navigation.setOnNavigationItemSelectedListener(mOnNavigationItemSelectedListener); fm.beginTransaction().add(R.id.main_container, fragment3, "3").hide(fragment3).commit(); fm.beginTransaction().add(R.id.main_container, fragment2, "2").hide(fragment2).commit(); fm.beginTransaction().add(R.id.main_container,fragment1, "1").commit(); } private BottomNavigationView.OnNavigationItemSelectedListener mOnNavigationItemSelectedListener = new BottomNavigationView.OnNavigationItemSelectedListener() { @Override public boolean onNavigationItemSelected(@NonNull MenuItem item) { switch (item.getItemId()) { case R.id.navigation_home: fm.beginTransaction().hide(active).show(fragment1).commit(); active = fragment1; return true; case R.id.navigation_dashboard: fm.beginTransaction().hide(active).show(fragment2).commit(); active = fragment2; return true; case R.id.navigation_notifications: fm.beginTransaction().hide(active).show(fragment3).commit(); active = fragment3; return true; } return false; } };` username_1: @username_0, hi 👋 Hopefully, you are well. I take a look at the problem. I tried to reproduce similar behaviour with default android components, so I've changed `ExpandableBottomBar` to ```kotlin <com.google.android.material.bottomnavigation.BottomNavigationView android:id="@+id/bottomNavigation" android:layout_width="0dp" android:layout_height="wrap_content" app:backgroundTint="#fff" app:menu="@menu/navigation_menu" app:layout_constraintBottom_toBottomOf="parent" app:layout_constraintLeft_toLeftOf="parent" app:layout_constraintRight_toRightOf="parent" /> ``` and also used NavigationUI is an alternative for native android view. ```kotlin NavigationUI.setupWithNavController(bottomNavigation, navigationController) ``` Unfortunately, the behaviour you mentioned below is not followed by material design components. As it's not a native behaviour, I cannot implement it in this library. The users will not expect such behaviour at this component. Also, I've made a small video of `BottomNavigationView` ![device-2020-07-19-223556](https://user-images.githubusercontent.com/14966197/87883605-44751500-ca11-11ea-8603-d9f8582f521c.gif) PS: also, I looked at the provided sample. The main highlight here: they don't use `NavigationUI`, but do use a custom extension (`BottomNavigationView#setupWithNavController) to provide required behaviour. Kind regards, <NAME> Status: Issue closed username_0: @username_1 hello I understand what you want to say, it is not necessary to add on this library, but in my suggestion this feature you can see in maximum apps, **_you can check from your side also, those who use bottom navigation in the app_**. **coz some time user don't want to recreate the fragment again and again**, **_user always want to use it from where he/she left_**, i hope you can understand, you can add this thing in your lib, **_if user want to use this functionality in app, they can use, if not they can not use_** Thanks and Regards <NAME>ick username_1: Good morning, @username_0 👋 Thank you for provided information, I really appreciate it. However, as I wrote below - it's not a default behavior for alternative native android widget. Moreover, it looks like as an application layer responsibility to cache fragments. I can recommend you to look at fragmentFactory - probably, it will cover you case. Best wishes, Alex username_0: @username_1 same2u very thanks from suggestion i will defiantly R&D on that (Fragment factory), but i can handle this with giving example of **_(MainActivity)_** in the previous comment, it is very good technique to handle this thing, but coder need to handle this code without progress dialog, **_coz progress dialog overlap the fragment screens_**. means if you use progress dialog in second screen, it will show in your first screen, coz above technique create the all fragments in ONCE, may be it give help to some user, they read are comments. because we can also use this technique in ExpandableBottomBar, i check it already, coder need to use bottomBar.setOnItemSelectedListener(new Function2<View, ExpandableBottomBarMenuItem, Unit>() { @Override public Unit invoke(View view, ExpandableBottomBarMenuItem expandableBottomBarMenuItem) { // code goes here return Unit.INSTANCE; } }); If you're using Java 8 syntax, then lambda expression will be enough here bottomBar.setOnItemSelectedListener((view, item) -> { Log.d(TAG, "selected: " + item.toString()); return null; }); Thanks and regards <NAME>
serde-rs/serde
255114785
Title: Annoying pattern in Elasticsearch JSON makes serde_derive less useful. Question: username_0: I wasn't sure exactly where to file this as it cuts across serde, serde_derive, and serde_json. I ran into a problem that required writing a custom `Serialize` impl for my type and it's a fairly common pattern in some JSON APIs and particularly Elasticsearch so I wanted to check I wasn't missing a nicer way of handling this. The Elasticsearch Query type I was implementing was [Term Query](https://www.elastic.co/guide/en/elasticsearch/reference/current/query-dsl-term-query.html). The problem is that rather than the document field being in a key-value pair like `field: "user"`, it's lifted into a key which them has the term value as the value. So this, ``` { "query": { "term" : { "user" : "Kimchy" } } } ``` Represents a query which returns documents which contain the term "Kimchy" in the document field "user". This is a somewhat common pattern in the ES API. My (working) attempt is below: https://github.com/username_0/duke/blob/master/src/lib.rs#L42-L78 I'd rather not have to do this over-and-over-and-over ad infinitum. Did I do the right thing? Is there an easier option? Should I make my own macro? Sorry for bothering you but seemed like it'd be best to ask the devs. Answers: username_1: I would factor out that logic into a single serialize function, maybe something like this. ```rust extern crate serde; #[macro_use] extern crate serde_derive; #[macro_use] extern crate serde_json; #[derive(Serialize)] #[serde(rename_all = "snake_case")] enum Query { Term(#[serde(with = "query")] TermQuery), Range(#[serde(with = "query")] RangeQuery), } #[derive(Serialize)] struct TermQuery { field: String, value: String, #[serde(skip_serializing_if = "Option::is_none")] boost: Option<Boost>, } #[derive(Serialize)] struct RangeQuery { field: String, gte: u64, lte: u64, } #[derive(Serialize)] struct Boost(f64); mod query { use serde::ser::{Serialize, Serializer, Error}; use serde_json::{self, Value}; pub fn serialize<T, S>(query: &T, serializer: S) -> Result<S::Ok, S::Error> where T: Serialize, S: Serializer { let value = serde_json::to_value(query).map_err(Error::custom)?; let mut map = match value { Value::Object(map) => map, _ => panic!("bug: query must be a JSON object"), }; let field = match map.remove("field") { Some(field) => field, None => panic!("bug: query must have a field called `field`"), }; let k = match field { Value::String(s) => s, _ => panic!("bug: `field` in query must be a string"), }; json!({ k: map }).serialize(serializer) } } fn main() { let query = Query::Term(TermQuery { field: "status".to_owned(), value: "urgent".to_owned(), boost: Some(Boost(2.0)), }); println!("{}", serde_json::to_string(&query).unwrap()); } ``` Status: Issue closed
PollBuddy/PollBuddy
573000876
Title: Settings page (student) Question: username_0: **Please describe what has to be done** - [ ] Design Complete/Approved - [ ] Code Created - [ ] Code approved/merged **Additional context** See sitemap and slide 22 Answers: username_0: (oops forgot to hit submit sorry) Probably a list of classes, if you're in charge of the class, you could change things like name, details, add people, remove people, etc. If you're a student, you could leave the class or see read-only info about it. username_0: Actually this feels like stuff that belongs elsewhere. We already have a list of groups ("classes"), a way to change their details, now we have an issue for leaving groups. Not very settings-like. We could consider changing the Account page to Settings with an Account section, and add stuff like light/dark theme toggle if/when we get there, but for now it's probably not necessary to have. Status: Issue closed
sheinsight/shineout
446434976
Title: Iconfont 本地文件使用,API文档未给出说明 Question: username_0: ![image](https://user-images.githubusercontent.com/26628346/58072681-0bee6700-7bd3-11e9-999c-be898165300e.png) 这样写产生 `ES lint` 报错 Answers: username_1: 因为不同的项目webpack 的配置都不太一样,所以文档中并没有写明,后续会考虑加进去的. 使用本地的 css 时, 推荐通过在 Icon 中不传url 的方式. 自己将所需要的字体样式文件引入到项目中即可. [https://sheinsight.github.io/shineout/1.3.x/cn/components/Icon#heading-3-Icon-](icon) username_0: ok. Status: Issue closed
NII-cloud-operation/Jupyter-LC_notebook_diff
328804292
Title: Cannot show a MergeView when the DiffView is open multiple times Question: username_0: When the DiffView is open multiple times, a MergeView is not responding to clicks on cells. Then the console shows the error below. ``` codemirror.js:353 Uncaught TypeError: Cannot read property 'first' of null at new Display (codemirror.js:353) at new CodeMirror$1 (codemirror.js:7783) at CodeMirror$1 (codemirror.js:7771) at new CodeMirror.MergeView (merge.js?v=20180603144208:568) at Function.CodeMirror.MergeView (merge.js?v=20180603144208:536) at DiffView.showMergeView (jupyter-notebook-diff.js?v=20180603144208:224) at HTMLDivElement.<anonymous> (jupyter-notebook-diff.js?v=20180603144208:329) at HTMLDivElement.dispatch (jquery.min.js:3) at HTMLDivElement.r.handle (jquery.min.js:3) ```<issue_closed> Status: Issue closed
ianstormtaylor/slate
708123908
Title: Runtime change of editor causes a crash Question: username_0: The issue seems to be caused by not including [editor] as a dependency in plenty of places. Answers: username_1: The editor object stores the editor state, and needs to be persisted for the editor's lifetime. `useMemo` doesn't guarantee that it's value will be persisted, and the slate docs should probably be updated to use a different hook. Here's a fork of your codesandbox that uses `useRef` and a lazy initializor to store the editor object for the entire editor lifetime: https://codesandbox.io/s/slate-reproductions-forked-1m9cv?file=/index.tsx The react docs go into more depth on persisting values using hooks: https://reactjs.org/docs/hooks-faq.html#how-to-create-expensive-objects-lazily @ianstormtaylor I can open a PR to update the docs if you want. Below are alternatives to `useMemo` that will persist the editor object without dropping it: ```tsx // Using useState const [editor] = useState(() => withReact(createEditor())); // Using useRef with an initializor function const editor = useRef<ReactEditor | null>(null); const getEditor = useCallback((): ReactEditor => { if(editor.current === null) editor.current = withReact(createEditor()); return editor.current; }, []) ``` Side node: the same bug occurs when using NextJS fast refresh. username_1: Another alternative is to write a `useEditor` hook that takes an editor creator function and does the above for you like so: ```tsx function useEditor<T extends Editor>(createEditor: () => T): T { const editor = useRef<T | null>(null); if(editor.current === null) editor.current = createEditor() return editor.current } function MySlateEditor() { const editor = useEditor(() => withReact(createEditor())) return <Slate editor={editor}> ... </Slate> } ``` username_2: There are a bunch of memoized handlers in the `Slate` and `Editable` components that close over `editor` but don't list `editor` in the memo keys, so if you pass a different editor to the same component instances many things break. You can work around it by forcing the components to remount (explicitly setting a `key` property on the `Slate` component that changes when the editor changes). But it seems like Slate should properly handle a change of editor, right?
lovell/sharp
405672204
Title: the file built with BlobBuilder is not working Question: username_0: At the front end, file built with 'new Blob()', get buffer after uploading `const sharpBuffer = await sharp(buffer).resize().....;` This code works fine but....built with BlobBuilder(I tested in android 4.3),isn't working, found error: Input buffer contains unsupported image format The front end uses the following method to convert, and upload the file using Ajax ``` base64ToBlob: function (base64, contentType, cb) { const bytes = window.atob(base64); const buffer = new ArrayBuffer(bytes.length); const bufferArray = new Uint8Array(buffer); for (let i = 0; i < bytes.length; i++) { bufferArray[i] = bytes.charCodeAt(i); } let blob; try { blob = new Blob([buffer], {type: contentType}); } catch (e) { window.BlobBuilder = window.BlobBuilder || window.WebKitBlobBuilder || window.MozBlobBuilder || window.MSBlobBuilder; if(e.name === 'TypeError' && window.BlobBuilder){ const blobBuilder = new BlobBuilder(); blobBuilder.append(buffer); blob = blobBuilder.getBlob(contentType); } } cb(blob); }, ``` Status: Issue closed Answers: username_1: Hello, this question doesn't really relate to sharp and according to https://developer.mozilla.org/en-US/docs/Web/API/BlobBuilder the use of BlobBuilder is considered obsolete. Perhaps try StackOverflow to see if there's anyone else still using it - good luck!
irods/irods
118867567
Title: Upgrade Documentation - recommend more specific notes please! Question: username_0: I'm starting to look at this, and I'm not entirely au fait with V4, so apologies for what I am sure are going to be stupid questions, but I'm trying to come from a new upgrader's perspective and writing what I see missing in the documentation.. :-) In the upgrade manual; There appears to be an assumption that the upgrade from run-in-place for 3.3.1 is going to a packaged installation. Is run-in-place to run-in-place supported, and if so, what are the changes to the procedure? The document also appears to assume a move to JSON syntax. My memory (which may be faulty), was that JSOn wasn't mandatory until 4.2? Assuming that I did want to do a upgrade from run-in-place to packaged install, however; *step 2* says; ``` Make a backup of the iCAT database & configuration files: core.re, core.fnm, core.dvm, etc. ``` However, _etc_ covers a lot of sins! Can a full list of the files to be backed up be provided please? *Step 10* says ``` If necessary, migrate 3.3.x in-place iCAT database to the system database installation. ``` This infers that iRODS 4.1.7 assumes the database is on the local host. This may not be what you mean. I'm assuming it means something like; "make a copy of your 3.3.1 iCAT database for your 4.1.7", and assumes that you will not be upgrading the database in-situ, which you may not be (and if this is contra-indicated, please provide reasons why (or link to github issues if a succinct summary isn't appropriate) *Step 12* says ``` Manually update any changes to 'core.re' and 'server_config.json'. Keep in mind immediate replication rules (acPostProcForPut, etc.) may be superceded by your new resource composition ``` Update from what to what? At this point in the manual, the user may not be familiar enough with V4 to know what will break! Replication note is a good example, but a list of rules to be aware of would be very helpful. How does one convert the existing config to JSON, is there a guide or section in the manual? Searching for JSON didn't help - is it as simple as updating the server_config.json that is put into /etc/irods/server_config.json ? Which bits should be changed, and which can be kept the same? Any gotchas that could be referred to (github issues fine!) *step 13* says `` Run ./packaging/setup_irods.sh (recommended) OR Manually update all 4.0+ configuration files given previous 3.3.x configuration (.irodsEnv, .odbc.ini DSN needs to be set to either 'postgres', 'mysql', or 'oracle'). The automatic ./packaging/setup_irods.sh script will work only with the system-installed database server. ``` Are .irodsEnv, .odbc.ini the only files that need to be changed? How would I know? Again the reference to a system installed database server - does this mean locally? Where are the reference notes for doing this with all the database plugins - have there been upgrade notes written for all of them (I'm particularly concerned with Oracle, of course, but if the default is postgress, then the MySQL users may also want this.. *Step 16* says ``` On all resource servers in the same Zone, install and setup 4.0+. Existing configuration details should be ported as well ('server.config', 'core.re', Vault permissions). ``` As above a list of files to update and pointers ass to how to cover to JSON appreciated. *step 20* says ``` Sunset 3.3.x server(s) ``` I'm not sure what this means; didn't we upgrade the iCAT and all the iRES in steps 15 and 16? Apologies for the stream of consciousness, please ask for clarification as required! Answers: username_1: There are no plans to document a migration to 4.x with `--run-in-place` (https://github.com/irods/irods/issues/2968) I'll work on a more specific list of configuration files for Step 2. The database can be an external database in Step 10, will update that. Steps 12 and 13 demand more attention, but specificity is hard there - every installation is different. The configuration files are converted if the new JSON files are not found. Once found, they will not be converted again. Step 20 is just referring to the old code that was running in a regular user's namespace, usually with a local instance of a database. This step is a placeholder for 'do whatever you want to delete the old installation, it is no longer needed'. I see now that 'server' might be read as a machine, rather than a piece of software. I'll update it. username_0: Thanks for all of this. I think I've got confused with the run-in-place and packaged installs because I don't see the packaged installs taking the run-in-place configs, that has to be done manually. I think this document refers to a run-in-place to run-in-place upgrade, or a packaged to packed upgrade. While I think this is entirely sensible it wasn't obvious to me starting down the road initially, so might be worth adding in a statement about what circumstances this upgrade doc is supposed to cover. Feel free to use the instructions I mailed Jason that we plan to use for upgrading from 3.3.1 to 4.x as a template for a once in a lifetime 3-4 migration doc, though! Status: Issue closed
redisson/redisson
373040770
Title: Default values for local cache in RedissonSpringLocalCachedCacheManager Question: username_0: What are default values for these properties? Are these values from LocalCachedMapOptions::defaults()? ``` localCacheOptions: evictionPolicy: "LRU" reconnectionStrategy: "CLEAR" syncStrategy: "UPDATE" writeMode: "WRITE_THROUGH" cacheSize: 1000 timeToLiveInMillis: 300000 maxIdleInMillis: 300000 ``` Answers: username_1: No, seems config instance had been changed further in code. username_0: Oh, I see, found it Status: Issue closed
aws-amplify/amplify-cli
485907767
Title: Cannot Create a trigger on a DynamoDB created with the CLI Question: username_0: **Describe the bug** I tried to create a trigger from a DynamoDB table created from the CLI but I got some errors on the console **To Reproduce** 1. `amplify add api`: and create a simple model (im my case I create this: ``` type Lead @model { id: ID! fullname: String! email: AWSEmail! phone: String! serviceType: Service comment: String } enum Service { RESIDENCIAL COMERCIAL INDUSTRIAL } ``` 1. `amplify add function`: create a simple function. add whatever name. no edit needed 1. `amplify push` 1. `amplify api console` 1. go to the triggers tab and create a new trigger from an existing Lambda Function 1. in the popup, Select the previous function created and leave the other inputs as is. 1. click on create nad you will see the error below ``` Cannot access stream arn:aws:dynamodb:eu-west-1:115238827609:table/Lead-xwwfwne24nd43feacszaupqof4-staging/stream/2019-08-25T15:08:37.527. Please ensure the role can perform the GetRecords, GetShardIterator, DescribeStream, and ListStreams Actions on your stream in IAM. (Service: AWSLambda; Status Code: 400; Error Code: InvalidParameterValueException; Request ID: 4f297881-6498-49c3-b237-97e902972409) ``` **Expected behavior** I expect the trigger to be created and let me edit/update my function on my editor and push it via the amplify cli. **Screenshots** <img width="1680" alt="Screen Shot 2019-08-27 at 6 02 34 PM" src="https://user-images.githubusercontent.com/725120/63788084-2c763d80-c8f5-11e9-9224-39cb11217b3b.png"> <img width="1680" alt="Screen Shot 2019-08-27 at 5 51 37 PM" src="https://user-images.githubusercontent.com/725120/63788102-326c1e80-c8f5-11e9-8e16-7e108217c1c9.png"> <img width="899" alt="Screen Shot 2019-08-27 at 2 50 39 PM" src="https://user-images.githubusercontent.com/725120/63789202-30a35a80-c8f7-11e9-9105-ed0e59c525b0.png"> **Desktop (please complete the following information):** Model Name: MacBook Pro Model Identifier: MacBookPro11,3 Processor Name: Intel Core i7 Processor Speed: 2.5 GHz Number of Processors: 1 Total Number of Cores: 4 L2 Cache (per Core): 256 KB L3 Cache: 6 MB Memory: 16 GB Google Chrome: Version: 76.0.3809.100 Obtained from: Identified Developer Last Modified: 8/6/19, 2:29 AM Kind: Intel 64-Bit (Intel): Yes **Additional context** I tried it with a function created from the cli and with another functionc reated directly in the console. on both cases I got the same error so my assumption is that cloudformation is not doing something on the DynamoDB and prevent this to work properly. Answers: username_0: thanks for the solution @username_1 ! I'm not too familiar with all AWS tools, can you give me a more detailed step by step? - where do I need to add this code you mentioned? - do I need to change something in the console? - can everything be done via de CLI and this snippet? thanks! username_1: @username_0 Sorry for the late response: 1. You would need to modify this via the console for tables managed by the GraphQL transformer, but this should soon be addressed out of the box - as a part of the CLI - as a part of this PR - https://github.com/aws-amplify/amplify-cli/pull/2463 2. In the meanwhile, you can go the Lambda console - go to your Lambda function trigger -> Find out the Lambda Execution Role Name -> Attach the policy mentioned above as a part of the Lambda execution role in the IAM console. 3. Answered in 1. Status: Issue closed
seazon/FeedMe
827158176
Title: Images in articles makeuseof.com do not show Question: username_0: The images in (the article)[https://www.makeuseof.com/why-you-should-use-linux-networking-tools-in-windows-with-wsl/] or such articles from the website do not show when phrased by FeedMe. Answers: username_1: will check it.
xmartlabs/Eureka
740054710
Title: font colour change in PickerInlineRow Question: username_0: PickerInlineRow font colour is no longer working as of todays apple iOS patch. where do I optionally set it? what I mean is, it's gone from darkgrey on a slate background (we opted for the slate via `$0.cellSetup({ (cell, row) in cell.backgroundcolor= self.slateBakground}))` to faint text on slate and it is hard to read Answers: username_0: Hi I just want to re-iterate there is no way whatsoever to chagne the PickerInlineRow. I spent 4 hours yesterday until 2am trying everything. YOU CAN do: - colour of label - background of selected value - textcolor of picker scroll text (options) - selected value text colour WHEN IN up/down animation "mode" (this is also the selected value colour while the picker options are visible) but not the resting selected value colour why is this? It has completely ruined my application look and feel. iOS altered the grey for the resting selected value font colour and it's almost invisible and the only way was to make all the pickers backgrounds different - which looks horrible and non uniform username_0: ![Uploading Picker.gif…]() As you can see the word "Manual" would be unreadable if the background was the same as "Tag" (which it used to be) the colour of the word "Manual" changes to black (as intended on tap) but the flaw is that the word "Manual" should be light grey on dark grey, I've had to inverse the colours to make it readable as of yesterdays iOS update forcing pale text everywhere It looks hideous that inline spinners are now totally unlike the rest of the UI username_1: Does this not work for you? ```swift .cellUpdate({ (cell, row) in cell.detailTextLabel?.textColor = .green }) ``` username_0: Thanks! I had tried cell.textBale? only I did not see that despite looking about 12 times through the entire list, if I may say, now it's been brought to my attention surely it's the other way round? `DetailText` is surely the label, describing the item, whereas `text` is the text value? username_1: `detailTextLabel` and `textLabel` are variables of UITableViewCell. There are different possible layouts so textLabel won't always be on the left and detailTextLabel on the right. You can check that out in the Apple documentation username_0: Silly move from Apple. Who wouldn't extend past that ? They live in a world of their own those guys. Honestly
SSAFY-5th-GwanJu-4C-Algorithms/Algorithm_basic
1009904923
Title: [성은][힙][9월3주][PGMS] 이중우선순위큐 Question: username_0: 이거 저희 백준에서 풀었던거랑 똑같이 풀었습니다! ```JAVA import java.util.TreeMap; class Solution { public int[] solution(String[] operations) { TreeMap<Integer,Integer> treeMap = new TreeMap<Integer,Integer>(); for(String str : operations){ String[] now = str.split(" "); char operation = now[0].charAt(0); int data = Integer.valueOf(now[1]); if(operation == 'I') { treeMap.put(data, treeMap.getOrDefault(data,0) + 1); } else { if(treeMap.isEmpty()) continue; if(data == -1) { int minKey = treeMap.firstKey(); if(treeMap.get(minKey) == 1) { treeMap.remove(minKey); } else { treeMap.put(minKey, treeMap.get(minKey) - 1); } } else { int maxKey = treeMap.lastKey(); if(treeMap.get(maxKey) == 1) { treeMap.remove(maxKey); }else { treeMap.put(maxKey, treeMap.get(maxKey) - 1); } } } } int[] answer = new int[2]; if(treeMap.isEmpty()) { answer[0] = 0; answer[1] = 0; } else { answer[0] = treeMap.lastKey(); answer[1] = treeMap.firstKey(); } return answer; } } ``` 다시복습 TreeMap자료구조가 laskKey()가 최댓값, firstKey()가 최솟값을 반환하더라구요 ![image](https://user-images.githubusercontent.com/78025547/135118546-ad9da465-1b69-4978-a03d-5313287774b8.png) Answers: username_1: 오 treemap 자료구조 신기해여 ``` treeMap.put(minKey, treeMap.get(minKey) - 1); treeMap.put(maxKey, treeMap.get(maxKey) - 1); ``` 요고 두개는 왜 처리하는거에여? Status: Issue closed
fog/fog
134027556
Title: [openstack]Is that making associate and disassociate in floating-ip objet is useful? Question: username_0: I found that `associate` and `disassociate` a public_ip with a port is not in object `model/floating_ip`, but in `requests/` there are services `associate_floating_ip` to `disassociate_floating_ip` to do the thing. https://github.com/fog/fog/blob/master/lib/fog/openstack/models/network/floating_ip.rb#L6-L37 https://github.com/fog/fog/blob/master/lib/fog/openstack/requests/network/associate_floating_ip.rb https://github.com/fog/fog/blob/master/lib/fog/openstack/requests/network/disassociate_floating_ip.rb So do you think is useful to put them into class `floating_ip`? I'm up to do this, if you are agree :) Answers: username_1: @username_0 yeah, I think that sounds reasonable and useful. Just let me know if you have questions or anything I can help with. username_2: There's quite a lot of that in the OpenStack code (requests available but not exposed on models) so please feel free to fix :+1: username_0: please check the PR #3850, it should do the thing. please also tell me the way you want me to improve :) username_1: closed by #3850 Status: Issue closed
bigeasy/islander
505595194
Title: Replace Monotonic with Paxos. Question: username_0: We where using the promise comparison functions in Monotonic. I've wanted to move away from Monotonic now that JavaScript has `BigInt`. I've only moved so far away. Keeping the hex value pair, but implementing the comaparison as `BigInt`. Retaining only bits of Monotonic that I was using in for Paxos.<issue_closed> Status: Issue closed
wooclap/moodle-mod_wooclap
953864584
Title: Reset not implemented Question: username_0: When resetting a Moodle course that has a Wooclap activity the page shows a message telling that this activity hasn't implemented that feature. Answers: username_1: Hello @username_0 , indeed our plugin does not support that feature of Moodle yet. Can you give us a bit more context? How do you use that feature? Is Wooclap not supporting it a blocker for how you want to use our plugin? username_0: Hi @username_1 The course reset functionality is usually used when you want to reuse the course for a new edition of it, typically with new students, and each activity offers (or not) some actions that could be done over the instances of that activity (the forum offers to delete conversations, the assignments offers to delete submissions, etc.): https://docs.moodle.org/en/Reset_course To support that functionality the plugin needs to describe the functions detailed here (even if it doesn't reset data at all): https://docs.moodle.org/dev/Implementing_Reset_course_functionality_in_a_module I don't know what kind of actions should be done within Wooclap module instances when reseting a course, but you can support this functionality without defining any actions, and in that way you will avoid that message showing up on every course reset with Wooclap modules on it (it's not a blocker but a enhancement proposal). username_1: Thanks for the context, I've added this enhancement to our internal roadmap. We'll post a new message here when we have more news to share.
activeadmin/activeadmin
20274055
Title: Filtering with Authorization returns unauthorized records Question: username_0: Using `CanCan` to authorize specific records works well, but filtering doesn't work as expected. I will explain via examples to be more easy (at least to me) describe the issue: I have something like this model relationship: ``` ruby class Item < ActiveRecord::Base belongs_to :category end ``` And my `CanCan`'s `Ability` configures the authorization as: ``` ruby can :read, Category, id: allowed_items can :manage, Item, category_id: allowed_items ``` When I go to Categories index page everything works well, filtering only allowed items. But I have a filter by category where shows a select box. The first bug is that the category filter shows all categories (equivalent to `Category.all`) instead of filtering out the unallowed ones. The second, and IMO worst problem is that when I filter by one unallowed category, it shows the unallowed items, bypassing the configured authorization scopes that initially worked. If someone needs a more substation code to be able to reproduce the issue, I will be glad to provide. And if someone is able to help me with the AA internals, I can investigate the issue by myself. And sorry about my limited english. Thanks for the attention :smile: Answers: username_1: @username_0, I'm not familiar with cancan, but I think reason of this was wrong cancan ability declaration. Can you confirm this? username_2: This issue is very old and we haven't received quality feedback in a long time. Please open a new issue with reproduction steps if this is still happening. Status: Issue closed
johnste/finicky
475131453
Title: Sign/Notarize releases Question: username_0: I just installed Finicky via `brew cask install finicky` and it turns out that Finicky.app is not signed for GateKeeper and not notarized with Apple. ``` $ codesign -dvvv /Applications/Finicky.app Executable=/Applications/Finicky.app/Contents/MacOS/Finicky Identifier=net.kassett.finicky Format=app bundle with Mach-O thin (x86_64) CodeDirectory v=20100 size=1572 flags=0x2(adhoc) hashes=42+5 location=embedded Hash type=sha256 size=32 CandidateCDHash sha256=3ad68f568cec6ee63e87ec4968fbda51d630c06e Hash choices=sha256 CDHash=3ad68f568cec6ee63e87ec4968fbda51d630c06e Signature=adhoc Info.plist entries=26 TeamIdentifier=not set Sealed Resources version=2 rules=13 files=24 Internal requirements count=0 size=12 ``` This results in window like that shown on first start: <img width="420" alt="screenshot_2019 07 31_085340" src="https://user-images.githubusercontent.com/868842/62213374-c38cbb80-b370-11e9-92c9-0013f9159a51.png"> I can safely right click and select open to use the app, but it would be nice if this extra step won't be required. Few links: * [Notarizing Your App Before Distribution](https://developer.apple.com/documentation/security/notarizing_your_app_before_distribution) * [Signing Your Apps for Gatekeeper](https://developer.apple.com/developer-id/) I understand this might be time consuming, but should be free to create a developer id to sign and/or notarize the app. Thanks! Answers: username_1: As far as I know signing the application requires an Apple developer license which has a cost attached to it (around $100/year I think) Since I'm not making money from finicky I'd rather not add any costs to develop it other than my time. If I'm not mistaken the only difference is that it can't be distributed through the app store and there's a warning displayed when you launch the app for the first time. username_0: Apple Developer account is free, as far as I know. You will need to get your developer certificate and it should be straight forward. Paid account required if you plan to publish app in the App Store. * [Upload a macOS app to be notarized](https://help.apple.com/xcode/mac/current/#/dev88332a81e) username_1: Okay, thank you for correcting me. :) I'll look into signing the application. I'm on summer vacation right now so progress is a bit slower than usual. username_1: I've been looking into this some more today, and (unless I am mistaken) to generate a developer certificate I need to enroll with the Apple Developer Program, which is 999 SEK (~100 USD) per year. For now this will have to wait. ![image](https://user-images.githubusercontent.com/886051/63185749-219cec80-c05b-11e9-8448-dce62c5cb814.png) ![image](https://user-images.githubusercontent.com/886051/63185800-35485300-c05b-11e9-92d8-3e29faaa3e03.png) username_2: Signing/notarizing the app would obv be ideal, but by saying, "..Apple soon will require all macOS Catalina apps to be notarized' is terribly misleading, imo. I would interpret that as I will no longer have the ability to run 3rd party/unsigned apps, and by removing the option to "Allow apps from unknown sources" in the UI it would appear that is correct. Devs releasing apps, such as finicky, should not be required to sign up and pay for an Apple Developer account. This is just step 1 of the process. It wouldn't surprise me if in the future actually won't be to run 3rd party/unsigned apps at all. This change, obviously centered around "security" of it's users. While a large portion of macOS users will be less vulnerable, some of the users who _need_ to run these types of apps are going to be more vulnerable by completely disabling Gatekeeper with `sudo spctl --master-disable`. /rant @username_1 'preciate all the work you've put into finicky. It flies under the radar since you don't actually interact with the app when it's doing it's thing the control it give you/ability to coral work versus personal items is soooo underrated. username_3: Afaik you can still run unsigned apps using ctrl-click > open, like you used to, no need to completely disable gatekeeper. username_4: Note that Little Snitch will put up a not signed warning as well when Finicky looks up a url (for example because of expanding shortend urls) that is intercepted.
google/ExoPlayer
693493368
Title: Does exoplayer has a feature like bitrate converter for streaming mp3 file Question: username_0: This is not an issue and I just ask this question to clarify my doubt. I just want to know, If I have stored a 320kbps mp3 in firebase cloud storage, Can I let the users **choose** the **bitrate** of the song (Either 64kbps, 128kbps, or 320kpbs) via **exoplayer**? Or should I have to store 3 different bitrates mp3 files in storage and stream them? Answers: username_1: What would be the purpose of choosing the bitrate? To download fewer bytes? ExoPlayer is a client-side library and can operate on content once the content arrives at the client device. If you want to change to bitrate of the hosted mp3 file in order to transfer less/more bytes, this needs to happen at the remote server, and that's outside ExoPlayer's functionalities. For clarity, ExoPlayer has some audio processing facilities (for example look at (ResamplingAudioProcessor)[https://github.com/google/ExoPlayer/blob/r2.11.8/library/core/src/main/java/com/google/android/exoplayer2/audio/ResamplingAudioProcessor.java]) that operate on the audio content once it is available on the device. I don't think this is for the use-case in question. Closing this issue, feel free to re-open if the answer did not provide enough information. Status: Issue closed
samuelhorwitz/phosphorescence
500319114
Title: Desktop Safari has a massive rendering bug Question: username_0: We need to figure this out even though it's Safari's fault. When switching to the constellation view and then back to playlist view, the playlist is rendered disastrously until you resize or hover over parts that force a re-render. No idea why. This needs to be fixed though, probably something to do with the canvas and Safari's decision making over redrawing sections of the screen used by the canvas.<issue_closed> Status: Issue closed
hsa-fuzzing/test
644639499
Title: Test issue Question: username_0: ### Description Description: Access violation near NULL on source operand Short description: SourceAvNearNull (16/22) Explanation: The target crashed on an access violation at an address matching the source operand of the current instruction. This likely indicates a read access violation, which may mean the application crashed on a simple NULL dereference to data structure that has no immediate effect on control of the processor. ### Command /home/qdl/fuzzing/vermont-fuzz/vermont -f testconfig.xml ### Vermont-Config [https://github.com/username_0/test/blob/master/1593007238.3800216/testconfig.xml](https://github.com/username_0/test/blob/master/1593007238.3800216/testconfig.xml) ### Input File [https://github.com/username_0/test/blob/master/1593007238.3800216/testpcap.pcap](https://github.com/username_0/test/blob/master/1593007238.3800216/testpcap.pcap) ### Stack |address|module|symbol| |--|--|--| |0x591db7|/home/qdl/fuzzing/vermont-fuzz/vermont|IpfixCsExporter::performShutdown| |0x5cbb5f|/home/qdl/fuzzing/vermont-fuzz/vermont|Module::shutdown| |0x4ee4b3|/home/qdl/fuzzing/vermont-fuzz/vermont|CfgHelper<IpfixCsExporter,| |0x47c373|/home/qdl/fuzzing/vermont-fuzz/vermont|ConfigManager::shutdown| |0x47891b|/home/qdl/fuzzing/vermont-fuzz/vermont|main| ### Disassembly |address|text| |--|--| |5840287|call 0x591a90 <IpfixCsExporter::closeFile()>| |5840292|jmp 0x591db0 <IpfixCsExporter::performShutdown()+80>| |5840294|mov edi,0x6a8d30| |5840299|call 0x5ee6f0 <__sanitizer_cov_trace_pc_guard>| |5840304|mov rdi,QWORD PTR [rbx+0x118]| |5840311|mov rax,QWORD PTR [rdi]| |5840314|xor esi,esi| |5840316|call QWORD PTR [rax+0x18]| |5840319|mov rax,QWORD PTR fs:0x28| |5840328|cmp rax,QWORD PTR [rsp+0x8]| ### Faulting Frame |address|module|symbol| |--|--|--| |5840311|/home/qdl/fuzzing/vermont-fuzz/vermont|IpfixCsExporter::performShutdown| ### Faulting Instruction |address|text| |--|--| |5840311|mov rax,QWORD PTR [rdi]| ### Registers |name|value| |--|--| |rax|13378894672732964096| |rbx|17541104| |rcx|14550| |rdx|0| [Truncated] |fs|0| |gs|0| ### Compiler-Version 1.2.1 ### Compiler-Flags -f -g ### System linux (uname -rop) ### Memory 8gb ### CPU - **Arch:** x86 - **Cores:** 4 - **MHz:** 3500
apache/trafficserver
198213088
Title: CID 1367515, CID 1367514, CID 1367513: ts_lua plugin Question: username_0: ``` *** CID 1367515: Error handling issues (CHECKED_RETURN) /plugins/experimental/ts_lua/ts_lua_server_response.c: 354 in ts_lua_server_response_set_version() 348 GET_HTTP_CONTEXT(http_ctx, L); 349 350 TS_LUA_CHECK_SERVER_RESPONSE_HDR(http_ctx); 351 352 version = luaL_checklstring(L, 1, &len); 353 CID 1367515: Error handling issues (CHECKED_RETURN) Calling "sscanf" without checking return value (as is done elsewhere 61 out of 67 times). 354 sscanf(version, "%2u.%2u", &major, &minor); 355 356 TSHttpHdrVersionSet(http_ctx->server_response_bufp, http_ctx->server_response_hdrp, TS_HTTP_VERSION(major, minor)); 357 358 return 0; ``` and ``` *** CID 1367514: Error handling issues (CHECKED_RETURN) /plugins/experimental/ts_lua/ts_lua_client_request.c: 903 in ts_lua_client_request_set_version() 897 ts_lua_http_ctx *http_ctx; 898 899 GET_HTTP_CONTEXT(http_ctx, L); 900 901 version = luaL_checklstring(L, 1, &len); 902 CID 1367514: Error handling issues (CHECKED_RETURN) Calling "sscanf" without checking return value (as is done elsewhere 61 out of 67 times). 903 sscanf(version, "%2u.%2u", &major, &minor); 904 905 TSHttpHdrVersionSet(http_ctx->client_request_bufp, http_ctx->client_request_hdrp, TS_HTTP_VERSION(major, minor)); 906 907 return 0; 908 } ``` and ``` *** CID 1367513: Error handling issues (CHECKED_RETURN) /plugins/experimental/ts_lua/ts_lua_client_response.c: 370 in ts_lua_client_response_set_version() 364 GET_HTTP_CONTEXT(http_ctx, L); 365 366 TS_LUA_CHECK_CLIENT_RESPONSE_HDR(http_ctx); 367 368 version = luaL_checklstring(L, 1, &len); 369 CID 1367513: Error handling issues (CHECKED_RETURN) Calling "sscanf" without checking return value (as is done elsewhere 61 out of 67 times). 370 sscanf(version, "%2u.%2u", &major, &minor); 371 372 TSHttpHdrVersionSet(http_ctx->client_response_bufp, http_ctx->client_response_hdrp, TS_HTTP_VERSION(major, minor)); 373 374 return 0; 375 } ``` Answers: username_0: Meh, there's already a PR for this, nm. Status: Issue closed
xsahil03x/giffy_dialog
852708417
Title: Is their a list of publicly available flare animations? Question: username_0: I don't have the experience to develop great animations is their a list of free animations available anywhere? Answers: username_1: On Rive's official site they have a place where you can browse. Just check what type of license is attached to it. A lot of them are CC BY 4.0 meaning you can use them in your own work as long as you give credit to the original author. https://flare.rive.app/explore/popular/trending/all
microsoft/Microsoft.Unity.Analyzers
625030421
Title: Use TryGetComponent to prevent memory allocation Question: username_0: Recommended on Twitter as part of #unitytips https://twitter.com/BinaryImpactG/status/1265191692531044352?s=20 #### Problem statement Users want to check if a component exists on a GameObject. #### Proposed solution When using GetComponent and null checking the result, TryGetComponent can be used instead and prevents memory allocation. Answers: username_1: This one is interesting because it requires knowing the version of Unity, or at least validating in the referenced Unity assembly that this method actually exists. username_2: `TryGetComponent` is defined in `Component` and `GameObject`, starting with `Unity 2019.2` Perhaps we can use Roslyn's `Project.ParseOptions.PreProcessorSymbolNames` and search for `UNITY_2019_2_OR_NEWER` ? But this will only work for projects generated from Unity (not handcrafted c# libs). We also need to upgrade the test infrastructure to use preprocessor symbols like this: ```csharp project = project .WithParseOptions(((CSharpParseOptions)project.ParseOptions) .WithPreprocessorSymbols("UNITY_2019_2_OR_NEWER")); ``` username_1: @username_2 I think the safest is just to look at the symbols of the referenced assembly from the current context. username_3: Any update on this? Status: Issue closed
sveltejs/eslint-plugin-svelte3
811252697
Title: Wrong no-unsafe-member-access for reactive assignments and store subscriptions Question: username_0: #87 uncovered some limitations when using type-aware rules. Some of these are fixed by #88 , but not all of them. Reactive assignments and store subscriptions will fail: ```svelte <script lang="ts"> import { writable } from 'svelte/store'; const store = writable([]); $store.length; // wrong no-unsafe-member-access error $: assignment = []; assignment.length; // wrong no-unsafe-member-access error // You can work around this by doing let another_assignment: string[]; $: another_assignment = []; another_assignment.length; // OK </script> ``` This is because the transformation just prepends generated `let X` statements above the code, with no specific type set to it. For it to work correctly, we would need to do transformations inline: - `$: assignment = ..` --> `let assignment = ..` (but only if user did not declare the `let` himself) - `const store = writable(...)` --> `const store = writable(..);let $store = get_store_value(store)` (where `get_store_value` is a fake function returning the `Value` in `Store<Value>`) This would be similar to how we transform code in `svelte2tsx` for the intellisense features. The hard part is getting the line mappings correctly afterwards. Answers: username_0: This should be fixed for stores, for reactive statements it's not working yet.
jaagameister/learning
136045967
Title: Create 100 most common nouns and Illustrate them Answers: username_1: ​​Some things that should probably be baked into the spreadsheet in eye-catching style as reminders: - We should pick objects attached to nouns that use common letters with regular spellings. This may differ some from language to language. - Swapna can exercise some "artist's choice", regardless. - For Kannada, we'll use the spoken colloquial vernacular throughout, in direct contravention of the convention (if any) that only the high language should appear in writing. Cf. http://karnatique.blogspot.in/2008/03/concept-of-h-and-l-kannada-disease.html Status: Issue closed
mvdoc/psiturk-docker
258979710
Title: nginx gives 400 bad request when using psiturk with jspsych Question: username_0: @username_1 I'm getting a Bad Request error 400 when using psiturk with jspsych. It works fine if I bypass nginx and access gunicorn directly, so there's something happening with nginx. I tried looking up on google and changed a bunch of parameters (mostly related to increasing the buffer size), but it still fails. I have created a minimal example that fails with nginx, which you can download here: https://www.dropbox.com/s/kk5q8vo0qh6wzq2/nginx_fail.zip Enabling debugging for nginx, this is the error I get: ``` 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy status 400 "400 BAD REQUEST" 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy header: "Server: gunicorn/19.4.5" 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy header: "Date: Tue, 19 Sep 2017 21:54:34 GMT" 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy header: "Connection: close" 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy header: "Content-Type: text/html" 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy header: "Content-Length: 192" 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy header done 2017/09/19 21:54:34 [debug] 6#6: *1 HTTP/1.1 400 BAD REQUEST Server: nginx/1.13.5 Date: Tue, 19 Sep 2017 21:54:34 GMT Content-Type: text/html Content-Length: 192 Connection: keep-alive ``` And this is the complete log ``` 2017/09/19 21:54:34 [debug] 6#6: *1 http process request line 2017/09/19 21:54:34 [debug] 6#6: *1 http request line: "GET /complete?uniqueId=debugRD825H:debug02HM1Q HTTP/1.1" 2017/09/19 21:54:34 [debug] 6#6: *1 http uri: "/complete" 2017/09/19 21:54:34 [debug] 6#6: *1 http args: "uniqueId=debugRD825H:debug02HM1Q" 2017/09/19 21:54:34 [debug] 6#6: *1 http exten: "" 2017/09/19 21:54:34 [debug] 6#6: *1 posix_memalign: 000000A63D7362B0:4096 @16 2017/09/19 21:54:34 [debug] 6#6: *1 http process request header line 2017/09/19 21:54:34 [debug] 6#6: *1 http header: "Host: 127.0.0.1" 2017/09/19 21:54:34 [debug] 6#6: *1 http header: "Connection: keep-alive" 2017/09/19 21:54:34 [debug] 6#6: *1 http header: "Upgrade-Insecure-Requests: 1" 2017/09/19 21:54:34 [debug] 6#6: *1 http header: "User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_12_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.91 Safari/537.36" 2017/09/19 21:54:34 [debug] 6#6: *1 http header: "Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,*/*;q=0.8" 2017/09/19 21:54:34 [debug] 6#6: *1 http header: "Referer: http://127.0.0.1/exp?hitId=debugOD8KFK&assignmentId=debug02HM1Q&workerId=debugRD825H&mode=debug" 2017/09/19 21:54:34 [debug] 6#6: *1 http header: "Accept-Encoding: gzip, deflate, br" 2017/09/19 21:54:34 [debug] 6#6: *1 http header: "Accept-Language: en-US,it-IT;q=0.8,it;q=0.6,en;q=0.4" 2017/09/19 21:54:34 [debug] 6#6: *1 http header done 2017/09/19 21:54:34 [debug] 6#6: *1 generic phase: 0 2017/09/19 21:54:34 [debug] 6#6: *1 rewrite phase: 1 2017/09/19 21:54:34 [debug] 6#6: *1 test location: "/" 2017/09/19 21:54:34 [debug] 6#6: *1 using configuration "/" 2017/09/19 21:54:34 [debug] 6#6: *1 http cl:-1 max:1048576 2017/09/19 21:54:34 [debug] 6#6: *1 rewrite phase: 3 2017/09/19 21:54:34 [debug] 6#6: *1 post rewrite phase: 4 2017/09/19 21:54:34 [debug] 6#6: *1 generic phase: 5 2017/09/19 21:54:34 [debug] 6#6: *1 generic phase: 6 2017/09/19 21:54:34 [debug] 6#6: *1 generic phase: 7 2017/09/19 21:54:34 [debug] 6#6: *1 access phase: 8 2017/09/19 21:54:34 [debug] 6#6: *1 access phase: 9 2017/09/19 21:54:34 [debug] 6#6: *1 access phase: 10 2017/09/19 21:54:34 [debug] 6#6: *1 post access phase: 11 2017/09/19 21:54:34 [debug] 6#6: *1 generic phase: 12 2017/09/19 21:54:34 [debug] 6#6: *1 try files handler 2017/09/19 21:54:34 [debug] 6#6: *1 http script var: "/complete" 2017/09/19 21:54:34 [debug] 6#6: *1 trying to use file: "/complete" "/var/www/exp/nginx_fail/complete" 2017/09/19 21:54:34 [debug] 6#6: *1 http script var: "/complete" [Truncated] 2017/09/19 21:54:34 [debug] 6#6: *1 http upstream process header 2017/09/19 21:54:34 [debug] 6#6: *1 malloc: 000000A63D7372C0:4096 2017/09/19 21:54:34 [debug] 6#6: *1 recv: eof:0, avail:1 2017/09/19 21:54:34 [debug] 6#6: *1 recv: fd:12 155 of 4096 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy status 400 "400 BAD REQUEST" 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy header: "Server: gunicorn/19.4.5" 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy header: "Date: Tue, 19 Sep 2017 21:54:34 GMT" 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy header: "Connection: close" 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy header: "Content-Type: text/html" 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy header: "Content-Length: 192" 2017/09/19 21:54:34 [debug] 6#6: *1 http proxy header done 2017/09/19 21:54:34 [debug] 6#6: *1 HTTP/1.1 400 BAD REQUEST Server: nginx/1.13.5 Date: Tue, 19 Sep 2017 21:54:34 GMT Content-Type: text/html Content-Length: 192 Connection: keep-alive ``` Do you have any clue on how to fix this? Answers: username_0: Ping @username_1 and perhaps @username_2? username_1: Hi @username_0. Yes, sorry, this is in on my radar. I'm hoping to get a chance to look at it this weekend. I've been a bit swamped this week. Sorry! Status: Issue closed username_0: Thanks @username_1, I managed to debug it and solved the problem. Updating psiturk.js with the latest version did it. The reason was that at the end of the experiment with the old psiturk.js, psiturk would request ``` self.completeHIT = function() { self.teardownTask(); // save data one last time here? window.location= self.taskdata.adServerLoc + "?uniqueId=" + self.taskdata.id; } ``` instead of (new version) ``` self.completeHIT = function() { self.teardownTask(); // save data one last time here? window.location= self.taskdata.adServerLoc + "?uniqueId=" + self.taskdata.id + "&mode=" + self.taskdata.mode; } ``` so because of the missing `&mode=`, nginx wouldn't serve the page (I don't know why though, would be curious why to know why). I'm gonna close this. Thanks again! username_1: It looks like [that change was committed by @username_2](https://github.com/NYUCCL/psiTurk/commit/eeb56fc161e7181edc7360f2b159d45a3e1c666c), so he can probably answer better. But it looks like this is an indicator of whether you're in live or debug mode. Maybe without this indicator, it was trying to save data to the live version of things while you were in debug mode? username_2: It worked when you used gunicorn because psiturk.js is loaded dynamically, from the up-to-date pip repo. But when you use nginx it serves it statically without asking flask/gunicorn to load it, because you've copied it into a static folder (why did you copy it into a static dir like that?). Flask/gunicorn explodes when you don't pass `mode` because the [route tries to access it](https://github.com/NYUCCL/psiTurk/blob/255bbe1515b7f2ab20209b8da950f58a1f0557b0/psiturk/experiment.py#L621-L643) username_0: I have no idea ¯\_(ツ)_/¯ probably a residual of when I was playing around...
TorXakis/TorXakis
954524522
Title: Abstraction of Events Question: username_0: In TorXakis, one currently can use synchronous events to link different abstraction layers. E.g the event `Key_Press` might be linked to a `Key_Down` at a lower abstraction layer (that also contains `Key_Up`) by `Key_Press | Key_Down`. I wonder whether * `rename` (i.e. `Key_Down renames Key_Press`), or * `inheritance` of events (i.e. `Key_Down is Key_Press`); might be alternatives that we should support as well. Or maybe we should only support one of these. And maybe enforce that events are instantaneous and can't happen synchronous.
fcrepo/Fedora-API-Test-Suite
362766128
Title: All versioning tests assume client-managed versioning Question: username_0: The fedora spec mentions different models for versioning, including server-managed and client-managed as two possibilities. Yet the test suite seems to check only the client-managed interactions; in fact, if the client-managed versioning interactions aren't supported, the tests fail. It seems that an implementation that has spec-compliant server-managed versioning would fail all the versioning tests here.<issue_closed> Status: Issue closed
MicrosoftDocs/azure-docs
380348560
Title: Multi-Geo minimum requirements Question: username_0: Minimum of 2500 licenses requirements mentioned in this document is incorrect. Please refer to Microsoft Product Terms for current minimum, being 5000 licenses. Would be good to have page updated. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: b285f539-0487-0de5-24ef-89f773a1081d * Version Independent ID: b4f26f73-6b46-9703-692b-abe1f0d5f825 * Content: [Azure Active Directory Connect sync: Configure preferred data location for Multi-Geo capabilities in Office 365](https://docs.microsoft.com/en-us/azure/active-directory/hybrid/how-to-connect-sync-feature-preferreddatalocation#enable-synchronization-of-preferred-data-location) * Content Source: [articles/active-directory/hybrid/how-to-connect-sync-feature-preferreddatalocation.md](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory/hybrid/how-to-connect-sync-feature-preferreddatalocation.md) * Service: **active-directory** * GitHub Login: @billmath * Microsoft Alias: **billmath** Answers: username_1: @username_0 Thanks for your feedback! We will investigate and update as appropriate. Status: Issue closed username_1: @username_0 @billmath I made a pull request to revise this. @billmath please review PR #57888 for your approval.
MicrosoftDocs/OfficeDocs-SkypeForBusiness
625573087
Title: O/M365 F3 Question: username_0: [Voer hier feedback in] Office 365 and M365 **F3** are not listed, while Office 365 F1 is. Please update --- #### Documentdetails ⚠ *Dit gedeelte niet bewerken. Het is vereist om problemen te koppelen tussen docs.microsoft.com en GitHub.* * ID: 35057f71-792b-a4e2-a2ee-a4908c11f3a4 * Version Independent ID: 7e23f067-1dcd-16f0-3f9a-ff8d13c1f659 * Content: [What is Microsoft 365 Business Voice? - Microsoft Teams](https://docs.microsoft.com/nl-nl/MicrosoftTeams/business-voice/whats-business-voice) * Content Source: [Teams/business-voice/whats-business-voice.md](https://github.com/MicrosoftDocs/OfficeDocs-SkypeForBusiness/blob/live/Teams/business-voice/whats-business-voice.md) * Service: **msteams** * GitHub Login: @username_4 * Microsoft Alias: **username_4** Answers: username_1: @username_0 Thank you for submitting feedback and contributing to the docs. We are currently investigating this. username_2: @username_3 Hello Lana, Can you kindly confirm if we have to add _Office 365 F3_ as this article states? https://docs.microsoft.com/en-us/office365/servicedescriptions/microsoft-365-business-voice-service-description Thank you username_3: Reassigning to @username_4. Status: Issue closed username_4: article updated
nbr23/youtube-dl-server
1027884469
Title: yt-dlp Update issue Question: username_0: For whatever reason **yt-dlp** from pip requires `gcc` and `musl-dev` packages to install now. So these are available here: https://github.com/username_1/youtube-dl-server/blob/47deb2f9ae363d968732146922c8f3652ffbe2df/Dockerfile#L8 But are not in the final docker file: https://github.com/username_1/youtube-dl-server/blob/47deb2f9ae363d968732146922c8f3652ffbe2df/Dockerfile#L20 Easy solution, add those packages to line 20. Alternative possibility, use a different install/upgrade scheme to update eg `yt-dlp -U` after making those packages available for initial build. Answers: username_1: Thanks for the report @username_0, this should be fixed with https://github.com/username_1/youtube-dl-server/commit/4855344e1ebe094e1c7e464891df49b3b1e951e4 Status: Issue closed
oppia/oppia
407842189
Title: Refactor image storage on Oppia Question: username_0: Currently, images are being stored as ```<exploration_id>/assets/images```. For a more general storage structure, the following changes are required: - The URL for storage should be changed as ```exploration/<exploration_id>/assets/images``` to go with all other structures and the changes that may be introduced later. - The images, as well as audio files, need to be migrated in the GCS and the URL will be changed from existing explorations. - The image uploading function needs to be refactored to become more generic by accepting ```<activity_id>``` and ```<activity_type>``` instead of just ```<exploration_id>```. (This issue was disappeared due to deletion of a *WickedBrat's* account ) Answers: username_1: I have seen a one-off job for this issue, are you sure this issue is not closed? username_0: I just received a report of the open issues that were created by WickedBrat from @username_2. This issue was listed in that. username_1: @username_3, any update on this? username_2: Pretty sure it's not fixed - I talked with @username_4 about this recently and I don't think anyone is currently working on it, though it's important. @username_4, is this correct? username_3: Just a quick comment I believe the first bullet is fixed. username_4: As of now, all 3 bullet points are done w.r.t the backend implementation, though in the frontend still image uploads are only for explorations. So, generalizing the RTE to upload images in all structures is pending, username_5: @username_4 are you still working on this? Status: Issue closed username_4: This issue is completed, closing.
callimero/vindriktning2adafruitio
971165837
Title: 'AdafruitIO_WiFi' does not name a type Question: username_0: when compiling I got an 'AdafruitIO_WiFi' does not name a type error. Do I miss an #include file? Thank you Answers: username_1: You need to install some libraries: Using the board manager install "Adafruit IO Arduino" and say yes if it wants to install the other needed ones.
platformio/platformio-home
1130605010
Title: Could not initialize project Question: username_0: PIO Core Call Error: "The current working directory C:\\Users\\Mesacrash\\Documents\\PlatformIO\\Projects\\Compile_2 will be used for the project.\r\n\r\nThe next files/directories have been created in C:\\Users\\Mesacrash\\Documents\\PlatformIO\\Projects\\Compile_2\r\ninclude - Put project header files here\r\nlib - Put here project specific (private) libraries\r\nsrc - Put project source files here\r\nplatformio.ini - Project Configuration File\r\n\n\nError: Processing genericSTM32F103RE (platform: ststm32; board: genericSTM32F103RE; framework: arduino)\r\n--------------------------------------------------------------------------------\r\nVerbose mode can be enabled via `-v, --verbose` option\r\nCONFIGURATION: https://docs.platformio.org/page/boards/ststm32/genericSTM32F103RE.html\r\nPLATFORM: ST STM32 (15.2.0) > STM32F103RE (64k RAM. 512k Flash)\r\nHARDWARE: STM32F103RET6 72MHz, 64KB RAM, 512KB Flash\r\nDEBUG: Current (blackmagic) External (blackmagic, cmsis-dap, jlink, stlink)\r\nPACKAGES: \r\n - framework-arduinoststm32 4.20100.211028 (2.1.0) \r\n - framework-cmsis 2.50700.210515 (5.7.0) \r\n - toolchain-gccarmnoneeabi 1.90201.191206 (9.2.1)\r\nError: Missing PlatformIO build script C:\\Users\\Mesacrash\\.platformio\\packages\\framework-arduinoststm32\\tools\\platformio\\platformio-build.py!\r\n========================== [FAILED] Took 2.80 seconds =========================="
ros-industrial-consortium/bezier
165333378
Title: bezier_application: missing dependency on 'fanuc_m20ia_moveit_config'? Question: username_0: Just tried to launch `bezier_application bezier_application_m20ia.launch surfacing_mode:=true mesh_cad:=plane/plane_defect.ply`, which fails with a `ResourceNotFound` exception complaining about `fanuc_m20ia_moveit_config`. That makes sense, since I don't have that package installed. `rosdep` did not ask me to install it. Is that dependency left out on purpose? Otherwise it might be nice to add it to the manifest. Answers: username_1: The dependency is missing and that is an error. Requiring this will also require `fanuc_m20ia_support` and all other fanuc packages indirectly right? username_0: yes, it will automatically pull in all the dependencies as well. username_0: Not 'all others' btw, just the bits that `fanuc_m20ia_moveit_config` needs (ikfast plugin, driver, support and resources). username_0: Btw: adding a `roslaunch` testing test to your package would automatically discover these things. See the tests in `fanuc_m10ia_support` [here](https://github.com/ros-industrial/fanuc/blob/e632abcb4c8f223879884da56fe412a6d488066a/fanuc_m10ia_support/CMakeLists.txt#L9-L13). Status: Issue closed
CityScope/CS_Cooper-Hewitt
380984084
Title: [Nov 15] Stress Testing Whole System [Simulation+Projection+Slider+Scanning] Question: username_0: Now that most tools are reaching matureness, this is a place to consolidate issues regarding stress-testing the system. This will help keeping track of all things related to all/some tools not playing nice together. Continues https://github.com/CityScope/CS_Cooper-Hewitt/issues/82 and https://github.com/CityScope/CS_Cooper-Hewitt/issues/81 and will serve for new issues arising on this matter. Answers: username_0: ### 82 As I have explained on issue #81, the System slows down dramatically when Unity Gets into the scene. Solutions: Improve framerate or other heavy stuff from unity, Have two computers, Others... username_1: You can also try to reduce a bit the number of Agent in processing to see if it has any impact. But I am not sure it will change a lot as now the number of agent are not very high but maybe username_2: Hi. Yesterday, the crash happened when the slider in Unity got into the scene; but Processing and the scanner were still working in a fine way. I am not 100% about anything in life, however, I can tell that Unity is very heavy and slows the machine a lot and that as soon as I clicked the "play" the slider stop working. It can be a coincidence, for sure, for sure, for sure, but... We should test all @username_4 options username_0: @username_2 @username_4 @username_3 @Carsonsmuts Before we start changing settings/code inside each app, a simple workflow would be to combinatorially run pairs and then triplets of apps at the same time and see what affects what. I.e: 1. Scanner+Slider 2. Slider+Unity 3. Unity+Sim 4. Sim+ Slider+ Scanner 5. unity+slider+scanner 6. etc.... at a certain point, something will show up to indicate where is the FPS drop happens username_3: If this statement is true, we could distribute the GPU load by displaying the Processing on two projectors only (instead of across all four display) and the Unity on another projector (the 4K one). @username_4 @username_1 username_4: Even minimizing the vram that is used by full screen Processing won’t help the situation, since it Unity consumes full vram by itself. @Carsonsmuts and Jason, Taka was talking about a VM solution, yet there the process will be convoluted and time consuming... For reference, Vtd for video cards weren’t supported in the current machine. username_2: In my opinion, and after thinking a bit in the options that we have, the realistic solutions today are: 1) Miracle: To split the Unity and processing into the two different GPUs. Jason thinks that it is a big challenge that either a specific software or a virtual machine can only achieve. Not possible from the IOs as far as we know. 2) The possible one: Reduce the vram consumption of UNITY. Unity, by itself, is consuming 99% of the GPU. Processing is using around 35% to 40% of the GPU. Reducing processing (the 2 screens instead of 4 = fewer pixels) is only an option if Unity can reduce his needs by 70%, then we can squeeze Processing. If not, it doesn't make sense to spend time on reducing processing. BTW: Reducing the output of unity to low quality does not reduce the needs of the GPU. Actually, it is the opposite. So other directions are needed. 3) Hardware solution: 2 computers physically connected. I do not contemplate this option. This is a software issue, not a hardware issue. It is too late for this, we need to rethink the whole table hardware and software. It complicates everything and it solves very little. 4) Buying a supercomputer from NASA: Same than above (?) 5) Deploy it with the lag and the slowness: The curator at the Cooper doesn't really care about the lag (she say that it is OK for her). Since this is an art piece, lag and slowness can be part of the "artistic" part 6) Dropping the ball in some parts: This may be the best solution. Being honest with our work usually is the best option, even when the work of a lot of people will be impacted and may be not showed. We can drop the ball on: - Dynamic interaction with buildings. Keeping processing and the slider (It is a mobility table so the streets are a priority): We can have the lego buildings glued to the table, the slider operating with processing that works smoothly showing the 2 worlds, and 3 options for the top projection: A) No projection. Safer and easier option. We eliminate the "F" cable to the top projector and all the setting in the ceiling. Simpler and efficient solution. B) We can have just a video from the top with the buildings that shift (or not) from good to bad in a loop. C) Link the video with the processing so the slider cannot be manipulated and we have a loop in the whole table (upgrade of members week). -Drop the processing, and to have only UNITY from the top: the Same setting, but with static agents and dynamic buildings. when you move the building you see the change in the videos, etc. This solution kills the mobility concept and keeps the complexity in the hardware. Personally, I never saw the interaction between the scanner and the buildings in Unity, and I am not sure about the lag and smoothness of physical interaction with unity -Drop the ball in all the interaction: We deploy a video in a loop like in Members week. -Drop the ball and do not deploy in the Cooper. username_2: My apologies about the latest note. "Stress Stress Testing Whole System" should be the place to show my own personal stress. Status: Issue closed username_1: Closed it become to generic for a specific issue and I guess it's is now fixed 'see #101)
approvals/ApprovalTests.Java
981893064
Title: Test failures in concurrent execution Question: username_0: JUnit has an experimental parallel test execution [feature](https://junit.org/junit5/docs/current/user-guide/#writing-tests-parallel-execution). When trying to use it in ApprovalsTests test suite, some failures occur. Failures may be observed by adding the file `src/test/resources/junit-platform.properties` with the following items ``` junit.jupiter.execution.parallel.enabled=true junit.jupiter.execution.parallel.mode.default = concurrent ``` ... in modules: - approvaltests - approvaltests-tests - approvaltests-util-tests Answers: username_1: Thank you for your contribution. Status: Issue closed
sphinx-doc/sphinx
1088465727
Title: autodoc_unqualified_typehints does not work well with autodoc_typehints="description" Question: username_0: ### Describe the bug autodoc_unqualified_typehints does not work well with autodoc_typehints="description". ### How to Reproduce ``` autodoc_unqualified_typehints = True autodoc_typehints = "description" ``` ### Expected behavior Typehints in info-field-list are also changed to the short style when autodoc_unqualifed_typehints enabled. ### Your project N/A ### Screenshots _No response_ ### OS Mac ### Python version 3.10.1 ### Sphinx version HEAD of 4.x ### Sphinx extensions autodoc ### Extra tools _No response_ ### Additional context _No response_<issue_closed> Status: Issue closed
dotnet/sdk
984850252
Title: [6.0 RC1] F5 debugging ASP.NET Core broken in Visual Studio due to HTTP 404 Question: username_0: I'm working on an ASP.NET Core 6 app using daily builds of .NET 6 RC1, and this morning when updating from an older daily build to today's latest, F5 debugging the app in Visual Studio is no longer working as when the browser is launcher I just get an HTTP 404 error. Launching with Visual Studio Code works as expected, as does `dotnet run`, as well as UI tests for the app using `WebApplicationFactory<T>`. This leads me to a hypothesis that there's an issue with the browser reload (or related) functionality that Visual Studio loads from the SDK as part of the debugging experience. I couldn't spot anything obvious in the Git history looking at the `release/6.0.1xx` branch, but there's been a few changes in this area in the last 10 days (#20079, #20131, #20267). It's of course possible that a newer build of Visual Studio 2022 is needed that isn't yet publically available, but I thought I'd open this in case it's an actual regression. I switched to and from the versions below and observed it alternating between working and not. LKG installer/SDK version: `6.0.0-rc.1.21420.16` Broken installer/SDK version: `6.0.0-rc.1.21430.26` Visual Studio version: Microsoft Visual Studio Enterprise 2022 Preview (64-bit) 17.0.0 Preview 3.1 /cc @pranavkm Answers: username_0: Doing more debugging around this trying to work out what was going on, I noticed something curious. If I had Visual Studio and Visual Studio Code open at the same time, if Visual Studio built the application and I debugged it, I'd get the problem. If I then launched it from Visual Studio Code it would work. If I then launched the debugger from Visual Studio again, it would not rebuild the application because it was up-to-date, then it would work as expected. If I then forced a build in Visual Studio and debugged it again, it would once again be broken. Comparing the bin and obj folders after each IDE compiled it yielded the following differences: | **File** | **Visual Studio Size** | **Visual Studio Code Size** | |:--|:-- |:--| | `Application.dll` | 203kb | 112kb | | `Application.pdb` | 83kb | 48kb | | `ref/Application.dll` | 50kb | 35kb | Comparing the two DLLs in ILSpy, I noticed this difference in the assembly attributes: ## Visual Studio ![image](https://user-images.githubusercontent.com/1439341/132124722-ae2359b2-b58a-4359-b136-448dad2aba3b.png) ## Visual Studio Code ![image](https://user-images.githubusercontent.com/1439341/132124744-76399580-2b8f-4845-a91a-ee0595833b62.png) It would appear that the Razor Views aren't being compiled into the application DLL, which then causes the views to fail to load with a 404. If I add `<UseRazorSourceGenerator>false</UseRazorSourceGenerator>` to the .csproj file and rebuild in Visual Studio, the application launches and works as expected. username_0: I believe this was caused by #20247. With `<LangVersion>preview</LangVersion>` the application works as expected. With `<LangVersion>latest</LangVersion>` the application fails to compile the Razor views into the assembly. username_0: Confirmed that this works as expected with .NET 6 RC1 and Visual Studio 2022 17.0 preview 4. Status: Issue closed
pato-ontology/pato
210880731
Title: New term request: fibrotic Question: username_0: New term request: fibrotic parent: PATO_0000025 composition suggested def: Fibrosis is the formation of excess fibrous connective tissue in an organ or tissue in a reparative or reactive process. def xref: https://en.wikipedia.org/wiki/Fibrosis Answers: username_1: Do you want to make a PR giving yourself an ID range?
micrometer-metrics/micrometer
542920824
Title: CustomMeterRegistry LoggingMeterRegistry null pointerException Question: username_0: I am trying to implement custom MeterRegistryCustomizer but i am getting below error any leads? Thanks in Advance :-) **Dependency Details** springBootVersion = '2.0.9.RELEASE' compile("org.springframework.boot:spring-boot-starter-actuator") compile("io.micrometer:micrometer-spring-legacy:1.3.2") compile group: 'io.micrometer', name: 'micrometer-core', version: '1.3.2' **Code:** import io.micrometer.core.instrument.Clock; import io.micrometer.core.instrument.Counter; import io.micrometer.core.instrument.DistributionSummary; import io.micrometer.core.instrument.Meter; import io.micrometer.core.instrument.MeterRegistry; import io.micrometer.core.instrument.cumulative.CumulativeCounter; import io.micrometer.core.instrument.distribution.DistributionStatisticConfig; import io.micrometer.core.instrument.logging.LoggingMeterRegistry; import io.micrometer.core.instrument.logging.LoggingRegistryConfig; import lombok.RequiredArgsConstructor; import org.springframework.boot.actuate.autoconfigure.metrics.MeterRegistryCustomizer; import org.springframework.context.annotation.Bean; import org.springframework.context.annotation.Configuration; import java.time.Duration; import static java.util.Objects.nonNull; //import io.micrometer.spring.autoconfigure.MeterRegistryCustomizer; @Configuration @RequiredArgsConstructor public class MetricConfig { private final String deliverymanName; private final Clock micrometerClock; @Bean public MeterRegistryCustomizer<MeterRegistry> meterRegistryCustomizer() { return registry -> registry.config().commonTags("deliveryman", deliverymanName); } @Bean public MeterRegistry meterRegistry() { return new CustomMeterRegistry(LoggingRegistryConfig.DEFAULT, micrometerClock); } public class CustomMeterRegistry extends LoggingMeterRegistry { CustomMeterRegistry(LoggingRegistryConfig config, Clock micrometerClock) { super(config, micrometerClock); } @Override protected DistributionSummary newDistributionSummary(Meter.Id id, DistributionStatisticConfig distributionStatisticConfig, double scale) { [Truncated] at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1065) at org.springframework.beans.factory.support.ConstructorResolver.resolveAutowiredArgument(ConstructorResolver.java:819) at org.springframework.beans.factory.support.ConstructorResolver.createArgumentArray(ConstructorResolver.java:725) ... 24 more Caused by: java.lang.NullPointerException at io.micrometer.core.instrument.MeterRegistry.getMappedId(MeterRegistry.java:542) at io.micrometer.core.instrument.MeterRegistry.registerMeterIfNecessary(MeterRegistry.java:528) at io.micrometer.core.instrument.MeterRegistry.registerMeterIfNecessary(MeterRegistry.java:522) at io.micrometer.core.instrument.MeterRegistry.gauge(MeterRegistry.java:258) at io.micrometer.core.instrument.Gauge$Builder.register(Gauge.java:190) at io.micrometer.core.instrument.binder.system.ProcessorMetrics.bindTo(ProcessorMetrics.java:86) at org.springframework.boot.actuate.autoconfigure.metrics.MeterRegistryConfigurer.lambda$addBinders$1(MeterRegistryConfigurer.java:80) at java.util.ArrayList.forEach(ArrayList.java:1257) at org.springframework.boot.actuate.autoconfigure.metrics.MeterRegistryConfigurer.addBinders(MeterRegistryConfigurer.java:80) at org.springframework.boot.actuate.autoconfigure.metrics.MeterRegistryConfigurer.configure(MeterRegistryConfigurer.java:62) at org.springframework.boot.actuate.autoconfigure.metrics.MeterRegistryPostProcessor.postProcessAfterInitialization(MeterRegistryPostProcessor.java:64) at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.applyBeanPostProcessorsAfterInitialization(AbstractAutowireCapableBeanFactory.java:431) at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.initializeBean(AbstractAutowireCapableBeanFactory.java:1698) at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:573) ... 34 more Answers: username_1: @username_0 Thanks for the report! Just based on the comment on your code, you seem to try to customize `DistributionStatisticConfig`. You can do it via `MeterFilter`. See https://github.com/username_1/sample-micrometer-spring-boot/blob/a67edc8a8be9842e594383041967207dacd4dfec/src/main/java/com/username_1/sample/config/MetricsConfig.java#L39-L44 As a side node, `io.micrometer:micrometer-spring-legacy` is only for Spring Boot 1.x, so you shouldn't use it with Spring Boot 2.x. Status: Issue closed username_2: In addition to the example @username_1 has shown, if you are using Spring Boot, you can make a `MeterFilter` bean and it will be applied to MeterRegistry beans that Spring Boot is aware of, [as mentioned in the Spring Boot documentation](https://docs.spring.io/spring-boot/docs/current/reference/htmlsingle/#production-ready-metrics-per-meter-properties). So for example: ```java @Bean public MeterFilter customizeDistributionConfig() { return new MeterFilter() { @Override public DistributionStatisticConfig configure(Meter.Id id, DistributionStatisticConfig config) { return DistributionStatisticConfig.builder().expiry(Duration.ofHours(1)).build().merge(config); } }; } ``` Let us know if that doesn't solve the issue you were facing.
firebase/FirebaseUI-iOS
260657563
Title: Error Domain=FUIAuthErrorDomain Code=3 "(null)" UserInfo={FUIAuthErrorUserInfoProviderIDKey=phone} Question: username_0: I was trying to implement Firebase Phone Auth UI into my app. Added following pods to my podfile: pod 'Firebase' pod 'Firebase/Auth' pod 'FirebaseUI/Phone' I set the project on Firebase and write the code. #### Here the Code in the main ViewController: @property (strong, nonatomic) FIRAuth *auth; @property (strong, nonatomic) FUIAuth *authUIVC; self.auth = [FIRAuth auth]; self.auth.languageCode = @"it"; self.authUIVC = [FUIAuth defaultAuthUI]; self.authUIVC.delegate = self; FUIPhoneAuth *phoneAuth = [[FUIPhoneAuth alloc] initWithAuthUI:self.authUIVC]; [phoneAuth signInWithPresentingViewController:self]; -(void)authUI:(FUIAuth *)authUI didSignInWithUser:(FIRUser *)user error:(NSError *)error { if (error) { #### Log Output Error Domain=FUIAuthErrorDomain Code=3 "(null)" UserInfo {FUIAuthErrorUserInfoProviderIDKey=phone} } } If I implement the method "[[FIRPhoneAuthProvider provider] verifyPhoneNumber:@"+391234.."" it works. Please help me.<issue_closed> Status: Issue closed
duckdb/duckdb
1147317671
Title: [C++] Add BYTE_STREAM_SPLIT encoder and decoder for Parquet Files feature request Question: username_0: Dear Community, First of all thank you for the great work you have done with duckdb. We are generating a large amount of data and we them on parquet files. It is very important for us that we could use the BYTE_STREAM_SPLIT as it really helps on float value type compressing. In the [apache/arrow](https://github.com/apache/arrow) community they have already introduced this [feature](https://github.com/apache/arrow/commit/25fd97b81a65df7aca5d5f3ce482a1126bb01b83) and I was wondering if you plan to or are already working in this feature?
open-mpi/ompi
778852424
Title: OpenMPI 4.1.0 fails with pmix 4.0.0 Question: username_0: Thank you for taking the time to submit an issue! ## Background information OpenMPI 4.1.0 PMIX 4.0.0 Debian Unstable (and ubuntu) When built against external 4.0.0 , we see segfaults on simple hello world tests: ``` autopkgtest [06:32:43]: test hello2: [----------------------- [ci-359-a03b0931:01932] *** Process received signal *** [ci-359-a03b0931:01932] Signal: Segmentation fault (11) [ci-359-a03b0931:01932] Signal code: Invalid permissions (2) [ci-359-a03b0931:01932] Failing at address: 0x5632cd610c70 [ci-359-a03b0931:01932] [ 0] /lib/x86_64-linux-gnu/libc.so.6(+0x3bd00)[0x7fcc11a84d00] [ci-359-a03b0931:01932] [ 1] [0x5632cd610c70] [ci-359-a03b0931:01932] *** End of error message *** Segmentation fault autopkgtest [06:32:46]: test hello2: -----------------------] autopkgtest [06:32:46]: test hello2: - - - - - - - - - - results - - - - - - - - - - ``` For the moment we have reverted to internal pmix (3.2.2?) Answers: username_1: @username_2 @username_3 Have you guys heard about this before, perchance? @username_0 Can you send us the full build details? E.g., `./configure` line for both PMIx and Open MPI. username_2: @username_0 I looked into this and found the issue - a configure code that is looking for `major version == 3` instead of `major version >= 3`. Should be a trivial change - @username_3 said he would fix it later today. Thanks for the report and sorry for the problem. username_3: Thanks for letting us know. I filed fixes for the two latest release streams: * `v4.0.x` branch in PR #8338 * `v4.1.x` branch in PR #8337 * This will need to go back to PRRTE as well. I'll update this comment with the PR link once I file it. username_0: Thanks. I can now build openmpi 4.1.0 with the external pmix 4.0 with this patch, but still have issues due to a bug in pix 4.0.0 ( https://github.com/open-mpi/ompi/issues/8323 ). Because of this, openmpi in Debian Unstable is currently configured to use the internal pmix 3.* but appears to be fragile in the presence of pmix 4.0 on the system ( mpich and slurm are also configured to use the external pmix). (pmix is built as a shared library) username_2: @username_0 Can you please expand on this statement? Are you just referring to https://github.com/openpmix/openpmix/issues/2003 or is there some other problem? username_0: Ok, with an apparent fix i've now reverted to openmpi 4.1.0 usng external pmix 4.0.0 on Debian Unstable. By "fragile" I mean if openmpi is built to use internal pmix, while another copy of pmix (both libpmix.so.2) exists on the system, albeit with different paths, bugs are likely to ensue - picking up the wrong library at build or runtime. I'm not sure that openmpi/openpmix should be doing anything (more) about this - I think its a Debian issue to ensure that only one libpmix.so.2 exists on the machine at a time. Status: Issue closed username_1: Merged into the v4.1.x branch -- fixed!
PlaidWeb/Publ
513649036
Title: Add `WWW-Authenticate:` headers to access-upgradable pages Question: username_0: If there is no active user, and an attempt is made to access an `entry` or `view` object where a user could potentially see more content, add an authentication request header to the response The actual header might make sense to add in an `@app.after_request` function. Answers: username_0: storing the flag as `flask.g.needs_token` and any error message as `flask.g.token_error` Status: Issue closed
DIYgod/RSSHub
476624561
Title: 微信公众号获取失败,但其他的都没问题。 Question: username_0: ![image](https://user-images.githubusercontent.com/15061002/62433791-88bbc680-b768-11e9-816f-417812c52609.png) ![image](https://user-images.githubusercontent.com/15061002/62433797-8f4a3e00-b768-11e9-93a9-b6117fcbe54b.png) Answers: username_1: 有梯子么?? username_2: 提 issue 请遵循模板 username_3: https://docs.rsshub.app/install/#%E4%BB%A3%E7%90%86%E9%85%8D%E7%BD%AE Status: Issue closed
php1301/DoAnReactJS
542433350
Title: displayName alert (on localhost not on hosting) Question: username_0: displayName alert (not affect to performance when running build on host) Khi login thành công sẽ cảnh báo displayName cannot find -> thật sự không ảnh hưởng lắm đến luồng đi của ứng dụng khi thực hiện các chức năng thao tác của trang về sau như mua vé, yêu thích, lịch sử phim, lịch sử vé,... chỉ cần bấm nút close là được ![image](/uploads/9c028f95c1fc019e506298d9931ce516/image.png)
MicrosoftDocs/azure-docs
852975693
Title: NSG flow logs are not available Question: username_0: This log doesn't appear to be available any more: NetworkSecurityGroupFlowEvent I assume that flow logs have to use Network Watcher and can't be directly enabled through diagnostic settings. Please confirm and remove that line from the document if that is correct. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 847615aa-75d5-01d1-c64b-8ce385c14397 * Version Independent ID: fad240b7-b122-7f3a-d185-43ad9747b72a * Content: [Azure Monitor Resource Logs supported services and categories - Azure Monitor](https://docs.microsoft.com/en-us/azure/azure-monitor/essentials/resource-logs-categories#microsoftnetworkvirtualnetworkgateways) * Content Source: [articles/azure-monitor/essentials/resource-logs-categories.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/azure-monitor/essentials/resource-logs-categories.md) * Service: **azure-monitor** * Sub-service: **essentials** * GitHub Login: @username_1 * Microsoft Alias: **robb** Answers: username_1: I have no indication from my backend systems reports that this log has been removed. According to my information, it still can be routed through diagnostic settings. What makes you think it's been removed @username_0 ? Did it disappear from a UI screen? username_0: Yep, it's not available in the Portal. (on this screen or the detailed setting screen) ![image](https://user-images.githubusercontent.com/34466884/113958335-c2629180-97d5-11eb-8bdb-ea1e14689413.png) username_0: I thought flow logs had to be collected from Network Watcher. There is a separate Portal blade for configuring those logs. This doc on the logs for NSGs only covers the two that are visible. https://docs.microsoft.com/en-us/azure/virtual-network/virtual-network-nsg-manage-log username_1: #label:"CreateAzMonWIFeb2021" username_2: Thank you for your contribution to the Azure Monitor documentation. We have created a workitem for this issue in our internal backlog database. We will prioritize it and make any necessary changes as we are able. #please-close Status: Issue closed
streamnative/pulsar-flink
1039112651
Title: [BUG] job canceled,Failed to cancel the pulsar, Failed to remove cursor or TopicRange Question: username_0: **Describe the bug** flink version 1.13.1 pulsar version 2.5.0 pulsar-flink-connector_2.11 version 1.13.1.0 run with flink standalone cluster flink job is a source connector job, when one of taskmanager restart cause oom, the job canceled,but throw one ERROR,: ``` ERROR org.apache.flink.streaming.connectors.pulsar.FlinkPulsarSource [] - Failed to cancel the pulsar Fetcher java.lang.RuntimeException: Failed to remove cursor or TopicRange[topic=persistent://xxxxx/xxxx/xxxx-partition-2, key-range=SerializableRange{range=[0,65535]}] at org.apache.flink.steraming.connectors.pulsar.inernal.PulsarMetadataReader.removeCursor(PulsarMetadataReader.java:278) ...... Caused by: org.apache.pulsar.client.admin.PUlsarAdminException:java.lang.InterruptedException ......6 more ``` when start the flink job again ,i got this ``` org.apache.pulsar.clint.api.PulsarClientException$ConsumerBusyException: Exclusive consumer is already connected ``` **To Reproduce** Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** If applicable, add screenshots to help explain your problem. **Additional context** Add any other context about the problem here. Answers: username_1: Can you upgrade your pulsar version? The connector 1.13.1.0 is designed to be used on Pulsar 2.7.0 or above. username_1: No response from the reporter, close this. @username_0 You can PING me if you still suck with this issue. Status: Issue closed
mlhubber/mlhub
426315473
Title: Support git repositories beyond github including bitbucket and gitlab Question: username_0: Currently only github repositories are supported as source of MLHub packages. Any git repository should be supported, including bitbucket and gitlab and independently run git repositories. Answers: username_1: A curated list of source code hosting facilities can be found at [Comparison of source-code-hosting facilities -- Wikipedia](https://en.wikipedia.org/wiki/Comparison_of_source-code-hosting_facilities) username_1: I collected a list of URL patterns and APIs for GitHub, GitLab and Bitbucket that are useful: https://github.com/username_1/tips/blob/master/github/compose-github-links.md username_1: - [X] Support for GitLab - [ ] Support for Bitbucket - [ ] Generalization for other git repos - [ ] Support for pull/merge request - [ ] GitLab - [ ] Bitbucket username_0: For now closing this - mostly implemeneted and working for gitlab and bitbucket. What was the pull/merge issue? Create new issue if required. Status: Issue closed
OriginTrail/ot-node
1143220502
Title: V6 Performance Documentation and potential mem-leak Question: username_0: ## Issue description This is intended as a feedback to the developers for the performance, as well as a potential bug I have found. **Documentation** I'm running the following setup on my Amazon Cloud: <img width="532" alt="image" src="https://user-images.githubusercontent.com/35450966/154702278-7e8c2e0f-78dd-4458-804d-e99b6b9a2070.png"> My AWS Lightsail Instance with the 4 nodes uses an UBUNTU 20.4 TLS instance with the given size <img width="158" alt="image" src="https://user-images.githubusercontent.com/35450966/154702655-5d6c4081-6ebd-4003-809c-24f941481e2c.png"> Directly after starting my OT-Nodes, this is the memory consumption of the four nodes running on the machine Attributes comma separated are: Process ID, user , %mem , command <img width="975" alt="image" src="https://user-images.githubusercontent.com/35450966/154702853-9a0ac6e8-9985-41d6-8547-bd4b397c56c3.png"> Each node takes roughly 153MB of RAM ( lets take 7.5% of 2048 MB) **Problem** I've been hammering my 4 nodes in 2 waves for roughly two hours. I don't have memory monitoring activated, so I cannot show a diagram of them RAM usage yet. After 2 hours of hammering jobs to those four nodes, my VM crashed. I've analyzed the behaviour and saw the following: **CPU** <img width="674" alt="image" src="https://user-images.githubusercontent.com/35450966/154704480-8062425c-5b3a-49f2-b3ae-3c902292ae32.png"> It looks like the CPU usage has risen within 10 mins to 80%, within the next 10 mins to 100% and then crash the EC2 instance. **RAM** As I said, I don't have mem usage graph yet, as lightsail is not as good integrated into AWS yet, but I can see, that the memory consumption of each node rises from the starting 6-7% up to 20% within a timespan of a few hours. This is the screenshot of the four nodes running for like 2 hours of constant publishing. Process ID, user , %mem , command <img width="1030" alt="image" src="https://user-images.githubusercontent.com/35450966/154705899-895c81e5-d130-44d0-ac50-eef7272fa4f5.png"> Since neither database, nor blazegraph are running on the node, it looks like there is some memory problem within the OT-Node at the moment. Do I analyze that correctly? Do you have another idea, what it could be? If this is just regular behaviour, because the consumption is indeed intended to double across the usage then nevermind this ticket and just take it as a performance documentation :) ## Expected behavior The OT-Node RAM usage does not triple/quad in 2 hours, but stay constant ## Actual behavior Triples until my tiny machine runs OOM/CPU ## Steps to reproduce the problem 1. Realize the architecture as mentioned above 2. Run 4 Nodes 3. Publish jobs each few seconds for a limited amount of time ## Specifications - Node version: Latest v6.0.0-beta.1.23 - Platform: Ubuntu 20.4 TLS - Node wallet: 0x03405Ce6eD71642EA50b0F6073c113f6Ea7149B6 - Node libp2p identity: Many different, do you need them? ## Contact details - Email: hansi1337 at gmail dot com - Discord: angrymob ## Error logs ## Disclaimer Please be aware that the issue reported on a public repository allows everyone to see your node logs, node details, and contact details. If you have any sensitive information, feel free to share it by sending an email to [<EMAIL>](<EMAIL>). Answers: username_0: The issue still persists with the latest 1.29 testnet version, thus I will migrate to the 20$ instance on AWS and check whether the memory leak is still persisting or will memory consumption will reach a plateau.
ionic-team/ionic-cli
348787473
Title: None Question: username_0: This error is totally normal when you reload the page: <img width="952" alt="image" src="https://user-images.githubusercontent.com/236501/43850991-1b237928-9aff-11e8-9bb2-387921e3b719.png"> The webpack-dev-server has to re-establish the connection because it gets killed when the page refreshes. Answers: username_0: @username_1 As for your other issue about the page being blank, I'm not able to reproduce it. username_1: @username_0 ya, for the 'ionic serve' issue, you can try to reload multiple times, you will notice sometimes you saw the browser show the default text, sometimes show totally white blank page username_2: I Had this problem today and fix it with this solution The solution was to add the following line in angular.json under architect>serve>options: "disableHostCheck": true i hope it will help you
gitskarios/Gitskarios
119390094
Title: Android notifications stays disabled Question: username_0: Heya, love the app. Bit since a few versions ago, the notifications option is no longer working for me. Under Android's Accounts, gitskarios is listed, but has no option to enable/disable. And when I enable notifications in the app itself, the next time I launch Gitskarios, the toggle is just set to disabled again. So I don't get the notifications anymore.. Answers: username_1: Happens to me as well, please fix username_2: Going to fix it today username_1: Thanks man, awesome work on the app 👍 username_2: Fixed Status: Issue closed username_0: Thank you! Marvellous. username_1: :+1:
elastic/observability-docs
1086040523
Title: [Request] Question: username_0: ## Description Starting 8.0, the alert workflow status filter (open / acknowledge / close) is replaced with the alert status filter (active / resolved). Documentation needs to be updated accordingly. The alert workflow status is a manual status management scheme requiring users to update and manage the status of each alert. In parallel, each Observability alert also has an active / resolved status that is automatically assigned and updated through its lifecycle. Separately, the Cases workflow also offers a status management scheme, along with further collaboration and case management. To simplify and streamline status management, alert workflow status filter in the Alerts view UI is being replaced with the alert status filter. Users wanting to do status and incident management are encouraged to peruse the Cases UI and workflow. ## Collaboration - The product team will provide the initial content and the docs team will edit / review **Contact Person:** @fkanout @username_0 ## Suggested Target Release v8.0 ## Stakeholders @jasonrhodes @username_0 Answers: username_0: Related to https://github.com/elastic/kibana/issues/117686 Status: Issue closed
smourier/DirectN
1058327899
Title: DWriteCore interfaces Question: username_0: Hi, It would be possible to add C# interfaces of new DWriteCore API? Answers: username_1: Hi, I guess so, I've not looked at it yet, the Windows App SDK has just been release in version 1.0. In fact, there may be other API from it to add. I'll check that when I have some time. Status: Issue closed username_1: Just a quick note: I've only seen a new function and a new enum value, all interfaces definitions seem the same. username_0: Docs says dwrite_core.h define a symbol that “enables” some interfaces in dwrite3.h, I didn’t dig into yet
NumEconCopenhagen/projects-2020-1001numpyarrays
615434746
Title: Feedback on model project Question: username_0: 1. The best part of the project was: (explain what and why) I think that solow_graph_3 with the widgets is quite dirty. You use some sexy coding tricks using time as a factor, true eye porn. There might be a wet spot in my pants after I've played a bit with those widgets on that graph 'watch file at the buttom'. 2. The hardest part of the project to understand was: (explain what) It's actually hard to tell what solow_graph_x shows from a theoretical point of view. A lot of the solow graph is Solow diagrams, I'm a bit confused what you wanted to show, is it a steady state? Is it how the virus develop through time? or just a 'solow_graph'? 3. This part of the project could be better documented: (explain what) You could pretty much everywhere in your project use a bit more text. What are you calculation/showing and why. 4. An idea for an improvement/clarification could be: (explain what and why) In part 1: Choose carefully your graphs, honestly I think you have a bit too many that just show the same, they are just without and widgets. Just as well I think it could be nice for the reader that you argued a bit around the graphs what do they show? steady_state_capital_func what is the idea of having that one, I have no clue? In part 2 (with the virus): you have introduced some extra equations in your model, write a bit more text, again the graph twice.(even though it's a nice one). Again a bit more explanation or arguing would do good. 5. An idea for an extension could be: (explain what and why): text/explanations. try and be short and simple (less is more). Thanks you, for giving me my daily graph modeling porn. <img width="352" alt="Screenshot 2020-05-10 at 19 09 35" src="https://user-images.githubusercontent.com/61465597/81505733-cd751f00-92f1-11ea-80ea-74dff1419403.png">
pcurrier/KoboToolbox-GoogleApps-Scripts
1127873916
Title: TypeError: Cannot read property 'init' of undefined Question: username_0: I have followed all the steps for getting data from kobo toolbox to google sheet. However when I click on (Import Kobo Toolbox Data into Sheet), I get the above error. Any assistance will be highly appreciated.
storybookjs/storybook
1032365884
Title: React 17 + Webpack 5: ArgsTable doesn't pick up component's propTypes Question: username_0: **To Reproduce** [This repo](https://github.com/username_0/sb-repro-argstable) captures the problem. It is a basic webpack5 setup (`npx sb@latest init --builder webpack5`), with the `Button.stories.mdx` file added to show that `<ArgsTable>` doesn't pick up on `<Button>`s prop types. **System** ``` Environment Info: System: OS: macOS 11.6 CPU: (8) arm64 Apple M1 Binaries: Node: 16.5.0 - ~/.nvm/versions/node/v16.5.0/bin/node Yarn: 1.22.10 - ~/.nvm/versions/node/v16.5.0/bin/yarn npm: 7.24.0 - ~/.nvm/versions/node/v16.5.0/bin/npm Browsers: Chrome: 95.0.4638.54 Firefox: 93.0 Safari: 15.0 npmPackages: @storybook/addon-actions: ^6.3.12 => 6.3.12 @storybook/addon-essentials: ^6.3.12 => 6.3.12 @storybook/addon-links: ^6.3.12 => 6.3.12 @storybook/builder-webpack5: ^6.3.12 => 6.3.12 @storybook/manager-webpack5: ^6.3.12 => 6.3.12 @storybook/react: ^6.3.12 => 6.3.12 ```
everyday-as/gmodstore-issues
607883579
Title: Creating/editing/deleting labels throw a 500 Question: username_0: ## Expected Behavior Being able to create/edit/delete a label ## Actual Behavior Nothing happens, 500 is thrown ## Steps to Reproduce the Problem 1. https://www.gmodstore.com/moderator/tickets/tags/edit 2. Create, edit or delete a label 3. Watch as the network tab shows a 500 ## Specifications - GmodStore version (see footer): 5.2.8 - Example url: https://www.gmodstore.com/moderator/tickets/tags/edit - Browser: Google Chrome - Event ID: N/A Answers: username_0: Can't replicate Status: Issue closed
guswns1659/JuraJura
699089806
Title: [2기] 2차 주간 목표 Question: username_0: ### 로그인 OAuth, JWT 구현 - [ ] OAuth를 이용한 로그인 - [ ] JWT 구현 - [ ] AOP를 이용한 로그인 검증 Answers: username_0: ## 회고 ### 잘한 점 ### 잘못한 점 - 이번 주 내내 너무 놀기만 했다. 다시 마음을 다잡아야 할 필요가 있다 - 알고리즘 문제 푸는데 너무 오래 걸린다. 시간을 조금 더 정해놓고 하는게 좋을 것 같다 username_1: ### 피드백 고생하셨습니다. 사실 로그인이 쉬워보이지만 어려운 분야인듯 합니다. 괜히 보안이 높은 연봉을 받는게 아닌가 싶습니다. 한단계 씩 나아가면 로그인에 대해서 이해할 거라 생각합니다. 저는 아예 4주동안을 로그인만 공부했습니다 하하하 알고리즘은 너무 성급하게 풀지 않아도 될 듯합니다. 긴 시간 보다는 최대 1시간을 잡고 풀다가 안되면 답변 보시고 다시 풀어보는건 어떨까요? Status: Issue closed
sympy/sympy
478617413
Title: Excess Tuples in FunctionCall AST Node Question: username_0: 'fun2((2, 3), (2, 3))' ``` The code printers print only one set of arguments if called for a declared object using p. But, if we explicitly assign the same node to some other variable and use code printers on that, it prints multiple sets of arguments Answers: username_1: @username_0 yes, that's definitely a bug. Hopefully fixed in gh-17384. Status: Issue closed
immersive-web/webxr
922126667
Title: Provide statistics to help guide performance Question: username_0: One of the hardest parts of WebXR is how to tune for good performance. It is hard for authors to gauge how much processing power is available to them so they end up optimizing their code for the device they have on hand. We should give them the tools they need so they (or framework authors) need so they can dynamically change parameters of their experience. We've added support for framebuffer scaling, ffr and variable framerate but unless an author produces code that can't keep up producing frames they can't figure out how much extra processing the system has. I looked at the [Compute pressure](https://oyiptong.github.io/compute-pressure/) API but I'm unsure if that will help since it doesn't seem applicable to VR Headsets. As a browser developer, what can we provide authors to help them out in this area? Total time spend in JS and rendering? Total idle time between Raf calls? Answers: username_1: I believe current browser Performance audit provides decent information on how much time is spent on JS and on rendering, also gives an idea on idle between frames. With nature of various platforms, they might have other processes out there, which for developer is not known. Means "available" idle, might be actually used by other processes. VRAM have been always a very major performance drivers. But this perhaps is more WebGL related. One thing definitely would be great to know, is how much time is spent by underlying WebXR systems to profile VR/AR, especially when using additional features. For example: optical hand tracking, plane detection, light estimation, etc. It would be useful to know of how much is spent by WebXR systems, so developer can take it into an account when using available hardware budget. username_0: At least on Quest, those run on different dedicated cores so they shouldn't interfere. Idle time on the between Raf calls is something that we can report and our OS also reports load on the CPU and GPU. If people are interested, I could expose these as an experiment to see if they can be used to tune performance. I'm also wondering what authors do with WebGL experiences because the problems seem similar. username_1: In PlayCanvas, we expose various profiling tools: [ministats](https://developer.playcanvas.com/en/user-manual/optimization/mini-stats/), [launcher profiler](https://developer.playcanvas.com/en/user-manual/optimization/profiler/), some [internal logging](https://github.com/playcanvas/engine/blob/master/src/framework/stats.js), to expose: 1. Render, update, shadows, times 2. VRAM (internal counters) usage, and split by types of resources: textures, buffers, framebuffers 3. Shader compilation times (as they are sync in WebGL by default). 4. Draw calls - this is important, as knowing your budget for target platforms helps to develop for it, even when developing not on the target platform. 5. Overdraw - we don't report that but have done custom profilers for specific projects, to identify overdraw for fragment shaders, this is very useful when optimizing with heavy fragment shaders. 6. More engine-related things: number of shaders, materials, shadows rendering, UI rendering, culling time, geometry batcher, etc. Things I personally believe is key for successful (in general) WebGL content for a non-specifically targeted audience: 1. **Loading times** - faster is better, download only what is needed for specific application state, clever bundling of assets, atlases for small textures (UI, sprites, etc.). 2. **UX** - being simple and accessible. Often app looks great, and can even load fast, but if UX is bad - people simply go away. Fewer clicks/touch/interactions, and more value is better. Hiding gameplay behind menus - is a bad practice on the web. 3. **Garbage collection** - this is often missed, but in the web due to automatic GC, allocating resources every frame, will lead to GC stalls, which can get out of hand quickly, so coders have to re-use resources, use pools for objects, and do not allocate/destroy anything as possible. 4. **Performance** - this goes to major few things: draw calls, VRAM usage, GC, shader complexity (can either not fit in some limited platforms like iOS, or just be slow due to complexity). username_0: Thank you! How do you know that the rendering takes too long? Do you look by knowing the frame rate and seeing that the Raf starts slowing down? I'm specifically looking for a way to optimize render performance for experiences that already apply the best practices you listed. username_2: For background, the [dynamic viewport scaling](https://immersive-web.github.io/webxr/explainer.html#dynamic-viewport-scaling) feature lets UAs provide a [recommendedViewportScale](https://immersive-web.github.io/webxr/#dom-xrview-recommendedviewportscale) value as a hint to applications. In the Chromium implementation, that's based on an internal [estimate of GPU utilization](https://source.chromium.org/chromium/chromium/src/+/main:device/vr/public/mojom/vr_service.mojom;l=872;drc=cb50206712c40d5b83ce2c96fdf97a5969fce5ae): ``` // If nonzero, an estimate of how much of the available render time budget // was used for GPU rendering for the most recent measured frame. A value // above 1.0 means that the application is dropping frames due to GPU load, // and a value well below 1.0 means that GPU utilization is low. This is // intended to be used as input for renderer-side adaptive viewport sizing. // A value of zero means the ratio is unknown and must not be used. ``` Would it be useful to provide such a value directly to applications? It would need to be appropriately quantized to ensure that it can't be abused to extract fine-grained timing information, and some care is needed to make the value useful across platforms if it's based on inexact heuristics. Maybe it would be more appropriate to use enums (UTILIZATION VERY_LOW/LOW/NOMINAL/HIGH/VERY_HIGH) instead of numeric values? Applications can to some extent detect and deal with CPU bottlenecks by observing rAF timing, but that doesn't help if overall performance is limited by GPU performance, and as far as I know there's currently no good way to get this information apart from developer tools applied to a specific device. username_0: Indeed. I've observed that developers are hand tuning their applications for a particular device and then the experience is sub-par on other devices. username_2: Sorry, I was being sloppy with phrasing here. Yes, if an application is bottlenecked by excessive GPU load, the system will drop frames, and the app can detect that based on the interval between successive rAF calls. However, if the GPU is underutilized, there's no clear way to detect this other than trying to selectively increase GPU workload until the system starts dropping frames. Conversely, if the bottleneck is CPU load, the app shouldn't try to use tuning methods that try to reduce GPU work, for example reducing the pixel count by decreasing the viewport scale wouldn't have any benefit. username_0: True, although an author could calculate this by timing the rAF call. username_0: /tpac discuss how to provide feedback to guide performance username_3: Would this be implemented as calls that allow a developer to tune the system or an event (or more) to report when certain thresholds are exceeded? The first case is better to building the system but may cause some overhead if used live. The second case (events) can be great as long as the thresholds are correctly set.
smahbod2014/YGO
189025795
Title: Implement proper combat logic Question: username_0: - Deal the difference in damage between two attack position monster - Implement attacking defense position monsters - Implement suiciding into another attack position monster - Implement stalemates between two equal power attack position monstesr<issue_closed> Status: Issue closed
XX-net/XX-Net
203558114
Title: XX-NET配置AppID无效 Question: username_0: # 基于最新稳定版:3.2.8 # 状况说明 昨天提示超额,去配置了几个 AppID,配置完成了,今天一看是无效的,而且今天就去了油管看了2个视频,不到1G,有5个以前了AppID超额了 # 状态页 ![default](https://cloud.githubusercontent.com/assets/16161582/22360921/b0cae4fc-e48f-11e6-9f2f-8223321cbcb1.png) ![default](https://cloud.githubusercontent.com/assets/16161582/22360930/c67c14ec-e48f-11e6-86f9-94c3fdbb8588.png) Answers: username_1: https://github.com/XX-net/XX-Net/issues/4720 Status: Issue closed
dencold/attainment-web
288443703
Title: Completed projects still show up in project listing Question: username_0: When a project is completed it shouldn't show up in the general project list. We'll want to filter those out. However, we'll probably want to have an "archived projects" list somewhere so we can go back and bring projects out of an archived state if we need to.<issue_closed> Status: Issue closed
broadinstitute/cromwell
230079116
Title: Revamp `object` Question: username_0: @username_2 commented on [Fri May 05 2017](https://github.com/broadinstitute/wdl/issues/109) **Needs Refinement** We want to be able to define types for the values of objects. One suggestion was something like the following (note `struct` is using as a possible replacement for `object`, see below): struct MyType { o_f: File x: Array[String] } MyType foo = read_object(...) It will coerce to the types it expects and if it can't that's a failure. Open questions: - Do we make a new construct (e.g. `struct` above), or replace objects - If replace, who (if anyone) is currently using `object` - What's the right syntax, regardless of the name of the construct. This needs focus grouping. Answers: username_1: @username_2 Is there an internal use case for this on your end or a community use case? I have yet to come across an instance where one of the other more concrete types will not suffice. username_2: @username_1 This came from @username_3 - I don't believe we've seen any requests/use cases outside of that, but that could also just be because it's so unwieldy now that people aren't bothering at the moment. For history's sake, `object` was one of the tidbits that we added in the earliest days of WDL to reflect some of our personal experience with previous systems, but since they werne't really being used the whole thing became a dusty corner. username_3: Patrick -- thanks for looking at this. I'd like to use Objects to group together related items so we can pass them as a single sample to processes. This enables things like batch or tumor/normal calling where you have multiple samples together and want to bundle each of items with each other. I've got a mock up of how I'd use it in a real somatic workflow here: https://github.com/bcbio/test_bcbio_cwl/blob/master/somatic/somatic-wdl/main_somatic.wdl The issue with current objects is that the items are typed so platforms can't identify files for localizing them. Happy to provide more details if it would help. username_4: I am currently experimenting with wdl4s object support in [dxWDL](https://github.com/dnanexus-rnd/dxWDL). The first use case is for representing JSON objects; it turns out that the syntax is very similar. I don't know if that is a coincidence. Presently, I am having trouble with the WDL object typing. For example, the workflow below: ``` workflow wf_complex { Object z = {"a": 3, "b": 1} output { Int sum = z.a + z.b } } ``` When evaluated with Cromwell ``` java -jar cromwell-29.jar run wf_complex.wdl ``` reports: ``` "outputs": { "wf_complex.sum": 31 } ``` However, I would think `a` and `b` are integers, and the addition result should be `4`. Clearly, Cromwell thinks these are strings, and the addition concatenates `3` to `1` instead. Had `z` been a JSON object, the values would have been interpreted as integers, as intended. 1) Can the experts shed light on the situation? 2) How do I create a WDL object literal? username_1: @Chapman I understand your issue. I encountered the Same thing when doing multi-sample workflows a while ago (since then I have only been writing single samples). I think I got around it by using a combination or multiple arrays and typed Maps. Not necessarily ideal, especially since at the time there was no `range()` function. For a use case like that, it seems like a `struct` type makes total sense. I feel like `Object` is mainly useful as an internal type. @username_4, I'm not 100% certain and @username_2 can correct me if I am wrong, but I believe I read somewhere that all members in an object are interpreted as strings. Therefore what you are seeing is string concatenation first, and then a type coercion to an int. If you have a statically typed object, I would suggest using the `Map` type. username_4: @username_1 thanks, that clarifies this mysterious object behavior. For what it is worth, I like structs better, because they are statically typed, and are easier to program with. username_2: @username_4 It's definitely not a coincidence that `object` is like json objects. while we didn't have a real use case when we first designed WDL, the handwaving was that it was a json object-like thing @username_3 @username_1 @username_4 @username_0 @vdauwera - considering `object` as-is isn't working for anyone (and is likely implemented incorrectly in Cromwell/WDL4s anyways) and we seem to be having a moment here, this seems like a good opportunity to rethink what this construct *should* be. It doesn't seem like we're far from a consensus anyways. Any suggestions on best scheme for collaborating/brainstorming this (assuming ppl are willing) .... here on github, shared google doc, email? username_4: @username_2 It would be great to collaborate and come up with a good solution. In my opinion, we can continue the discussion here on github. That is my two cents. username_1: @username_2 Github seems like a good enough place for now! At least for the initial discussion of what the the best type of construct would be. It would be good to identify all of the current uses of the `Object` type, both internally to cromwell / wdl, and externally in users workflows. Any replacement should address all of the current uses while adding the functionality we need. username_3: Jeff, Ohad, Patrick and all; Discussing on GitHub works great for me. From my side having equivalent functionality to CWL records (http://www.commonwl.org/v1.0/CommandLineTool.html#CommandInputRecordSchema) would be perfect for me. Essentially it's a set of typed key/value pairs, so the current objects plus typing. CWL allows nested records but having a flat structure is fine with me if that's easier for implementation and support. Happy to provide any more details that would be helpful. Thanks again for looking at this. username_5: +1 on this :) I think the idea of `struct` would be very useful to potentially enhance/replace the Object type. May want the order of the type and variable name to be consistent with other WDL, e.g. instead of: ``` struct MyType { o_f: File x: Array[String] } ``` ``` struct MyType { File of Array[String] x } ``` IMHO, the word `struct` is nice since it pays homage to C and it seems like a fairly nice correspondence. May want to consider `tuple` as well. Finally, I think a nice benefit of this is that it could have nice correspondence with records in CWL and thus potentially a nice representation in the WOM to handle both. username_2: @username_1 re use cases I wouldn't be surprised if no one internally is using `object` but I'll try to investigate username_2: Although now that I've said that I know that we *do* have a use case where something like this is being requested. They want a typed set of key/value pairs, but the thing that they really want is to be able to define some boundaries (e.g. "Foo" is a number between 1 and 10) and to have the static analysis fail to validate the workflow if one of these are a workflow input and the values are wrong. Now that I type that out, having refinement types in WDL seems like a bad path to be going down. I should verify that's *really* what they want or if I read too much into an example they gave. username_1: @username_2 if the use case is really intended for validation criteria on objects which the user sets, I feel the same as you, that this is an abstraction that should not be handled withing wdl/cromwell. While I understand the use case (we also have toyed around with the idea of this as a feature request) it adds unnecessary boundaries to object types that should be handled at the level of execution and not job submition. I think what might be of use in these instances, for users (like myself) is using the parameter meta more efficiently to define in writing what constitutes valid entries. Going back to the idea of objects as typed key Value pairs, I still think this is a valid idea, that has real use cases and purposes. In many cases data must be paired with other corresponding datasets and values. In a scatter operation having these types of structured objects would greatly simplify how we can group data together username_6: This is in Cromwell now Status: Issue closed
dantecatalfamo/agenda-html
1088497224
Title: Suggestion: Use non-terminal Emacs Question: username_0: Hi! Actually I do what you have done here in a slightly different way. I export my agenda in non-terminal Emacs to an SVG file. So it contains all the fancy category icons. Then I use the exported SVG file as my homepage. I use [this](https://github.com/username_0/dotfiles/blob/master/doom/config.el#L791) function to export it and here is the[ final result](https://www.reddit.com/r/emacs/comments/moc6dw/my_orgagenda_on_my_phone/). My method has the advantage of having icons. But the method used here, using tmux, has its own advantages. I mean I could not find a way to do my export headless and create a cron job for it. The idea here to use tmux and a shell script to export the agenda and then quit Emacs is really nice. Sorry that this is not a real rigorous suggestion. I thought maybe bringing it up here could lead to some new ideas. Thanks Answers: username_0: I did this using `xvfb-run` here: https://github.com/username_0/dotfiles/blob/master/bin/sync-agenda-svg
CGATOxford/UMI-tools
499384076
Title: Computation of PCR duplication per library - Question: username_0: Hi, I am using umitools in a pipeline for total RNASeq libraries. When deduplicating i have the following code: `umi_tools dedup -I "${align_out_dir}"/Aligned.sortedByCoord.out.bam -S "${outDir}"/umi_deduplicated.bam --multimapping-detection-method=NH --output-stats="${outDir}"/deduplicated --paired --log="${outDir}"/deduplication.log` My goal is to compute the level of PCR duplication per library. When the stat files are produced from **dedup** i am analysing the files "deduplicated_per_umi.tsv" and when i take a look at the sums of the columns **total_counts_pre** and **total_counts_post** i get the exact same values. See here: ![image](https://user-images.githubusercontent.com/6638031/65765204-ef85ac80-e127-11e9-8b56-994920373e5d.png) Does this mean that there is actually no deduplication done in my data or am I doing something wrong ? Thanks for the help Answers: username_1: Hi @username_0 - Sorry for the complete absence of a reply to your question. Did you get to the bottom of the above. From the stat file, it does indeed appear that no duplicates were identified which is odd. How did you extract the UMIs and do they look reasonable? username_1: @username_0 I'm closing this issue due to inactivity. For related problems, please open a new issue and reference this one. Status: Issue closed
ContinuumIO/anaconda-issues
284883972
Title: Navigator Error Question: username_0: ## Main error An unexpected error occurred on Navigator start-up<br>psutil.AccessDenied (pid=769) ## Traceback ``` Traceback (most recent call last): File "/Users/huyhn/anaconda/lib/python3.6/site-packages/psutil/_psosx.py", line 293, in wrapper return fun(self, *args, **kwargs) File "/Users/huyhn/anaconda/lib/python3.6/site-packages/psutil/_psosx.py", line 356, in cmdline return cext.proc_cmdline(self.pid) PermissionError: [Errno 13] Permission denied During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/Users/huyhn/anaconda/lib/python3.6/site-packages/anaconda_navigator/exceptions.py", line 75, in exception_handler return_value = func(*args, **kwargs) File "/Users/huyhn/anaconda/lib/python3.6/site-packages/anaconda_navigator/app/start.py", line 108, in start_app if misc.load_pid() is None: # A stale lock might be around File "/Users/huyhn/anaconda/lib/python3.6/site-packages/anaconda_navigator/utils/misc.py", line 384, in load_pid cmds = process.cmdline() File "/Users/huyhn/anaconda/lib/python3.6/site-packages/psutil/__init__.py", line 701, in cmdline return self._proc.cmdline() File "/Users/huyhn/anaconda/lib/python3.6/site-packages/psutil/_psosx.py", line 306, in wrapper raise AccessDenied(self.pid, self._name) psutil.AccessDenied: psutil.AccessDenied (pid=769) ``` ## System information ``` python: 3.6.2 language: en os: Darwin;17.3.0;Darwin Kernel Version 17.3.0: Thu Nov 9 18:09:22 PST 2017; root:xnu-4570.31.3~1/RELEASE_X86_64;x86_64;i386 version: 1.6.2 platform: osx-64 qt: 5.6.2 pyqt: 5.6.0 conda: 4.3.27 ``` Answers: username_1: **See issue #1984 for more information on how to fix this.** --- Closing as duplicate of #1984 --- Please remember to update to the latest version of Navigator to include the latest fixes. Open a terminal (on Linux or Mac) or the Anaconda Command Prompt (on windows) and type: ``` $ conda update anaconda-navigator $ conda update navigator-updater ``` Status: Issue closed
google/fswalker
456240546
Title: Windows build Question: username_0: 👋 this looks like a great tool! Surely, it's probably mostly used on Linux, but is there a reason (apart being not in focus) that it doesn't build on Windows? Walking across the devices and the comparison of the platform-dependent fields might need another implementation. Is this desirable, or are there insurmountable hurdles? ``` %GOPATH%\src\github.com\google\fswalker (master -> origin) λ go build # github.com/google/fswalker .\walker.go:129:30: undefined: syscall.Stat_t .\walker.go:251:36: undefined: syscall.Stat_t %GOPATH%\src\github.com\google\fswalker (master -> origin) λ go version go version go1.12.5 windows/amd64 ``` Answers: username_1: We've actually never tried this on Windows. The error posted here should be relatively easy to fix. Maybe we could hook this up with appveyor or something to run regular Windows builds as well. username_2: Any updates regarding windows version? username_1: Sorry, I won't have time to work on this in the near future. Releasing for someone else to pick up.
jupyterlab/jupyterlab
411053990
Title: No way to find out how to select next occurence (multiple cursors, ie CMD+D) Question: username_0: **Describe the bug** No way to select next occurence (you can see how this behavior works in the first animation on https://www.sublimetext.com/). **Expected behavior** Key map or command palette mentions it, let's me rebind it, there is documentation around editor commands somewhere. **Desktop (please complete the following information):** - OS: macOS Mojave - Browser Chrome 71 - JupyterLab 0.35.4 **Additional context** This is an essential editing functionality, which is supported by the underlying CodeMirror editor (afair). It's included in Sublime Text, Atom, VS Code, even traditional IDEs. It's a massive productivity boost. JupyterLab should seriously care about its text-editing experience. Answers: username_1: This works in the text editor when the text editor keymap is set to "sublime" (see the settings menu). It appears that the notebook does not use that setting for the editor keymap, unfortunately. username_1: A relevant comment about the notebook using the text editor keymap is at https://github.com/jupyterlab/jupyterlab/issues/3885#issuecomment-412294469. username_2: ++ for this feature, even it it uses a different keybinding
click-contrib/sphinx-click
329076198
Title: Warning when multiple options take the same envvar Question: username_0: Our use case is to have an option which is common to multiple commands. This option has an `envvar` parameter. This gives a warning. I think that this is a valid, not warning-worthy error. Perhaps `sphinx-click` should differentiate the built environment variable blocks somehow. A small reproducible case follows: `hello.py`: ```python import click @click.command() @click.option('--foo', envvar='EXAMPLE') def A(foo): pass @click.command() @click.option('--foo', envvar='EXAMPLE') def B(foo): pass ``` `source/index.rst`: ```rst .. click:: hello:A :prog: hello .. click:: hello:B :prog: hello ``` Building the docs shows the following error: ``` hello:1: WARNING: Duplicate explicit target name: "cmdoption-hello--foo". /Users/dev/Workspace/sphinx-envvar/source/index.rst:3: WARNING: Duplicate ID: "envvar-EXAMPLE". ``` Status: Issue closed Answers: username_1: Seems there was another issue here. I've resolved this in 1.4.1.
pytorch/pytorch
1030140671
Title: CUDAgraph error while capturing Question: username_0: ## 🐛 Bug <!-- A clear and concise description of what the bug is. --> I found bug while capturing the resnet50 model using CUDAgraph ## Code ```python mport torch [0/94] import torchvision model = torchvision.models.resnet50(num_classes=1000) model = model.cuda() model.train() #torch.backends.cudnn.benchmark = True # Define a loss function and an optimizer #loss_fn = torch.nn.CrossEntropyLoss() loss_fn = torch.nn.MSELoss() optimizer = torch.optim.SGD(model.parameters(), lr=0.1) # Prepare a dummy input input_shape = [64, 3, 224, 224] output_shape = [64,1000] dummy_input = torch.randn(*input_shape, device='cuda') dummy_output = torch.randn(*output_shape, device='cuda') output = model(dummy_input) # warmup # Uses static_input and static_target here for convenience, # but in a real setting, because the warmup includes optimizer.step() # you must use a few batches of real data. s = torch.cuda.Stream() s.wait_stream(torch.cuda.current_stream()) with torch.cuda.stream(s): for i in range(3): optimizer.zero_grad(set_to_none=True) output = model(dummy_input) loss = loss_fn(output, dummy_output) loss.backward() optimizer.step() torch.cuda.current_stream().wait_stream(s) torch.cuda.synchronize() # capture g= torch.cuda.CUDAGraph() # Sets grads to None before capture, so backward() will create # .grad attributes with allocations from the graph's private pool s.wait_stream(torch.cuda.current_stream()) optimizer.zero_grad(set_to_none=True) with torch.cuda.stream(s): g.capture_begin() static_y_pred = model(dummy_input) static_loss = loss_fn(static_y_pred, dummy_output) static_loss.backward() optimizer.step() g.capture_end() ``` [Truncated] return self._forward_impl(x) File "/home/with1015/anaconda3/envs/torch11/lib/python3.7/site-packages/torchvision/models/resnet.py", line 238, in _forward_impl x = self.layer2(x) File "/home/with1015/anaconda3/envs/torch11/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1106, in _call_impl return forward_call(*input, **kwargs) File "/home/with1015/anaconda3/envs/torch11/lib/python3.7/site-packages/torch/nn/modules/container.py", line 141, in forward input = module(input) File "/home/with1015/anaconda3/envs/torch11/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1106, in _call_impl return forward_call(*input, **kwargs) File "/home/with1015/anaconda3/envs/torch11/lib/python3.7/site-packages/torchvision/models/resnet.py", line 128, in forward out = self.conv2(out) File "/home/with1015/anaconda3/envs/torch11/lib/python3.7/site-packages/torch/nn/modules/module.py", line 1106, in _call_impl return forward_call(*input, **kwargs) File "/home/with1015/anaconda3/envs/torch11/lib/python3.7/site-packages/torch/nn/modules/conv.py", line 446, in forward return self._conv_forward(input, self.weight, self.bias) File "/home/with1015/anaconda3/envs/torch11/lib/python3.7/site-packages/torch/nn/modules/conv.py", line 443, in _conv_forward self.padding, self.dilation, self.groups) RuntimeError: captures_underway == 0INTERNAL ASSERT FAILED at "../c10/cuda/CUDACachingAllocator.cpp":1224, please report a bug to PyTorch. cc @mcarilli Answers: username_1: Hey @username_0 did you ever figure out what cause this?
matplotlib/matplotlib
150909397
Title: Animation with blit broken Question: username_0: Matplotliv 1.5.1 on Linux (Debian) [should have animations fixed in 1.5.1](https://github.com/matplotlib/matplotlib/commit/78bb82cec63f6f8adb5aa2e2ea62c99cca77249d) but I still get an error for what used to work. Example: ```` %matplotlib inline from pylab import * import matplotlib.animation import matplotlib.patches fig, ax = plt.subplots(figsize=(13, 8)) def update(frame): ax.add_patch(matplotlib.patches.Rectangle(numpy.array([0, 0]), 1, 1)) anim = matplotlib.animation.FuncAnimation(fig, update, frames=10, blit=True) # blit=True !! anim.save('anim.mp4', fps=20, writer='avconv', codec='libx264') ```` Gives: ```` python3.5/site-packages/matplotlib/animation.py in _draw_frame(self, framedata) 1212 self._drawn_artists = self._func(framedata, *self._args) 1213 if self._blit: -> 1214 for a in self._drawn_artists: 1215 a.set_animated(self._blit) TypeError: 'NoneType' object is not iterable ```` Answers: username_1: For blitting to work properly in a FuncAnimation, the artists being updated must also be returned by the updating function, IIRC. username_0: Oh! Does that mean returning the `ax`? A good error message would be very useful. Something like: Blit requires the update function to return the artist being updated (e.g. XXX) but the function returned `None`. username_1: No, not `ax`. The patches you created. I think there has been some fix-ups in master, but I think it might be more in the direction of turning blitting off in the situation where the function returns None. username_0: Yes it works without blit. Not sure why I put it in the first place and the output looks the same; may be faster. It did use to work though as it is. username_1: Right, blitting is merely an optimization approach (though it does have some nasty side-effects in certain edge-cases). Further, our code used to be a fair amount more relaxed about this, but changes for v1.5 made things more strict, which revealed bugs like this. If you don't need the speed increases that blitting gives you, then feel free to turn that off (which is the default). Status: Issue closed username_2: blitting used to only be used for updating an interactive figure, if you are only saving a file blitting gets you nothing (we fully re-render every frame when saving to file).
discordjs/discord.js
290228558
Title: HELP: message.replace Question: username_0: Hi, i have this: client.on("message", (message) => { if (message.content.startsWith ('-lis ')) { var p1 = message.content.replace("-lis ", "") const embed = { "title": "<:Lis:404390869576450048> **Life Is Strange**", "color": 14396413, "fields": [ { "name": p1 + '´ information', "value": "[**Click here to open the character' profile**](http://pt-br.life-is-strange.wikia.com/wiki/" + p1 + ')', "inline": true } ] } message.channel.send({embed}); } }); It send's this: ![image](https://user-images.githubusercontent.com/35117083/35188579-caccb280-fe2f-11e7-9a10-4a5cc1ed9cee.png) I want the bot send all character's name but on link just the 1st name, like that: ![image](https://user-images.githubusercontent.com/35117083/35188617-c7dd1cbc-fe30-11e7-9922-bcd960db6761.png) Status: Issue closed Answers: username_1: __The issue tracker is only for bug reports and enhancement suggestions. If you have a question, please ask it in the [Discord server](https://discord.gg/bRCvFy9) instead of opening an issue – you will get redirected there anyway.__
facebook/react-native
377111238
Title: Strengthening Flow Types for Core Components Question: username_0: Many of these types are used to validate what we send to native. If JS defines a function that expects a string but Native calls it with a number, applications can crash. We'd like to fix that by removing all references to `any`, `Function`, or `Object`. While we ideally want to remove all references of these files from the codebase, there are a lot, so we should prioritize the ones that are used in the props type definitions for components. 😄 If your are able to remove all of these weak types from the file, try to change `@flow` at the top of the file to `@flow strict-local`. That should ensure that weak types can't come back to these files in the future. [Step 1](https://github.com/facebook/react-native/issues/21342), [Step 2](https://github.com/facebook/react-native/issues/21485), and [Step 3](https://github.com/facebook/react-native/issues/21581) helped prepare our components for this step. # How to submit quality PRs Since many of these weak types are used at the boundary between JS and native, you will likely need to read through the native code for these components to see how these props are being used. Unfortunately, Flow will likely pass with invalid types. Paying close attention to what iOS and Android expects will be helpful to ensure the types are accurate. I also urge those that submit PRs for this to help out with reviewing the PRs for this issue from other contributors. Code review is a great opportunity to learn and improve your own code as well as make sure everyone is on the same page and consistent. If you find tips that would have helped you investigate and improve the types, commenting on this issue with those tips would be appreciated. Help each other. ❤️ Also note that since you are improving these types you will likely help catch a bunch of bugs at Facebook (and elsewhere) where code isn't handling the types correctly. This means that PRs will likely take longer to land then the other issues like this we have asked for help on. This is a good thing, it is direct impact on the stability of React Native projects and catching bugs. # The files The following is a list of files that I'd like to address first. If you want to take one of these files please comment on this issue with the file name so that others don't work on it as well and waste work. There are plenty of files to go around. 😄 Also, `TextProps.js` and `ViewPropTypes.js` are probably the files with the most changes necessary. I don't really expect those to be done by one person. Feel free to type a few and send a PR. Once that PR is landed someone else can take it on and type a few more. - [ ] Libraries/Text/TextProps.js - [ ] Libraries/Components/CheckBox/CheckBox.android.js - [ ] Libraries/Components/DatePicker/DatePickerIOS.ios.js - [ ] Libraries/Components/DatePickerAndroid/DatePickerAndroid.android.js - [ ] Libraries/Components/DatePickerAndroid/DatePickerAndroid.ios.js - [ ] Libraries/Components/DrawerAndroid/DrawerLayoutAndroid.android.js - [ ] Libraries/Components/Keyboard/Keyboard.js - [ ] Libraries/Components/Picker/Picker.js - [ ] Libraries/Components/Picker/PickerAndroid.android.js - [ ] Libraries/Components/Picker/PickerIOS.ios.js - [ ] Libraries/Components/RefreshControl/RefreshControl.js - [ ] Libraries/Components/ScrollResponder.js - [ ] Libraries/Components/ScrollView/__mocks__/ScrollViewMock.js - [ ] Libraries/Components/ScrollView/InternalScrollViewType.js - [ ] Libraries/Components/ScrollView/ScrollView.js - [ ] Libraries/Components/SegmentedControlIOS/SegmentedControlIOS.ios.js - [ ] Libraries/Components/Slider/Slider.js - [ ] Libraries/Components/StaticContainer.react.js - [ ] Libraries/Components/StatusBar/StatusBar.js - [ ] Libraries/Components/TextInput/TextInput.js - [ ] Libraries/Components/TimePickerAndroid/TimePickerAndroid.android.js - [ ] Libraries/Components/TimePickerAndroid/TimePickerAndroid.ios.js - [ ] Libraries/Components/Touchable/TouchableBounce.js - [ ] Libraries/Components/Touchable/TouchableHighlight.js - [ ] Libraries/Components/Touchable/TouchableNativeFeedback.android.js - [ ] Libraries/Components/Touchable/TouchableOpacity.js - [ ] Libraries/Components/Touchable/TouchableWithoutFeedback.js - [ ] Libraries/Components/View/ViewPropTypes.js - [ ] Libraries/Components/ViewPager/ViewPagerAndroid.android.js - [ ] Libraries/Components/WebView/WebView.ios.js <details> <summary> All the warnings for these files: </summary> ``` Libraries/Text/TextProps.js 109:23 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort 110:22 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort [Truncated] 183:32 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort 194:39 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort 198:30 warning The `Object` type is too generic and could lead to hard to debug errors. Please use only as a last resort 199:30 warning The `Object` type is too generic and could lead to hard to debug errors. Please use only as a last resort Libraries/Components/ViewPager/ViewPagerAndroid.android.js 26:14 warning The `Object` type is too generic and could lead to hard to debug errors. Please use only as a last resort 50:19 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort 60:31 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort 68:21 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort Libraries/Components/WebView/WebView.ios.js 57:11 warning The `any` type is too generic and could lead to hard to debug errors. Please use only as a last resort 58:9 warning The `any` type is too generic and could lead to hard to debug errors. Please use only as a last resort 59:16 warning The `any` type is too generic and could lead to hard to debug errors. Please use only as a last resort 62:14 warning The `Object` type is too generic and could lead to hard to debug errors. Please use only as a last resort 659:26 warning The `any` type is too generic and could lead to hard to debug errors. Please use only as a last resort ``` </details> Answers: username_1: I'll take `Libraries/Components/DatePicker/DatePickerIOS.ios.js` username_2: I'll start with `Libraries/Components/ScrollView/__mocks__/ScrollViewMock.js` username_0: Many of these types are used to validate what we send to native. If JS defines a function that expects a string but Native calls it with a number, applications can crash. We'd like to fix that by removing all references to `any`, `Function`, or `Object`. While we ideally want to remove all references of these files from the codebase, there are a lot, so we should prioritize the ones that are used in the props type definitions for components. 😄 If your are able to remove all of these weak types from the file, try to change `@flow` at the top of the file to `@flow strict-local`. That should ensure that weak types can't come back to these files in the future. [Step 1](https://github.com/facebook/react-native/issues/21342), [Step 2](https://github.com/facebook/react-native/issues/21485), and [Step 3](https://github.com/facebook/react-native/issues/21581) helped prepare our components for this step. # How to submit quality PRs Since many of these weak types are used at the boundary between JS and native, you will likely need to read through the native code for these components to see how these props are being used. Unfortunately, Flow will likely pass with invalid types. Paying close attention to what iOS and Android expects will be helpful to ensure the types are accurate. I also urge those that submit PRs for this to help out with reviewing the PRs for this issue from other contributors. Code review is a great opportunity to learn and improve your own code as well as make sure everyone is on the same page and consistent. If you find tips that would have helped you investigate and improve the types, commenting on this issue with those tips would be appreciated. Help each other. ❤️ Also note that since you are improving these types you will likely help catch a bunch of bugs at Facebook (and elsewhere) where code isn't handling the types correctly. This means that PRs will likely take longer to land then the other issues like this we have asked for help on. This is a good thing, it is direct impact on the stability of React Native projects and catching bugs. # The files The following is a list of files that I'd like to address first. If you want to take one of these files please comment on this issue with the file name so that others don't work on it as well and waste work. There are plenty of files to go around. 😄 Also, `TextProps.js` and `ViewPropTypes.js` are probably the files with the most changes necessary. I don't really expect those to be done by one person. Feel free to type a few and send a PR. Once that PR is landed someone else can take it on and type a few more. - [ ] Libraries/Text/TextProps.js - [ ] Libraries/Components/CheckBox/CheckBox.android.js - [ ] Libraries/Components/DatePicker/DatePickerIOS.ios.js - [ ] Libraries/Components/DatePickerAndroid/DatePickerAndroid.android.js - [ ] Libraries/Components/DatePickerAndroid/DatePickerAndroid.ios.js - [ ] Libraries/Components/DrawerAndroid/DrawerLayoutAndroid.android.js - [ ] Libraries/Components/Keyboard/Keyboard.js - [ ] Libraries/Components/Picker/Picker.js - [ ] Libraries/Components/Picker/PickerAndroid.android.js - [ ] Libraries/Components/Picker/PickerIOS.ios.js - [ ] Libraries/Components/RefreshControl/RefreshControl.js - [ ] Libraries/Components/ScrollResponder.js - [ ] Libraries/Components/ScrollView/__mocks__/ScrollViewMock.js - [ ] Libraries/Components/ScrollView/InternalScrollViewType.js - [ ] Libraries/Components/ScrollView/ScrollView.js - [ ] Libraries/Components/SegmentedControlIOS/SegmentedControlIOS.ios.js - [ ] Libraries/Components/Slider/Slider.js - [ ] Libraries/Components/StaticContainer.react.js - [ ] Libraries/Components/StatusBar/StatusBar.js - [ ] Libraries/Components/TextInput/TextInput.js - [ ] Libraries/Components/TimePickerAndroid/TimePickerAndroid.android.js - [ ] Libraries/Components/TimePickerAndroid/TimePickerAndroid.ios.js - [ ] Libraries/Components/Touchable/TouchableBounce.js - [ ] Libraries/Components/Touchable/TouchableHighlight.js - [ ] Libraries/Components/Touchable/TouchableNativeFeedback.android.js - [ ] Libraries/Components/Touchable/TouchableOpacity.js - [ ] Libraries/Components/Touchable/TouchableWithoutFeedback.js - [ ] Libraries/Components/View/ViewPropTypes.js - [ ] Libraries/Components/ViewPager/ViewPagerAndroid.android.js - [ ] Libraries/Components/WebView/WebView.ios.js <details> <summary> All the warnings for these files: </summary> ``` Libraries/Text/TextProps.js 109:23 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort 110:22 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort [Truncated] 183:32 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort 194:39 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort 198:30 warning The `Object` type is too generic and could lead to hard to debug errors. Please use only as a last resort 199:30 warning The `Object` type is too generic and could lead to hard to debug errors. Please use only as a last resort Libraries/Components/ViewPager/ViewPagerAndroid.android.js 26:14 warning The `Object` type is too generic and could lead to hard to debug errors. Please use only as a last resort 50:19 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort 60:31 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort 68:21 warning The `Function` type is too generic and could lead to hard to debug errors. Please use only as a last resort Libraries/Components/WebView/WebView.ios.js 57:11 warning The `any` type is too generic and could lead to hard to debug errors. Please use only as a last resort 58:9 warning The `any` type is too generic and could lead to hard to debug errors. Please use only as a last resort 59:16 warning The `any` type is too generic and could lead to hard to debug errors. Please use only as a last resort 62:14 warning The `Object` type is too generic and could lead to hard to debug errors. Please use only as a last resort 659:26 warning The `any` type is too generic and could lead to hard to debug errors. Please use only as a last resort ``` </details> username_0: Silly bot. username_2: I'll take `Libraries/Components/RefreshControl/RefreshControl.js` username_1: I'll take `Libraries/Components/DatePicker/DatePickerIOS.ios.js` , `Libraries/Components/DatePickerAndroid/DatePickerAndroid.ios.js` username_3: If anyone needs help with figuring out the callback/return types of native components feel free to ask here, and we can help out! 👍 username_4: I'd like to take `Libraries/Components/DrawerAndroid/DrawerLayoutAndroid.android.js` username_1: I'll take `Libraries/Components/Keyboard/Keyboard.js` username_2: @username_3 How would you type `requireNativeComponent` ? username_0: Don’t worry about requireNativeComponent. We have another project internally working to remove that function and instead provide an accurate type. username_4: @username_0 I don't see `Object` anywhere on line 121 in `CheckBox.android.js`. What I do see is `any` on line 81 but it's in reference to `requireNativeComponent`, which you said above we shouldn't worry about. username_0: Whoops. Looks like [another PR](https://github.com/facebook/react-native/commit/28de61e9f0a46a2389503b8bf4e1b33f333ed0e4#diff-7f3d990a8a1579593beee654dfc118fe) from @username_3 landed after I made that list which solved that line in Checkbox. Sorry for the churn. I'll mark that off the list. username_2: I'll take `Libraries/Components/StaticContainer.react.js` username_2: I'll take `Libraries/Text/TextProps.js` username_2: I could take `Libraries/Components/Slider/Slider.js` as well username_5: If no one has tackled `Libraries/Components/ScrollView/InternalScrollViewType.js`, I could take a swing at it username_6: I'll take `Libraries/Utilities/deepFreezeAndThrowOnMutationInDev.js` username_7: I'll take Libraries/Components/StatusBar/StatusBar.js username_8: I'll take ` Libraries/Components/Touchable/TouchableBounce.js` ! username_9: I'll take `Libraries/Components/TimePickerAndroid/TimePickerAndroid.android.js` & `Libraries/Components/TimePickerAndroid/TimePickerAndroid.ios.js` username_6: I'll take `Libraries/Components/WebView/WebView.ios.js` username_2: I'll take `Libraries/Components/Touchable/TouchableOpacity.js` username_3: I'll take `Libraries/Components/View/ViewPropTypes.js` username_10: I'll take `Libraries/Components/Touchable/TouchableWithoutFeedback.js` ~~ username_2: I'll take `Libraries/Components/Touchable/TouchableHighlight.js` username_11: I took ` Libraries/Components/DrawerAndroid/DrawerLayoutAndroid.android.js` username_12: Hey y'all! I took Libraries/Components/ScrollView/ScrollView.js username_13: I'll take `Libraries/Components/ScrollResponder.js` username_2: I'll take `Libraries/Components/Touchable/TouchableBounce.js` username_1: I'll take `Libraries/Components/SegmentedControlIOS/SegmentedControlIOS.ios.js` username_14: I'll take ` Libraries/Components/Touchable/TouchableNativeFeedback.android.js` username_1: I think the fix is not needed in `Libraries/Components/SegmentedControlIOS/SegmentedControlIOS.ios.js` Maybe #21888 fixed this issue 👍 username_2: I'll take `Libraries/Components/TextInput/TextInput.js` username_0: @username_2, TextInput has an open PR to convert it to an es6 class by @username_3. You might have merge conflicts if you change more things in that file before it lands. username_3: Yeah, you might want to wait - when it lands, however, there will still need to be work done on it to make it pass `flow strict-local` however! 👍 username_2: @username_0 @username_3 Alright I'll wait username_15: I'll take `Libraries/Components/ScrollView/ScrollView.js` username_2: I'll take `Libraries/Components/StatusBar/StatusBar.js` username_3: @username_2 The ES6 conversion had to be rolled back, so you should be able to take this on. username_15: I'll take `Libraries/Components/StatusBar/StatusBar.js` username_16: Happy to take `Libraries/Text/TextProps` username_2: @username_16 There's this PR opened #22122 Have you seen issues in it ? Thanks username_16: Oh, I haven't noticed this. Will try to find something different then ;) username_2: Hey, I was wondering, isn't `Libraries/Components/View/ViewPropTypes.js` done ? 😄 username_0: I think it is close but there are one or two callsites that need strengthening in that file still. username_2: Thanks @username_0 username_0: We still have one file to go here but I have created another issue with the next step of this project. I'd love any help we can get. You all have been amazing. ❤️ https://github.com/facebook/react-native/issues/22990 username_17: Anybody can to explain me why `import type {SyntheticEvent} from 'CoreEventTypes';` works? Why flow understund `CoreEventTypes` alias ? Because `module.system.haste.paths.whitelist=<PROJECT_ROOT>/Libraries/.*` in `.flowconfig`? But `.flowconfig` of react native project containts `module.system.haste.paths.whitelist=<PROJECT_ROOT>/node_modules/react-native/Libraries/.*` and I can't using alias `CoreEventTypes`. Why? Status: Issue closed username_18: And we are done with the last PR by @danibonilha. Thanks everyone for helping out with this effort to make our flow types stricter!
pycket/pycket
427278018
Title: Error when trying make setup-local-racket Question: username_0: Hi Caner, thanks a lot for implementing setup-local-racket. I unfortunately only now got to trying it and am getting the below error. Any clue what could be causing this? ``` $ make setup-local-racket Downloading Racket --2019-03-30 12:59:12-- http://www.cs.utah.edu/plt/snapshots/current/installers/racket-current-x86_64-linux-precise.sh Resolving www.cs.utah.edu (www.cs.utah.edu)... 192.168.3.11 Connecting to www.cs.utah.edu (www.cs.utah.edu)|192.168.3.11|:80... connected. HTTP request sent, awaiting response... 200 OK Length: 122883352 (117M) [text/x-sh] Saving to: ‘racket-current-x86_64-linux-precise.sh’ racket-current-x86_64-linux-precise.sh 100%[=================================================================================================>] 117.19M 941KB/s in 2m 8s 2019-03-30 13:01:22 (937 KB/s) - ‘racket-current-x86_64-linux-precise.sh’ saved [122883352/122883352] Installing Racket Telling Racket about Pycket chmod 755 racket-current-x86_64-linux-precise.sh ./racket-current-x86_64-linux-precise.sh --in-place --dest racket This program will extract and install Racket v7.2.0.11. Note: the required diskspace for this installation is 522M. Checking the integrity of the binary archive... ok. Unpacking into "/home/username_0/projects/pycket/racket" (Ctrl+C to abort)... Done. Installation complete. rm -f racket-current-x86_64-linux-precise.sh racket/bin/raco pkg install -t dir pycket/pycket-lang/ || \ racket/bin/raco pkg update --link pycket/pycket-lang raco setup: version: 7.2.0.11 raco setup: platform: x86_64-linux [3m] raco setup: target machine: racket raco setup: installation name: snapshot raco setup: variants: 3m raco setup: main collects: /home/username_0/projects/pycket/racket/collects raco setup: collects paths: raco setup: /home/username_0/projects/pycket/racket/collects raco setup: main pkgs: /home/username_0/projects/pycket/racket/share/pkgs raco setup: pkgs paths: raco setup: /home/username_0/projects/pycket/racket/share/pkgs raco setup: /home/username_0/.racket/snapshot/pkgs raco setup: links files: raco setup: /home/username_0/projects/pycket/racket/share/links.rktd raco setup: /home/username_0/.racket/snapshot/links.rktd raco setup: main docs: /home/username_0/projects/pycket/racket/doc raco setup: --- updating info-domain tables --- [13:01:30] raco setup: updating: /home/username_0/.racket/snapshot/share/info-cache.rktd raco setup: --- pre-installing collections --- [13:01:30] raco setup: --- installing foreign libraries --- [13:01:30] raco setup: --- installing shared files --- [13:01:30] raco setup: --- compiling collections --- [13:01:30] raco setup: --- parallel build using 4 jobs --- [13:01:30] raco setup: 3 making: <pkgs>/pycket-lang raco setup: --- creating launchers --- [13:01:38] raco setup: --- installing man pages --- [13:01:38] raco setup: --- building documentation --- [13:01:38] raco setup: --- installing collections --- [13:01:39] raco setup: --- post-installing collections --- [13:01:39] raco pkg install -t dir pycket/pycket-lang/ || \ raco pkg update --link pycket/pycket-lang racket/collects/raco/compiled/raco_rkt.zo::1: read (compiled): wrong version for compiled code compiled version: 7.2.0.11 expected version: 6.11 context...: standard-module-name-resolver /home/username_0/projects/pycket/racket/collects/raco/main.rkt: [running body] racket/collects/raco/compiled/raco_rkt.zo::1: read (compiled): wrong version for compiled code compiled version: 7.2.0.11 expected version: 6.11 context...: standard-module-name-resolver /home/username_0/projects/pycket/racket/collects/raco/main.rkt: [running body] Makefile:139: recipe for target 'setup-racket-for-old-pycket' failed make: *** [setup-racket-for-old-pycket] Error 1``` Answers: username_1: This seems to be happening because the zo files (the compiled bytecodes) for the `pycket-lang` are compiled previously with a Racket 6.11, so now the new raco tries to use those zo files but they have an older version. So try removing the `compiled` directory under `pycket-lang` and it should work. Sat, Mar 30, 2019, 8:35 AM <NAME> < username_0: Unfortunately that didn't help, the error is the same. Maybe some part of the Makefile is using my system racket? username_0: Yes, that seems to be it, if I uninstall my system racket, it complains about racket and raco not found. so some part of the Makefile needs to use the newly installed racket explicitly. username_0: this fixes it for me, does taht make sense?: ``` diff --git a/Makefile b/Makefile index 902b8668..72ba9d87 100644 --- a/Makefile +++ b/Makefile @@ -131,13 +131,13 @@ download-and-install-racket: $(eval export PLTHOME=$(shell pwd)/racket) $(eval export PLTCOLLECTS=$(shell pwd)/racket/collects) $(info Telling Racket about Pycket) - racket/bin/raco pkg install -t dir pycket/pycket-lang/ || \ - racket/bin/raco pkg update --link pycket/pycket-lang + ./racket/bin/raco pkg install -t dir pycket/pycket-lang/ || \ + ./racket/bin/raco pkg update --link pycket/pycket-lang # Use the one below for non-local Racket builds setup-racket-for-old-pycket: - raco pkg install -t dir pycket/pycket-lang/ || \ - raco pkg update --link pycket/pycket-lang + ./racket/bin/raco pkg install -t dir pycket/pycket-lang/ || \ + ./racket/bin/raco pkg update --link pycket/pycket-lang clean-racket: rm -rf racket ``` username_2: Yes, that looks right (although the first half of that diff probably doesn't do anything). Status: Issue closed