repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
mlr-org/mlrCPO
324384518
Title: cpoSelect with statements Question: username_0: Is there a CPO where I can just write something like `cpoSelect(sel.statement = function(x) is.numeric(x) && !any(is.na(x)))` and sel.statement just gets lapplied on the columns? Answers: username_1: Currently the way to select columns is only via `cpoSelect(type = "numeric")`; dropping columns with missing values is not supported right now. I will try to make this nicer, probably together with "slices" (see #38) username_1: In the mean time, a quick and dirty implementation of this would be ```R library("BBmisc") cpoSelectFunctional = makeCPO("select.functional", pSS(selector: funct), dataformat = "df.features", properties.adding = c("numerics.sometimes", "factors.sometimes", "ordered.sometimes", "missings.sometimes"), cpo.train = { vlapply(data, selector) }, cpo.retrafo = { data[control] }) ``` username_0: Thanks. That's a nice start. If you want, you can close this issue.
GoogleCloudPlatform/python-docs-samples
197052818
Title: _Rendezvous of RPC that terminated with (StatusCode.CANCELLED, Cancelled!) Question: username_0: I was running the following file: /python-docs-samples/speech/grpc/transcribe_streaming.py as following: python transcribe_streaming.py It works fine (I get transcript of my speech) until I press Ctrl+C. When I press Ctrl+C I get the following error: ^CTraceback (most recent call last): File "transcribe_streaming.py", line 234, in <module> main() File "transcribe_streaming.py", line 225, in main listen_print_loop(recognize_stream) File "transcribe_streaming.py", line 172, in listen_print_loop for resp in recognize_stream: File "/usr/local/lib/python2.7/dist-packages/grpc/_channel.py", line 344, in next return self._next() File "/usr/local/lib/python2.7/dist-packages/grpc/_channel.py", line 335, in _next raise self grpc._channel._Rendezvous: <_Rendezvous of RPC that terminated with (StatusCode.CANCELLED, Cancelled!)> I believe that this error shouldn't be happening because of (in the main() try section): except face.CancellationError: # This happens because of the interrupt handler pass So my question is what is the right way to stop transcribing? If I just do recognize_stream.cancel() I get these _Rendezvous StatusCode.CANCELLED errors even though there is except face.CancellationError exception handler. Is this a bug or I'm doing something wrong? Answers: username_1: Yeah - I just saw this too. I believe the library got updated and now the exception that's being thrown is different... fix forthcoming... username_0: Ok, thanks! Status: Issue closed
dyama/mruby-siren
248297750
Title: mruby-siren と Fusion360 の連帯の可能性 Question: username_0: Autodesk社のFusion360と言うCADをご存知でしょうか? こちらのCADは非常に多くのフォーマットのCADデータを インポートすることが可能なのですが、サーフェス・ソリッドのみ で点・線が欠落してしまいます。 https://forums.autodesk.com/t5/fusion-360-ri-ben-yu/igesrain-kabu-detanoinpoto/m-p/6606029#M3144 https://forums.autodesk.com/t5/fusion-360-ri-ben-yu/stp-xing-shi-dengnofairuwo-dumi-rumi-shuki-chushiwo-xingtta-shini-dian-xiannadono-yao-suga-xiaoeteiru/m-p/7218766#M6156 Fusion360もAPIは、かなり公開されており、 http://help.autodesk.com/view/fusion360/ENU/?guid=GUID-dc7ae251-e060-4d87-b6b8-e7f78abc0777 これらの問題を "Sirenを利用すれば解決出来るのでは?" と思ったのですが、素人レベルで可能なものか・・・と感じております。 Siren自身の魅力は、バイナリ版であれば導入が非常に簡単なため、 多くのユーザーにも受け入れやすいものと感じているのですが、 mRuby↔Python間でデータのやり取りが可能なものか?どうか?  もよくわからない状態なのが正直なところです…。 Answers: username_1: Fusion360 を触ったことがないですが、スケッチの概念があったり、[Curve3D クラス](http://help.autodesk.com/view/fusion360/ENU/?guid=GUID-283c1e3d-2260-4d5b-ae9e-f84b8f6a0ca9)があるところを見ると Fusion360 自体はしっかりした構造を持っていて、パーサ側の機能がないようですね。 siren で IGES ファイルを読み込み、Fusion360 向けの Python スクリプトを出力する mruby スクリプトを書いてみました。 ``` #!siren # coding: utf-8 iges = "sample.iges" py = "sample.py" # 評価用の IGES ファイルを生成 def make_sample_iges(path) points = [[0, 0, 0], [1, 0, 0], [1, 1, 0], [2, 1, 0], [2, 2, 0]] wire = Siren.polyline points Siren.save_iges wire, path end # IGES ファイルがなければ生成 unless File.exist? iges make_sample_iges iges end # IGES ファイルの読み込み model = Siren.load_iges iges # 書き出すスクリプトファイルを開く file = File.open(py, "w") file.puts "import adsk.core" # 例えば、モデル中の直線要素を処理する model.edges.select{|e| e.curve.is_a? Siren::Line }.each_with_index do |e, i| if e.infinite? # 無限直線の場合 c = e.curve file.write "res#{i} = InfiniteLine3D.create(" file.write "Point3D.create(#{e.pos.x}, #{e.pos.y}, #{e.pos.z}), " file.puts "Vector3D.create(#{c.dir.x}, #{c.dir.y}, #{c.dir.z}))" else # 有限直線(線分)の場合 file.write "res#{i} = Line3D.create(" file.write "Point3D.create(#{e.sp.x}, #{e.sp.y}, #{e.sp.z}), " file.puts "Point3D.create(#{e.tp.x}, #{e.tp.y}, #{e.tp.z}))" end end # スクリプトファイルを閉じる file.close ``` 例は直線だけですが、同じ要領で[円弧](http://help.autodesk.com/view/fusion360/ENU/?guid=GUID-fef681fd-ca49-4fce-b9bd-6e663021d2db)や [NURBS 曲線](http://help.autodesk.com/view/fusion360/ENU/?guid=GUID-ee0fb092-7a5c-4596-902e-fff01c7d50b1)も対応できそうです。 実行すると ``` S0000001 ,,31HOpen CASCADE IGES processor 7.1,13HFilename.iges, G0000001 16HOpen CASCADE 7.1,31HOpen CASCADE IGES processor 7.1,32,308,15,308,15,G0000002 ,1.,2,2HMM,1,0.01,15H20170807.123552,1E-007,2.,7Hyamadai,,11,0, G0000003 [Truncated] 102,4,3,5,7,9; 0000001P0000001 110,0.E+000,0.E+000,0.E+000,1.,0.E+000,0.E+000; 0000003P0000002 110,1.,0.E+000,0.E+000,1.,1.,0.E+000; 0000005P0000003 110,1.,1.,0.E+000,2.,1.,0.E+000; 0000007P0000004 110,2.,1.,0.E+000,2.,2.,0.E+000; 0000009P0000005 S 1G 4D 10P 5 T0000001 ``` のような評価用 IGES ファイル `sample.iges` が生成され、それを読み込み ``` import adsk.core res0 = Line3D.create(Point3D.create(0, 0, 0), Point3D.create(1, 0, 0)) res1 = Line3D.create(Point3D.create(1, 0, 0), Point3D.create(1, 1, 0)) res2 = Line3D.create(Point3D.create(1, 1, 0), Point3D.create(2, 1, 0)) res3 = Line3D.create(Point3D.create(2, 1, 0), Point3D.create(2, 2, 0)) ``` といった Python スクリプト `sample.py` を出力します。 Fusion360 を動かせる環境が手元にないため、チェックはしていませんが、考え方はこれでいいはずだと思います。 ただ、 Fusion360 側をしっかり調べればもっとスマートな方法も見つかるかもしれません。 username_0: 早速ありがとうございます。(速いです・・・) 全く理解できていませんが、時間が出来次第勉強させていただきます。 username_0: エクスポートする為の Fusion360 → Siren は、Sirenスプリクトを書き出し → 実行で出来そうだな と感触は持っていたのですが、インポートの Siren → Fusion360 は、どうすれば良いのだろう? と思っておりました。 試してはいませんが、モジュールとして書き出したPythonスプリクトをインポート すれば、実行できそうな気がしてきました。 色々とアドバイスありがとうございます。 username_0: みすぼらしいものですが、公開まで漕ぎ着けました。 https://forums.autodesk.com/t5/fusion-360-ri-ben-yu/3dcad-zhong-jianfairuno3dna-dian-qu-xianwoinpotosurusupurikuto/td-p/7370346
CDRH/api
434884899
Title: Look into possibility of using multiple language analyzers on content? Question: username_0: Right now we're splitting content into fields like `*_t_es` and `*_t_en` when we need to carry out a search with specific stemming, etc. Look into possibility of using this ES functionality: https://www.elastic.co/guide/en/elasticsearch/guide/current/mixed-lang-fields.html#_analyze_multiple_times We would want to compare with Solr possibilities so that we know if this would be a feature that could be lost if we change backends.
minbrowser/min
236904149
Title: Page magnification resets randomly on Windows Question: username_0: As in title. After opening page (FB especially) and magnifying it to proper level it stays on this level for a while, and then resets to previous level. May be related to duble finger scrolling (sometimes d-f scrolling triggers reset). Answers: username_1: I think you're correct that this is related to scrolling; scrolling with the ctrl key pressed will zoom in or out. We could possibly look into making this less sensitive. username_0: Thanks for reply :) I've checked it double and here are the results: every type of scrolling resets page magnification, no matter it's d-f, down arrow, PgDn. It appears especially on FB page, but not only. Browser also can't remember magnification of page and set same among tabs, I have to change it on every tab manually, it's quite annoying. username_1: I've opened another issue for this: #372. username_0: Win 10 Creators Update - youtube.com , after clicking notification bell, not connected to scrolling - google.com serch results page, randomly, also after clicking on grafic results tab, - scholar.google.pl results page, after clicking on "Od 2016" on left panel, not connected to scrolling - czasgentlemanow.pl/2017/06/kiedy-przestrzegac-zasad/ after clicking in link and going back to this page magnification is reseted (nothing, when link is opened in new tab) - facebook.com - it appears when loading new content eg. in newsfeed username_1: The 2nd, 3rd and 4th issues are because zoom is reset whenever a new page is loaded, which I've added an issue for in #372. Otherwise, I can't reproduce this. It looks like scrolling will only change the zoom if ```e.metaKey``` is true (https://github.com/minbrowser/min/blob/master/js/webview/swipeEvents.js#L72), which on windows is the windows key (and actually I think this code has a bug; this is supposed to be the ctrl key). Is there any way the windows key could be pressed when you are scrolling? username_0: No, it's not possible. Ctrl zooming works well. I think all this issue is #372. PS I have no idea about code so far :D username_1: OK, I'll close this and leave #372 open then. Status: Issue closed
IntergalacticAvenger/myRPG
800889417
Title: test Question: username_0: <img width="413" alt="Screen Shot 2021-02-03 at 10 14 36 PM" src="https://user-images.githubusercontent.com/53759586/106840062-66f91380-666d-11eb-9d98-36bbff0d107f.png"> <img width="358" alt="Screen Shot 2021-02-03 at 10 14 55 PM" src="https://user-images.githubusercontent.com/53759586/106840065-6791aa00-666d-11eb-8d9b-a9031f83732a.png"> <img width="225" alt="Screen Shot 2021-02-03 at 10 15 16 PM" src="https://user-images.githubusercontent.com/53759586/106840066-6791aa00-666d-11eb-96f3-b4cd6de6c516.png"> <img width="281" alt="Screen Shot 2021-02-03 at 10 15 29 PM" src="https://user-images.githubusercontent.com/53759586/106840067-6791aa00-666d-11eb-90b5-c9d4f02c87bf.png"> <img width="393" alt="Screen Shot 2021-02-03 at 10 15 50 PM" src="https://user-images.githubusercontent.com/53759586/106840068-6791aa00-666d-11eb-85e1-5fdf61e746e9.png"><issue_closed> Status: Issue closed
tj/consolidate.js
36607225
Title: Layout support for handlbars Question: username_0: Hi, would you accept a pull request to add layout support for handlebars? Are you opposed to adding layout support to consolidate.js? I am trying to get something like this working: ```js app.get("/home", function(req, res) { res.render("home", { layout: "main", title: "hello world", test: test }); }); ``` Thanks. Status: Issue closed Answers: username_1: I think it's better to support features exposed by the template engine and not add additional functionality in consolidate. Check out [this handlebars layouts library](https://github.com/shannonmoeller/handlebars-layouts) if you're interested in adding layouts to handlebars.
OBOFoundry/OBOFoundry.github.io
479882335
Title: ontology table is missing from homepage Question: username_0: http://obofoundry.org/ - the page is missing the table with the ontologies, the page is loading for me in Chrome and Firefox, but the able is missing ![image](https://user-images.githubusercontent.com/6722114/62903514-c3e56900-bd17-11e9-9482-a0c7582d8679.png) Maybe related to #1022 Answers: username_1: Sorry, I should have clarified. It's the table that doesn't load, not the whole page. So yes, this is the same as #1022. :) username_0: I will close this ticket then :) Status: Issue closed
googleads/googleads-php-lib
256416213
Title: v29.0.0 installation problem Question: username_0: When I try to install v29.0.0 like before versions, I got below errors. Can you help me please? I installed with steps as mentioned https://github.com/googleads/googleads-php-lib root@nod4 [/home/www]# php composer.phar require googleads/googleads-php-lib Do not run Composer as root/super user! See https://getcomposer.org/root for details Using version ^29.0 for googleads/googleads-php-lib ./composer.json has been updated Loading composer repositories with package information Updating dependencies (including require-dev) Your requirements could not be resolved to an installable set of packages. Problem 1 - Installation request for googleads/googleads-php-lib ^29.0 -> satisfiable by googleads/googleads-php-lib[29.0.0]. - Conclusion: remove google/auth v0.11.1 - Conclusion: don't install google/auth v0.11.1 - googleads/googleads-php-lib 29.0.0 requires google/auth ^1.0.0 -> satisfiable by google/auth[v1.0, v1.0.1]. - Can only install one of: google/auth[v1.0, v0.11.1]. - Can only install one of: google/auth[v1.0.1, v0.11.1]. - Installation request for google/auth (locked at v0.11.1) -> satisfiable by google/auth[v0.11.1]. Installation failed, reverting ./composer.json to its original content. Answers: username_1: Hello, Could you please share what the `composer.json` of your project looks like? Cheers, Knack username_1: And does this flag (--update-with-dependencies) for composer help? username_0: Error updated ; php composer.phar require googleads/googleads-php-lib Do not run Composer as root/super user! See https://getcomposer.org/root for details Using version ^29.0 for googleads/googleads-php-lib ./composer.json has been updated Loading composer repositories with package information Updating dependencies (including require-dev) Your requirements could not be resolved to an installable set of packages. Problem 1 - The requested package googleads/googleads-php-lib No version set (parsed as 1.0.0) is satisfiable by googleads/googleads-php-lib[No version set (parsed as 1.0.0)] but these conflict with your requirements or minimum-stability. Problem 2 - The requested package google/auth (installed at v0.11.1, required as ^1.0.0) is satisfiable by google/auth[v0.11.1] but these conflict with your requirements or minimum-stability. Installation failed, reverting ./composer.json to its original content. composer.json like below { "name": "googleads/googleads-php-lib", "description": "Google Ads APIs Client Library for PHP (AdWords and DFP)", "require": { "php": ">=5.5.9", "ext-openssl": "*", "ext-soap": "*", "google/auth": "^1.0.0", "guzzlehttp/guzzle": "^6.0", "guzzlehttp/psr7": "^1.2", "monolog/monolog": "^1.17.1", "phpdocumentor/reflection-docblock": "^3.0.3", "symfony/serializer": "^2.8.0 || ^3.0.3" }, "require-dev": { "php": ">=5.5.17", "phpunit/phpunit": "^4.8" }, "suggest": { "php-64bit": ">=5.5.9" }, "homepage": "https://github.com/googleads/googleads-php-lib", "license": "Apache-2.0", "autoload": { "psr-4": { "Google\\AdsApi\\": "src/Google/AdsApi/" } }, "authors": [ { "name": "Google", "homepage": "https://github.com/googleads/googleads-php-lib/contributors" } ] } username_1: Could you remove your `composer.json` and try again? Probably the error was caused b your `composer.json`. It specifies v28.0.0 but your command demands v29.0.0. In v29.0.0, we have updated the requirement of google/auth to v1.0.0, which conflicts with that of older versions. Knack username_0: Thank you very much for your help after deleting composer.json with this code it worked successfully. `php composer.phar require googleads/googleads-php-lib --update-with-dependencies` Status: Issue closed
karaage0703/karaage-ai-book
776424504
Title: 文字の重複? Question: username_0: **ページ数** P.34 きゅうり農家の画像の上 **内容** 「AI」という文字が重複している 以下、本文 AIを用いて、きゅうりの~作業を、AIにより実現しています。 **コメント** 個人的に気になったのですが、意図的でしたらすみません m(__)m Answers: username_1: @username_0 さん 貴重なご意見、ありがとうございます。確かに「AIを用いて、きゅうりの等級(ランク)を選別する作業を、AIにより実現しています。」より「AIを用いて、きゅうりの等級(ランク)を選別する作業を実現しています。」の方が自然ですね。 @KazumaAndoh さん 意味は通じますが、直した方が自然かと思います。正誤表への追記はいかがいたしましょうか? Status: Issue closed username_1: [正誤表](https://github.com/username_1/karaage-ai-book/blob/master/ERRATA.md)に追記しましたのでcloseします。
composer/composer
538645704
Title: composer-setup.exe identified as known malware via Carbon Black Question: username_0: Associates in network are flagged when working with composer-setup.exe SHA 6a0b75feaf3823a1054274a362bdd92b9687b265c0f0ff782741853c7995cdd8 as Known Malware via the networks Carbon Black protections. Answers: username_1: Obviously we haven't included any malware in this. It does use the latest version of Inno Setup (6.03), so perhaps this explains the false positive. Source code: https://github.com/composer/windows-setup Status: Issue closed username_1: https://www.virustotal.com/gui/file/6a0b75feaf3823a1054274a362bdd92b9687b265c0f0ff782741853c7995cdd8/detection Nothing is detected on VirusTotal now (having initially shown 5 false-positives, which is much more than the usual none or one).
CocoaPods/CocoaPods
771627863
Title: Install CocoaPods macOS BigSur Apple M1 MacBook Pro Question: username_0: <!-- ℹ Please fill out this template when filing an issue. All lines beginning with an ℹ symbol instruct you with what info we expect. Before you start, are you using the latest CocoaPods release? A lot changes with Xcode releases that are not backwards compatible. Not an issue about the CocoaPods command line app? Please file an issue in the appropriate repo - https://github.com/CocoaPods Issues are for feature requests, and bugs; questions should go to Stack Overflow Using CocoaPods <= 0.39: https://blog.cocoapods.org/Sharding/ Using Xcode 10.1: Requires CocoaPods 1.6.0 or above. Issue with Nanaimo not loading: Please run `[sudo] gem uninstall nanaimo` and remove all but the latest version. Issues with `pod search`? Try deleting your cache `rm -rf ~/Library/Caches/CocoaPods`first. --> * [ ] I've read and understood the [*CONTRIBUTING* guidelines and have done my best effort to follow](https://github.com/CocoaPods/CocoaPods/blob/master/CONTRIBUTING.md). # Report ## What did you do? ℹ Please replace these two lines with what you did. e.g. Run `pod install` ## What did you expect to happen? ℹ Please replace these two lines with what you expected to happen. e.g. Install all pod dependencies correctly. ## What happened instead? ℹ Please replace these two lines with of what happened instead. e.g. Pod A is missing the subspec B for target C. ## CocoaPods Environment ℹ Please replace these two lines with the output of `pod env`. e.g. via `pod env | pbcopy` ## Project that demonstrates the issue ℹ Please link to a project we can download that reproduces the issue. You can delete this section if your issue is unrelated to build problems, i.e. it's only an issue with CocoaPods the tool. Answers: username_1: Hy, you can find a solution to your problem here: https://github.com/CocoaPods/CocoaPods/issues/9907#issuecomment-729980327 username_2: That or many other issues related to ffi gem here. We've kept one issue open https://github.com/CocoaPods/CocoaPods/issues/9907
joopert/nad_receiver
775008438
Title: typeerror cannot concatenate 'str' and 'int' objects Question: username_0: I had some time to put the 0.2.0 in a custom component for home assistant. This resulted in the issue for me when setting the volume. Home assistant passes this as an int while in the new code it must be passed as a str. In the old code we converted it to a str https://github.com/username_0/nad_receiver/blob/0.1.0/nad_receiver/__init__.py#L37 and in the new code we don't https://github.com/username_0/nad_receiver/blob/master/nad_receiver/__init__.py#L46 To be backwards compatible we should convert it to str again? @gladhorn any suggestion for this? In my custom component I just fixed it with str(value) Answers: username_1: I want to remember that some platform sent e.g. -20.5 (not the telnet platform). But this fix should work also for that case, but I wonder, is not the 'type' information then incorrect. I guess I would correct the issue @ main_volume to better show where the unexpected input is received. main_volume then also has incorrect type information. Do many other commands also have unexpected input ? Then go with the current patch. My 5-cents username_0: There was at least one other method which also had some typeerror. I was thinking about the type information as well, so I think your way would be cleaner. I will look into which other method(s) have to be converted to str and do a new commit. Thanks. Status: Issue closed
facebookresearch/mmf
756626735
Title: Where is the specific code of TextVQA Question: username_0: Hi, FAIR. Thank you for opening the code. I want to study and reimplement the code of TextVQA.However, I'm not familiar with this mmf system. I can't find the specific module like data loader, preprocessing, extract_image_featue... So, I need help where is the reference fo LoRRA. Thank you! Answers: username_1: Hi, - Code for TextVQA dataset is available in https://github.com/facebookresearch/mmf/tree/master/mmf/datasets/builders/textvqa - Model code for LoRRA is available in https://github.com/facebookresearch/mmf/blob/master/mmf/models/lorra.py - Required configs for the TextVQA and LoRRA are in https://github.com/facebookresearch/mmf/tree/master/projects/lorra Let us know if something is confusing or not clear. username_0: @username_1 Hi, username_1. Thank you for your kind reply. I have some questions. I succeeded to run the extract_features_vmb.py by looking at many issues in your repository. 1. Could you provide the final extracted image features ? it is because it takes so many times since my GPU resources are not enough. 2. Could I know the part of preprocessing the question tokens and OCR tokens? 3. Could I get the pre-trained LoRRA model? 4. Could you let me know how to run the test and evaluation? Sorry for the many questions ... I'm so confusing the integrated repository system. Thank you :) username_1: 1. If you download TextVQA dataset using MMF, it will automatically download the features you are looking for. 2. For preprocessing, please check the config you want in the folder I shared above and look for the processors configuration. Map this to processors present in `mmf/datasets/processors` folder. 3. We don't have pretrained model for LoRRA directly usable with current mmf as it was superseded by better model M4C which we do have inside MMF. 4. Please check the documentation for M4C on how to run M4C. https://mmf.sh/docs/projects/m4c username_0: @username_1 Thank you. It was very helpful. I closed this issue :+1: Status: Issue closed username_2: closing as the author of this issue suggested.
deezer/spleeter
527623429
Title: [Bug] Tuple formatting incorrectly included in output directory name Question: username_0: ## Step to reproduce ```python from spleeter.spleeter.separator import Separator separator = Separator('spleeter:2stems') filein = 'GilScottHeron_WeAlmostLostDetroit.mp3' fileout = './stems' separator.separate_to_file(filein, fileout, codec='mp3') ``` ## Output Output directory name : `./stems/('GilScottHeron_WeAlmostLostDetroit', '.mp3')/` Expected output: `./stems/GilScottHeron_WeAlmostLostDetroit/` ## Environment <!-- Fill the following table --> | | | | ----------------- | ------------------------------- | | OS | MacOS | | Installation type | `pip` | | RAM available | 8 GB | | Hardware spec | CPU: 3.2 GHz Intel Core i5, GPU: NVIDIA GeForce GT 755M 1 GB | ## Additional context The reason for this bug is [line 124 in `separator.py`](https://github.com/deezer/spleeter/blob/85ff00797f6c615c62885793923eca952e9e791f/spleeter/separator.py#L124). There needs to be a `[0]` added after the output of `splitext` so that the directory name is created from a `string`, not a `tuple`. Status: Issue closed Answers: username_1: Thanks for spotting the bug and proposing fix :). We will integrate your PR ASAP ! username_0: Great! Thanks for making this awesome package available!
KeplerGO/ScientificOpportunities
377970221
Title: Investigating Flicker noise across the Kepler/K2 sample Question: username_0: As discussed in #17 and raised by @kstassun, "Flicker" noise is an excellent way to characterize stellar properties using photometric time series. Flicker noise occurs due to the granulation on the surface of a star, and can be used to measure properties such as the surface gravity. It is possible to identify Flicker noise both in the frequency domain (where it is a "background" to asteroseismic acoustic oscillations) and by fitting in the time domain. From @ktstassun in #17: `...Bastien et al. (2013, 2016) have shown that from the Kepler long-cadence light curves it is possible to extract the granulation "flicker" that correlates very strongly with stellar surface gravity and thus provides a means for measuring stellar surface gravity with a precision of ~0.1 dex, for stars with logg > 2.5. More recent work demonstrates that the addition of metallicity as a term in the fit enables the surface gravity precision to be improved to ~0.05 dex (Corsaro et al. 2017, Tayar et al. 2018). Application of these methods to the full Kepler + K2 data set holds the promise of enabling the determination of precise stellar properties for stars far beyond the original Kepler footprint. Finally, overlap of the Kepler/K2 sample with upcoming TESS observations should enable calibration and extension of the granulation "flicker" methodology to TESS stars across the entire sky (see, e.g., Stassun et al. 2018).` Applying an analysis of Flicker noise across the entire Kepler sample, perhaps capitalizing on new methods in machine learning, would provide a catalog of robust, independently determined stellar properties using solely the Kepler sample. Answers: username_0: If anyone has any further comments on this, including how many stars may be amenable to analysis of Flicker noise, and how many of those may or may not have already been analyzed, please do comment here.
Framinus/roam
279174401
Title: User Name and Profile Pic missing from review detail page. Question: username_0: Every post should show the username and profile pic, just like they do on the cities page. Status: Issue closed Answers: username_0: Just copied over mini-profile div from the cities page with slight style modifications and it worked! need to fix width on full post view, but otherwise good.
awslabs/aws-sdk-ios-samples
275121761
Title: How to implement custom authentication using only Mobile Number Question: username_0: Hi, I would like to implement user login using just mobile number (password less). Could you please advice how I could achieve this. I am specifically looking for sample code for iOS swift or at least guidance on this. Signin/Signup flow 1. User submits his mobile number 2. User will receive pass code 3. Upon entering the passcode, the user shall be signed in Answers: username_1: I have encounter the same issue as well. Any sample code on this? I saw Android sdk already support custom challenge. username_1: After many trial and error, I have been able to sign in with custom challenge using the sdk. I have documented it here: https://medium.com/@username_1/aws-cognito-user-pools-with-mobile-sdk-for-ios-using-custom-challenge-5d40a06a3b07 username_2: I am having the same issue as well, @username_1 the link you've posted is expired, can you please give the example code to login from mobile number..? username_2: FYI user.getSession is giving an error with code=20 message="invalid pass or username" username_3: Hi @username_1, I am facing the same issue while sign in with mobile number using custom challenge. Can you please provide any sample regarding this? username_1: @username_3 @username_2 Please checkout the updated link here: https://medium.com/@username_1/aws-cognito-user-pools-with-mobile-sdk-for-ios-using-custom-challenge-df2c3b163d3d username_4: Hello @username_0, sorry for the delay. Do you still need help with the issue?
VoltDB/voltdb-client-go
377810529
Title: Passing VoltTable as an argument to stored procedure Question: username_0: Hi, Based on the documentation a stored procedure can accept these types as arguments: Integer types | byte, short, int, long, Byte, Short, Integer, and Long Floating point types | float, double, Float, Double Fixed decimal types | BigDecimal String and binary types | String and byte[] Timestamp types | org.voltdb.types.TimestampType java.util.Date, java.sql.Date, java.sql.Timestamp **VoltDB type | VoltTable** The VoltTable type is unexported in this client so it cannot be used for stored procedure invocations. Based on the code the encoder may not even handle this argument. I'd be very grateful if You could share some information on this matter. Thank You, Best Regards, Roland Answers: username_1: Hi Roland, The wire protocol supports serialization for the basic data types for input parameters, plus VoltTable for results coming back from the database. Since the java client library uses some of the same code as the database, it can serialize a VoltTable to the byte code used by the wire protocol. Therefore, at some point it was decided to allow VoltTable to be supported as input to a stored procedure. It is very rarely used to my knowledge. I'm not sure if any other client libraries support the creation of a VoltTable from code, and the serialization of a VoltTable onto the wire to be sent to the database as the parameter to a stored procedure invocation. Perhaps the python client can do it, I'm not certain. I think for most of the other client libraries, it's just not implemented. Best regards, Ben username_0: I understand, thank You, Ben! Status: Issue closed
TarikHuber/react-most-wanted
355696655
Title: Can't create new components that take advantage of react-intl Question: username_0: Hi! First off, impressive work, especially for my current usecase, I already write a react app with react-intl but my webpack build didn't work well enough and this projects fits perfectly. I wanted to attach a Header and Footer component and I'm getting the following error: ` <IntlProvider> needs to exist in the component ancestry` Normally i would see the provider but I coudn't find it anywhere on the src I'm adding the Header component to the index inside the router, as it should accept components out side the switch, everything seems fine until I try to use the intl `injectIntl(Header)` and add the references required Is there a configuration I'm missing? Thanks in advantage Answers: username_1: Hi @username_0 , Thx 😄 The IntlProvider is set in the `rwm-shell` Root component. That one is in the App component we use in our project so everything outside of that will have no IntlProvider. Here are two possible solutions for you: 1 - Try to use your Header and Footer inside the App component of `rmw-shell` and with the routes switch. Maybe you can explain more what your goal is. 2 - Add a IntlProvider on your own outside the App component. Downside of this one is that you have to manage the language settings and messages on your own. But from the `react-intl` prespective it is no problem to have more Providers inside each other. Let me know if this could help. Status: Issue closed username_2: Hey @username_1, unfortunately if I change the landingPage to `lazy(() => import('../pages/Home/Home')),` in `config.js` it returns the Error `Error: [React Intl] Could not find required `intl` object. <IntlProvider> needs to exist in the component ancestry.`. I know the cause of this error but don't understand how to fix it. The shell's code is quite hard to read for somebody who is not familiar with node username_1: Hi. Don't make the landing page async. It is the first page to load and an asyn does not make sense. After the first load it is already offline first and after that it doesn't matter anyway. If the landing page is to big you can splitt it but the first part should be sync. The demo page of rmw js madr like that. We even load the rest only on a scroll event ;) username_2: @username_1 I did not make it async, it was already async as I downloaded RMW. `config.js` ```Javascript pages: { LandingPage: import('../pages/Home/Home'), PageNotFound: lazy(() => import('../pages/PageNotFound/PageNotFound')), } ``` does not work as well :( username_1: That is not async. Only the one with lazy is async. username_2: @username_1 Ok, but how do I change the landingpage to eg. Home? username_1: Can you explain more your use case? username_2: @username_1 I added a `<Redirect.../>` to `/home` in the landing page and set every route in `routes.js` to an `unauthorizedRoute`
EarthSystemDiagnostics/cpt-picarr
492071438
Title: UI testing Question: username_0: At the moment, bugs occur frequently and can be introduced easily because there is no comprehensive UI testing with `shinytest`. I suggest adding one shinytest-file per page. - [ ] add an example project that is used for testing to the repo - [ ] make it possible to override the BASE_PATH using an environment variable (used for testing) - [ ] add the shinytest scripts Answers: username_0: It is unclear whether or not this applies to our app. We use modules and helper files, however we also have a file `app.R` that brings all the application logic together. Looking further I found [this issue](https://github.com/rstudio/shinytest/issues/3) where someone is encountering issues when trying to use shinytest with modules. The maintainers did not answer whether shinytest can work with modules. I also found [this blog post](https://www.r-bloggers.com/shiny-modules/) that makes it seem like shinytest and modules can be used together. In the blog post they use a different directory structure (app.R in inst/app/; modules in R/). Maybe using the different directory structure can solve our issue. ## Conclusion It is unclear if we can use `shinytest` for UI testing. Changing the directory structure may help. We can also [open an issue](https://github.com/rstudio/shinytest/issues) and ask for help from the maintainers. username_0: Update: I opened an issue in the shinytest repository (https://github.com/rstudio/shinytest/issues/283) username_0: @username_1 I have not received an answer from the shinytest maintainers, yet. It will be difficult to find the time to experiment with ui testing to try and somehow make it work in the limited time we have left. Therefore I would suggest to not do any ui testing. Quite a few functions are already being tested with testthat, we will have to rely on manual testing to make sure the UI works. Is that fine with you or would you prefer to take the time and try to get shinytest to work? username_1: I am fine with no doing any UI testing now. Status: Issue closed
kurodakazumichi/NoJobDungeon
655224513
Title: プレイヤーが攻撃して敵が攻撃をくらって死ぬ処理 Question: username_0: - プレイヤーが攻撃ボタンを押す - プレイヤーが攻撃の動きをする - 攻撃をくらった敵が「痛い!」みたいな演出をする - 攻撃を食らった敵が死ぬ 攻撃力とかHPとか細かい事は後で考えるとして 大まかな流れの実装 Answers: username_0: - [ ] PlayerManager.HasOnMovePlayerとEnemyManager.HasOnMoveEnemy HasOnMoveXXXXといいつつ、実際にはアイドル状態になってない奴が存在するかどうかの判定なので HasActivePlayer、HasActiveEnemyの方に変える - [x] DungeonSceneのStateMachineに設定しているUpdate処理について 大半が動いているプレイヤー、敵がいるかどうかの判定で共通処理に出来そうな雰囲気 しかし遷移先のPhaseは異なるので完全に共通化はできないので とりあえず現状維持でいこう。 - [ ] IAttackableの内容は精査して作り替えていく事になるだろう - [ ] Enemyに実装しているhpやisAcceptAttackの処理のリファクタリング hpなどはStatusクラスなど別クラスにまとめて持たせる形にしていきたい Status: Issue closed
ShammyLevva/FTAnalyzer
1142757567
Title: Custom Filters not working Question: username_0: v8.5.2.0 On the Main Lists Tab, Individuals Report Looking at Budgie Code or Relation to Root ![image](https://user-images.githubusercontent.com/29726075/154659178-ddcdffaf-9932-4c41-9094-b84f40e08dee.png) When you use the custom filter i.e. Begins with, and enter a value, it produces 0 results This snapshot is from v8.5.0.0beta3 where the image is clearer. ![image](https://user-images.githubusercontent.com/29726075/154659930-3f6c59ba-470b-413b-9760-b55cd989dc0d.png)
innoveit/react-native-ble-manager
316071204
Title: Issue running with react-native-navigation Question: username_0: - react-native-ble-manager v6.2.9 - react-native 0.54.4 - Android 8.0.0 I am somewhat new to react-native so be kind. I was able to get the react-native-ble-manager example working on my local hardware with relatively few issues. From there I wanted to integrate this into my current app; I of course used the example application as a template to accomplish this. What I noticed is that on Android the BleManager.start({showAlert: false}); would result in: `TypeError: Cannot read property 'start' of undefined` Even though it was specifically imported at the top of the view. I also noticed that when I would console log NativeModules, BleManager was not one of the modules that was included in the list. Note these issues do not occur when running iOS (in emulation granted). When exploring a bit deeper the main difference between the example app and my app is I'm using react-native-navigation by Wix (https://wix.github.io/react-native-navigation/#/). I'm using this so I can have a login screen that then navigates to a tabbed application. This requires MainApplication to extend a different class: `package com.macaw; import android.app.Application; import com.facebook.react.ReactNativeHost; import com.facebook.react.ReactPackage; import com.facebook.react.shell.MainReactPackage; import com.facebook.soloader.SoLoader; import com.reactnativenavigation.NavigationApplication; import it.innove.BleManagerPackage; import java.util.Arrays; import java.util.List; public class MainApplication extends NavigationApplication { private final ReactNativeHost mReactNativeHost = new ReactNativeHost(this) { @Override public boolean getUseDeveloperSupport() { return BuildConfig.DEBUG; } @Override protected List<ReactPackage> getPackages() { return Arrays.<ReactPackage>asList( new MainReactPackage(), new BleManagerPackage() ); } @Override protected String getJSMainModuleName() { return "index"; } }; @Override public ReactNativeHost getReactNativeHost() { return mReactNativeHost; } @Override [Truncated] super.onCreate(); SoLoader.init(this, /* native exopackage */ false); } @Override public boolean isDebug(){ return BuildConfig.DEBUG; } @Override public List<ReactPackage> createAdditionalReactPackages(){ return null; } } ` I'm wondering if this could perhaps be causing the issue when running on Android. Are these two npm mutually exclusive? Is there a way to get them to play nicely together? Thoughts? Thanks, Clark Answers: username_1: Hi, I think you only miss something in the installation, try again. Status: Issue closed
ecederstrand/exchangelib
413873188
Title: can't pickle _thread.lock objects Question: username_0: Am trying to process emails (exchangelib messages) parallely using Python multiprocessing but got "can't pickle _thread.lock objects" error. How to release this lock? Answers: username_0: Installing exchangelib 1.12.2 fix the problem. Now, I get new error: "AttributeError: 'FileAttachment' object has no attribute '__dict__'. username_1: Please post some example code that causes this error. username_0: Sample code ... Aim: to process emails parallely def func(emails): do sth # emails = {emailId:email} # emailId: 1 ... 100 for hundred emails # emails: exchangelib message jobs = [] nthreads = 4 jobEach = len(emails)//nthreads firstEmailId = list(emails.keys())[0] lastEmailId = list(emails.keys())[-1] for i in range(nthreads): if i == nthreads - 1: emailIds = [(jobEach*i)+firstEmailId,lastEmailId] emailsProcessor = {emailId:emails[emailId] for emailId in emailIds} else: emailIds = [(jobEach*i)+firstEmailId,(jobEach*(i+1))+firstEmailId] emailsProcessor = {emailId:emails[emailId] for emailId in emailIds} p = multiprocessing.Process(target=func, args=(emailsProcessor)) jobs.append(p) # print(emailIds) for p in jobs: p.start() for p in jobs: p.join() Error message (not full) : ... File "C:\Users\~\AppData\Local\Continuum\anaconda3\lib\multiprocessing\reduction.py", line 60, in dump ForkingPickler(file, protocol).dump(obj) File "C:\Users\~\AppData\Local\Continuum\anaconda3\lib\site-packages\exchangelib\attachments.py", line 188, in __getstate__ state = self.__dict__.copy() AttributeError: 'FileAttachment' object has no attribute '__dict__' username_1: The issue happens when the `Message` items are being serialized and de-serialized when they are passed to the processes. In general with multiprocessing, it's much better to pass around simple strings and integers instead of full objects, and then load the full objects within the process run method. One reason is to avoid trying to pickle something that cannot be pickled. Another reason is to avoid passing potentially huge amounts of data around in IPC channels. For example, attachments can be very large. In your case, you should query message IDs in your main thread and then fetch the full items in the run method. So, do something like: ```python from multiprocessing import Pool pool = Pool(processes=10) ids = account.inbox.all().values('id', 'changekey') def worker(item_ids): items = account.fetch(item_ids) # Do something with the items pool.map(worker, ids, chunksize=100) ``` That said, it is a bug that attachments cannot be pickled. I have now pushed a fix and added a test for this bug. username_0: Thanks. Since I don't want all inbox to be fetched, I did this: account = ... def worker(item_ids): items = account.fetch(item_ids) for item in items: print(item.subject) if __name__ == '__main__': SomeFolder = ... dateToCheck1 = EWSDateTime(2018, 5, 10, 10, 00, 00,0, pytz.UTC) dateToCheck2 = EWSDateTime(2018, 5, 10, 20, 15, 00,0, pytz.UTC) pool = Pool(processes=4) ids = SomeFolder.filter(datetime_received__range=(dateToCheck1,dateToCheck2)).values('id', 'changekey')[:5] pool.map(worker, ids, chunksize=5) but throws this error (runs fine with out the multiprocessing): ... return self._map_async(func, iterable, mapstar, chunksize).get() File "~\AppData\Local\Continuum\anaconda3\lib\multiprocessing\pool.py", line 644, in get raise self._value TypeError: unhashable type: 'slice' username_1: You may need to materialize the list of IDs: ```python pool.map(worker, list(ids), chunksize=10) ``` username_0: Thanks. List didn't work for me but user defined iterator slice using tuple works. Status: Issue closed
Mercury1089/2018-robot-code
294134032
Title: Tune PID Values Question: username_0: ## GOAL Tune PID values for when it is used on carpet. ## ACCEPTABLE CRITERIA Tune PID for the following commands: - [ ] RotateRelative - [ ] DriveDistance - [ ] MoveOnPath Answers: username_1: Just marking this done.... Status: Issue closed
kamuiroeru/nitac-nenpo-latex
538207653
Title: Warningが消えない Question: username_0: ``` Package caption Warning: Unsupported document class (or package) detected, usage of the caption package is not recommended. ``` キャプションのフォントが **ゴシック** じゃないとダメって指定なので、仕方なく ***captionパッケージ*** 使ってるんだけど、それが非推奨なのでワーニングが出てるみたいです。 直せたら直したいが、めんどくさいので放置…
madhawav/YOLO3-4-Py
341991359
Title: SIGSEGV when creating a Detector Question: username_0: Hi ! First, thank you for your work ! I was first using the darknet python script, but with really poor performances. As yours seems better I am trying to use it but get a segfault each time I try to instantiate a Detector. I have first installed your lib through PyPi, but it did not work. Then building it from source with my own build of Darknet, was successful when installing, but not working on execution, with always a segfault error. Apparently after running valgrind it may concern memory allocation ... ``` Access not within mapped region at address 0x0 ==18723== at 0x16254939: free_network (in /home/pbreton/.local/lib/python3.6/site-packages/__libdarknet/libdarknet.so) ==18723== by 0x15FE2FF8: __pyx_pf_9pydarknet_8Detector_6__dealloc__ (pydarknet.cpp:3157) ==18723== by 0x15FE2FF8: __pyx_pw_9pydarknet_8Detector_7__dealloc__ (pydarknet.cpp:3142) ==18723== by 0x15FE2FF8: __pyx_tp_dealloc_9pydarknet_Detector(_object*) (pydarknet.cpp:5861) ==18723== by 0x15FE6FC7: __pyx_tp_new_9pydarknet_Detector(_typeobject*, _object*, _object*) (pydarknet.cpp:5847) ==18723== by 0x5591FD4: ??? (in /usr/lib/libpython3.6m.so.1.0) ==18723== by 0x558A64B: _PyObject_FastCallDict (in /usr/lib/libpython3.6m.so.1.0) ==18723== by 0x5552B41: ??? (in /usr/lib/libpython3.6m.so.1.0) ==18723== by 0x5522859: _PyEval_EvalFrameDefault (in /usr/lib/libpython3.6m.so.1.0) ==18723== by 0x5552FC7: PyEval_EvalCodeEx (in /usr/lib/libpython3.6m.so.1.0) ==18723== by 0x552240B: PyEval_EvalCode (in /usr/lib/libpython3.6m.so.1.0) ==18723== by 0x55F2213: ??? (in /usr/lib/libpython3.6m.so.1.0) ==18723== by 0x55F50AD: PyRun_FileExFlags (in /usr/lib/libpython3.6m.so.1.0) ==18723== by 0x55F5294: PyRun_SimpleFileExFlags (in /usr/lib/libpython3.6m.so.1.0) ``` Do you have any idea ...? Thank you ... Answers: username_0: Everything is fine concerning the command line, I checked myself and manually everything works. The `pkg-config` is successfully returning the correct opencv libs : ```-I/usr/include/opencv``` ```-lopencv_stitching -lopencv_superres -lopencv_videostab -lopencv_aruco -lopencv_bgsegm -lopencv_bioinspired -lopencv_ccalib -lopencv_dnn_objdetect -lopencv_dpm -lopencv_face -lopencv_photo -lopencv_freetype -lopencv_fuzzy -lopencv_hdf -lopencv_hfs -lopencv_img_hash -lopencv_line_descriptor -lopencv_optflow -lopencv_reg -lopencv_rgbd -lopencv_saliency -lopencv_stereo -lopencv_structured_light -lopencv_phase_unwrapping -lopencv_surface_matching -lopencv_tracking -lopencv_datasets -lopencv_text -lopencv_dnn -lopencv_plot -lopencv_xfeatures2d -lopencv_shape -lopencv_video -lopencv_ml -lopencv_ximgproc -lopencv_calib3d -lopencv_features2d -lopencv_highgui -lopencv_videoio -lopencv_flann -lopencv_xobjdetect -lopencv_imgcodecs -lopencv_objdetect -lopencv_xphoto -lopencv_imgproc -lopencv_core``` The problem seems to be concerning the way python is accessing the libraries. Status: Issue closed username_0: Ok it is truly my fault. I used 3 parameters instead of 4, forgetting that the third is a 0. Thus with 3 bytes as parameters it was logically crashing. Thank you for your time,
HodorNV/ALOps
568269846
Title: How-to avoid generating an artifact when doing a build for a specific branch Question: username_0: Is there an option that I can add in one of the build steps in the yaml file to specify that the artifact(s) should **not** be created if the following criteria is met: - Branch = develop - Build reason = Pull Request build Answers: username_1: I don't think there is - you could create a dedicated build pipeline, obviously ;-). If you work with yaml-templates, it's rather efficient, still (reusing yaml-configurations in multiple pipelines..). username_0: ok Status: Issue closed
jooby-project/jooby
760514090
Title: The jooby-apt should not require jaxrs jars Question: username_0: For annotation processing you don't actually need the annotation jars to be on the classpath/annotation class path. jooby-apt has a provided dependency on `javax.ws.rs` which unfortunately has the negative side effect pulling that into your classpath. Consequently you can very easily accidentally import `javax.ws.rs.GET` instead of jooby's. The only place that JAX RS classes are actually accessed for annotation processing is here https://github.com/jooby-project/jooby/blob/2.x/modules/jooby-apt/src/main/java/io/jooby/apt/Annotations.java but that is just to get the classname. I recommend just hardcoding the string and then making a unit test to check if the canonical names match the actual canonical javax.ws.rs (I mean its unlikely those names will ever change but I guess if you want to be safe). Because of the above you also don't need jooby-apt to have a provided dependency on `jakarta.ws.rs-api` since those who use the jars will have it on their classpath anyway.<issue_closed> Status: Issue closed
NativeScript/NativeScript
1149054743
Title: Slow Build times on MacBook Pro M1 32GB RAM Question: username_0: ### Issue Description After migrating my projects to a new MacBook Pro M1 my builds seem very slow. NS Doctor:- ✔ Your ANDROID_HOME environment variable is set and points to correct directory. ✔ Your adb from the Android SDK is correctly installed. ✔ The Android SDK is installed. ✔ A compatible Android SDK for compilation is found. ✔ Javac is installed and is configured properly. ✔ The Java Development Kit (JDK) is installed and is configured properly. ✔ Xcode is installed and is configured properly. ✔ xcodeproj is installed and is configured properly. ✔ CocoaPods are installed. ✔ CocoaPods update is not required. ✔ CocoaPods are configured properly. ✔ Your current CocoaPods version is newer than 1.0.0. ✔ Python installed and configured correctly. ✔ The Python 'six' package is found. ✔ Xcode version 13.2.1 satisfies minimum required version 10. ✔ Getting NativeScript components versions information... ✔ Component nativescript has 8.1.5 version and is up to date. ✔ Component @nativescript/core has 8.1.5 version and is up to date. ✔ Component @nativescript/ios has 8.1.0 version and is up to date. ✔ Component @nativescript/android has 8.1.1 version and is up to date. My previous MacBook Pro was intel based with 16GB RAM and was much quicker building. I've cleaned the project, updated plugins etc but build time are still slower. ### Reproduction _No response_ ### Relevant log output (if applicable) _No response_ ### Environment OS: macOS 12.2.1 CPU: (8) x64 Apple M1 Pro Shell: /bin/zsh node: 14.11.0 npm: 7.11.2 nativescript: 8.1.5 # android java: 1.8.0_172 ndk: Not Found apis: 25, 27, 28 build_tools: 25.0.2, 27.0.3, 28.0.3 system_images: - android-28 | Google APIs Intel x86 Atom # ios xcode: 13.2.1/13C100 cocoapods: 1.11.2 python: 2.7.18 [Truncated] "reflect-metadata": "0.1.13", "rxjs": "6.5.4", "rxjs-compat": "6.5.4", "zone.js": "0.10.2" }, "devDependencies": { "@nativescript/android": "~8.1.1", "@nativescript/ios": "8.1.0", "@nativescript/types": "~8.1.1", "@nativescript/webpack": "~5.0.0-beta.14", "typescript": "~4.0.0" } ``` ### Please accept these terms - [X] I have searched the [existing issues](https://github.com/NativeScript/NativeScript/issues) as well as [StackOverflow](https://stackoverflow.com/questions/tagged/nativescript) and this has not been posted before - [X] This is a bug report - [X] I agree to follow this project's [Code of Conduct](https://github.com/NativeScript/NativeScript/blob/master/tools/notes/CONTRIBUTING.md#coc) Answers: username_1: Everything 1st party in NativeScript works without rosetta, so definitely try native node and running without rosetta to see if that makes a difference. username_0: Updated Node and now have v17.6.0 Running ns doctor does not complete and stops at getting Nativescript components No issues were detected. ✔ Xcode is installed and is configured properly. ✔ xcodeproj is installed and is configured properly. ✔ CocoaPods are installed. ✔ CocoaPods update is not required. ✔ CocoaPods are configured properly. ✔ Your current CocoaPods version is newer than 1.0.0. ✔ Python installed and configured correctly. ✔ The Python 'six' package is found. ✔ Xcode version 13.2.1 satisfies minimum required version 10. ✔ Getting NativeScript components versions information... ns doctor stops at the line above. when entering ns --version this error occurs 8.1.5 ⠼ Checking for updates...Error: Command failed: npm view nativescript dist-tags --json node:internal/modules/cjs/loader:936 throw err; ^ Error: Cannot find module '../lib/cli.js' Require stack: - /usr/local/lib/node_modules/npm/bin/npm-cli.js at Function.Module._resolveFilename (node:internal/modules/cjs/loader:933:15) at Function.Module._load (node:internal/modules/cjs/loader:778:27) at Module.require (node:internal/modules/cjs/loader:999:19) at require (node:internal/modules/cjs/helpers:102:18) at Object.<anonymous> (/usr/local/lib/node_modules/npm/bin/npm-cli.js:2:1) at Module._compile (node:internal/modules/cjs/loader:1097:14) at Object.Module._extensions..js (node:internal/modules/cjs/loader:1151:10) at Module.load (node:internal/modules/cjs/loader:975:32) at Function.Module._load (node:internal/modules/cjs/loader:822:12) at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:77:12) { code: 'MODULE_NOT_FOUND', requireStack: [ '/usr/local/lib/node_modules/npm/bin/npm-cli.js' ] } Node.js v17.6.0 at Errors.failWithOptions (/usr/local/lib/node_modules/nativescript/lib/common/errors.js:157:27) at Errors.fail (/usr/local/lib/node_modules/nativescript/lib/common/errors.js:130:21) at NodePackageManager.<anonymous> (/usr/local/lib/node_modules/nativescript/lib/node-package-manager.js:104:30) at Generator.throw (<anonymous>) at rejected (/usr/local/lib/node_modules/nativescript/lib/node-package-manager.js:12:65) ⠼ Checking for updates...% username_2: @username_0 seems like you have an incomplete installation, as that error is coming straight from npm itself. Most likely you updated node and is still using different versions of npm or a dirty npm package folder (from the old installation). I suggest switching to nvm instead of using a global node installation to avoid these issues. username_0: Getting the same error whether I use Terminal with or without Rosetta. Not sure how this manifested itself as I can no longer compile apps. Compilation stops at the last line of the error above: - If increasing the memory doesn't solve the issue, it's most probably a bug in the TypeScript or EsLint.
freeCodeCamp/testable-projects-fcc
491063523
Title: Pomodoro Clock User Stories update Question: username_0: But actually, to pass test case 8, I need to update the time-left content too with the new value. So please, modify to help fellow coders to get a clear idea about they have to do. Thanks for your time & effort. #### Browser Information <!-- Describe your workspace in which you are having issues--> * Browser Name, Version: * Operating System: * Mobile, Desktop, or Tablet: #### Your Code / Link to Your Pen <!-- Paste relevant code in here, or a link to your pen which would be most helpful --> ``` ``` #### Screenshot <!-- Add a screenshot of your issue --> Status: Issue closed Answers: username_1: For more visibility, please open issues on the [main repo](https://github.com/freeCodeCamp/freeCodeCamp/issues) for these. Thanks and happy coding 🎉
Vector35/binaryninja-api
405771851
Title: “Error initializing database: Data too large” when saving Analysis Database for a large binary Question: username_0: Binary Ninja Version: 1.1.1470 Platform: macOS 10.14.3, 32G RAM When loading a large binary (about 81MB) with Binary Ninja, it takes about 18GB of RAM. Furthermore, saving Analysis Database for this binary ends on the following error: ![8f075d3e350b84f3a2dff3362984cbd7](https://user-images.githubusercontent.com/1687847/52134556-798a7d80-264c-11e9-9715-f15481a20974.jpg) Any idea how to successfully save the database? I can provide/upload the binary if you need. Answers: username_1: Unfortunately this is a known issue that I guess we don't already have a bug for. Large files fail to save due to some considerations SQLite's maximum size limitations. There is a SQLite feature that exists we're just currently not using it. username_0: Hi. Thanks for your response. Is there any ETA for the fix? Status: Issue closed username_2: Fixed as of 1.1.1745-dev.
gchq/gaffer-experimental
1042472478
Title: Run ui tests against npm 16 Question: username_0: Currently the CI runs UI tests against node 14 only: https://github.com/gchq/gaffer-experimental/blob/16bcbe46e0700417ce16840455892483e4db4eb5/.github/workflows/ci.yaml#L32 Would be nice if it used a matrix to test multiple versions of node, including, like the [ci in gaffer-tools](https://github.com/gchq/gaffer-tools/blob/60ebc8bb6c61e114a24c381ee998e42bd0d15553/.github/workflows/continuous-integration.yaml#L62-L69). As well as this, there are lots of warnings when running `npm install` and it would be nice if these were cleaned up.<issue_closed> Status: Issue closed
open-telemetry/opentelemetry-python
647697973
Title: Proposal: Use GitHub Actions for CI/CD Question: username_0: # Consistent CI/CD Repo Issue Template ### Title Proposal: Use GitHub Actions for CI/CD ### Description This issue is in reference to [issue 398](https://github.com/open-telemetry/community/issues/398) posted on the community repository. In this issue, we are proposing that all OpenTelemetry repositories *consider* using GitHub Actions as their CI provider in order to maintain consistency across the various language repositories. The overall proposal was discussed in the OpenTelemetry maintainers SIG meeting. @trask has been assigned as the mentor for the project. |Repository |CI Provider |Automated Build and Test |Code Coverage |Automated Performance Testing |Automated Deployment |Automated Docs Deployment | |--- |--- |--- |--- |--- |--- |--- | |Python |Travis/CircleCI |[x] |[x] |[] |[x] |[x] | The justification and benefits are enumerated in the issue on the community repository and are pasted here as well for convenience: ## Proposal We propose that all languages consider using the same CI provider. This would create a more consistent development process and make it easier for developers to contribute to multiple language libraries. We suggest that provider be GitHub Actions. Here’s why: ### **Ease-of-Use** CircleCI will automatically run when pull requests and commits are issued against the repository. But if a contributor forks the repository, unless they set up an account with the CI provider and link it to their forked repository, CI will not be activated and tests will not be run automatically. In contrast, GitHub Actions works out of the box on a forked repository and can be easily configured to run a test workflow each time a commit is issued. This would help individual contributors test their code and ensure code quality before submitting a pull request against the repository. ### **Transparency** Current CI providers such as CircleCI and Travis allow anyone to view the console output when building and running tests but the test results can not be seen anywhere on the GitHub repository. To view this testing output: You need go to a different website, navigate a different user interface, and then sift through thousands of lines of console output. This is not a seamless developer experience. In contrast, using GitHub Actions would provide all testing output directly on the repository’s GitHub page, which would help contributors to find, read, and use the test output to maintain code quality. ### **Control** GitHub Actions’ integration with other GitHub features means you can have finer control over the CI pipeline. For example, certain workflows can be set to only run on a new release. Workflows can even be used to close stale issues and pull requests. ### Recommendation We recommend that we consider using one consistent CI provider, GitHub Actions, which provides an integrated and seamless developer experience for all contributors. ### Example Please see [this example](https://github.com/open-telemetry/opentelemetry-cpp/blob/master/.github/workflows/ci.yml) that the C++ repository has adopted for the above reasons. ### Next Steps This issue shall serve as a place for discussion about this proposal. Could a maintainer please assign this issue to us if approved? Answers: username_1: @username_0 I took a brief look at github actions and it seemed not to support older versions of Python out of the box. Have you had a chance to look at whether we can support the full matrix of versions we're testing today? One of the other benefits of using CircleCI was the ability to cache tox environments, is this something that's achievable with github actions? I'd love to see an example of this. See PR https://github.com/open-telemetry/opentelemetry-python/pull/828 for more details on the circle CI configuration. username_2: The [reference](https://docs.github.com/en/actions/reference/software-installed-on-github-hosted-runners) only lists the following as preinstalled on Ubuntu 20.04 LTS: ``` Python: Python 2.7.18 Python 3.5.9 Python 3.6.11 Python 3.7.8 Python 3.8.5 PyPy: PyPy 2.7.13 [PyPy 7.3.1 with GCC 7.3.1 20180303 (Red Hat 7.3.1-5)] PyPy 3.6.9 [PyPy 7.3.1 with GCC 7.3.1 20180303 (Red Hat 7.3.1-5)] ``` But [actions/python-versions](https://github.com/actions/python-versions) seems to support the following: ```console $ curl -s https://raw.githubusercontent.com/actions/python-versions/main/versions-manifest.json | jq ".[].version" "3.9.0-beta.5" "3.9.0-beta.4" "3.8.5" "3.8.4" "3.8.3" "3.8.2" "3.8.1" "3.8.0" "3.7.8" "3.7.7" "3.7.6" "3.7.5" "3.6.11" "3.6.10" "3.6.9" "3.6.8" "3.6.7" "3.5.9" "3.5.4" "3.4.10" "3.3.7" "2.7.18" "2.7.17" ``` Not sure about caching. username_2: @username_1 I have not gotten it to work as expected yet (see PoC #984), but caching should work, according to the [migration guide](https://docs.github.com/en/actions/migrating-to-github-actions/migrating-from-circleci-to-github-actions#caching). username_2: As an additional argument for switching to GitHub Actions: After setting up my CircleCI on my fork I noticed that the checks on my pull requests did not complete. Apparently, if a user follows their own fork, CircleCI does not run these jobs on the main repository, even for pull requests. This results in the checks on the pull requests forever _waiting for status to be reported_. See: [Why aren't pull requests triggering jobs on my organization?](https://support.circleci.com/hc/en-us/articles/360008097173-Why-aren-t-pull-requests-triggering-jobs-on-my-organization-) Status: Issue closed
meowtec/Imagine
256444169
Title: Compilation error Question: username_0: on `npm run dev` run get ``` modules/renderer/store/reducer.ts(58,49): error TS2345: Argument of type '{ [x: string]: ((state: ITaskItem[], action: Action<ITaskAddPayloadItem[]>) => ITaskItem[]) | ((s...' is not assignable to parameter of type 'ReducerMap<ITaskItem[], ITaskItem[]>'. Index signatures are incompatible. Type '((state: ITaskItem[], action: Action<ITaskAddPayloadItem[]>) => ITaskItem[]) | ((state: ITaskItem...' is not assignable to type 'Reducer<ITaskItem[], ITaskItem[]> | ReducerNextThrow<ITaskItem[], ITaskItem[]>'. Type '(state: ITaskItem[], action: Action<string[]>) => ITaskItem[]' is not assignable to type 'Reducer<ITaskItem[], ITaskItem[]> | ReducerNextThrow<ITaskItem[], ITaskItem[]>'. Type '(state: ITaskItem[], action: Action<string[]>) => ITaskItem[]' has no properties in common with type 'ReducerNextThrow<ITaskItem[], ITaskItem[]>'. modules/renderer/store/reducer.ts(122,55): error TS2345: Argument of type '{ [x: string]: ((state: IGlobals, action: Action<string>) => { activeId: string | undefined; upda...' is not assignable to parameter of type 'ReducerMap<IGlobals, IGlobals>'. Index signatures are incompatible. Type '((state: IGlobals, action: Action<string>) => { activeId: string | undefined; updateInfo?: IUpdat...' is not assignable to type 'Reducer<IGlobals, IGlobals> | ReducerNextThrow<IGlobals, IGlobals>'. Type '(state: IGlobals, action: Action<string>) => { activeId: string | undefined; updateInfo?: IUpdate...' is not assignable to type 'Reducer<IGlobals, IGlobals> | ReducerNextThrow<IGlobals, IGlobals>'. Type '(state: IGlobals, action: Action<string>) => { activeId: string | undefined; updateInfo?: IUpdate...' has no properties in common with type 'ReducerNextThrow<IGlobals, IGlobals>'. ``` Answers: username_1: TypeScript and TSLint should use fixed versions. 3ef0dc9 Now pull and try again. username_0: it fix, thanks Status: Issue closed
wojtekmaj/react-calendar
580566520
Title: programmatic reset (question) Question: username_0: Hi Using v3.0.0 Is there a way to reset the value or range programmatically? Meaning, removing the selection/s, associated tile classes, etc Thank you Answers: username_1: v3.0.0 can be both controlled and uncontrolled. If you need to control the value from the outside, controlled would be the best approach. Then it's as easy as clearing value prop. username_0: Hi, Is there documentation and / or examples of what "controlled would be the best approach" means? Thank you username_0: To be more specific, I used `value={rangeValue}` and change it onClickDay and onChange, that works, but if I programmatically set `rangeValue=null` is doesn't remove the tile classes, so it all looks as selected. username_0: Sorry, my fault, using the following just works: `const [rangeValue, setRangeValue] = useState(null);` Thank you again Status: Issue closed
flutter/flutter
343944029
Title: How do I prohibit font size from changing with the font size of the mobile phone? Question: username_0: Can I set a maximum and minimum font interval to prevent page elements from being confused? ![simulator screen shot - iphone x - 2018-07-24 at 16 43 52](https://user-images.githubusercontent.com/12829578/43127391-f44c7528-8f61-11e8-8d4a-2cb3f107811f.png) ![simulator screen shot - iphone x - 2018-07-24 at 16 44 46](https://user-images.githubusercontent.com/12829578/43127392-f48300ac-8f61-11e8-881d-a3b588516adb.png) ![simulator screen shot - iphone x - 2018-07-24 at 16 44 56](https://user-images.githubusercontent.com/12829578/43127393-f4b8cbd8-8f61-11e8-829a-404b722733d6.png) Answers: username_1: Please consider asking support questions in one of the other channels listed at http://flutter.io/support . username_1: You could wrap your app with a custom `MediaQuery` where you override the default https://docs.flutter.io/flutter/widgets/MediaQueryData/textScaleFactor.html See also - https://docs.flutter.io/flutter/dart-ui/Window/textScaleFactor.html - https://docs.flutter.io/flutter/dart-ui/Window/onTextScaleFactorChanged.html Please consider asking support questions in one of the other channels listed at http://flutter.io/support . Status: Issue closed username_2: MaterialApp( builder: (BuildContext context, Widget child) { return MediaQuery( data: MediaQuery.of(context).copyWith(textScaleFactor: textScaleFactor), child: child, ); }, title: 'Home Page', ); This should do the trick~ username_3: Isn't making `textScaleFactor: 1.0` much easier than using `MediaQuery` stuff? For example, the following will prohibit scaling based on system's accessibility settings: ```dart Text( 'hello', textScaleFactor: 1.0, ), ``` username_4: No, I think you should use `MediaQuery` Because of your method doesn't change textField label sizes and etc
PIVX-Project/PIVX
604230852
Title: [GUI][Visual bug] Pull down menu / transactions disapears. Question: username_0: <!--- Remove this description and sections that do not apply --> This issue tracker is only for technical issues related to PIVX Core. General PIVX questions and/or support requests and are best directed to the [PIVX Discord](https://discord.pivx.org). ### Describe the issue Please refer to the first screenshot below and note all 9 pull-down Transactions Types. Randomly select one Transactions Type, followed by another, and a third/forth/fifth. At some point the pull-down menu disappears (refer to the second screenshot below). The only way to recover (to my knowledge) is to close/open the wallet. ### Can you reliably reproduce the issue? #### If so, please list the steps to reproduce below: I have reproduced this issue twice today. Randomly select a Transaction Type in the pull-down menu and ensure the proper information is displayed. Repeat several times (4-6) with different Transaction Types. At some point the pull-down menu will disappear. ### Screenshots. If the issue is related to the GUI, screenshots can be added to this issue via drag & drop. ![ice_screenshot_20200421-212341](https://user-images.githubusercontent.com/59299109/79905514-ac1dc300-8416-11ea-91fa-49d53b05b532.png) ![ice_screenshot_20200420-001502](https://user-images.githubusercontent.com/59299109/79905485-9f996a80-8416-11ea-9f31-3d913bf50e9d.png) ### What version of PIVX Core are you using? List the version number/commit ID, and if it is an official binary, self compiled or a distribution package. Self compiled 4.1.0 ### Machine specs: - OS: Windows 10 - CPU: i9 - RAM: 8GB - Disk size: 512 GB - Disk Type (HD/SDD): SDD Answers: username_1: possibly related to #1172 username_2: This was solved in #2073, closing. Status: Issue closed
open-austin/iced-coffee
138799743
Title: [Code Across] Follow up emails with more info Question: username_0: - Google Slides - Link to Slack - Link to Github issues with Code Across tag - Upcoming events Answers: username_1: Can I get a link to the slides? I want to post them to Meetup. username_0: https://drive.google.com/open?id=1aWNB5wWr9ZIjwvMNEZR2uoGJcK9zDKuFw0_B1JIhFBk username_2: Here is the view only link: https://docs.google.com/presentation/d/1aWNB5wWr9ZIjwvMNEZR2uoGJcK9zDKuFw0_B1JIhFBk/edit?usp=sharing Status: Issue closed username_0: <img width="475" alt="screen shot 2016-03-07 at 9 18 34 am" src="https://cloud.githubusercontent.com/assets/5697474/13573515/a83e3802-e445-11e5-988e-fc0501d72e83.png">
Redbility/paper-lightbox
240404668
Title: fix demo when doing polymer serve Question: username_0: minor thing but when doing `polymer serve` on this element to do some investigation it comes up 'not found'. I know that you can add demo to the path to get it to show up but still. ![2017-07-04_08-19-47](https://user-images.githubusercontent.com/329735/27829891-d71c8984-6091-11e7-878c-3929e8a713e0.png)
kukko/web-wombat
415075674
Title: Is the repo open for contribution? Question: username_0: Hey! Thanks for the presentation yesterday, well done! Are you accepting PRs? What is the style you are using for commits? I'm not yet familiar with the tags used: fix/enhancement/progress. Cheers, Balázs Answers: username_1: Hello! Thanks for your attendance yesterday! Yeah, I tottally accept PRs. In th beginning, I would merge pull requests, which is not conflicting with my future plans, about this repo. So If you want to make a feature, please contact with me, to prevent needless coding, which will not be merged into the main repo. Keep coding! Your sincerely, username_1. Status: Issue closed username_0: Thanks for the answer! I opened a PR for some minor typo fixes, I was guessing that can go without a separate issue beforhand. Let's see how that goes. :) Closing this issue.
microsoft/CCF
731242732
Title: Upgrade Evercrypt to 0.3.0 Question: username_0: Evercrypt has its first official release, we should upgrade to that: https://github.com/project-everest/hacl-star/releases/tag/v0.3.0 Thanks @wintersteiger for the heads up! Answers: username_0: #1903 revealed that the sha256 implementation we use from Evercrypt now is slower than mbedtls. @wintersteiger helped me take a look at why, and that's because we no longer get the fast assembly version, which is only enabled for chips that have SHA extensions (something no modern Intel CPUs feature, and no SGX-enabled CPUs ever supported). One of the things we want to watch out for in the upgrade is that we do get a fast SHA implementation. Status: Issue closed
tekHudson/DruidBarClassic
502940918
Title: "Bliz-Like" Text Display Inconsistencies and Suggested Improvements Question: username_0: DruidBarTextLeft displays with a slightly incorrect x axis offset. Changing Line 50 of `DruidBar.xml` from `<AbsDimension x="5" y="0"/>` to `<AbsDimension x="7" y="0"/>` fixes this. `DruidBarTextRight` is set to the value of `ManaValues()` on Line 342 of `DruidBar.lua`, which returns both the current mana and maximum mana separated by a slash. On the Blizzard UI only the current mana is shown. Changing Line 342 from `DruidBarTextRight:SetText(ManaValues());` to `DruidBarTextRight:SetText(floor(DruidBarKey.currentmana));` fixes this, suggest making a new way to handle the requests for mana values by the UI so that current mana can be retrieved separately, or separated from maximum mana. Before Image: https://i.imgur.com/l3vynL8.png After Image: https://i.imgur.com/QYafDzN.png Answers: username_1: Meged with #17 Status: Issue closed
uber/baseweb
524971737
Title: [Notification] Force line break in message Question: username_0: <!--- Provide a general summary of the issue in the Title above --> <!--- Provide a codesandbox that reproduces your issue - you can fork one from https://baseweb.design/ --> Example at https://codesandbox.io/s/notification-qiwp0 ## Current Behavior Inserting `\\n`in the string doesn't line break the message. <!--- Describe what happens instead of the expected behavior. --> ## Expected Behavior Should give a line break at the position where I input `\\n`. <!--- Describe what should happen. --> ## Your Environment <!--- Include as many relevant details about the environment with which you experienced the bug. --> | Tech | Version | | ------- | ------- | | Base UI | v9.17.0 | | React | v16.11.0 | | browser | Chrome | - [x] I have searched the [issues](https://github.com/uber/baseweb/issues) of this repository and believe that this is not a duplicate. Answers: username_1: You can do this https://codesandbox.io/s/notification-dls2y?fontsize=14&hidenavigation=1&theme=dark Use it with caution though. If the message is determined using user input, there's a possibility the user can inject arbitrary scripts that could compromise security. username_0: Realised that you could do it like in this codesandbox: https://codesandbox.io/s/notification-jde02 No need for the `dangerouslySetInnerHTML`. Status: Issue closed username_2: I believe a more correct and cleaner way would be to use an override and use `white-space: pre-line` instead of `white-space: pre`. Something like: ``` overrides: { Body: { style: { whiteSpace: 'pre-line', }, }, }, ``` This will also work in a `toaster` where you cannot use a function with a child `div`.
hbz/link-templates
220933464
Title: TH Köln OPAC link is broken Question: username_0: For example https://katalog.bibl.fh-koeln.de/webOPACClient/start.do?Query=0010=%22HT001045964%22 Answers: username_0: Also Duisburg-Essen, see http://primo.ub.uni-due.de/primo_library/libweb/action/dlSearch.do?vid=UDE&institution=UDE&search_scope=localude&bulkSize=10&lang=ger&indx=1&onCampus=false&query=any,contains,HT002391361 . Other example: http://lobid.org/resources/HT002526560# username_0: We should also test the other organizations. username_0: TH Köln oviously switched to DigiBib, see https://www.th-koeln.de/hochschulbibliothek/katalog_24926.php. :-) Status: Issue closed
jlippold/tweakCompatible
556596714
Title: `libcolorpicker` working on iOS 13.3.1 Question: username_0: ``` { "packageId": "org.thebigboss.libcolorpicker", "action": "working", "userInfo": { "arch32": false, "packageId": "org.thebigboss.libcolorpicker", "deviceId": "iPhone10,5", "url": "http://cydia.saurik.com/package/org.thebigboss.libcolorpicker/", "iOSVersion": "13.3.1", "packageVersionIndexed": true, "packageName": "libcolorpicker", "category": "Development", "repository": "BigBoss", "name": "libcolorpicker", "installed": "1.6.7", "packageIndexed": true, "packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.", "id": "org.thebigboss.libcolorpicker", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.1.5", "shortDescription": "color picker library for developers", "latest": "1.6.7", "author": "PixelFire", "packageStatus": "Unknown" }, "base64": "<KEY> "chosenStatus": "working", "notes": "" } ```<issue_closed> Status: Issue closed
neuronsimulator/nrn
934890000
Title: Hoc Object deletion not occurring with combination of deleted Section, expect_err, and PointProcess. Question: username_0: ## Context The following code does not delete IClamp[0] on exit from foo. ``` from neuron import h from neuron.expect_hocerr import expect_err, set_quiet set_quiet(False) s = h.Section() def foo(): ic = h.IClamp(s(.5)) h.delete_section(sec=s) expect_err('ic.amp = .1') #del ic # problem goes away if this and following statements are uncommented #print (locals()) foo() h.allobjects() ``` prints: ``` CHECKING: ic.amp = .1 NEURON: point process not located in a section near line 0 objref hoc_obj_[2] ^ No value IClamp[0] with 1 refs ParallelContext[0] with 1 refs ``` Note: ignore the fact that ParallelContext[0] exists. That is created by the expect_hocerr module. If the last two lines of def foo(): are uncommented then we get the correct output: ``` CHECKING: ic.amp = .1 NEURON: point process not located in a section near line 0 objref hoc_obj_[2] ^ No value {} ParallelContext[0] with 1 refs ``` ### Overview of the issue I find expect_err to be very useful for testing error messages. However if, because of its use, objects are not properly deleted then that can cause problems during later tests with ```python -m pytest ...``` ### Expected result/behavior IClamp[0] should not exist whether or not the last two lines of foo are commented. Answers: username_0: When experiencing this problem, a general work around is to ```del``` all the offending local variables at the end of the function and then call ```locals()```. It remains a puzzle why ```del``` is needed but an even greater puzzle why the mere call to ```locals()``` is also needed.
MicrosoftDocs/azure-docs
610897969
Title: Under Linux section - Configure agent communication typo Question: username_0: Command line should only be in the following format azcmagent connect --resource-group "&lt;resourceGroupName&gt;" --tenant-id "&lt;tenantID&gt;" --location "&lt;regionName&gt;" --subscription-id "&lt;subscriptionID&gt;" --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 6036c6de-a2e8-ffd7-50e9-44ae32bf1682 * Version Independent ID: f7ab9abc-abd0-09e1-4b64-9ec304495ee5 * Content: [Connect hybrid machines to Azure from the Azure portal](https://docs.microsoft.com/en-us/azure/azure-arc/servers/onboard-portal#feedback) * Content Source: [articles/azure-arc/servers/onboard-portal.md](https://github.com/Microsoft/azure-docs/blob/master/articles/azure-arc/servers/onboard-portal.md) * Service: **azure-arc** * Sub-service: **azure-arc-servers** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte** Answers: username_1: @username_0 Thanks for the feedback! We are currently investigating and will update you shortly. username_2: @username_0 Updated the referred document with the required changes. The changes will go live within 24 hours. Thanks #please-close Status: Issue closed
lasso-js/lasso
238003224
Title: Mention Sass in the description Question: username_0: When I see this, I notice you don't mention Sass, which over 70% of web devs use (Less is down to sub-5%, and Stylus is still kicking it at around the 1% mark, and Marko is another internal eBay project that isn't well known). I immediately assumed you didn't support Sass and realized Lasso is just some internal-use eBay thing, and they must be stuck using outdated Less, or have some weird affiliation with the creator of Stylus or something. Clearly not a product for general use. I was just about to close the tab when I decided to Ctrl+F for `sass` and saw a plugin for it. Anyways, the way you associate yourselves to other technologies effects the perception people have of your project. Namedropping obscure projects isn't a good way to obtain adopters. I would either remove the example all together, or replace them with more commonly used tools (like Sass in this instance). But in a broader light, just be more aware of the associations people will make and how it will cause them to make assumptions about your project.
PHP-DI/PHP-DI
355916971
Title: Support for variadic arguments Question: username_0: PHP-DI does not currently support variadic arguments for autowiring. It will only inject the first element to the variadic parameter in a constructor, silently discarding the rest. For example for this class: ```PHP class UserProfile implements Timeline\Engine { public function __construct(DbConnection $dbConnection, TimelineModule ...$modules) { ``` cannot be autowired by: ```PHP Engine::class => \DI\autowire(UserProfile::class) ->constructor( \DI\get(DbConnection::class), \DI\get(Module\CasinoFollow::class), \DI\get(Module\UserFollow::class), \DI\get(Module\Review::class) ), ``` but must be instead autowired by: ```PHP UserProfile::class => function(ContainerInterface $c) { return new UserProfile( $c->get(DbConnection::class), $c->get(Module\CasinoFollow::class), $c->get(Module\UserFollow::class), $c->get(Module\Review::class) ); }, ``` which in turn reduces the efficiency of caching I believe and is harder to debug. Answers: username_1: Thanks for the detailed issue! Can you confirm that: ```php Engine::class => \DI\autowire(UserProfile::class) ->constructor( \DI\get(DbConnection::class), \DI\get(Module\CasinoFollow::class), \DI\get(Module\UserFollow::class), \DI\get(Module\Review::class) ), ``` doesn't work and that it injects only `$dbConnection` and the first module? username_0: Indeed, it only injects the first module: https://i.imgur.com/BANmom5.png username_0: I would like to start working on this, but I don't feel so accustomed to PHP-DI internals yet. Do you have any tips @username_1 for me? Where should I start? I have written some preliminary tests for this behavior. username_1: @username_0 that's great! I suggest you have a look here: https://github.com/PHP-DI/PHP-DI/blob/master/src/Definition/Resolver/ParameterResolver.php#L51 You can see that we do `foreach` on parameters returned by PHP's reflection. So a variadic parameter will be interpreted as 1 parameter only. Try to write unit tests but more importantly functional tests in here: https://github.com/PHP-DI/PHP-DI/blob/master/tests/IntegrationTest/Definitions/CreateDefinitionTest.php username_1: @username_0 yes that is awesome thanks! I had a quick look last week but because of conferences I didn't have enough time to review it carefully (this PR is more complex that other simple PRs :) ) I'll try to have a look at it this week. username_2: Need this feature too. username_2: Is it possible to make it work the same way with constructorParameter() method too? To be able to extend definition for one param only? Example: ``` class MyClass { public function __construct(Foo $foo, Bar $bar, Baz $baz, ...$args) {} } MyClass::class => DI\autowire()->constructorParameter('args', ...['foo', 'bar', 'baz', 'etc']), ``` username_2: Not quite what I would like to do (-: Now I can avoid defining other constructor arguments and have them autowired only this way: ``` class MyClass { public function __construct(Foo $foo, Bar $bar, Baz $baz, $args) {} } [ MyClass::class => DI\autowire() ->constructorParameter('args', ['a', 'b', 'c', 'etc']) ] ``` But I would like to type-hint $args: `public function __construct(Foo $foo, Bar $bar, Baz $baz, string ...$args) {}` And define it like this: ``` [ MyClass::class => DI\autowire() ->constructorParameter('args', ...['a', 'b', 'c', 'etc']) ] ``` username_0: You could do it this way (with the patch merged): ```PHP class MyClass { public function __construct(Foo $foo, Bar $bar, Baz $baz, string ...$args) {} } MyClass::class => DI\autowire() ->constructorParameter(3, 'a') ->constructorParameter(4, 'b') ->constructorParameter(5, 'b') ``` Parameters 0 to 2 (types of Foo, Bar and Baz) would be autowired automatically without defining them as you'd expect with PHP-DI. I do not personally prefer this ```PHP MyClass::class => DI\autowire() ->constructorParameter('args', ...['a', 'b', 'c', 'etc']) ``` to work as you described, because that would show the user in the constructorParameter method signature that multiple parameters are valid to any constructor parameter, when the situation with PHP is that only the last parameter can be variadic. For example: user might think based on method signature and documentation that this would be valid: ```PHP MyClass::class => DI\autowire() ->constructorParameter('bar', ...['a', 'b', 'c', 'etc']) ->constructorParameter('args', ...['a', 'b', 'c', 'etc']) ``` When obviously the part where bar is defined would not work. I understand that you'd perhaps want to use the parameter name instead of the argument number for clarity, but I don't think it's a good idea. username_1: I had a quick look at this and I think it might be worth implementing the same rules as PHP 8's named parameters: https://wiki.php.net/rfc/named_params (over all the PHP-DI features) The integration with variadics is clearly defined (and constrained) in there, so it could be both a good lead to follow, as well as minimize confusion and friction down the road.
github-vet/rangeloop-pointer-findings
774906043
Title: yzs981130/oyashirosama2: controllers/testjob_controller.go; 24 LoC Question: username_0: [Click here to see the code in its original context.](https://github.com/yzs981130/oyashirosama2/blob/c0c27bdd26dccfbd67eafa87bcfada4a6e0bc7f7/controllers/testjob_controller.go#L121-L144) <details> <summary>Click here to show the 24 line(s) of Go which triggered the analyzer.</summary> ```go for i, job := range childJobs.Items { _, finishedType := isJobFinished(&job) switch finishedType { case "": // ongoing activeJobs = append(activeJobs, &childJobs.Items[i]) case schedulev1.JobFailed: failedJobs = append(failedJobs, &childJobs.Items[i]) case schedulev1.JobComplete: successfulJobs = append(successfulJobs, &childJobs.Items[i]) } scheduledTimeForJob, err := getScheduledTimeForJob(&job) if err != nil { log.Error(err, "unable to parse schedule time for child job", "job", &job) continue } if scheduledTimeForJob != nil { if mostRecentTime == nil { mostRecentTime = scheduledTimeForJob } else if mostRecentTime.Before(*scheduledTimeForJob) { mostRecentTime = scheduledTimeForJob } } } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` The following dot graph describes paths through the callgraph that could lead to a function calling a goroutine: no paths found; call may have ended in third-party code; stay tuned for diagnostics ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: c0c27bdd26dccfbd67eafa87bcfada4a6e0bc7f7
Mangodream01/ISMI_project
771965468
Title: Dataset Question: username_0: Hello Michel, I have walked through your code on segmenting liver, liver cancer and spleen and have found it to well presented and informative. However, in trying to replicate the work, I have realized that, the data set is not provided neither is any link provided to the it. Could you please assist me by providing either the link to the data or the data set itself? Thanks and looking forward to your rapid response. Answers: username_1: Hi @username_0, We used data from the Medical Segmentation Decathlon (http://medicaldecathlon.com/). You can download the data there, we used the data from task 3 (Liver) and task 9 (Spleen). Hope all is clear and goodluck! username_0: Hello @username_1 , Thanks for your rapid reply and assistance. It is indeed helpful. Cheers! Status: Issue closed
vklochan/python-logstash
64301310
Title: SSL Support? Question: username_0: Hi I want to know if python-logstash has SSL support, because i cannot find any way to define custom CA certificate. Thanks Answers: username_1: Same issue here, TLS/SSL support would be nice. username_0: ping username_2: I had a quick look at the code to see if this was possible. It looks to me like the easiest way to achieve this is to implement `makeSocket()` in the `TCPLogstashHandler` class as has been done in @klynch's [logstash_handler](https://github.com/klynch/python-logstash-handler/blob/master/logstash_handler/__init__.py#L17) library. @vklochan would you be likely to accept a patch that does something like that? username_3: Ping for @vklochan The possibility to encrypt logs is a super useful feature! I think the one proposed by @username_2 is an easy-winning patch... If you want I can help out username_4: +1 please.
rancher-sandbox/rancher-desktop
1082974421
Title: [macOS] Stuck in Waiting for Kubernetes after update to 0.7 Question: username_0: ### Rancher Desktop Version 0.7 ### Rancher Desktop K8s Version 1.21.7 ### What operating system are you using? macOS ### Operating System / Build Version macOS Monterey ### What CPU architecture are you using? arm64 (Apple Silicon) ### Windows User Only _No response_ ### Actual Behavior After manually updated my app from 0.7.0-beta to 0.7.0 (replace /Applications/Rancher\ Desktop.app), it stuck on Waiting for Kubernetes API... forever. I check the k3s.log and got `invalid node ip, unable to get global unicast ip from interface name: can't find ip for interface rd0.` My active network interface is `en0`, and checkout that interface in file `~/Library/Application\ Support/rancher-desktop/lima/_config/networks.yaml` is `en0`, which means no need to change. I stop and restart Rancher Desktop but nothing fixed. ### Steps to Reproduce Update the app version to 0.7.0. ### Result Stuck on `Waiting for Kubernetes API...`. ### Expected Behavior Start K3S successful. ### Additional Information Here I paste the file content: ```yaml paths: vdeSwitch: /opt/rancher-desktop/bin/vde_switch vdeVMNet: /opt/rancher-desktop/bin/vde_vmnet varRun: /private/var/run/rancher-desktop-lima sudoers: /private/etc/sudoers.d/rancher-desktop-lima group: staff networks: shared: mode: shared gateway: 192.168.205.1 dhcpEnd: 192.168.205.254 netmask: 255.255.255.0 bridged: mode: bridged interface: en0 host: mode: host gateway: 192.168.206.1 dhcpEnd: 192.168.206.254 netmask: 255.255.255.0 ``` And log files [log.tar.gz](https://github.com/rancher-sandbox/rancher-desktop/files/7733385/log.tar.gz) Status: Issue closed Answers: username_0: Well that actually works, after I switch an wifi ssid. username_1: For those who may run into this problem, like I did. I am not using wifi. I am using an ethernet adapter, en6. I ended up changing `rd0` to `en6` in ~/Library/Application Support/rancher-desktop/lima/_config/0.yaml ~/Library/Application Support/rancher-desktop/lima/0/lima.yaml
briansmith/ring
794560623
Title: PBKDF2 and other APIs should be redesigned to avoid `out: &mut [u8]` parameters. Question: username_0: Here's the current PBKDF2 API: ``` pub fn derive( algorithm: Algorithm, iterations: NonZeroU32, salt: &[u8], secret: &[u8], out: &mut [u8], ) { ``` This was one of the first APIs written for *ring*, about 5 years ago, and it should be improved. The API uses `out: &,mut [u8]` because it wants to support `alloc`-less environments and it doesn't want to depend on crates like `SmallVec` or similar. We should use what we've learned about Rust and how Rust has improved to fix this to be something roughly like this: ``` pub fn derive( algorithm: Algorithm, iterations: NonZeroU32, salt: &[u8], secret: &[u8], ) -> T where T: Default + Extend<u8> { ``` See also issue #987 about allowing `secret` to be encapsulated.
tc39/proposal-temporal
978510576
Title: Use cases and sample code for multi-Temporal-type string parser Question: username_0: To better think through #1751, I wanted to understanding the long tail of ISO-string-parsing use cases better. I also wanted to understand how well `from` methods support those parsing use cases. Here's my sense of those cases: * **Impose stricter control on string inputs.** e.g. require minutes to be specified in ISO time strings, or require the time to be included to be specified in a date/time string * **Clarify user intent.** e.g. differentiating Z strings vs. local+offset strings in `ZonedDateTime.from` * **Handle partially-invalid/partially-valid ISO strings.** Temporal doesn't support this case today, because AFAIK invalid ISO strings throw even if some parts are valid. * **Generically parse ISO strings for use in non-Temporal code.** In this case, the user wants a parser to chop up an ISO string into parts so they can do whatever they want with those parts (e.g. custom parsing, display to users, etc.) Temporal doesn't support this case today. Are there others? To validate Temporal's `from` methods for those first two use cases, as an exercise I wrote a simple parser with the goal of relying solely on `from` methods of various Temporal types to extract information about the original string. ### Issues Found This effort helped me discover some gaps in Temporal's support for the first two parsing use-cases above, where "gaps" means that `from` was insufficient and I had to look inside the string was required to get info I was interested in. #1765 seems like a clear bug in the polyfill and/or spec. Maybe #1766 too. The other two seem like use cases we could support with new `from` options in a V2 if there's enough demand. * PlainTime will parse anything, e.g. 01-01 or 2020-01-01 or 2020-01 - #1765 * Can't distinguish default ISO calendar from user appending [u-ca=iso8601] - #1766 * Can't distinguish Z from +00:00 when parsing strings with TZ annotations - #1767 * Can't determine if the original string had minutes or smaller time units - #1769 ### Parser Prototype I figured it'd be useful to file this issue to provide sample code for others who may want to do something similar in the future. Some quirks: * I worked around the top 3 issues above using simple string parsing. The workarounds I added aren't bulletproof, but if we choose not to fix those issues then it may be worth using better regexes for those cases. * I was too lazy to work around #1769. This could be an exercise for the reader. 😄 * Time zone IDs, calendar IDs, and offsets are normalized to their canonical format. There's no way using the output of `from` to now what the original input was. I didn't think this was important enough to make into the list of issues above. * Date-only strings can be parsed by PlainDateTime and ZonedDateTime, but the output of the parser doesn't include time units because someone using a multi-type parser probably wants to know that the input was only a date. * The output includes properties like `monthCode` and `era` even though those are never in an ISO string. This seemed to be the right thing to enable the resulting object to be used as a property bag to initialize a Temporal type like PlainMonthDay. * Speaking of PlainMonthDay, it has no `month` property so the result of parsing `01-01` has no `month` property either, only `monthCode`, `day`, `plainMonthDay`, and (if calendar annotation is present) a `calendar` property. * Even though ISO is the default calendar when parsing strings, no `calendar` property is returned unless an annotation is present in the string. This lets callers to know if the string has a calendar annotation or not. * `offset` is returned as a string, which may be "Z" or a numeric offset. * As discussed in #1751, if an instant string (no TZ annotation) uses a "Z" offset, my opinion is that the components are meaningless, for the same reason that we don't offer `year`, `month`, `day` etc. properties on Instant. The code below therefore limits the output for Instant Z string to just `{instant, offset: "Z"}`. **Parser Code** ```js /** * Parse an ISO string into Temporal instances and their components: date units, * time units, offset, calendar, and time zone * * Issues found: * * #1765 - PlainTime will parse anything, e.g. 01-01 or 2020-01-01 or 2020-01 * * #1766 - Can't distinguish default ISO calendar from user appending [u-ca=iso8601] * * #1767 - Can't distinguish Z from +00:00 when parsing strings with TZ annotations * * #1769 - No way to know if minutes and smaller units were present in a time string * * @param s * @returns An object with: * * a property for each Temporal type parsed successfully, e.g. `zonedDateTime`, * `plainDate`, `plainTime` * * the following properties from those successfully-parsed instances: * 'year`, `eraYear`, `era`, `month`, `monthCode`, `week`, `day`, `hour`, * `minute`, `second`, `millisecond`, `microsecond`, `nanosecond' * * a `calendar` property if a bracketed calendar annotation was in the string * * a `timeZone` property if a bracketed time zone annotation was in the string * * an `offset` property ("Z" or numeric offset string) if an offset was in the string */ function parse(s) { const types = [ [Truncated] "plainMonthDay": "01-01", "year": 2020, "month": 1, "monthCode": "M01", "day": 1 }, "2020-01": { "plainYearMonth": "2020-01", "year": 2020, "month": 1, "monthCode": "M01" }, "01-01": { "plainMonthDay": "01-01", "monthCode": "M01", "day": 1 }, "bogus": {} } ```
OpenFn/Miracle-Feet
1010823797
Title: Turn on/off SMS alerts based on `Account.Status__c` field Question: username_0: ## Background, context, and business value The clinics (Salesforce `Account`-s) MiracleFeets works with can be temporarily or permanently suspended, which affects if patients (SF `Contact`-s) should receive SMSs, and which ones. Clininc status is set in SF. Depending on the clinic's status in SF, we want to deactivate some/all SMS-s for patients associated with that clinic. ## The specific request, in as few words as possible We have 2 different flows based on clinic status which can be handled in the same job: Flow #1 ([data flow diagram](https://lucid.app/lucidchart/9454d9ca-7c35-482d-b9e9-0e41284d1281/edit?page=3kldohKkc3rC#)) 1. Query Salesforce `Account` records where "Status" = "Previously Supported " 2. Find all `Contact` records for each of the `Account` records 3. Disable all SMS for these Contacts. Flow #2 1. Query Salesforce `Account` records where "Status" = "Temporarily Suspended" 2. Find all `Contact` records for each of the `Account` records 3. Disable Alert [#17, 18, 19](https://docs.google.com/spreadsheets/d/1quhQJgQkVRC8oObDzkwgnnm-Rov5BGOW85I4YqcNV0I/edit?pli=1#gid=262234774&range=112:114) for these Contacts. ## state.json Either provide state directly, or link to a file. If sensitive information should be in state, redact it and provide instructions for where it can be found. ```json { "configuration": { "username": "abc", "password": "<PASSWORD>" }, "data": { "a": 1 }, "cursor": "2020-01-19 00:00:00" } ``` ```json { "configuration": ["SEE LAST PASS: 'client cred'"], "data": { "a": 1 }, "cursor": "2020-01-19 00:00:00" } ``` ## adaptor List the adaptor to be used for this job. If changes must be made to the adaptor, explain why existing functions dont work and specify the new API you'd like from a helper function. ```md There is no "upsert" in postgres. I'd like an API where I can provide the table, the UUID, and some data to upsert. Like this: upsert('some_table', 'some_column', state.data.records); ``` ## expression.js In pseudocode, either in the current job expression or in a new file, describe as best you can what changes need to be made ```js each( [Truncated] ``` ## output.json Either provide the output you'd like, or describe it in terms of final state and side effects. ### side effects 1. upsert new records to postgres ### output.json ```json { "configuration": {}, "data": { "statusCode": 200 }, "references": { "a": 1 } } ``` Answers: username_1: @username_0 For step 3 (Disable SMSs)... how will this be done in Infobip? A SF query like this will return a list of Contacts... but is there more info we need from SF in order to find the corresponding contacts in Infobip? Do Infobip Contacts maybe have the CommCare `case_id` or some other externalId we need in order to build the `bulkId` to match on? ``` Select Id, Account from Contact WHERE Account.Status__c = "Previously Supported" ``` username_0: @username_1 We would need to cycle through all alert `[bulkId](https://docs.google.com/spreadsheets/d/1quhQJgQkVRC8oObDzkwgnnm-Rov5BGOW85I4YqcNV0I/edit?pli=1#gid=262234774&range=M:M)`-s constructed from the `case_id`. Same method as we use when a patient opts out of all SMS-s. username_1: @username_0 2 test clinics and 3 test patients have been setup in sandbox. @username_2 moving to the bottom of the backlog. When you get to this, keep us posted on questions... and Rita will also be around tomorrow afternoon for a quick chat if needed. username_2: @username_0 I have tried to fetch the bulkIds above you scheduled but they are not found on postman username_0: @username_2 I scheduled the following bulkIds: `casting_intro-3-1234567` -> `case_id: 1234567` `bracing_day-6-345678` -> `345678` `bracing_night-6-7777777` -> `7777777` `visitBefore-234567-2021-12-24` username_2: I pushed draft for the 2 jobs in the meantime. The blocker for now is to get the visit date for some bulkIds. username_0: @username_2 I added a visit in Infobip and SF: `visitBefore-345678-2021-12-24` for caseId `345678` to access it: ``` Select Id, Next_Visit_Date__c from Visit_new__c where Patient__r.CommCare_Case_ID__c = ${case_id} AND Next_Visit_Date__c !=null ``` This should return a list of `Next_Visit_Date__c` values for a selected patient username_2: @username_0 I made the update. username_0: @username_2 getting a syntax(?) error with disableSMS.js, can you check? ```Line 1: fn(state => { ^^ Function not available. ``` see run https://www.openfn.org/projects/pdbznd/runs/0615c2fa-66f8-7cbc-adc8-33c501cce5ca username_0: @username_2 Now we're using `language-http v3.1.11`. I tested with Infobip credentials and set up a new one too to test, but same error https://www.openfn.org/projects/pdbznd/runs/0615c474-e7d8-72ca-aec3-8ad59448ceef username_2: @username_0 feel free to re-run after update on the credentials. I don't have access to it username_2: @username_0 @username_1 find the run here. I add more log to separate for more precision: https://www.openfn.org/projects/pdbznd/runs/0615d873-d4c3-77ea-930a-ade52bfd0428 username_1: @username_0 please close whenever you've tested this
llimllib/limbo
151907713
Title: New user can't execute bot command Question: username_0: The bot is running and then a new user is being added to the team. When the new user is trying to execute the bot command he/she will trigger the KeyError because `server.slack.server.users` hasn't been updated when the user got added. -> https://github.com/username_1/limbo/blob/master/limbo/limbo.py#L131 Answers: username_1: ooh good catch, I hadn't considered that case! Thanks so much for reporting, fixing it will be a top priority when I get time to work on this. username_2: I'd encountered this before but wasn't sire what caused it. I knew a restart fixed it. A similar issue happens when you add the bot to a new private group on the first time its triggered. Mind if I have a go at fixing this? username_1: Not at all, have at it username_2: Looking at the offending code, it appears to be because slackrtm pulls the list of users originally and as such the server object has a static list of users in `server.slack.server.users` When a new user enters, slackrtm doesn't update the list of users, which is giving the key error. The immediately obvious way to fix it would be to make slackrtm update the users first when this KeyError occurs, try again and then if it happens again perform the debugging. Having said that, what purpose does this line serve? (from `limbo/limbo.py ln 131`) `msguser = server.slack.server.users[event["user"]]` Can this be replaced with just `msguser = event["user"]`? As far as I can see this would server the same purpose and solve the issue, or is msguser used elsewhere that I can't find? username_1: I think this is the correct answer! We no longer need to check that the user is in the slackrtm roster, though we should fix that too because plugins could depend on that behavior. username_2: As `msguser` isn't used at all I've changed the code to the following ``` if "user" not in event: logger.debug("event {0} has no user".format(event)) return ``` I'll make the PR soon. username_2: Ok, I deviated a bit and kept some of the code closer to what it was, and I've made slackrtm update at the same time as fixing this issue (and another issue) username_1: Fixed in https://github.com/username_1/limbo/commit/6be28e4c973780922b90eb3b02b522c5b4f6e5ff Status: Issue closed
NLog/NLog.Extensions.Logging
1004617361
Title: JSON configuration always throws when throwConfigExceptions is true. Question: username_0: **NLog version**: NLog 5 preview 1 **Platform**: .NET Core 5 **Current NLog config** ```json { "NLog": { "throwConfigExceptions": true } } ``` It looks like `SetNLogElementSettings` attempts to validate the key/value pair produced by the configuration abstraction below. ![LoadConfig](https://user-images.githubusercontent.com/9957114/134315628-cf7d3b4e-15d0-44ed-9152-ce282af34bdb.png) The `name` key/value pair throws in `SetNLogElementSettings` as it is unknown. I am adding NLog to a .net core generic host using `.UseNLog()` which may be the problem? The resulting exception is: ```LOG NLog.NLogConfigurationException: Unrecognized value 'name'='NLog' for element 'NLog' at NLog.Config.LoggingConfigurationParser.SetNLogElementSettings(ILoggingConfigurationElement nlogConfig) at NLog.Config.LoggingConfigurationParser.LoadConfig(ILoggingConfigurationElement nlogConfig, String basePath) at NLog.Extensions.Logging.NLogLoggingConfiguration.LoadConfigurationSection(IConfigurationSection nlogConfig) at NLog.Extensions.Logging.NLogLoggingConfiguration..ctor(IConfigurationSection nlogConfig, LogFactory logFactory) at NLog.Extensions.Logging.RegisterNLogLoggingProvider.<>c__DisplayClass1_0.<TryLoadConfigurationFromSection>b__0(ISetupLoadConfigurationBuilder configBuilder) at NLog.SetupBuilderExtensions.LoadConfiguration(ISetupBuilder setupBuilder, Action`1 configBuilder) at NLog.Extensions.Logging.RegisterNLogLoggingProvider.TryLoadConfigurationFromSection(NLogLoggerProvider loggerProvider, IConfiguration configuration) at NLog.Extensions.Hosting.ConfigureExtensions.CreateNLogLoggerProvider(IServiceProvider serviceProvider, IConfiguration configuration, NLogProviderOptions options) at NLog.Extensions.Logging.RegisterNLogLoggingProvider.<>c__DisplayClass0_1.<TryAddNLogLoggingProvider>b__2(IServiceProvider provider, IConfiguration cfg, NLogProviderOptions opt) at NLog.Extensions.Logging.RegisterNLogLoggingProvider.<>c__DisplayClass0_0.<TryAddNLogLoggingProvider>b__4(IServiceProvider serviceProvider) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitFactory(FactoryCallSite factoryCallSite, RuntimeResolverContext context) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSiteMain(ServiceCallSite callSite, TArgument argument) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitCache(ServiceCallSite callSite, RuntimeResolverContext context, ServiceProviderEngineScope serviceProviderEngine, RuntimeResolverLock lockType) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitRootCache(ServiceCallSite singletonCallSite, RuntimeResolverContext context) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSite(ServiceCallSite callSite, TArgument argument) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitConstructor(ConstructorCallSite constructorCallSite, RuntimeResolverContext context) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSiteMain(ServiceCallSite callSite, TArgument argument) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitCache(ServiceCallSite callSite, RuntimeResolverContext context, ServiceProviderEngineScope serviceProviderEngine, RuntimeResolverLock lockType) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitRootCache(ServiceCallSite singletonCallSite, RuntimeResolverContext context) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSite(ServiceCallSite callSite, TArgument argument) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitConstructor(ConstructorCallSite constructorCallSite, RuntimeResolverContext context) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSiteMain(ServiceCallSite callSite, TArgument argument) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitCache(ServiceCallSite callSite, RuntimeResolverContext context, ServiceProviderEngineScope serviceProviderEngine, RuntimeResolverLock lockType) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitRootCache(ServiceCallSite singletonCallSite, RuntimeResolverContext context) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSite(ServiceCallSite callSite, TArgument argument) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitConstructor(ConstructorCallSite constructorCallSite, RuntimeResolverContext context) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSiteMain(ServiceCallSite callSite, TArgument argument) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitCache(ServiceCallSite callSite, RuntimeResolverContext context, ServiceProviderEngineScope serviceProviderEngine, RuntimeResolverLock lockType) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitRootCache(ServiceCallSite singletonCallSite, RuntimeResolverContext context) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSite(ServiceCallSite callSite, TArgument argument) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.Resolve(ServiceCallSite callSite, ServiceProviderEngineScope scope) at Microsoft.Extensions.DependencyInjection.ServiceLookup.DynamicServiceProviderEngine.<>c__DisplayClass1_0.<RealizeService>b__0(ServiceProviderEngineScope scope) at Microsoft.Extensions.DependencyInjection.ServiceLookup.ServiceProviderEngine.GetService(Type serviceType, ServiceProviderEngineScope serviceProviderEngineScope) at Microsoft.Extensions.DependencyInjection.ServiceLookup.ServiceProviderEngine.GetService(Type serviceType) at Microsoft.Extensions.DependencyInjection.ServiceProvider.GetService(Type serviceType) at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService(IServiceProvider provider, Type serviceType) at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService[T](IServiceProvider provider) at Microsoft.Extensions.Hosting.HostBuilder.Build() at Program.Main(String[] args) ``` Answers: username_1: @username_0 Thank you for reporting this issue. The fix will be included in NLog 5.0 preview-2 See also https://github.com/NLog/NLog.Extensions.Logging/pull/531 Status: Issue closed username_1: NLog.Extensions.Logging 5.0 preview 2 is now available: https://www.nuget.org/packages/NLog/5.0.0-preview.2
urbit/urbit
641680134
Title: Bridge Error Question: username_0: TypeError: Cannot create property 'message' on string 'Node error: {"code":-32000,"message":"already known"}' at renderAdditionalInfo (https://bridge.urbit.org/static/js/main.b6a701ac.chunk.js:2322:385) at PassportTransfer (https://bridge.urbit.org/static/js/main.b6a701ac.chunk.js:2322:1914) at dh (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:159072:7) at Jh (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:159548:7) at mj (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:161821:86) at jj (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:160809:11) at Z (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:160666:15) at $i (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:160525:16) at https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:158251:17 at push.exports.unstable_runWithPriority (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:162946:12)
openshift/origin
289784475
Title: Strange error when getting logs for terminating pod Question: username_0: 3.9 master, using json-file driver with docker 1.12.6-68 I had just deleted this pod. I get this when I call logs briefly ``` $ oc logs -n kube-system sts/prometheus -c prometheus failed to get container status {"" ""}: rpc error: code = OutOfRange desc = EOF ``` @openshift/sig-pod Answers: username_0: I'll move this to a bug if this isn't already known username_0: @username_1 if you can have someone at your team look at this - not sure what we're doing wrong, but looks broken. username_1: @username_2 PTAL username_1: xref https://bugzilla.redhat.com/show_bug.cgi?id=1537237 username_0: I *think* I opened a bug, but can't find the tab anymore. username_2: Upstream issued filed here: https://github.com/kubernetes/kubernetes/issues/59296 Upstream PR posted here: https://github.com/kubernetes/kubernetes/pull/59297
openshift/origin
413071996
Title: Is the Openshift API swagger.json exposed anywhere? Question: username_0: I know there are swagger endpoint docs accessible at `OPENSHIFT_URL/openapi/v2` and also `OPENSHIFT_URL/swagger.json` for the Kubernetes related API's. Is there also an equivalent endpoint for the Openshift api in swagger format. I see in this repo, there is this document, https://raw.githubusercontent.com/openshift/origin/master/api/swagger-spec/openshift-openapi-spec.json, but i'm not really sure if that is exposed as an endpoint in my Openshift Cluster. Answers: username_1: @username_0, as near as I can tell, for release-3.11 that file is generated by this script: https://github.com/openshift/origin/blob/release-3.11/hack/update-generated-swagger-spec.sh Out of curiosity, what were you planning on doing with spec file? username_0: We have a Node.js module, [openshift-rest-client](https://github.com/nodeshift/openshift-rest-client), that we created and have been adding functionality "manually". we are looking to refactor the library so it generates the api automatically, instead of us adding things in manually. There is a nice [kubernetes node module](https://github.com/godaddy/kubernetes-client), which does something similar. One of the ways it can load the api is querying the running clusters `/openapi/v2` or `/swagger.json` endpoint. If there isn't a similar endpoint for the Openshift api's, i can just use this file, which i've started to do, https://github.com/nodeshift/openshift-rest-client/pull/113 username_1: @username_0, that looks quite interesting. My own interest in this is using the OpenShift OpenAPI specs to create REST API docs. I'll let you know if I learn anything helpful. It does look like a running cluster is queried to generate the spec for 3.11 here: https://github.com/openshift/origin/blob/b6db8e68d78723f943eba6710ab6adc8d4254a2f/hack/update-generated-swagger-spec.sh#L40
ATMartin/TIY-Homework
59117632
Title: 2015-02-26 Question: username_0: # ember-crud ## Description A CRUD app using Ember.js ## Objectives ### Learning Objectives After completing this assignment, you should: - Understand how to separate data/syncing logic from presentation/interaction and application state using Ember.js Routes, Controllers, and Templates. - Understand how to properly write CRUD functionality using Ember.js actions. ### Performance Objectives After completing this assignment, you should be able to: * Use ember-cli to quickly generate the files needed to set up an application with multiple routes, controllers, and templates. ## Details ### Deliverables * A repo containing an ember-cli project ### Requirements * No JSHint warnings or errors ## Normal Mode Write an Ember.js application that allows you to create, update, read, and delete a resource. The assignment is open ended, but some examples of things you could make: - A blog - A bookmarking application - A contact list You may use either the localStorage adapter or Parse. ### Parse (preferred) - Use an [ember initializer](http://emberjs.com/api/classes/Ember.Application.html#method_initializer) to set up your API keys on $.ajaxSetup, which you can generate with ember-cli - Use $.ajax in your route's `model` and `actions` to GET/POST/PUT/DELETE ### localStorage - Set up the [localStorage adapter](https://www.npmjs.com/package/ember-localstorage-adapter) - Use `this.store.find(thing)`, `this.store.createRecord({})`, etc. as seen in the Todo tutorial. ## Hard Mode Implement a filtering route. For example, if you made a bookmarking app, create a route that allows you to view the bookmarks with a specific tag. ## Additional Resources * Read [the Ember guides](http://emberjs.com/guides/). Try to get through "Enumerables", but at least read through "Routing". Answers: username_1: https://github.com/username_1/TIY-HW-21-EmberCRUD Status: Issue closed username_1: At the moment, this has all CRUD features working, and has a working route to filter by category. However, I'm trying to render category names out to the app index route and am having no luck due to what I think is an implicit controller. The majority of my logic is controller-based, but I'd also like to refactor that to the route so I can be more comfortable with the MVVM vs MVC paradigm here. I'm gonna keep hacking on it with fresh eyes this morning, and if I can get the category list working then I plan on adding unique images (all placeholders right now) and auth to it next. Thanks! username_0: I think the problem is that there is only ever one active controller for a template, and it's the child-most controller (if that makes sense). So allCategories would only be available at `/`. In order to make it available elsewhere, you could add the following to that particular controller: ```js needs: ['rentit'], allCategories: Ember.computed.alias('controllers.rentit.allCategories') ``` That might not be the best way, but it's the first way that comes to mind.
hlminh2000/lifeBalance
321420554
Title: Fix com.facebook.react.common.JavascriptException in ExceptionsManagerModule.java line 56 Question: username_0: ### Version 1.0(1525833228) ### ### Stacktrace ### com.facebook.react.modules.core.ExceptionsManagerModule.showOrThrowError (ExceptionsManagerModule.java:56); com.facebook.react.modules.core.ExceptionsManagerModule.reportFatalException (ExceptionsManagerModule.java:40); com.facebook.react.bridge.JavaMethodWrapper.invoke (JavaMethodWrapper.java:374); com.facebook.react.bridge.JavaModuleWrapper.invoke (JavaModuleWrapper.java:162); com.facebook.react.bridge.queue.NativeRunnable.run (NativeRunnable.java); com.facebook.react.bridge.queue.MessageQueueThreadHandler.dispatchMessage (MessageQueueThreadHandler.java:31); com.facebook.react.bridge.queue.MessageQueueThreadImpl$3.run (MessageQueueThreadImpl.java:194); ### Reason ### com.facebook.react.common.JavascriptException ### Link to App Center ### * [https://appcenter.ms/users/username_0-pyo4/apps/lifeBalance/crashes/groups/c51b37d817263cebfbf6d06c18a78556eff56f6a](https://appcenter.ms/users/username_0-pyo4/apps/lifeBalance/crashes/groups/c51b37d817263cebfbf6d06c18a78556eff56f6a)
juliema/aTRAM
221874071
Title: Don't assume we have write access to directories Question: username_0: Use case: If you're running data built by another user then you may not have access to the --work-dir directory. Change it to use a prefix for input only (trim the names) and make an --output-dir option for the temp files.<issue_closed> Status: Issue closed
didykivoice/ride-my-way
443811633
Title: Code Formatting Question: username_0: @didykivoice Do not use more than one blank line in your html code ``` <span id="slide-line"></span> </nav> <div id="box" class="box-1"> <div id="boxtext" class="boxtext-1"> <h1>Ride-MyWay</h1> <p>fORLoop week-01 Challnge 02 UI Designing</p> <img src="img/picha.jpg" width="100%" height="70%" alt="pic"> <p>Ping a driver to get your ride</p> </div> </div> <footer class="fixed_footer"> ``` Also make sure before commit your code is formatted with 2 or 4 spaces for code indent. Good work
openebs/openebs
284908119
Title: Resolve maya-nodebot dependency on admin privilege(sudo) Question: username_0: <!-- This form is for bug reports and feature requests ONLY! --> <!-- Thanks for filing an issue! Before hitting the button, please answer these questions.--> ## Is this a BUG REPORT or FEATURE REQUEST? Choose one: BUG REPORT or FEATURE REQUEST FEATURE REQUEST <!-- If this is a BUG REPORT, please: - Fill in as much of the template below as you can. If you leave out information, we can't help you as well. If this is a FEATURE REQUEST, please: - Describe *in detail* the feature/behavior/change you'd like to see. In both cases, be ready for followup questions, and please respond in a timely manner. If we can't reproduce a bug or think a feature already exists, we might close your issue. If we're wrong, PLEASE feel free to reopen it and explain why. --> **What happened**: kubeconfig file location needs to be passed as flag variable and that can't be handled when maya-nodebot runs as daemonset. **What you expected to happen**: Use incluster config kubernetes api. If it fails then try to use kubeconfig flag. **How to reproduce it (as minimally and precisely as possible)**: Running maya-nodebot start should work, without kubeconfig flag. <!-- **Environment**: - kubectl get nodes - kubectl get pods --all-namespaces - kubectl get services - kubectl get sc - kubectl get pv - kubectl get pvc - OS (e.g. from /etc/os-release): - Kernel (e.g. `uname -a`): - Install tools: - Others: --> Answers: username_1: Add to the implementation details for https://github.com/openebs/node-disk-manager/pull/1
versionone/component-library
471101096
Title: Switch doesn't behave as expected when passing default state Question: username_0: While working within CTM @mickeyrogers and I observed some unexpected behavior. The switch itself worked fine unless we passed state to it for the `checked` prop. After spending a couple hours on it we tried a regular checkbox and it worked fine. Answers: username_1: Can you express what the expected behavior was? If you pass a checked prop then the switch should reflect exactly that state independent of further user interaction. Controlled vs. uncontrolled React components. Can you verify this is still an issue? username_0: Can you clarify "switch should reflect exactly that state independent of further user interaction"? **Expected behavior:** When I pass a default value of `true` to the checked prop then the switch should show that it's checked. When the user clicks the switch to turn it off then the switch should show that it's unchecked. username_2: Talked with Matt, the source of his issue was not the switch. Closing Status: Issue closed
ChalkyBrush/roshpit-bug-tracker
298726443
Title: Arkimus - Glyph of Mana Shield Question: username_0: Even if damage instance is absorbed, mana still consumes ![20180220225015_1](https://user-images.githubusercontent.com/32621322/36445931-5ee299ce-1688-11e8-8547-91de5de47e0d.jpg) ![20180220225016_1](https://user-images.githubusercontent.com/32621322/36445933-5ffc0890-1688-11e8-9ebc-8a4e90b71393.jpg) Answers: username_0: I know, mana shield effect comes before any mitigations, right? username_1: yea so it needs to be moved to final operation so this is in fact a bug (not fixed yet just commenting) Status: Issue closed username_1: fixed next patch
stripe/bonsai
125048467
Title: Consider removing implicit/default Layouts Question: username_0: Since an efficient layout is such a crucial part of successfully using Bonsai, it might be better to have folks create their own, rather than providing implicit instances. Look into what it would take to remove the implicits and just use builders. Answers: username_1: I definitely generally agree. I think the future will likely have people using some sort of Generic-style builder that'll remove a lot of the complexity in constructing instances in the common case (eg `Layout.generic[MyLabelType]`). Also, serialization (eg on-disk) is a thing we need to start thinking about. It's unclear to me if we want to conflate in-memory layout with on-disk, but there are some obvious benefits in load time if we do (can even imagine an on-disk, memory-mapped back `Vec`). However, in that case we *really* need to be sure we don't accidentally swap out the `Layout` being used, so there can be no generic implicits.
sailfishos/sailfish-secrets
311495134
Title: Misleading package name: libsailfishcryptoplugin Question: username_0: You would think that `libsailfishcryptoplugin` contains an actual crypto plugin, but it's just the QML plugin. This issue is a reminder to myself to change the name of this package, and also `libsailfishsecretsplugin`.
osmlab/osm-community-index
316044208
Title: OSM Peru 🇵🇪 Question: username_0: <!-- This template can help you add your community resource to the index. Fill in whatever sections you want, or feel free to ignore it! --> ## Community Resource Name Openstreetmap Peru ### What is it? - Homepage - Mail list - Matrix - Telegram - Twitter - Facebook ### URL link to the resource URL links to the resources - Homepage OSMPE: http://osmpe.ourproject.org/ - Mail list OSMPE: http://lists.openstreetmap.org/listinfo/talk-pe - Matrix: https://matrix.to/#/#osmpe:matrix.org - Telegram: https://telegram.me/osmPe - Twitter: https://twitter.com/osmpe - Facebook: https://www.facebook.com/Osmpe ### Points of contact @osm-pe Team on GIthub ### Bounding polygon - Relation: https://www.openstreetmap.org/relation/288247 - Geojson file: https://gist.github.com/username_0/d80c5396b0d8efd22be69e768806a6e7 ### Description (optional) <!-- example: "Join our group here: {url}" --> ### Languages spoken (optional) "es" ### Country codes (optional) "pe" ### Add an emoji flag (optional) 🇵🇪<issue_closed> Status: Issue closed
jaegertracing/jaeger-idl
1162225713
Title: Unclear documentation for max, minDuration for traces endpoint Question: username_0: Must all spans in a trace be larger than `minDuration`, if set? Must only one span in a trace fit within the min and max duration parameters? It's not really obvious what the intention is here. Answers: username_1: The zipkin files are just a copy from the Zipkin project, we're not going to change them. Status: Issue closed username_0: OK - can you point me to the right place to file a ticket? username_1: probably here https://github.com/openzipkin/zipkin-api
nuclio/nuclio
644188369
Title: Blocked by github for misbehaving after restarting too many times Question: username_0: When nuclio dashboard starts it fetches function templates from github. Restarted it too many times, and got blocked by github for misbehaving Am using minikube setup from https://nuclio.io/docs/latest/setup/minikube/getting-started-minikube/ See logs below from dashboard pod that has status crashloopbackoff: Running in parallel Starting dashboard Starting nginx 20.06.23 08:46:36.210 [37m dashboard.platform[0m [32m(D)[0m Using kubeconfig {"kubeconfigPath": ""} 20.06.23 08:46:36.212 [37mrd.platform.docker.runner[0m [32m(D)[0m Executing {"command": "docker version"} 20.06.23 08:46:36.241 [37mrd.platform.docker.runner[0m [32m(D)[0m Command executed successfully {"output": "Client: Docker Engine - Community\n Version: 18.09.6\n API version: 1.39\n Go version: go1.10.8\n Git commit: 481bc77\n Built: Sat May 4 02:33:34 2019\n OS/Arch: linux/amd64\n Experimental: false\n\nServer: Docker Engine - Community\n Engine:\n Version: 19.03.8\n API version: 1.40 (minimum version 1.12)\n Go version: go1.12.17\n Git commit: afacb8b7f0\n Built: Wed Mar 11 01:30:32 2020\n OS/Arch: linux/amd64\n Experimental: false\n containerd:\n Version: v1.2.13\n GitCommit: 7ad184331fa3e55e52b890ea95e65ba581ae3429\n runc:\n Version: 1.0.0-rc10\n GitCommit: dc9208a3303feef5b3839f4323d9beb36df0a9dd\n docker-init:\n Version: 0.18.0\n GitCommit: fec3683\n", "stderr": "", "exitCode": 0} 20.06.23 08:46:36.250 [37m dashboard[0m [32m(D)[0m Fetching function templates from git repository {"templatesGitRepository": "https://github.com/nuclio/nuclio-templates.git", "templatesGitRef": "refs/heads/master"} 20.06.23 08:46:36.273 [37mitFunctionTemplateFetcher[0m [32m(D)[0m Fetching templates from git {"ref": "refs/heads/master"} Error - Get "https://github.com/nuclio/nuclio-templates.git/info/refs?service=git-upload-pack": dial tcp: lookup github.com on 10.96.0.10:53: server misbehaving .../gitfunctiontemplatefetcher.go:80 Call stack: Failed to initialize git repository .../gitfunctiontemplatefetcher.go:80 Failed to clone repository .../gitfunctiontemplatefetcher.go:58 Failed to fetch one of given templateFetchers .../dashboard/functiontemplates/repository.go:35 Failed to create repository out of given fetchers /nuclio/cmd/dashboard/app/dashboard.go:131 parallel: This job failed: /runners/dashboard.sh Exiting Answers: username_1: Hi @username_0, Not a very common issue as far as I can tell :\ Why were your pod "misbehaving" - in other words, whats the cause of the "too many" (?!) restarts? I'm guessing after you solved the root cause of the restarts, life went back to normal and all is well? username_2: Hey @username_3 , You may use the following envs on nuclio-dashboard - `NUCLIO_DASHBOARD_OFFLINE:true` - `NUCLIO_TEMPLATES_ARCHIVE_ADDRESS:none` - `NUCLIO_TEMPLATES_GIT_REPOSITORY:none` to achieve offline working. also look at https://github.com/nuclio/nuclio/issues/2021#issuecomment-757130215 to avoid pulling base images while building functions username_3: Thanks for the quick response. I had already circumvented the issue by creating an offline git repo and mounting the private CA pem file in the Nuclio container. Status: Issue closed
virtualstaticvoid/asdf-docker-compose
1016018273
Title: Download URL changed since version 2 Question: username_0: When installing docker-compose 2.0.1 using `asdf install docker-compose 2.0.1` the download fails and `~/.asdf/installs/docker-compose/2.0.1/bin/docker-compose` contains "Not found". In fact, the download URL has changed : - v1.29.2 : https://github.com/docker/compose/releases/download/1.29.2/docker-compose-Linux-x86_64 - v2.0.1 : https://github.com/docker/compose/releases/download/v2.0.1/docker-compose-linux-x86_64 Answers: username_1: See the [About update and backward compatibility][update] and [Where to get Docker Compose][install] pages of the `docker-compose` repository for further details. I will update this plugin to include this on the README and show a warning when running `asdf list-all docker-compose` or when a user attempts to install a version >= 2.0. [install]: https://github.com/docker/compose#where-to-get-docker-compose [update]: https://github.com/docker/compose#about-update-and-backward-compatibility
mbert/kubeadm2ha
301257824
Title: kube component can not do `Watch` when apiserver is set to master IP? Question: username_0: hi (: the default load balancing strategy of nginx is **rr**, so when a pod(sth like kube-proxy) do `Watch` action, it will print a lots of warning log message like W0301 02:10:52.929987 1 reflector.go:341] k8s.io/kubernetes/pkg/client/informers/informers_generated/internalversion/factory.go:85: watch of *core.Service ended with: very short watch: k8s.io/kubernetes/pkg/client/informers/informers_generated/internalversion/factory.go:85: Unexpected watch close - watch lasted less than a second and no items received How to deal with this issue? or just ignore it? Answers: username_1: The most obvious thing to try would be setting up load balancing with a strategy other than rr. As far as I understand, this is not supported by the free version NGINX (you'd have to upgrade to the expensive 'plus' edition). Hence something other than NGINX may be worth a try - at least in order to see whether the effect you describe disappears. I might take a look into this at some time, but this may take a while. If you have something in your sleeve I'd happily take a look at a PR :) username_1: Closing this because there does not seem to be a tweak to get this working with _nginx_. Hence if one wants to use the 'watch' commands, then _nginx_ load balancing cannot be used. Status: Issue closed
leolorenzoluis/xyz.MonacoEditorLoader
530887382
Title: Custom path for loading monaco doesn't work Question: username_0: I spent some time trying to pass in a custom path for the *loadMonacoEditor directive and I realized that no matter what I changed the path to, the path the service was trying to load from would not change. I noticed that the editor load logic actually happens in the constructor of the loader service: https://github.com/username_1/xyz.MonacoEditorLoader/blob/55c9ad2b12e808eabaf39b3b88105352ced21c58/lib/monaco-editor-loader/monaco-editor-loader.service.ts#L15 Even though the directive has an @Input that passes in the custom path, that doesn't actually ever activate since the loading logic executes in the constructor. By the time the input is set, the LoaderService has already attempted to load the editor. Additionally, the README suggests that it's the inclusion of the *loadMonacoEditor directive that loads the editor, but this isn't the case. ``` <monaco-editor *loadMonacoEditor></monaco-editor> ``` Even without the directive, the editor will load since the logic will be execute on construction of the service when any component imports the MonacoEditorLoaderModule. I propose that the editor service should have a load() method that is actually only called in the MonacoEditorLoaderDirective, after the @Inputs are set within the ngOnInit. Answers: username_1: How are you setting the custom path? I tested with custom path and it is working as expected. The life cycle is correct. When the input directive is set, the property is respected to whatever value you set. username_0: @username_1 Thanks for looking into this. I set the directive as specified in the README ``` <monaco-editor *loadMonacoEditor="'test/monaco'"></monaco-editor> ``` Perhaps there is a race condition. In your test, is the AMD loader being loaded via loaderScript? If so, onGotAmdLoader() might be executing after the Input is set - but that wouldn't be the case if the window.require was provided and onGotAmdLoader() fires immediately username_1: Ah makes sense now. I have my require as undefined. Is this what you propose? ``` import { Injectable, NgZone } from '@angular/core'; // tslint:disable-next-line:import-blacklist import { BehaviorSubject } from 'rxjs'; @Injectable() export class MonacoEditorLoaderService { isMonacoLoaded: BehaviorSubject<boolean> = new BehaviorSubject<boolean>(false); private _monacoPath = 'assets/monaco-editor/vs'; set monacoPath(value: any) { if (value) { this._monacoPath = value; this.load(); } } load = () => { // Load monaco console.log(this._monacoPath); (<any>window).require.config({ paths: { 'vs': this._monacoPath } }); (<any>window).require(['vs/editor/editor.main'], () => { this.ngZone.run(() => this.isMonacoLoaded.next(true)); }); }; constructor(private ngZone: NgZone) { // Load AMD loader if necessary if (!(<any>window).require) { const loaderScript = document.createElement('script'); loaderScript.type = 'text/javascript'; loaderScript.src = `${this._monacoPath}/loader.js`; loaderScript.addEventListener('load', this.load); document.body.appendChild(loaderScript); } } } ``` username_0: Yes, that should do the job. Thanks! username_1: @username_0 Published. Should be fixed on `8.0.13` Status: Issue closed
jfrimmel/cargo-valgrind
607027939
Title: Not working on unit tests Question: username_0: Hi. Really awesome project. I love how much simpler it is to check for memory leaks etc with this tool. However I'm not able to run this on my integration tests on a Fedora Linux machine. I'm not sure if Cargo changed in some incompatible way or if I have an unsupported version of valgrind? This is the software I'm running: ``` $ valgrind --version valgrind-3.15.0 $ cargo valgrind --version cargo-valgrind 1.3.0 $ cargo --version cargo 1.43.0 (3532cf738 2020-03-17) ``` And this is the issue I'm getting: ``` $ cargo valgrind --test all_tests ... error: Could not parse XML: custom: 'missing field `xwhat`' ``` Answers: username_0: It seems to work when running under Rust 1.39. `cargo +1.39.0 valgrind --test all_tests` does work. But only after a `cargo clean`. The builds with newer Rust seem to leave some metadata that `cargo-valgrind` does not understand. username_1: that would be great if I could run this on unit tests. username_2: Hello @username_0, can you provide me some test input for reproduction? Alternatively, you can try out the new version I've just updated. If you want to, you can try the new software using the following command: ```bash cargo install --git https://github.com/username_2/cargo-valgrind --branch custom-runner ``` username_0: I currently can't reproduce the issue. I can't remember exactly how I hit the problem, but I remember it being easy to get back then. I now have valgrind-3.16.1. So that could be why. Or newer Rust. I think I got the problem in this repo: https://github.com/username_0/loom-executor. It's the only one of my repos where I can find a `all_tests.rs`. username_3: I ran into this issue when using Valgrind 3.15.0 as packaged in the Ubuntu 20.04 repositories. The problem went away when I compiled Valgrind 3.16.1 on my own. username_4: Hello and *thank you very much* for this tool which is incredibly useful! I am encountering the same bug locally with valgrind 3.16.1. Therefore, I installed cargo-valgrind from the malformed-xml branch using: ``` $ cargo install --git https://github.com/username_2/cargo-valgrind --branch malformed-xml -- cargo-valgrind ``` I rerun the test and I now have the failing XML printed out! [valgrind.log](https://github.com/username_2/cargo-valgrind/files/6488984/valgrind.log) I hope this helps! username_2: Thank you for the log, that's helpful. I'll have a look and hope to fix it soon. username_5: Hi! I wanted to share the environment in which I’m encountering the same issue today, including sample project and Dockerfile to reproduce. I hope this helps! [valgrind_test.zip](https://github.com/username_2/cargo-valgrind/files/6577728/valgrind_test.zip)
mapbox/mapbox-navigation-android
338315228
Title: Issue in OffRouteDetector Question: username_0: **Android API:** N/A **Mapbox Navigation SDK version:** 0.15.0 ### Steps to trigger behavior 1. Start navigation 2. Move slowly along the route (walking pace can be enough), such that the distance to maneuver stays basically the same 3. Wait for 4 GPS readings ### Expected behavior - Navigation continues normally ### Actual behavior - Rerouting is triggered I believe this is caused by an error in OffRouteDetector.movingAwayFromManeuver(). it checks that: distancesAwayFromManeuver.peekLast() - distancesAwayFromManeuver.peekFirst() < MINIMUM_BACKUP_DISTANCE_FOR_OFF_ROUTE I believe there is an error here. - The queue will already be cleared (later in the code) if the distance is decreasing. - Therefore the queue from oldest to newest must be increasing (or it could remain constant) - peekLast() takes the oldest value, peekFirst() takes the newest, so peekLast() - peekFirst() will always be <= 0 - Which means this check will always be true I assume it is meant to check if the user has gone "more than 50m backwards" before triggering, in which case the check should be: peekFirst() - peekLast() > MINIMUM_BACKUP_DISTANCE_FOR_OFF_ROUTE Note that because distancesAwayFromManeuver are integers, even slightly decreasing distances will still trigger this - as long as they round to the same integer. e.g. distances of 210.40, 210.18, 210.34 and 209.94 will trigger this. As will distances of 200, 210, 220, 230 (which is backwards, but as per the above description, I assume is meant to be filtered). Also note the 4th distance seems to be pointless. It will always trigger then anyway because the "return true" happens before there is a chance to clear if it's back in the right direction. If that's intended, it's not clear why it doesn't just trigger on the previous call. Answers: username_1: Hey @username_0 👋 thanks for the great write-up here. We will take a look at this logic in-depth and get back to you here on this ticket. At first glance, all of your points look completely valid to me. username_2: I have the same problem. Thank you username_3: We've the same problem. username_4: I have same problem on v16. here is logcat: `8-17 17:14:31.778 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false 08-17 17:14:33.744 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false 08-17 17:14:34.827 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false 08-17 17:14:36.764 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: true 08-17 17:14:37.815 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false 08-17 17:14:38.874 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false 08-17 17:14:40.853 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false 08-17 17:14:42.788 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false 08-17 17:14:44.785 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: true 08-17 17:14:45.800 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false 08-17 17:14:46.820 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false 08-17 17:14:48.788 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false 08-17 17:14:49.848 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: true 08-17 17:14:52.904 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false 08-17 17:14:54.819 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false 08-17 17:14:55.823 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: ` after first offRoute it should print true for all logs but as you can see sometimes it detect user is not offroute and Rerouting is not triggered. username_4: **UPDATE** I think I found the reason and some added information for debuging: the problem would occur in last leg. the `checkForArrivalEvent` method in `NavigationEventDispatcher` consider it as Arrival and delete the listener so when offroute occur there is no listener to invoke reroute request username_1: @username_0 @username_4 do ya'll mind retesting with `0.20.0`? username_0: @username_1 If I get time I'll have a look. But we've actually taken the opportunity to build our own off route engine, because we wanted extra functionality. I'll have to see how easy it is to go back to the original. username_1: @username_0 okay great, thank you. @username_4 if you don't mind testing as well, I'd just like to confirm this is no longer and issue with the new logic. username_5: Hey @username_4 👋 were you able to retest? It'd be 💯 if you could report back confirming that the issue is gone. Thanks a lot! I also want to take the opportunity to let you know that we released version [`0.22.0`](https://github.com/mapbox/mapbox-navigation-android/releases/tag/v0.22.0) so I encourage you to upgrade because there are a bunch of improvements and new features plus lots of bug fixes since `0.16.0`. username_4: Hi sorry I didn't see the thread. I will test in tomorrow and let you know 🙏🏻 username_5: Sweet! Thanks for the update @username_4 🙇 username_1: Hey @username_4 @username_0 I'm going to go ahead and close this as resolved. @username_0 once we unblock you from upgrading and if you re-test and still see issues, please feel free to cut a new ticket and we will continue to dig on this. Thanks! Status: Issue closed username_1: @username_0 this should be right behind our current push for offline routing - I can't give a specific timeline but this is definitely high priority. We will be sure to update your ticket regarding this once we start moving. Thanks as always for your patience 👍
studydash/cards
936362956
Title: Release `Server 0.0.5` & `GUI 1.8.5` Question: username_0: ## Saturday Tonight, I released `Server 0.0.5` and `GUI 1.8.5` into `QA`. The new server build now adds a `Record` and increments `RecordCount` and `Streaks` on every new card insertion: ![image](https://user-images.githubusercontent.com/45280066/124371321-9f1d7600-dc4e-11eb-9950-7e60939e740d.png) And for the new GUI, I reorganized the `Members Pane`; it now looks like this: ![image](https://user-images.githubusercontent.com/45280066/124371329-b52b3680-dc4e-11eb-97a7-05520d44136b.png) <br /> ## Extra/Fun 🎈🎉: This is July 4th weekend, so Shaza and I visited a new bakery this afternoon and also checked out the Art Climb steps at the Cincinnati Art Museum. The weather was thankfully still cool; it was a good time! 🥳 ![image](https://user-images.githubusercontent.com/45280066/124371342-e4da3e80-dc4e-11eb-8ab5-65568a5271fb.png)
Hirnbix/captain-holetooth
183537477
Title: Missing translations Question: username_0: There appears to be support for adding translations to texts and labels, however it has not been fully implemented. This is just a reminder that it is an issue and should be implemented. :) Answers: username_1: Milestone goal: Add english and german username_1: Pushing this to a new milestone / project as it is unclear how much translation work is needed at all.
nss-evening-cohort-10/dinomate
491738243
Title: Create Wireframes Question: username_0: # User Story As a developer, when I do stuff I would like to see some wireframes # AC **WHEN** I open this ticket **THEN** I should see screenshots for each page # Dev Notes * Use Moquops to generate wireframes * paste them into this ticket Answers: username_0: Home Page: ![image](https://user-images.githubusercontent.com/8093647/64626759-fb494300-d3b3-11e9-97f4-880825e35dea.png) username_0: About Us: ![image](https://user-images.githubusercontent.com/8093647/64626788-0ac88c00-d3b4-11e9-9ecb-2d62d293f36c.png) username_0: Profile: ![image](https://user-images.githubusercontent.com/8093647/64626811-16b44e00-d3b4-11e9-9a8f-da36559f022a.png) Status: Issue closed
dask/distributed
282580628
Title: Move worker_info onto WorkerState? Question: username_0: @username_1 I'm curious why you chose not to move worker metadata like this onto the WorkerState object? Answers: username_1: Just that it was not relevant for the immediate goal of trying to cythonize the scheduler, and did not present any obvious difficulties that deserved to be tackled up front :-) But, yes, we should probably do it anyway. username_1: This is fixed, no? Status: Issue closed
gbif/portal-feedback
299466620
Title: Link to description of data validation rules in the data validator Question: username_0: **Link to description of data validation rules in the data validator** It would be helpful if the data validator (which by the way is a great tool) could provide links to the data validation rules / short description of what is expected of the different fields checked. - i.e. a short link to the proper documentation. This would help data publishers in faster find the out what is wrong in cases where this is not obvious . ----- User provided contact info: <EMAIL> System: Firefox 58.0.0 / Ubuntu 0.0.0 User: [See in registry](https://www.gbif.org/api/feedback/user/15a8d2e2150b6c740b608f88ace5269d:ee803319c81be99ceb64813a4b8c0226f0f69b9435fb3c4972cd094f767c6346342f3b684e4e97bfe4c5635c4d42af1f57517bacbf2c6d2398da9492a117dbf6) Referer: https://www.gbif.org/tools/data-validator/1516700048547 Window size: width 1663 - height 895 [API log](http://elk.gbif.org:5601/app/kibana?#/discover?_g=(refreshInterval:(display:Off,pause:!f,value:0),time:(from:'2018-02-22T18:47:00.712Z',mode:absolute,to:'2018-02-22T18:53:00.712Z'))&_a=(columns:!(_source),index:'prod-varnish-*',interval:auto,query:(query_string:(analyze_wildcard:!t,query:'response:%3E499')),sort:!('@timestamp',desc))) [Site log](http://elk.gbif.org:5601/app/kibana?#/discover?_g=(refreshInterval:(display:Off,pause:!f,value:0),time:(from:'2018-02-22T18:47:00.712Z',mode:absolute,to:'2018-02-22T18:53:00.712Z'))&_a=(columns:!(_source),index:'prod-portal-*',interval:auto,query:(query_string:(analyze_wildcard:!t,query:'response:%3E499')),sort:!('@timestamp',desc))) System health at time of feedback: OPERATIONAL<issue_closed> Status: Issue closed
linkerd/linkerd
296633206
Title: namerd admin interface hangs Question: username_0: ``` There are **no** log entries neither in the consul nor in the namerd log. **What you expected to happen**: Not to hang? :) **How to reproduce it (as minimally and precisely as possible)**: Configure as I did, curl /. **Anything else we need to know?**: Until I got the consul configuration right, the interface *did* load with an error telling me that no namespaces were found (as a side note, the documentation around namerd and its integration with linkerd and other services was not quite up to the level of what I got used to from linkerd's docs – I needed additional guidance otherwise I’d still be stumbling). **Environment**: - linkerd/namerd version, config files: namerd.yml: ```yaml --- admin: port: 9991 ip: 10.6.180.250 tls: certPath: /etc/ssl/certs/XXX.crt keyPath: /etc/ssl/private/XXX.pkcs8 namers: - kind: io.l5d.consul includeTag: true useHealthCheck: true storage: kind: io.l5d.consul token: XXX readConsistencyMode: stale pathPrefix: /namerd/dtabs interfaces: - kind: io.l5d.mesh ip: 10.6.180.250 port: 4321 telemetry: - kind: io.l5d.prometheus path: /admin/metrics/prometheus prefix: namerd_ ``` namerd: 1.3.5 consul: 1.0.1 - Platform, version, and config files (Kubernetes, DC/OS, etc): Ubuntu Xenial, since this is an isolation namerd issue, the rest shouldn’t matter. - Cloud provider or hardware configuration: LXC running on metal. More than 1 GB of RAM free. Answers: username_1: @username_0 thanks so much for filing this detailed report. These kinds of issues make debugging linkerd/namerd much easier. We will take a look at this issue as soon as we can. Also, if you have any suggestions on where you would love to see more detail in namerd's documentation, please do let us know! username_0: JFTR, just tested with consul 1.0.6 and the problem persists. username_2: @username_0 we're taking a look at this username_3: Hey @username_0, thanks for such a detailed report! I've been trying to reproduce this issue locally, and haven't been able to exactly. But I did notice something in your setup that seems potentially problematic, based on the contents of your consul KV store: ``` [ { "CreateIndex": 137788444, "Flags": 0, "Key": "namerd/dtabs/", "LockIndex": 0, "ModifyIndex": 137788444, "Value": null }, { "CreateIndex": 137788454, "Flags": 0, "Key": "namerd/dtabs/default", "LockIndex": 0, "ModifyIndex": 137801542, "Value": "<KEY>" } ] ``` The first key in that list is "namerd/dtabs/", which is also the folder where the second key, "namerd/dtabs/default", is located. When testing locally, if I create the first key first and the second key second, then namerd's admin UI doesn't hang, but it does tell me that there are no namespaces found: ![image](https://user-images.githubusercontent.com/9226/36238223-24549bb2-11b6-11e8-81ff-a9043c85f70a.png) If I use the consul API to delete the first key: ``` $ curl -X DELETE http://localhost:8500/v1/kv/namerd/dtabs/ true ``` And then bounce namerd, its admin interface successfully loads the namespace from the second key: ![image](https://user-images.githubusercontent.com/9226/36238419-55dc84aa-11b7-11e8-81d9-c6b777e78347.png) Am not positive you're experiencing the same issue, but can you try applying the same fix in your setup (delete the `namerd/dtabs/` key and restart namerd)? --- When playing around with this locally, I setup a docker-compose env that might be useful. You can find it here: [consul-issue.tar.gz](https://github.com/linkerd/linkerd/files/1726287/consul-issue.tar.gz) There are some instructions for reproducing the issue in the README. username_4: Wow, @username_3, great catch! I tried to repeat your steps and was able to reproduce the issue exactly as @username_0 reported it! In my case, as well as in @username_0's case, Consul has ACLs enabled with default policy set to "deny-all" and ACL permissions set to ``` key "namerd/dtabs/" { policy = "write" } ``` which may explain the difference in behavior we see. Let me see if I can figure out what causes it... username_0: I can confirm that deleting everything and creating `namerd/dtabs/default` in one step fixes the problem for me. Thanks everyone and good luck fixing. :) username_4: The cause of the issue as well as the fix turned out to be embarrassingly simple and _it looks like_ this issues has been in place since first implementation of the store that I pushed almost 2 years ago. 😞 username_3: @username_0 Thanks for confirming that the workaround works -- that's great! And huge thanks to @username_4 for putting together the fix. I tested the changes from #1816 in my local repro, and they do indeed fix the issue that I was seeing. Status: Issue closed username_2: W00t!
quarkusio/quarkus
860756670
Title: Controlling Service Account roles when using the Kubernetes and Kubernetes Client extension Question: username_0: ## Description It currently is impossible to disable the service account generation when using the Kubernetes en Kubernetes Client extension in a project. It gererates a service account and a rolebinding linked to view role. Sometimes this is not strict enough. It is possible to work around using Kustomize, but this way using the deployment feature of the Kubernetes extension is impossible. Being able to disable the Service Account generation will solve this, but that is not optimal as a Service Account and RoleBinding would have to be applied outside of the extension control. An other option might be to specify a role to bind to in stead of view. The role could either be provided using the kubernetes.yml in the source tree or be provided or it would need to be present on deployment already. Any check on the existence of the role should be configurable. Answers: username_1: cc @username_2 username_2: In theory people could specify the rbac configuration in src/main/kubernetes and make it as strict as they need to. Haven't tried it though... Now I can see a feature flag to disable the generation of rbac resources. Also we could point to an existing Role binding or even control the privileges. These are all valid approaches. username_0: @username_2 specifying the rbac configuration src/main/kubernetes does add them to the generated descriptors, but if the name is not the same as would be generated it is just added and the *-view rolebinding is still generated. A bit more control over how it all is generated seems a good idea to me.
google/filament
687325242
Title: ANDROID: How to eliminate lag and improve FPS. Question: username_0: I am referencing sample gltf viewer demo to load my model. everything is working fine except lagging. once model loads, the screen starts lagging so much. is there any way to tackle this issue? My model size is around 150 MB. Answers: username_1: The size of the glTF doesn't matter much, what matters is going to be: - How many triangles - How many materials - How large the textures - etc. You can optimize your glTF by trying to reduce all those values. A good rule of thumb would be to use ~40,000 triangles and textures that are 1024x1024 each. You can also: - Make sure the app doesn't render at native resolution (1080p or 720p is good enough on many devices) - Enable dynamic resolution - Turn off expensive features like bloom or SSAO Status: Issue closed username_2: @username_1 you suggested changing the render resolution in a couple of threads. Do you mean setting it via `this.surfaceView.holder.setFixedSize(760, 360)` ? username_2: Answering the initial question, removing unnecessary lights from the scene will improve performance drastically. My app was lagging bad and removing a light from the scene basically fixed the issue for me.
MicrosoftDocs/azure-docs
665706493
Title: Incorrect output Question: username_0: Hi - I've followed these steps exactly, but the output doesn't give you a CSV with the transformed data. Instead, it just prints out the text "iris_setosa.csv" into Cell A1 of a CSV in my output container - is there an error in the code or something that I'm missing? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 894562e1-2c07-2f75-b420-d843262a7f07 * Version Independent ID: 1b1fa915-19f0-653c-04d5-218ed4335398 * Content: [Run Python scripts through Data Factory - Azure Batch](https://docs.microsoft.com/en-us/azure/batch/tutorial-run-python-batch-azure-data-factory) * Content Source: [articles/batch/tutorial-run-python-batch-azure-data-factory.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/batch/tutorial-run-python-batch-azure-data-factory.md) * Service: **batch** * GitHub Login: @mammask * Microsoft Alias: **komammas** Answers: username_1: @username_0 Thank you for your feedback . We will investigate and update the thread further. username_2: @username_0 Apologies for the delay in response. I was checking with our team internally on the issue and it seems like there are some changes made. Can you try again and confirm if the issue is resolved for you or not. Thanks. username_2: @username_0 Hope the provided information is helpful. We will now close this issue. If there are further questions regarding this, please tag me in a comment. I will reopen it and we will continue the discussion. Status: Issue closed
uspki/policies
271756880
Title: Section 5.4.1 Question: username_0: **Organization / Program:** DoD <br>**Section:** 5.4.1 <br>**PDF Page:** 52<br>**PDF Line(s):** 1257<br>**Comment:** Items b and c are too broad and do not constitute clear policy. Revise to make them more concrete.<br><br>**Suggested Change:** Revise per comment Answers: username_0: Is it preferred to state this policy requirement in terms of 800-53 security controls inclusive of AC-6, AC-7, AU-2, AU-3 (etc)? Status: Issue closed
Shopify/semian
188339905
Title: Support for error percentage? Question: username_0: Hi guys, love the library, but I wanted to see if you would be open/interested in adding support for error percent in addition to the current absolute error threshold? I find dealing with error rates in terms of percentages much more flexible than absolute values. If you're open to it I would be interested in trying to tackle it and send you a pull request. However, I don't want to maintain a fork of the project, so only want to go down this path if it is likely to get integrated back into the main code line. Answers: username_1: The percentage would have to be over some sliding window of requests, wouldn't you have to configure the size of that window based on time or absolute size? username_0: Agreed, if I'm understanding the current code there's already a SlidingWindow used for tracking errors so I would imagine that we'd have to track the number of requests utilizing the same window so that we could do the proper comparison. Plenty of details to work out, but not work figuring out unless it would be considered generally useful. username_1: Honestly, I'm not sure this is worth it. I actually think it's harder to reason about errors in a window than it is for absolute errors in a row, and in our experience this has worked well. If you'd like to contribute, a huge, huge improvement to Semian would be a way to share the circuit breakers between processes in an IPC namespace. A SysV implementation of the circuit breaker using shared memory would be a massive improvement. username_0: Strange, I don't think it sounds that hard to reason about (assuming we expose the right information). However, I'll take that as a no for now. Let me know if you guys reconsider. I'll see how far we get utilizing absolute values. I'm just worried that we'll find ourselves having to do very different configurations for services that are called very different volumes. Status: Issue closed username_1: Why do you think the configuration would be different? We use the same between MySQL shards that take 10,000s of queries per second and services that do few requests per minute. Works great. username_0: That's interesting Simon as that's exactly the type of scenario I was worried would require very different configurations when using absolute values (although at this point we're planning on only using this for various HTTP based services not the database connection itself). If you don't mind can I ask what your configuration looks like? username_1: Absolutely. For MySQL shards we use this configuration: ```yaml semian: &semian tickets: <%= Resiliency.semian_ticket_count %> timeout: <%= Resiliency.semian_ticket_timeout %> success_threshold: 2 error_threshold: 3 error_timeout: 10 ``` `Resiliency.semian_ticket_count` is defined as the number of Unicorn or job workers on the box divided by two. We generally run around 48 workers per box, that's around 3 per core. `Resiliency.semian_ticket_timeout` is `2s` for job workers, and `0s` for web workers. The difference between a ticket count and a timeout is that timeouts will allow short bursts on a resource, whereas ticket counts is the absolute maximum concurrency. This is a problem for job workers with our workload, but not for web workers. With this configuration we run over a dozen shards with 10,000+ workers without issues. A single shard outage never has an impact on the latency of the entire platform. MySQL, given our throughput to it, is _by_ far the biggest threat. As for HTTP based services, we do very low throughput to them. This means that bulkheads are not really useful. Circuit breakers are quite useful though. For HTTP services our configuration looks like this: ``` SEMIAN_PARAMETERS = { tickets: 1000, success_threshold: 1, error_threshold: 3, error_timeout: 20 }.freeze ``` This effectively disabled bulkheads for them because we do low throughput to them (https://github.com/Shopify/semian/pull/89 allows just disabling them). It's fairly hard to reason about good ticket counts for them, and we don't see a scenario where it'd be a huge issue. It'll take a while for all the circuits to trigger, but given the throughput spending that capacity waiting for the timeouts is not really an issue as long as we converge towards failing fast. So they're not actually exactly the same, but the circuit breaker thresholds don't matter too much in our experience for these low latency services. We're much more concerned about the high throughput resources (MySQL, basically) than anything else. username_0: Thanks so much for sharing this! We will experiment with a variation of this for our candidate services. I appreciate your insight! username_1: Absolutely! Let me know if you have other issues. Another thing I'd love to see in core Semian is that HTTP can be circuit broken by default, intead of ad-hoc configuring every endpoint. The reason it's not is because if you talk to 10,000s of hosts like we do, you'd allocate way too many circuit breakers. The implementation needs to have a ring-buffer scoped by resource for that to not be a subtle memory leak. If you have similar issues, let me know.
uraimo/SwiftyGPIO
1084106996
Title: Segmentation Fault Question: username_0: ### Board Type RaspberryP4 ### Operating System raspIO Bullseye 64 bit ### Swift Version 5.5 ### Description When using `onRaising` to measure the rotational speed of a 5000 rpm fan, I'm getting segmentation faults
FreshPorts/freshports
842517091
Title: Two FreshSource issues Question: username_0: (Is this repo good for FreshSource issues?) Via <https://www.freshsource.org/>: 1. <http://docs.freebsd.org/mail/current/cvs-all.html> is obsolete, according to <https://lists.freebsd.org/mailman/listinfo> 2. <https://www.freshsource.org/news.php> does not appear to be RSS. I would have ignored point (1) but then, I vaguely recalled (2) the RSS feed working two or three days ago. This might be a confused recollection of a feed elsewhere (!) but I'm reporting it, just in case. Answers: username_1: Yes it is, and FreshSource should have a link to here. username_1: Fixed by adding adding different links, mentioning subversion, and git. See https://dev.freshsource.org username_1: This is what I get when I fetch it to a file: https://gist.github.com/username_1/c92807d0a1b9a1071a6cbdd25158dce4 Although the DTD is deprecated: https://validator.w3.org/feed/check.cgi?url=https%3A%2F%2Fwww.freshsource.org%2Fnews.php username_1: For the MIME type: `header("Content-Type: application/rss+xml; charset=UTF-8");` Please try again. :) username_0: I'm not getting it in Firefox. Debatably not the best choice, since Firefox dropped support for RSS, but I typically use this extension: * [Want My RSS](https://addons.mozilla.org/addon/want-my-rss/) – and today I enabled this: * [Smart RSS Reader](https://addons.mozilla.org/addon/smart-rss-reader/) (Off the top of your head, can you think of any browser in ports that has integral support for RSS?) username_1: Do you have any feeds which work in Firefox? If you do, tell me which ones and I'll compare to the FreshSource feed. username_0: Thanks, try https://freebsdfoundation.org/feed/ (referred from <https://freebsdfoundation.org/>) seems to work with Want My RSS username_1: @username_0 Please try https://dev.freshsource.org/news.php - I fear I was not clear about where the fix was located. username_0: Ah, got it now: ![image](https://user-images.githubusercontent.com/192271/113348824-074a7c00-932f-11eb-846a-6698c9d8f346.png) * to the left, the one feed that's found at https://dev.freshports.org/ * to the right, https://dev.freshsource.org/news.php username_1: Fixed in repo. username_1: @username_0 how is https://www.freshsource.org/news.php now? username_0: Sorry! I lost sight of this. ![image](https://user-images.githubusercontent.com/192271/147770183-0cada370-1cc9-47fd-86e8-2a4e6caad786.png) With that extension disabled: ![image](https://user-images.githubusercontent.com/192271/147770260-46ebefc2-ec3d-4d45-bc2a-0cb9a2cc1bfe.png)
rundeck/rundeck
418706201
Title: LDAPS not working on Rundeck 3.0.16 Question: username_0: **Describe the bug** Secure LDAP connection (LDAPS) cannot be established following the how-to under: https://docs.rundeck.com/docs/administration/security/authenticating-users.html#communicating-over-secure-ldap-ldaps `Error message: PKIX path validation failed` **My Rundeck detail** * Rundeck version: 3.0.16 * install type: rpm, * OS Name/version: rhel 7.5 * DB Type/version: postgres **To Reproduce** 1. : Do the steps to configure LDAPS described under: https://docs.rundeck.com/docs/administration/security/authenticating-users.html#communicating-over-secure-ldap-ldaps **Expected behavior** LDAPS connection can be established based on the stores described in /etc/rundeck/ssl/ssl.properties **Additional context** I had to manually pass the keystore and truststore as JVM Options: RDECK_JVM_OPTS="-Djavax.net.ssl.keyStore=${RDECK_CONFIG}/ssl/keystore -Djavax.net.ssl.keyStorePassword=${CERT_STORE_PASS} -Djavax.net.ssl.trustStore=${RDECK_CONFIG}/ssl/truststore -Djavax.net.ssl.trustStorePassword=${CERT_STORE_PASS}" Setting RUNDECK_WITH_SSL=true is not sufficient, even though the stores are being used by Rundeck, because HTTPS is working! Answers: username_1: I also had problems setting up LDAPs, as Java's way of dealing keyStore is a bit hard. I recommend using a proxy like stunnel to setup your ldaps.
edouardlicn/Ifxoss_ragemp
362860465
Title: Can not set Hospital Leader Question: username_0: When i try the command "/sethospitalleader" or "/sethospitalleader 1" then nothing happend the Database is not changed. And there is a error on the console. `(node:5356) UnhandledPromiseRejectionWarning: Unhandled promise rejection (rejection id: 1): TypeError: s.getAdminLvl is not a function (node:5356) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code. (node:5356) UnhandledPromiseRejectionWarning: Unhandled promise rejection (rejection id: 2): TypeError: s.getAdminLvl is not a function` Answers: username_1: this is not originally create by me that i don't have the ability to fix that bug.may be you ask the original author in readme.md better.i can update here when he fix. username_1: BTW,you make a barbershop but did not set the cord? username_1: if you not add cord i will delete it.for it will not work and become bug. Status: Issue closed
Unidata/thredds
129265748
Title: PointFeature.getLocation returns altitude in wrong units (Point Feature Types) Question: username_0: [Migrated from Jira](https://bugtracking.unidata.ucar.edu/browse/TDS-672) per Sean This issue concerns the Point Feature Types. Read in the attached nc/ncml combination (below). The data should be CF compliant to my knowledge. John looked at it too. Call the `getLocation()` method. This will return the altitude in kilometers, but is should be in meters. According to the docs: "double getAltitude() altitude in meters; missing = NaN". [28219.476.nc](https://files.zenhub.io/56a92d81e79663ef099dfceb) [28219.476.ncml](https://files.zenhub.io/56a92d81e79663ef099dfcea) Answers: username_1: I can confirm this behavior on both 4.6 and 5.0. `PointFeature.getLocation()` returns an `EarthLocation`. `EarthLocation.getAltitude()` claims that it returns altitude in meters, but its implementation class–`EarthLocationImpl`–makes no attempt to convert values that aren't already in meters. The altitude unit **is** read and made available via `DsgFeatureCollection.getAltUnits()`, but we're currently not doing a whole lot with it. The lazy solution is to edit `EarthLocation.getAltitude()`'s Javadoc to indicate that the unit is not necessarily meters, and can be found by calling `DsgFeatureCollection.getAltUnits()`. A better solution is to store the units in each `EarthLocation`, but that's significantly more work as we'd have to change every piece of code where an instance is created. username_2: I think we can force it to be in km using the udunits package. For example, we could first check if the unit is compatible with km, and if so, change it using the udunits package to get the correct scale factor and make it correct. I sort of get the feeling, especially in coordinate system classes, that we do not utilize the udunits package enough. username_1: Yeah, but not all vertical coordinates are compatible with meters. Some are even dimensionless. It's just a bad idea to guarantee meters from that method. username_2: Ok, here is what we came up with: * change docs to reflect reality * make sure we give users an idea of where to look for the units * All of this needs a major API review * Try to address this in 5.0 * Look at OGC geotoolkit thingy and investigate for THREDDS 6. username_1: Accomplished the first two bullets in #412. I only intend to address this in 5.0.0.
gaffneyc/heroku-buildpack-jemalloc
989412066
Title: Jemalloc no longer working Question: username_0: true ``` ![image](https://user-images.githubusercontent.com/2123767/132257744-7e01653e-60ac-448e-add9-aa550c52fc37.png) Answers: username_1: @username_0 There isn't a good way to see if jemalloc is enabled through RbConfig or inside a process. The best way at the moment (which needs to be added to the readme) is this: https://github.com/username_1/heroku-buildpack-jemalloc/issues/5#issuecomment-499932026 Give that a try and let me know if it's not actually enabled. Jemalloc's allocation algorithm does a better job of allocating objects in already claimed free memory rather than requesting more from the system so it's possible there is a change in your app which is causing it to request and retain a lot more memory Status: Issue closed username_0: You are correct. It seems to be enabled using this way of testing. I will investigate alternative routes. Thank you. P.S. Could you please document this: `MALLOC_CONF=stats_print:true ruby -e "exit"`
docker/for-mac
185979103
Title: Building OpenNetworkLinux yields unresponsive container Question: username_0: ### Expected behavior Container completes build. ### Actual behavior Container becomes unresponsive, i.e. * `docker ps` shows it is `Up` * `docker exec -it <cmd> <container_id>` hangs * `docker stop <container_id>` hangs (unable to stop it) ### Information ``` Docker for Mac: version: 1.12.3-beta29.2 (902414d) OS X: version 10.11.6 (build: 15G1108) logs: /tmp/BC71B46C-FACD-4D89-8C0D-B26FE75FBF4E/20161028-102715.tar.gz [OK] vmnetd [OK] dns [OK] driver.amd64-linux [OK] virtualization VT-X [OK] app [OK] moby [OK] system [OK] moby-syslog [OK] db [OK] env [OK] virtualization kern.hv_support [OK] slirp [OK] osxfs [OK] moby-console [OK] logs [OK] docker-cli [OK] menubar [OK] disk ``` ### Steps to reproduce the behavior - https://opennetlinux.org/docs/build ``` git clone https://github.com/opencomputeproject/OpenNetworkLinux cd OpenNetworkLinux make docker ``` Answers: username_1: Does Diagnose & Feedback from the :whale: menu work? Could you please run and upload the diagnostics and post your diagnostic ID back here? Status: Issue closed username_2: This issue has been inactive for more than 14 days while marked as `status/0-more-info-needed`. It is being closed due to abandonment. Please feel free to re-open with more information about the problem. MORE_INFO_EXPIRY_TIMEOUT
convox/rack
292313709
Title: ERROR: timeout starting process Question: username_0: CLI and web console returning this on all attempts to build or deploy. Rack version: 20180126132639 Health checks all 200 and nothing unusual in rack logs Status in AWS CloudFormation all looks fine, as per Troubleshooting Docs. Any idea what might be causing this? Answers: username_1: We see this all the time. Was it resolved for you somehow? username_2: Would generally mean that you build instance is not large enough and AWS is timing out trying to schedule the build container. username_1: Thank you @username_2! ℹ️ @tapajos
cloudmesh-community/hid-sp18-602
382637207
Title: swagger example is incomplete Question: username_0: Your Docker Glance example is not completed. Two things are missing: a) what is the abstraction to GLance (e.g. are there other clouds that have a similar service)\ b) the service that glance provides whith just a string attribute is certainly incomplete. However the focus is (a) https://github.com/cloudmesh-community/hid-sp18-602/blob/master/swagger-docker/openstack/swagger.yaml<issue_closed> Status: Issue closed