repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
mlr-org/mlrCPO | 324384518 | Title: cpoSelect with statements
Question:
username_0: Is there a CPO where I can just write something like `cpoSelect(sel.statement = function(x) is.numeric(x) && !any(is.na(x)))` and sel.statement just gets lapplied on the columns?
Answers:
username_1: Currently the way to select columns is only via `cpoSelect(type = "numeric")`; dropping columns with missing values is not supported right now. I will try to make this nicer, probably together with "slices" (see #38)
username_1: In the mean time, a quick and dirty implementation of this would be
```R
library("BBmisc")
cpoSelectFunctional = makeCPO("select.functional",
pSS(selector: funct),
dataformat = "df.features",
properties.adding = c("numerics.sometimes", "factors.sometimes",
"ordered.sometimes", "missings.sometimes"),
cpo.train = {
vlapply(data, selector)
},
cpo.retrafo = {
data[control]
})
```
username_0: Thanks. That's a nice start. If you want, you can close this issue. |
GoogleCloudPlatform/python-docs-samples | 197052818 | Title: _Rendezvous of RPC that terminated with (StatusCode.CANCELLED, Cancelled!)
Question:
username_0: I was running the following file:
/python-docs-samples/speech/grpc/transcribe_streaming.py
as following:
python transcribe_streaming.py
It works fine (I get transcript of my speech) until I press Ctrl+C. When I press Ctrl+C I get the following error:
^CTraceback (most recent call last):
File "transcribe_streaming.py", line 234, in <module>
main()
File "transcribe_streaming.py", line 225, in main
listen_print_loop(recognize_stream)
File "transcribe_streaming.py", line 172, in listen_print_loop
for resp in recognize_stream:
File "/usr/local/lib/python2.7/dist-packages/grpc/_channel.py", line 344, in next
return self._next()
File "/usr/local/lib/python2.7/dist-packages/grpc/_channel.py", line 335, in _next
raise self
grpc._channel._Rendezvous: <_Rendezvous of RPC that terminated with (StatusCode.CANCELLED, Cancelled!)>
I believe that this error shouldn't be happening because of (in the main() try section):
except face.CancellationError:
# This happens because of the interrupt handler
pass
So my question is what is the right way to stop transcribing? If I just do recognize_stream.cancel() I get these _Rendezvous StatusCode.CANCELLED errors even though there is except face.CancellationError exception handler. Is this a bug or I'm doing something wrong?
Answers:
username_1: Yeah - I just saw this too. I believe the library got updated and now the exception that's being thrown is different... fix forthcoming...
username_0: Ok, thanks!
Status: Issue closed
|
dyama/mruby-siren | 248297750 | Title: mruby-siren と Fusion360 の連帯の可能性
Question:
username_0: Autodesk社のFusion360と言うCADをご存知でしょうか?
こちらのCADは非常に多くのフォーマットのCADデータを
インポートすることが可能なのですが、サーフェス・ソリッドのみ
で点・線が欠落してしまいます。
https://forums.autodesk.com/t5/fusion-360-ri-ben-yu/igesrain-kabu-detanoinpoto/m-p/6606029#M3144
https://forums.autodesk.com/t5/fusion-360-ri-ben-yu/stp-xing-shi-dengnofairuwo-dumi-rumi-shuki-chushiwo-xingtta-shini-dian-xiannadono-yao-suga-xiaoeteiru/m-p/7218766#M6156
Fusion360もAPIは、かなり公開されており、
http://help.autodesk.com/view/fusion360/ENU/?guid=GUID-dc7ae251-e060-4d87-b6b8-e7f78abc0777
これらの問題を "Sirenを利用すれば解決出来るのでは?"
と思ったのですが、素人レベルで可能なものか・・・と感じております。
Siren自身の魅力は、バイナリ版であれば導入が非常に簡単なため、
多くのユーザーにも受け入れやすいものと感じているのですが、
mRuby↔Python間でデータのやり取りが可能なものか?どうか?
もよくわからない状態なのが正直なところです…。
Answers:
username_1: Fusion360 を触ったことがないですが、スケッチの概念があったり、[Curve3D クラス](http://help.autodesk.com/view/fusion360/ENU/?guid=GUID-283c1e3d-2260-4d5b-ae9e-f84b8f6a0ca9)があるところを見ると Fusion360 自体はしっかりした構造を持っていて、パーサ側の機能がないようですね。
siren で IGES ファイルを読み込み、Fusion360 向けの Python スクリプトを出力する mruby スクリプトを書いてみました。
```
#!siren
# coding: utf-8
iges = "sample.iges"
py = "sample.py"
# 評価用の IGES ファイルを生成
def make_sample_iges(path)
points = [[0, 0, 0], [1, 0, 0], [1, 1, 0], [2, 1, 0], [2, 2, 0]]
wire = Siren.polyline points
Siren.save_iges wire, path
end
# IGES ファイルがなければ生成
unless File.exist? iges
make_sample_iges iges
end
# IGES ファイルの読み込み
model = Siren.load_iges iges
# 書き出すスクリプトファイルを開く
file = File.open(py, "w")
file.puts "import adsk.core"
# 例えば、モデル中の直線要素を処理する
model.edges.select{|e| e.curve.is_a? Siren::Line }.each_with_index do |e, i|
if e.infinite?
# 無限直線の場合
c = e.curve
file.write "res#{i} = InfiniteLine3D.create("
file.write "Point3D.create(#{e.pos.x}, #{e.pos.y}, #{e.pos.z}), "
file.puts "Vector3D.create(#{c.dir.x}, #{c.dir.y}, #{c.dir.z}))"
else
# 有限直線(線分)の場合
file.write "res#{i} = Line3D.create("
file.write "Point3D.create(#{e.sp.x}, #{e.sp.y}, #{e.sp.z}), "
file.puts "Point3D.create(#{e.tp.x}, #{e.tp.y}, #{e.tp.z}))"
end
end
# スクリプトファイルを閉じる
file.close
```
例は直線だけですが、同じ要領で[円弧](http://help.autodesk.com/view/fusion360/ENU/?guid=GUID-fef681fd-ca49-4fce-b9bd-6e663021d2db)や [NURBS 曲線](http://help.autodesk.com/view/fusion360/ENU/?guid=GUID-ee0fb092-7a5c-4596-902e-fff01c7d50b1)も対応できそうです。
実行すると
```
S0000001
,,31HOpen CASCADE IGES processor 7.1,13HFilename.iges, G0000001
16HOpen CASCADE 7.1,31HOpen CASCADE IGES processor 7.1,32,308,15,308,15,G0000002
,1.,2,2HMM,1,0.01,15H20170807.123552,1E-007,2.,7Hyamadai,,11,0, G0000003
[Truncated]
102,4,3,5,7,9; 0000001P0000001
110,0.E+000,0.E+000,0.E+000,1.,0.E+000,0.E+000; 0000003P0000002
110,1.,0.E+000,0.E+000,1.,1.,0.E+000; 0000005P0000003
110,1.,1.,0.E+000,2.,1.,0.E+000; 0000007P0000004
110,2.,1.,0.E+000,2.,2.,0.E+000; 0000009P0000005
S 1G 4D 10P 5 T0000001
```
のような評価用 IGES ファイル `sample.iges` が生成され、それを読み込み
```
import adsk.core
res0 = Line3D.create(Point3D.create(0, 0, 0), Point3D.create(1, 0, 0))
res1 = Line3D.create(Point3D.create(1, 0, 0), Point3D.create(1, 1, 0))
res2 = Line3D.create(Point3D.create(1, 1, 0), Point3D.create(2, 1, 0))
res3 = Line3D.create(Point3D.create(2, 1, 0), Point3D.create(2, 2, 0))
```
といった Python スクリプト `sample.py` を出力します。 Fusion360 を動かせる環境が手元にないため、チェックはしていませんが、考え方はこれでいいはずだと思います。
ただ、 Fusion360 側をしっかり調べればもっとスマートな方法も見つかるかもしれません。
username_0: 早速ありがとうございます。(速いです・・・)
全く理解できていませんが、時間が出来次第勉強させていただきます。
username_0: エクスポートする為の Fusion360 → Siren は、Sirenスプリクトを書き出し → 実行で出来そうだな
と感触は持っていたのですが、インポートの Siren → Fusion360 は、どうすれば良いのだろう?
と思っておりました。
試してはいませんが、モジュールとして書き出したPythonスプリクトをインポート
すれば、実行できそうな気がしてきました。 色々とアドバイスありがとうございます。
username_0: みすぼらしいものですが、公開まで漕ぎ着けました。
https://forums.autodesk.com/t5/fusion-360-ri-ben-yu/3dcad-zhong-jianfairuno3dna-dian-qu-xianwoinpotosurusupurikuto/td-p/7370346 |
CDRH/api | 434884899 | Title: Look into possibility of using multiple language analyzers on content?
Question:
username_0: Right now we're splitting content into fields like `*_t_es` and `*_t_en` when we need to carry out a search with specific stemming, etc. Look into possibility of using this ES functionality:
https://www.elastic.co/guide/en/elasticsearch/guide/current/mixed-lang-fields.html#_analyze_multiple_times
We would want to compare with Solr possibilities so that we know if this would be a feature that could be lost if we change backends. |
minbrowser/min | 236904149 | Title: Page magnification resets randomly on Windows
Question:
username_0: As in title. After opening page (FB especially) and magnifying it to proper level it stays on this level for a while, and then resets to previous level. May be related to duble finger scrolling (sometimes d-f scrolling triggers reset).
Answers:
username_1: I think you're correct that this is related to scrolling; scrolling with the ctrl key pressed will zoom in or out. We could possibly look into making this less sensitive.
username_0: Thanks for reply :)
I've checked it double and here are the results: every type of scrolling resets page magnification, no matter it's d-f, down arrow, PgDn. It appears especially on FB page, but not only.
Browser also can't remember magnification of page and set same among tabs, I have to change it on every tab manually, it's quite annoying.
username_1: I've opened another issue for this: #372.
username_0: Win 10 Creators Update
- youtube.com , after clicking notification bell, not connected to scrolling
- google.com serch results page, randomly, also after clicking on grafic results tab,
- scholar.google.pl results page, after clicking on "Od 2016" on left panel, not connected to scrolling
- czasgentlemanow.pl/2017/06/kiedy-przestrzegac-zasad/ after clicking in link and going back to this page magnification is reseted (nothing, when link is opened in new tab)
- facebook.com - it appears when loading new content eg. in newsfeed
username_1: The 2nd, 3rd and 4th issues are because zoom is reset whenever a new page is loaded, which I've added an issue for in #372.
Otherwise, I can't reproduce this. It looks like scrolling will only change the zoom if ```e.metaKey``` is true (https://github.com/minbrowser/min/blob/master/js/webview/swipeEvents.js#L72), which on windows is the windows key (and actually I think this code has a bug; this is supposed to be the ctrl key). Is there any way the windows key could be pressed when you are scrolling?
username_0: No, it's not possible. Ctrl zooming works well. I think all this issue is #372.
PS I have no idea about code so far :D
username_1: OK, I'll close this and leave #372 open then.
Status: Issue closed
|
IntergalacticAvenger/myRPG | 800889417 | Title: test
Question:
username_0: <img width="413" alt="Screen Shot 2021-02-03 at 10 14 36 PM" src="https://user-images.githubusercontent.com/53759586/106840062-66f91380-666d-11eb-9d98-36bbff0d107f.png">
<img width="358" alt="Screen Shot 2021-02-03 at 10 14 55 PM" src="https://user-images.githubusercontent.com/53759586/106840065-6791aa00-666d-11eb-8d9b-a9031f83732a.png">
<img width="225" alt="Screen Shot 2021-02-03 at 10 15 16 PM" src="https://user-images.githubusercontent.com/53759586/106840066-6791aa00-666d-11eb-96f3-b4cd6de6c516.png">
<img width="281" alt="Screen Shot 2021-02-03 at 10 15 29 PM" src="https://user-images.githubusercontent.com/53759586/106840067-6791aa00-666d-11eb-90b5-c9d4f02c87bf.png">
<img width="393" alt="Screen Shot 2021-02-03 at 10 15 50 PM" src="https://user-images.githubusercontent.com/53759586/106840068-6791aa00-666d-11eb-85e1-5fdf61e746e9.png"><issue_closed>
Status: Issue closed |
tj/consolidate.js | 36607225 | Title: Layout support for handlbars
Question:
username_0: Hi, would you accept a pull request to add layout support for handlebars? Are you opposed to adding layout support to consolidate.js?
I am trying to get something like this working:
```js
app.get("/home", function(req, res) {
res.render("home", {
layout: "main",
title: "hello world",
test: test
});
});
```
Thanks.
Status: Issue closed
Answers:
username_1: I think it's better to support features exposed by the template engine and not add additional functionality in consolidate.
Check out [this handlebars layouts library](https://github.com/shannonmoeller/handlebars-layouts) if you're interested in adding layouts to handlebars. |
OBOFoundry/OBOFoundry.github.io | 479882335 | Title: ontology table is missing from homepage
Question:
username_0: http://obofoundry.org/ - the page is missing the table with the ontologies, the page is loading for me in Chrome and Firefox, but the able is missing

Maybe related to #1022
Answers:
username_1: Sorry, I should have clarified. It's the table that doesn't load, not the whole page. So yes, this is the same as #1022. :)
username_0: I will close this ticket then :)
Status: Issue closed
|
googleads/googleads-php-lib | 256416213 | Title: v29.0.0 installation problem
Question:
username_0: When I try to install v29.0.0 like before versions, I got below errors. Can you help me please?
I installed with steps as mentioned https://github.com/googleads/googleads-php-lib
root@nod4 [/home/www]# php composer.phar require googleads/googleads-php-lib
Do not run Composer as root/super user! See https://getcomposer.org/root for details
Using version ^29.0 for googleads/googleads-php-lib
./composer.json has been updated
Loading composer repositories with package information
Updating dependencies (including require-dev)
Your requirements could not be resolved to an installable set of packages.
Problem 1
- Installation request for googleads/googleads-php-lib ^29.0 -> satisfiable by googleads/googleads-php-lib[29.0.0].
- Conclusion: remove google/auth v0.11.1
- Conclusion: don't install google/auth v0.11.1
- googleads/googleads-php-lib 29.0.0 requires google/auth ^1.0.0 -> satisfiable by google/auth[v1.0, v1.0.1].
- Can only install one of: google/auth[v1.0, v0.11.1].
- Can only install one of: google/auth[v1.0.1, v0.11.1].
- Installation request for google/auth (locked at v0.11.1) -> satisfiable by google/auth[v0.11.1].
Installation failed, reverting ./composer.json to its original content.
Answers:
username_1: Hello,
Could you please share what the `composer.json` of your project looks like?
Cheers,
Knack
username_1: And does this flag (--update-with-dependencies) for composer help?
username_0: Error updated ;
php composer.phar require googleads/googleads-php-lib
Do not run Composer as root/super user! See https://getcomposer.org/root for details
Using version ^29.0 for googleads/googleads-php-lib
./composer.json has been updated
Loading composer repositories with package information
Updating dependencies (including require-dev)
Your requirements could not be resolved to an installable set of packages.
Problem 1
- The requested package googleads/googleads-php-lib No version set (parsed as 1.0.0) is satisfiable by googleads/googleads-php-lib[No version set (parsed as 1.0.0)] but these conflict with your requirements or minimum-stability.
Problem 2
- The requested package google/auth (installed at v0.11.1, required as ^1.0.0) is satisfiable by google/auth[v0.11.1] but these conflict with your requirements or minimum-stability.
Installation failed, reverting ./composer.json to its original content.
composer.json like below
{
"name": "googleads/googleads-php-lib",
"description": "Google Ads APIs Client Library for PHP (AdWords and DFP)",
"require": {
"php": ">=5.5.9",
"ext-openssl": "*",
"ext-soap": "*",
"google/auth": "^1.0.0",
"guzzlehttp/guzzle": "^6.0",
"guzzlehttp/psr7": "^1.2",
"monolog/monolog": "^1.17.1",
"phpdocumentor/reflection-docblock": "^3.0.3",
"symfony/serializer": "^2.8.0 || ^3.0.3"
},
"require-dev": {
"php": ">=5.5.17",
"phpunit/phpunit": "^4.8"
},
"suggest": {
"php-64bit": ">=5.5.9"
},
"homepage": "https://github.com/googleads/googleads-php-lib",
"license": "Apache-2.0",
"autoload": {
"psr-4": { "Google\\AdsApi\\": "src/Google/AdsApi/" }
},
"authors": [
{
"name": "Google",
"homepage": "https://github.com/googleads/googleads-php-lib/contributors"
}
]
}
username_1: Could you remove your `composer.json` and try again?
Probably the error was caused b your `composer.json`.
It specifies v28.0.0 but your command demands v29.0.0.
In v29.0.0, we have updated the requirement of google/auth to v1.0.0, which conflicts with that of older versions.
Knack
username_0: Thank you very much for your help after deleting composer.json with this code it worked successfully.
`php composer.phar require googleads/googleads-php-lib --update-with-dependencies`
Status: Issue closed
|
karaage0703/karaage-ai-book | 776424504 | Title: 文字の重複?
Question:
username_0: **ページ数**
P.34 きゅうり農家の画像の上
**内容**
「AI」という文字が重複している
以下、本文
AIを用いて、きゅうりの~作業を、AIにより実現しています。
**コメント**
個人的に気になったのですが、意図的でしたらすみません m(__)m
Answers:
username_1: @username_0 さん
貴重なご意見、ありがとうございます。確かに「AIを用いて、きゅうりの等級(ランク)を選別する作業を、AIにより実現しています。」より「AIを用いて、きゅうりの等級(ランク)を選別する作業を実現しています。」の方が自然ですね。
@KazumaAndoh さん
意味は通じますが、直した方が自然かと思います。正誤表への追記はいかがいたしましょうか?
Status: Issue closed
username_1: [正誤表](https://github.com/username_1/karaage-ai-book/blob/master/ERRATA.md)に追記しましたのでcloseします。 |
composer/composer | 538645704 | Title: composer-setup.exe identified as known malware via Carbon Black
Question:
username_0: Associates in network are flagged when working with composer-setup.exe SHA 6a0b75feaf3823a1054274a362bdd92b9687b265c0f0ff782741853c7995cdd8 as Known Malware via the networks Carbon Black protections.
Answers:
username_1: Obviously we haven't included any malware in this. It does use the latest version of Inno Setup (6.03), so perhaps this explains the false positive.
Source code: https://github.com/composer/windows-setup
Status: Issue closed
username_1: https://www.virustotal.com/gui/file/6a0b75feaf3823a1054274a362bdd92b9687b265c0f0ff782741853c7995cdd8/detection
Nothing is detected on VirusTotal now (having initially shown 5 false-positives, which is much more than the usual none or one). |
CocoaPods/CocoaPods | 771627863 | Title: Install CocoaPods macOS BigSur Apple M1 MacBook Pro
Question:
username_0: <!--
ℹ Please fill out this template when filing an issue.
All lines beginning with an ℹ symbol instruct you with
what info we expect.
Before you start, are you using the latest CocoaPods release?
A lot changes with Xcode releases that are not backwards compatible.
Not an issue about the CocoaPods command line app? Please file an issue in the appropriate repo - https://github.com/CocoaPods
Issues are for feature requests, and bugs; questions should go to Stack Overflow
Using CocoaPods <= 0.39: https://blog.cocoapods.org/Sharding/
Using Xcode 10.1: Requires CocoaPods 1.6.0 or above.
Issue with Nanaimo not loading:
Please run `[sudo] gem uninstall nanaimo` and remove all but the latest version.
Issues with `pod search`? Try deleting your cache `rm -rf ~/Library/Caches/CocoaPods`first.
-->
* [ ] I've read and understood the [*CONTRIBUTING* guidelines and have done my best effort to follow](https://github.com/CocoaPods/CocoaPods/blob/master/CONTRIBUTING.md).
# Report
## What did you do?
ℹ Please replace these two lines with what you did.
e.g. Run `pod install`
## What did you expect to happen?
ℹ Please replace these two lines with what you expected to happen.
e.g. Install all pod dependencies correctly.
## What happened instead?
ℹ Please replace these two lines with of what happened instead.
e.g. Pod A is missing the subspec B for target C.
## CocoaPods Environment
ℹ Please replace these two lines with the output of `pod env`.
e.g. via `pod env | pbcopy`
## Project that demonstrates the issue
ℹ Please link to a project we can download that reproduces the issue.
You can delete this section if your issue is unrelated to build problems,
i.e. it's only an issue with CocoaPods the tool.
Answers:
username_1: Hy, you can find a solution to your problem here:
https://github.com/CocoaPods/CocoaPods/issues/9907#issuecomment-729980327
username_2: That or many other issues related to ffi gem here. We've kept one issue open https://github.com/CocoaPods/CocoaPods/issues/9907 |
joopert/nad_receiver | 775008438 | Title: typeerror cannot concatenate 'str' and 'int' objects
Question:
username_0: I had some time to put the 0.2.0 in a custom component for home assistant. This resulted in the issue for me when setting the volume. Home assistant passes this as an int while in the new code it must be passed as a str.
In the old code we converted it to a str https://github.com/username_0/nad_receiver/blob/0.1.0/nad_receiver/__init__.py#L37 and in the new code we don't https://github.com/username_0/nad_receiver/blob/master/nad_receiver/__init__.py#L46
To be backwards compatible we should convert it to str again?
@gladhorn any suggestion for this? In my custom component I just fixed it with str(value)
Answers:
username_1: I want to remember that some platform sent e.g. -20.5 (not the telnet platform). But this fix should work also for that case, but
I wonder, is not the 'type' information then incorrect. I guess I would correct the issue @ main_volume to better
show where the unexpected input is received. main_volume then also has incorrect type information.
Do many other commands also have unexpected input ? Then go with the current patch.
My 5-cents
username_0: There was at least one other method which also had some typeerror.
I was thinking about the type information as well, so I think your way would be cleaner. I will look into which other method(s) have to be converted to str and do a new commit. Thanks.
Status: Issue closed
|
facebookresearch/mmf | 756626735 | Title: Where is the specific code of TextVQA
Question:
username_0: Hi, FAIR.
Thank you for opening the code.
I want to study and reimplement the code of TextVQA.However, I'm not familiar with this mmf system.
I can't find the specific module like data loader, preprocessing, extract_image_featue...
So, I need help where is the reference fo LoRRA.
Thank you!
Answers:
username_1: Hi,
- Code for TextVQA dataset is available in https://github.com/facebookresearch/mmf/tree/master/mmf/datasets/builders/textvqa
- Model code for LoRRA is available in https://github.com/facebookresearch/mmf/blob/master/mmf/models/lorra.py
- Required configs for the TextVQA and LoRRA are in https://github.com/facebookresearch/mmf/tree/master/projects/lorra
Let us know if something is confusing or not clear.
username_0: @username_1
Hi, username_1.
Thank you for your kind reply.
I have some questions.
I succeeded to run the extract_features_vmb.py by looking at many issues in your repository.
1. Could you provide the final extracted image features ? it is because it takes so many times since my GPU resources are not enough.
2. Could I know the part of preprocessing the question tokens and OCR tokens?
3. Could I get the pre-trained LoRRA model?
4. Could you let me know how to run the test and evaluation?
Sorry for the many questions ... I'm so confusing the integrated repository system.
Thank you :)
username_1: 1. If you download TextVQA dataset using MMF, it will automatically download the features you are looking for.
2. For preprocessing, please check the config you want in the folder I shared above and look for the processors configuration. Map this to processors present in `mmf/datasets/processors` folder.
3. We don't have pretrained model for LoRRA directly usable with current mmf as it was superseded by better model M4C which we do have inside MMF.
4. Please check the documentation for M4C on how to run M4C. https://mmf.sh/docs/projects/m4c
username_0: @username_1
Thank you. It was very helpful. I closed this issue :+1:
Status: Issue closed
username_2: closing as the author of this issue suggested. |
deezer/spleeter | 527623429 | Title: [Bug] Tuple formatting incorrectly included in output directory name
Question:
username_0: ## Step to reproduce
```python
from spleeter.spleeter.separator import Separator
separator = Separator('spleeter:2stems')
filein = 'GilScottHeron_WeAlmostLostDetroit.mp3'
fileout = './stems'
separator.separate_to_file(filein, fileout, codec='mp3')
```
## Output
Output directory name : `./stems/('GilScottHeron_WeAlmostLostDetroit', '.mp3')/`
Expected output: `./stems/GilScottHeron_WeAlmostLostDetroit/`
## Environment
<!-- Fill the following table -->
| | |
| ----------------- | ------------------------------- |
| OS | MacOS |
| Installation type | `pip` |
| RAM available | 8 GB |
| Hardware spec | CPU: 3.2 GHz Intel Core i5, GPU: NVIDIA GeForce GT 755M 1 GB |
## Additional context
The reason for this bug is [line 124 in `separator.py`](https://github.com/deezer/spleeter/blob/85ff00797f6c615c62885793923eca952e9e791f/spleeter/separator.py#L124). There needs to be a `[0]` added after the output of `splitext` so that the directory name is created from a `string`, not a `tuple`.
Status: Issue closed
Answers:
username_1: Thanks for spotting the bug and proposing fix :).
We will integrate your PR ASAP !
username_0: Great! Thanks for making this awesome package available! |
KeplerGO/ScientificOpportunities | 377970221 | Title: Investigating Flicker noise across the Kepler/K2 sample
Question:
username_0: As discussed in #17 and raised by @kstassun, "Flicker" noise is an excellent way to characterize stellar properties using photometric time series. Flicker noise occurs due to the granulation on the surface of a star, and can be used to measure properties such as the surface gravity. It is possible to identify Flicker noise both in the frequency domain (where it is a "background" to asteroseismic acoustic oscillations) and by fitting in the time domain.
From @ktstassun in #17:
`...Bastien et al. (2013, 2016) have shown that from the Kepler long-cadence light curves it is possible to extract the granulation "flicker" that correlates very strongly with stellar surface gravity and thus provides a means for measuring stellar surface gravity with a precision of ~0.1 dex, for stars with logg > 2.5. More recent work demonstrates that the addition of metallicity as a term in the fit enables the surface gravity precision to be improved to ~0.05 dex (Corsaro et al. 2017, Tayar et al. 2018). Application of these methods to the full Kepler + K2 data set holds the promise of enabling the determination of precise stellar properties for stars far beyond the original Kepler footprint. Finally, overlap of the Kepler/K2 sample with upcoming TESS observations should enable calibration and extension of the granulation "flicker" methodology to TESS stars across the entire sky (see, e.g., Stassun et al. 2018).`
Applying an analysis of Flicker noise across the entire Kepler sample, perhaps capitalizing on new methods in machine learning, would provide a catalog of robust, independently determined stellar properties using solely the Kepler sample.
Answers:
username_0: If anyone has any further comments on this, including how many stars may be amenable to analysis of Flicker noise, and how many of those may or may not have already been analyzed, please do comment here. |
Framinus/roam | 279174401 | Title: User Name and Profile Pic missing from review detail page.
Question:
username_0: Every post should show the username and profile pic, just like they do on the cities page.
Status: Issue closed
Answers:
username_0: Just copied over mini-profile div from the cities page with slight style modifications and it worked! need to fix width on full post view, but otherwise good. |
awslabs/aws-sdk-ios-samples | 275121761 | Title: How to implement custom authentication using only Mobile Number
Question:
username_0: Hi,
I would like to implement user login using just mobile number (password less).
Could you please advice how I could achieve this.
I am specifically looking for sample code for iOS swift or at least guidance on this.
Signin/Signup flow
1. User submits his mobile number
2. User will receive pass code
3. Upon entering the passcode, the user shall be signed in
Answers:
username_1: I have encounter the same issue as well. Any sample code on this? I saw Android sdk already support custom challenge.
username_1: After many trial and error, I have been able to sign in with custom challenge using the sdk. I have documented it here: https://medium.com/@username_1/aws-cognito-user-pools-with-mobile-sdk-for-ios-using-custom-challenge-5d40a06a3b07
username_2: I am having the same issue as well,
@username_1 the link you've posted is expired, can you please give the example code to login from mobile number..?
username_2: FYI user.getSession is giving an error with code=20 message="invalid pass or username"
username_3: Hi @username_1,
I am facing the same issue while sign in with mobile number using custom challenge.
Can you please provide any sample regarding this?
username_1: @username_3 @username_2 Please checkout the updated link here: https://medium.com/@username_1/aws-cognito-user-pools-with-mobile-sdk-for-ios-using-custom-challenge-df2c3b163d3d
username_4: Hello @username_0, sorry for the delay. Do you still need help with the issue? |
VoltDB/voltdb-client-go | 377810529 | Title: Passing VoltTable as an argument to stored procedure
Question:
username_0: Hi,
Based on the documentation a stored procedure can accept these types as arguments:
Integer types | byte, short, int, long, Byte, Short, Integer, and Long
Floating point types | float, double, Float, Double
Fixed decimal types | BigDecimal
String and binary types | String and byte[]
Timestamp types | org.voltdb.types.TimestampType java.util.Date, java.sql.Date, java.sql.Timestamp
**VoltDB type | VoltTable**
The VoltTable type is unexported in this client so it cannot be used for stored procedure invocations.
Based on the code the encoder may not even handle this argument.
I'd be very grateful if You could share some information on this matter.
Thank You,
Best Regards,
Roland
Answers:
username_1: Hi Roland,
The wire protocol supports serialization for the basic data types for input parameters, plus VoltTable for results coming back from the database. Since the java client library uses some of the same code as the database, it can serialize a VoltTable to the byte code used by the wire protocol. Therefore, at some point it was decided to allow VoltTable to be supported as input to a stored procedure. It is very rarely used to my knowledge.
I'm not sure if any other client libraries support the creation of a VoltTable from code, and the serialization of a VoltTable onto the wire to be sent to the database as the parameter to a stored procedure invocation. Perhaps the python client can do it, I'm not certain. I think for most of the other client libraries, it's just not implemented.
Best regards,
Ben
username_0: I understand, thank You, Ben!
Status: Issue closed
|
TarikHuber/react-most-wanted | 355696655 | Title: Can't create new components that take advantage of react-intl
Question:
username_0: Hi!
First off, impressive work, especially for my current usecase,
I already write a react app with react-intl but my webpack build didn't work well enough and this projects fits perfectly.
I wanted to attach a Header and Footer component and I'm getting the following error:
` <IntlProvider> needs to exist in the component ancestry`
Normally i would see the provider but I coudn't find it anywhere on the src
I'm adding the Header component to the index inside the router, as it should accept components out side the switch, everything seems fine until I try to use the intl `injectIntl(Header)` and add the references required
Is there a configuration I'm missing?
Thanks in advantage
Answers:
username_1: Hi @username_0 ,
Thx 😄
The IntlProvider is set in the `rwm-shell` Root component. That one is in the App component we use in our project so everything outside of that will have no IntlProvider. Here are two possible solutions for you:
1 - Try to use your Header and Footer inside the App component of `rmw-shell` and with the routes switch. Maybe you can explain more what your goal is.
2 - Add a IntlProvider on your own outside the App component. Downside of this one is that you have to manage the language settings and messages on your own. But from the `react-intl` prespective it is no problem to have more Providers inside each other.
Let me know if this could help.
Status: Issue closed
username_2: Hey @username_1,
unfortunately if I change the landingPage to `lazy(() => import('../pages/Home/Home')),` in `config.js` it returns the Error `Error: [React Intl] Could not find required `intl` object. <IntlProvider> needs to exist in the component ancestry.`. I know the cause of this error but don't understand how to fix it. The shell's code is quite hard to read for somebody who is not familiar with node
username_1: Hi. Don't make the landing page async. It is the first page to load and an asyn does not make sense. After the first load it is already offline first and after that it doesn't matter anyway. If the landing page is to big you can splitt it but the first part should be sync. The demo page of rmw js madr like that. We even load the rest only on a scroll event ;)
username_2: @username_1 I did not make it async, it was already async as I downloaded RMW.
`config.js`
```Javascript
pages: {
LandingPage: import('../pages/Home/Home'),
PageNotFound: lazy(() => import('../pages/PageNotFound/PageNotFound')),
}
```
does not work as well :(
username_1: That is not async. Only the one with lazy is async.
username_2: @username_1 Ok, but how do I change the landingpage to eg. Home?
username_1: Can you explain more your use case?
username_2: @username_1 I added a `<Redirect.../>` to `/home` in the landing page and set every route in `routes.js` to an `unauthorizedRoute` |
EarthSystemDiagnostics/cpt-picarr | 492071438 | Title: UI testing
Question:
username_0: At the moment, bugs occur frequently and can be introduced easily because there is no comprehensive UI testing with `shinytest`. I suggest adding one shinytest-file per page.
- [ ] add an example project that is used for testing to the repo
- [ ] make it possible to override the BASE_PATH using an environment variable (used for testing)
- [ ] add the shinytest scripts
Answers:
username_0: It is unclear whether or not this applies to our app. We use modules and helper files, however we also have a file `app.R` that brings all the application logic together.
Looking further I found [this issue](https://github.com/rstudio/shinytest/issues/3) where someone is encountering issues when trying to use shinytest with modules. The maintainers did not answer whether shinytest can work with modules.
I also found [this blog post](https://www.r-bloggers.com/shiny-modules/) that makes it seem like shinytest and modules can be used together. In the blog post they use a different directory structure (app.R in inst/app/; modules in R/). Maybe using the different directory structure can solve our issue.
## Conclusion
It is unclear if we can use `shinytest` for UI testing. Changing the directory structure may help. We can also [open an issue](https://github.com/rstudio/shinytest/issues) and ask for help from the maintainers.
username_0: Update: I opened an issue in the shinytest repository (https://github.com/rstudio/shinytest/issues/283)
username_0: @username_1 I have not received an answer from the shinytest maintainers, yet. It will be difficult to find the time to experiment with ui testing to try and somehow make it work in the limited time we have left. Therefore I would suggest to not do any ui testing. Quite a few functions are already being tested with testthat, we will have to rely on manual testing to make sure the UI works. Is that fine with you or would you prefer to take the time and try to get shinytest to work?
username_1: I am fine with no doing any UI testing now.
Status: Issue closed
|
kurodakazumichi/NoJobDungeon | 655224513 | Title: プレイヤーが攻撃して敵が攻撃をくらって死ぬ処理
Question:
username_0: - プレイヤーが攻撃ボタンを押す
- プレイヤーが攻撃の動きをする
- 攻撃をくらった敵が「痛い!」みたいな演出をする
- 攻撃を食らった敵が死ぬ
攻撃力とかHPとか細かい事は後で考えるとして
大まかな流れの実装
Answers:
username_0: - [ ] PlayerManager.HasOnMovePlayerとEnemyManager.HasOnMoveEnemy
HasOnMoveXXXXといいつつ、実際にはアイドル状態になってない奴が存在するかどうかの判定なので
HasActivePlayer、HasActiveEnemyの方に変える
- [x] DungeonSceneのStateMachineに設定しているUpdate処理について
大半が動いているプレイヤー、敵がいるかどうかの判定で共通処理に出来そうな雰囲気
しかし遷移先のPhaseは異なるので完全に共通化はできないので
とりあえず現状維持でいこう。
- [ ] IAttackableの内容は精査して作り替えていく事になるだろう
- [ ] Enemyに実装しているhpやisAcceptAttackの処理のリファクタリング
hpなどはStatusクラスなど別クラスにまとめて持たせる形にしていきたい
Status: Issue closed
|
ShammyLevva/FTAnalyzer | 1142757567 | Title: Custom Filters not working
Question:
username_0: v8.5.2.0
On the Main Lists Tab, Individuals Report
Looking at Budgie Code or Relation to Root

When you use the custom filter i.e. Begins with, and enter a value, it produces 0 results
This snapshot is from v8.5.0.0beta3 where the image is clearer.
 |
innoveit/react-native-ble-manager | 316071204 | Title: Issue running with react-native-navigation
Question:
username_0: - react-native-ble-manager v6.2.9
- react-native 0.54.4
- Android 8.0.0
I am somewhat new to react-native so be kind.
I was able to get the react-native-ble-manager example working on my local hardware with relatively few issues. From there I wanted to integrate this into my current app; I of course used the example application as a template to accomplish this.
What I noticed is that on Android the BleManager.start({showAlert: false}); would result in:
`TypeError: Cannot read property 'start' of undefined`
Even though it was specifically imported at the top of the view. I also noticed that when I would console log NativeModules, BleManager was not one of the modules that was included in the list. Note these issues do not occur when running iOS (in emulation granted).
When exploring a bit deeper the main difference between the example app and my app is I'm using react-native-navigation by Wix (https://wix.github.io/react-native-navigation/#/). I'm using this so I can have a login screen that then navigates to a tabbed application.
This requires MainApplication to extend a different class:
`package com.macaw;
import android.app.Application;
import com.facebook.react.ReactNativeHost;
import com.facebook.react.ReactPackage;
import com.facebook.react.shell.MainReactPackage;
import com.facebook.soloader.SoLoader;
import com.reactnativenavigation.NavigationApplication;
import it.innove.BleManagerPackage;
import java.util.Arrays;
import java.util.List;
public class MainApplication extends NavigationApplication {
private final ReactNativeHost mReactNativeHost = new ReactNativeHost(this) {
@Override
public boolean getUseDeveloperSupport() {
return BuildConfig.DEBUG;
}
@Override
protected List<ReactPackage> getPackages() {
return Arrays.<ReactPackage>asList(
new MainReactPackage(),
new BleManagerPackage()
);
}
@Override
protected String getJSMainModuleName() {
return "index";
}
};
@Override
public ReactNativeHost getReactNativeHost() {
return mReactNativeHost;
}
@Override
[Truncated]
super.onCreate();
SoLoader.init(this, /* native exopackage */ false);
}
@Override
public boolean isDebug(){
return BuildConfig.DEBUG;
}
@Override
public List<ReactPackage> createAdditionalReactPackages(){
return null;
}
}
`
I'm wondering if this could perhaps be causing the issue when running on Android. Are these two npm mutually exclusive? Is there a way to get them to play nicely together? Thoughts?
Thanks,
Clark
Answers:
username_1: Hi, I think you only miss something in the installation, try again.
Status: Issue closed
|
ecederstrand/exchangelib | 413873188 | Title: can't pickle _thread.lock objects
Question:
username_0: Am trying to process emails (exchangelib messages) parallely using Python multiprocessing but got "can't pickle _thread.lock objects" error. How to release this lock?
Answers:
username_0: Installing exchangelib 1.12.2 fix the problem. Now, I get new error: "AttributeError: 'FileAttachment' object has no attribute '__dict__'.
username_1: Please post some example code that causes this error.
username_0: Sample code ...
Aim: to process emails parallely
def func(emails):
do sth
# emails = {emailId:email}
# emailId: 1 ... 100 for hundred emails
# emails: exchangelib message
jobs = []
nthreads = 4
jobEach = len(emails)//nthreads
firstEmailId = list(emails.keys())[0]
lastEmailId = list(emails.keys())[-1]
for i in range(nthreads):
if i == nthreads - 1:
emailIds = [(jobEach*i)+firstEmailId,lastEmailId]
emailsProcessor = {emailId:emails[emailId] for emailId in emailIds}
else:
emailIds = [(jobEach*i)+firstEmailId,(jobEach*(i+1))+firstEmailId]
emailsProcessor = {emailId:emails[emailId] for emailId in emailIds}
p = multiprocessing.Process(target=func, args=(emailsProcessor))
jobs.append(p)
# print(emailIds)
for p in jobs:
p.start()
for p in jobs:
p.join()
Error message (not full) :
...
File "C:\Users\~\AppData\Local\Continuum\anaconda3\lib\multiprocessing\reduction.py", line 60, in dump
ForkingPickler(file, protocol).dump(obj)
File "C:\Users\~\AppData\Local\Continuum\anaconda3\lib\site-packages\exchangelib\attachments.py", line 188, in __getstate__
state = self.__dict__.copy()
AttributeError: 'FileAttachment' object has no attribute '__dict__'
username_1: The issue happens when the `Message` items are being serialized and de-serialized when they are passed to the processes.
In general with multiprocessing, it's much better to pass around simple strings and integers instead of full objects, and then load the full objects within the process run method. One reason is to avoid trying to pickle something that cannot be pickled. Another reason is to avoid passing potentially huge amounts of data around in IPC channels. For example, attachments can be very large.
In your case, you should query message IDs in your main thread and then fetch the full items in the run method. So, do something like:
```python
from multiprocessing import Pool
pool = Pool(processes=10)
ids = account.inbox.all().values('id', 'changekey')
def worker(item_ids):
items = account.fetch(item_ids)
# Do something with the items
pool.map(worker, ids, chunksize=100)
```
That said, it is a bug that attachments cannot be pickled. I have now pushed a fix and added a test for this bug.
username_0: Thanks. Since I don't want all inbox to be fetched, I did this:
account = ...
def worker(item_ids):
items = account.fetch(item_ids)
for item in items:
print(item.subject)
if __name__ == '__main__':
SomeFolder = ...
dateToCheck1 = EWSDateTime(2018, 5, 10, 10, 00, 00,0, pytz.UTC)
dateToCheck2 = EWSDateTime(2018, 5, 10, 20, 15, 00,0, pytz.UTC)
pool = Pool(processes=4)
ids = SomeFolder.filter(datetime_received__range=(dateToCheck1,dateToCheck2)).values('id', 'changekey')[:5]
pool.map(worker, ids, chunksize=5)
but throws this error (runs fine with out the multiprocessing):
...
return self._map_async(func, iterable, mapstar, chunksize).get()
File "~\AppData\Local\Continuum\anaconda3\lib\multiprocessing\pool.py", line 644, in get
raise self._value
TypeError: unhashable type: 'slice'
username_1: You may need to materialize the list of IDs:
```python
pool.map(worker, list(ids), chunksize=10)
```
username_0: Thanks. List didn't work for me but user defined iterator slice using tuple works.
Status: Issue closed
|
Mercury1089/2018-robot-code | 294134032 | Title: Tune PID Values
Question:
username_0: ## GOAL
Tune PID values for when it is used on carpet.
## ACCEPTABLE CRITERIA
Tune PID for the following commands:
- [ ] RotateRelative
- [ ] DriveDistance
- [ ] MoveOnPath
Answers:
username_1: Just marking this done....
Status: Issue closed
|
kamuiroeru/nitac-nenpo-latex | 538207653 | Title: Warningが消えない
Question:
username_0: ```
Package caption Warning: Unsupported document class (or package) detected, usage of the caption package is not recommended.
```
キャプションのフォントが **ゴシック** じゃないとダメって指定なので、仕方なく ***captionパッケージ*** 使ってるんだけど、それが非推奨なのでワーニングが出てるみたいです。
直せたら直したいが、めんどくさいので放置… |
madhawav/YOLO3-4-Py | 341991359 | Title: SIGSEGV when creating a Detector
Question:
username_0: Hi !
First, thank you for your work !
I was first using the darknet python script, but with really poor performances. As yours seems better I am trying to use it but get a segfault each time I try to instantiate a Detector.
I have first installed your lib through PyPi, but it did not work. Then building it from source with my own build of Darknet, was successful when installing, but not working on execution, with always a segfault error. Apparently after running valgrind it may concern memory allocation ...
```
Access not within mapped region at address 0x0
==18723== at 0x16254939: free_network (in /home/pbreton/.local/lib/python3.6/site-packages/__libdarknet/libdarknet.so)
==18723== by 0x15FE2FF8: __pyx_pf_9pydarknet_8Detector_6__dealloc__ (pydarknet.cpp:3157)
==18723== by 0x15FE2FF8: __pyx_pw_9pydarknet_8Detector_7__dealloc__ (pydarknet.cpp:3142)
==18723== by 0x15FE2FF8: __pyx_tp_dealloc_9pydarknet_Detector(_object*) (pydarknet.cpp:5861)
==18723== by 0x15FE6FC7: __pyx_tp_new_9pydarknet_Detector(_typeobject*, _object*, _object*) (pydarknet.cpp:5847)
==18723== by 0x5591FD4: ??? (in /usr/lib/libpython3.6m.so.1.0)
==18723== by 0x558A64B: _PyObject_FastCallDict (in /usr/lib/libpython3.6m.so.1.0)
==18723== by 0x5552B41: ??? (in /usr/lib/libpython3.6m.so.1.0)
==18723== by 0x5522859: _PyEval_EvalFrameDefault (in /usr/lib/libpython3.6m.so.1.0)
==18723== by 0x5552FC7: PyEval_EvalCodeEx (in /usr/lib/libpython3.6m.so.1.0)
==18723== by 0x552240B: PyEval_EvalCode (in /usr/lib/libpython3.6m.so.1.0)
==18723== by 0x55F2213: ??? (in /usr/lib/libpython3.6m.so.1.0)
==18723== by 0x55F50AD: PyRun_FileExFlags (in /usr/lib/libpython3.6m.so.1.0)
==18723== by 0x55F5294: PyRun_SimpleFileExFlags (in /usr/lib/libpython3.6m.so.1.0)
```
Do you have any idea ...?
Thank you ...
Answers:
username_0: Everything is fine concerning the command line, I checked myself and manually everything works. The `pkg-config` is successfully returning the correct opencv libs :
```-I/usr/include/opencv```
```-lopencv_stitching -lopencv_superres -lopencv_videostab -lopencv_aruco -lopencv_bgsegm -lopencv_bioinspired -lopencv_ccalib -lopencv_dnn_objdetect -lopencv_dpm -lopencv_face -lopencv_photo -lopencv_freetype -lopencv_fuzzy -lopencv_hdf -lopencv_hfs -lopencv_img_hash -lopencv_line_descriptor -lopencv_optflow -lopencv_reg -lopencv_rgbd -lopencv_saliency -lopencv_stereo -lopencv_structured_light -lopencv_phase_unwrapping -lopencv_surface_matching -lopencv_tracking -lopencv_datasets -lopencv_text -lopencv_dnn -lopencv_plot -lopencv_xfeatures2d -lopencv_shape -lopencv_video -lopencv_ml -lopencv_ximgproc -lopencv_calib3d -lopencv_features2d -lopencv_highgui -lopencv_videoio -lopencv_flann -lopencv_xobjdetect -lopencv_imgcodecs -lopencv_objdetect -lopencv_xphoto -lopencv_imgproc -lopencv_core```
The problem seems to be concerning the way python is accessing the libraries.
Status: Issue closed
username_0: Ok it is truly my fault. I used 3 parameters instead of 4, forgetting that the third is a 0. Thus with 3 bytes as parameters it was logically crashing.
Thank you for your time, |
HodorNV/ALOps | 568269846 | Title: How-to avoid generating an artifact when doing a build for a specific branch
Question:
username_0: Is there an option that I can add in one of the build steps in the yaml file to specify that the artifact(s) should **not** be created if the following criteria is met:
- Branch = develop
- Build reason = Pull Request build
Answers:
username_1: I don't think there is - you could create a dedicated build pipeline, obviously ;-). If you work with yaml-templates, it's rather efficient, still (reusing yaml-configurations in multiple pipelines..).
username_0: ok
Status: Issue closed
|
jooby-project/jooby | 760514090 | Title: The jooby-apt should not require jaxrs jars
Question:
username_0: For annotation processing you don't actually need the annotation jars to be on the classpath/annotation class path.
jooby-apt has a provided dependency on `javax.ws.rs` which unfortunately has the negative side effect pulling that into your classpath.
Consequently you can very easily accidentally import `javax.ws.rs.GET` instead of jooby's.
The only place that JAX RS classes are actually accessed for annotation processing is here https://github.com/jooby-project/jooby/blob/2.x/modules/jooby-apt/src/main/java/io/jooby/apt/Annotations.java but that is just to get the classname.
I recommend just hardcoding the string and then making a unit test to check if the canonical names match the actual canonical javax.ws.rs (I mean its unlikely those names will ever change but I guess if you want to be safe).
Because of the above you also don't need jooby-apt to have a provided dependency on `jakarta.ws.rs-api` since those who use the jars will have it on their classpath anyway.<issue_closed>
Status: Issue closed |
NativeScript/NativeScript | 1149054743 | Title: Slow Build times on MacBook Pro M1 32GB RAM
Question:
username_0: ### Issue Description
After migrating my projects to a new MacBook Pro M1 my builds seem very slow.
NS Doctor:-
✔ Your ANDROID_HOME environment variable is set and points to correct directory.
✔ Your adb from the Android SDK is correctly installed.
✔ The Android SDK is installed.
✔ A compatible Android SDK for compilation is found.
✔ Javac is installed and is configured properly.
✔ The Java Development Kit (JDK) is installed and is configured properly.
✔ Xcode is installed and is configured properly.
✔ xcodeproj is installed and is configured properly.
✔ CocoaPods are installed.
✔ CocoaPods update is not required.
✔ CocoaPods are configured properly.
✔ Your current CocoaPods version is newer than 1.0.0.
✔ Python installed and configured correctly.
✔ The Python 'six' package is found.
✔ Xcode version 13.2.1 satisfies minimum required version 10.
✔ Getting NativeScript components versions information...
✔ Component nativescript has 8.1.5 version and is up to date.
✔ Component @nativescript/core has 8.1.5 version and is up to date.
✔ Component @nativescript/ios has 8.1.0 version and is up to date.
✔ Component @nativescript/android has 8.1.1 version and is up to date.
My previous MacBook Pro was intel based with 16GB RAM and was much quicker building.
I've cleaned the project, updated plugins etc but build time are still slower.
### Reproduction
_No response_
### Relevant log output (if applicable)
_No response_
### Environment
OS: macOS 12.2.1
CPU: (8) x64 Apple M1 Pro
Shell: /bin/zsh
node: 14.11.0
npm: 7.11.2
nativescript: 8.1.5
# android
java: 1.8.0_172
ndk: Not Found
apis: 25, 27, 28
build_tools: 25.0.2, 27.0.3, 28.0.3
system_images:
- android-28 | Google APIs Intel x86 Atom
# ios
xcode: 13.2.1/13C100
cocoapods: 1.11.2
python: 2.7.18
[Truncated]
"reflect-metadata": "0.1.13",
"rxjs": "6.5.4",
"rxjs-compat": "6.5.4",
"zone.js": "0.10.2"
},
"devDependencies": {
"@nativescript/android": "~8.1.1",
"@nativescript/ios": "8.1.0",
"@nativescript/types": "~8.1.1",
"@nativescript/webpack": "~5.0.0-beta.14",
"typescript": "~4.0.0"
}
```
### Please accept these terms
- [X] I have searched the [existing issues](https://github.com/NativeScript/NativeScript/issues) as well as [StackOverflow](https://stackoverflow.com/questions/tagged/nativescript) and this has not been posted before
- [X] This is a bug report
- [X] I agree to follow this project's [Code of Conduct](https://github.com/NativeScript/NativeScript/blob/master/tools/notes/CONTRIBUTING.md#coc)
Answers:
username_1: Everything 1st party in NativeScript works without rosetta, so definitely try native node and running without rosetta to see if that makes a difference.
username_0: Updated Node and now have v17.6.0
Running ns doctor does not complete and stops at getting Nativescript components
No issues were detected.
✔ Xcode is installed and is configured properly.
✔ xcodeproj is installed and is configured properly.
✔ CocoaPods are installed.
✔ CocoaPods update is not required.
✔ CocoaPods are configured properly.
✔ Your current CocoaPods version is newer than 1.0.0.
✔ Python installed and configured correctly.
✔ The Python 'six' package is found.
✔ Xcode version 13.2.1 satisfies minimum required version 10.
✔ Getting NativeScript components versions information...
ns doctor stops at the line above.
when entering ns --version this error occurs
8.1.5
⠼ Checking for updates...Error: Command failed: npm view nativescript dist-tags --json
node:internal/modules/cjs/loader:936
throw err;
^
Error: Cannot find module '../lib/cli.js'
Require stack:
- /usr/local/lib/node_modules/npm/bin/npm-cli.js
at Function.Module._resolveFilename (node:internal/modules/cjs/loader:933:15)
at Function.Module._load (node:internal/modules/cjs/loader:778:27)
at Module.require (node:internal/modules/cjs/loader:999:19)
at require (node:internal/modules/cjs/helpers:102:18)
at Object.<anonymous> (/usr/local/lib/node_modules/npm/bin/npm-cli.js:2:1)
at Module._compile (node:internal/modules/cjs/loader:1097:14)
at Object.Module._extensions..js (node:internal/modules/cjs/loader:1151:10)
at Module.load (node:internal/modules/cjs/loader:975:32)
at Function.Module._load (node:internal/modules/cjs/loader:822:12)
at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:77:12) {
code: 'MODULE_NOT_FOUND',
requireStack: [ '/usr/local/lib/node_modules/npm/bin/npm-cli.js' ]
}
Node.js v17.6.0
at Errors.failWithOptions (/usr/local/lib/node_modules/nativescript/lib/common/errors.js:157:27)
at Errors.fail (/usr/local/lib/node_modules/nativescript/lib/common/errors.js:130:21)
at NodePackageManager.<anonymous> (/usr/local/lib/node_modules/nativescript/lib/node-package-manager.js:104:30)
at Generator.throw (<anonymous>)
at rejected (/usr/local/lib/node_modules/nativescript/lib/node-package-manager.js:12:65)
⠼ Checking for updates...%
username_2: @username_0 seems like you have an incomplete installation, as that error is coming straight from npm itself. Most likely you updated node and is still using different versions of npm or a dirty npm package folder (from the old installation). I suggest switching to nvm instead of using a global node installation to avoid these issues.
username_0: Getting the same error whether I use Terminal with or without Rosetta.
Not sure how this manifested itself as I can no longer compile apps. Compilation stops at the last line of the error above: -
If increasing the memory doesn't solve the issue, it's most probably a bug in the TypeScript or EsLint. |
freeCodeCamp/testable-projects-fcc | 491063523 | Title: Pomodoro Clock User Stories update
Question:
username_0: But actually, to pass test case 8, I need to update the time-left content too with the new value. So please, modify to help fellow coders to get a clear idea about they have to do.
Thanks for your time & effort.
#### Browser Information
<!-- Describe your workspace in which you are having issues-->
* Browser Name, Version:
* Operating System:
* Mobile, Desktop, or Tablet:
#### Your Code / Link to Your Pen
<!-- Paste relevant code in here, or a link to your pen which would be most helpful -->
```
```
#### Screenshot
<!-- Add a screenshot of your issue -->
Status: Issue closed
Answers:
username_1: For more visibility, please open issues on the [main repo](https://github.com/freeCodeCamp/freeCodeCamp/issues) for these. Thanks and happy coding 🎉 |
Vector35/binaryninja-api | 405771851 | Title: “Error initializing database: Data too large” when saving Analysis Database for a large binary
Question:
username_0: Binary Ninja Version: 1.1.1470
Platform: macOS 10.14.3, 32G RAM
When loading a large binary (about 81MB) with Binary Ninja, it takes about 18GB of RAM. Furthermore, saving Analysis Database for this binary ends on the following error:

Any idea how to successfully save the database?
I can provide/upload the binary if you need.
Answers:
username_1: Unfortunately this is a known issue that I guess we don't already have a bug for. Large files fail to save due to some considerations SQLite's maximum size limitations. There is a SQLite feature that exists we're just currently not using it.
username_0: Hi. Thanks for your response. Is there any ETA for the fix?
Status: Issue closed
username_2: Fixed as of 1.1.1745-dev. |
gchq/gaffer-experimental | 1042472478 | Title: Run ui tests against npm 16
Question:
username_0: Currently the CI runs UI tests against node 14 only:
https://github.com/gchq/gaffer-experimental/blob/16bcbe46e0700417ce16840455892483e4db4eb5/.github/workflows/ci.yaml#L32
Would be nice if it used a matrix to test multiple versions of node, including, like the [ci in gaffer-tools](https://github.com/gchq/gaffer-tools/blob/60ebc8bb6c61e114a24c381ee998e42bd0d15553/.github/workflows/continuous-integration.yaml#L62-L69).
As well as this, there are lots of warnings when running `npm install` and it would be nice if these were cleaned up.<issue_closed>
Status: Issue closed |
open-telemetry/opentelemetry-python | 647697973 | Title: Proposal: Use GitHub Actions for CI/CD
Question:
username_0: # Consistent CI/CD Repo Issue Template
### Title
Proposal: Use GitHub Actions for CI/CD
### Description
This issue is in reference to [issue 398](https://github.com/open-telemetry/community/issues/398) posted on the community repository. In this issue, we are proposing that all OpenTelemetry repositories *consider* using GitHub Actions as their CI provider in order to maintain consistency across the various language repositories.
The overall proposal was discussed in the OpenTelemetry maintainers SIG meeting. @trask has been assigned as the mentor for the project.
|Repository |CI Provider |Automated Build and Test |Code Coverage |Automated Performance Testing |Automated Deployment |Automated Docs Deployment |
|--- |--- |--- |--- |--- |--- |--- |
|Python |Travis/CircleCI |[x] |[x] |[] |[x] |[x] |
The justification and benefits are enumerated in the issue on the community repository and are pasted here as well for convenience:
## Proposal
We propose that all languages consider using the same CI provider. This would create a more consistent development process and make it easier for developers to contribute to multiple language libraries.
We suggest that provider be GitHub Actions. Here’s why:
### **Ease-of-Use**
CircleCI will automatically run when pull requests and commits are issued against the repository. But if a contributor forks the repository, unless they set up an account with the CI provider and link it to their forked repository, CI will not be activated and tests will not be run automatically.
In contrast, GitHub Actions works out of the box on a forked repository and can be easily configured to run a test workflow each time a commit is issued. This would help individual contributors test their code and ensure code quality before submitting a pull request against the repository.
### **Transparency**
Current CI providers such as CircleCI and Travis allow anyone to view the console output when building and running tests but the test results can not be seen anywhere on the GitHub repository. To view this testing output: You need go to a different website, navigate a different user interface, and then sift through thousands of lines of console output. This is not a seamless developer experience.
In contrast, using GitHub Actions would provide all testing output directly on the repository’s GitHub page, which would help contributors to find, read, and use the test output to maintain code quality.
### **Control**
GitHub Actions’ integration with other GitHub features means you can have finer control over the CI pipeline. For example, certain workflows can be set to only run on a new release. Workflows can even be used to close stale issues and pull requests.
### Recommendation
We recommend that we consider using one consistent CI provider, GitHub Actions, which provides an integrated and seamless developer experience for all contributors.
### Example
Please see [this example](https://github.com/open-telemetry/opentelemetry-cpp/blob/master/.github/workflows/ci.yml) that the C++ repository has adopted for the above reasons.
### Next Steps
This issue shall serve as a place for discussion about this proposal.
Could a maintainer please assign this issue to us if approved?
Answers:
username_1: @username_0 I took a brief look at github actions and it seemed not to support older versions of Python out of the box. Have you had a chance to look at whether we can support the full matrix of versions we're testing today?
One of the other benefits of using CircleCI was the ability to cache tox environments, is this something that's achievable with github actions? I'd love to see an example of this. See PR https://github.com/open-telemetry/opentelemetry-python/pull/828 for more details on the circle CI configuration.
username_2: The [reference](https://docs.github.com/en/actions/reference/software-installed-on-github-hosted-runners) only lists the following as preinstalled on Ubuntu 20.04 LTS:
```
Python:
Python 2.7.18
Python 3.5.9
Python 3.6.11
Python 3.7.8
Python 3.8.5
PyPy:
PyPy 2.7.13 [PyPy 7.3.1 with GCC 7.3.1 20180303 (Red Hat 7.3.1-5)]
PyPy 3.6.9 [PyPy 7.3.1 with GCC 7.3.1 20180303 (Red Hat 7.3.1-5)]
```
But [actions/python-versions](https://github.com/actions/python-versions) seems to support the following:
```console
$ curl -s https://raw.githubusercontent.com/actions/python-versions/main/versions-manifest.json | jq ".[].version"
"3.9.0-beta.5"
"3.9.0-beta.4"
"3.8.5"
"3.8.4"
"3.8.3"
"3.8.2"
"3.8.1"
"3.8.0"
"3.7.8"
"3.7.7"
"3.7.6"
"3.7.5"
"3.6.11"
"3.6.10"
"3.6.9"
"3.6.8"
"3.6.7"
"3.5.9"
"3.5.4"
"3.4.10"
"3.3.7"
"2.7.18"
"2.7.17"
```
Not sure about caching.
username_2: @username_1 I have not gotten it to work as expected yet (see PoC #984), but caching should work, according to the [migration guide](https://docs.github.com/en/actions/migrating-to-github-actions/migrating-from-circleci-to-github-actions#caching).
username_2: As an additional argument for switching to GitHub Actions:
After setting up my CircleCI on my fork I noticed that the checks on my pull requests did not complete. Apparently, if a user follows their own fork, CircleCI does not run these jobs on the main repository, even for pull requests. This results in the checks on the pull requests forever _waiting for status to be reported_.
See: [Why aren't pull requests triggering jobs on my organization?](https://support.circleci.com/hc/en-us/articles/360008097173-Why-aren-t-pull-requests-triggering-jobs-on-my-organization-)
Status: Issue closed
|
meowtec/Imagine | 256444169 | Title: Compilation error
Question:
username_0: on `npm run dev` run get
```
modules/renderer/store/reducer.ts(58,49): error TS2345: Argument of type '{ [x: string]: ((state: ITaskItem[], action: Action<ITaskAddPayloadItem[]>) => ITaskItem[]) | ((s...' is not assignable to parameter of type 'ReducerMap<ITaskItem[], ITaskItem[]>'.
Index signatures are incompatible.
Type '((state: ITaskItem[], action: Action<ITaskAddPayloadItem[]>) => ITaskItem[]) | ((state: ITaskItem...' is not assignable to type 'Reducer<ITaskItem[], ITaskItem[]> | ReducerNextThrow<ITaskItem[], ITaskItem[]>'.
Type '(state: ITaskItem[], action: Action<string[]>) => ITaskItem[]' is not assignable to type 'Reducer<ITaskItem[], ITaskItem[]> | ReducerNextThrow<ITaskItem[], ITaskItem[]>'.
Type '(state: ITaskItem[], action: Action<string[]>) => ITaskItem[]' has no properties in common with type 'ReducerNextThrow<ITaskItem[], ITaskItem[]>'.
modules/renderer/store/reducer.ts(122,55): error TS2345: Argument of type '{ [x: string]: ((state: IGlobals, action: Action<string>) => { activeId: string | undefined; upda...' is not assignable to parameter of type 'ReducerMap<IGlobals, IGlobals>'.
Index signatures are incompatible.
Type '((state: IGlobals, action: Action<string>) => { activeId: string | undefined; updateInfo?: IUpdat...' is not assignable to type 'Reducer<IGlobals, IGlobals> | ReducerNextThrow<IGlobals, IGlobals>'.
Type '(state: IGlobals, action: Action<string>) => { activeId: string | undefined; updateInfo?: IUpdate...' is not assignable to type 'Reducer<IGlobals, IGlobals> | ReducerNextThrow<IGlobals, IGlobals>'.
Type '(state: IGlobals, action: Action<string>) => { activeId: string | undefined; updateInfo?: IUpdate...' has no properties in common with type 'ReducerNextThrow<IGlobals, IGlobals>'.
```
Answers:
username_1: TypeScript and TSLint should use fixed versions.
3ef0dc9
Now pull and try again.
username_0: it fix, thanks
Status: Issue closed
|
wojtekmaj/react-calendar | 580566520 | Title: programmatic reset (question)
Question:
username_0: Hi
Using v3.0.0
Is there a way to reset the value or range programmatically? Meaning, removing the selection/s, associated tile classes, etc
Thank you
Answers:
username_1: v3.0.0 can be both controlled and uncontrolled. If you need to control the value from the outside, controlled would be the best approach. Then it's as easy as clearing value prop.
username_0: Hi,
Is there documentation and / or examples of what "controlled would be the best approach" means?
Thank you
username_0: To be more specific, I used `value={rangeValue}` and change it onClickDay and onChange, that works, but if I programmatically set `rangeValue=null` is doesn't remove the tile classes, so it all looks as selected.
username_0: Sorry, my fault, using the following just works:
`const [rangeValue, setRangeValue] = useState(null);`
Thank you again
Status: Issue closed
|
flutter/flutter | 343944029 | Title: How do I prohibit font size from changing with the font size of the mobile phone?
Question:
username_0: Can I set a maximum and minimum font interval to prevent page elements from being confused?



Answers:
username_1: Please consider asking support questions in one of the other channels listed at http://flutter.io/support .
username_1: You could wrap your app with a custom `MediaQuery` where you override the default https://docs.flutter.io/flutter/widgets/MediaQueryData/textScaleFactor.html
See also
- https://docs.flutter.io/flutter/dart-ui/Window/textScaleFactor.html
- https://docs.flutter.io/flutter/dart-ui/Window/onTextScaleFactorChanged.html
Please consider asking support questions in one of the other channels listed at http://flutter.io/support .
Status: Issue closed
username_2: MaterialApp(
builder: (BuildContext context, Widget child) {
return MediaQuery(
data: MediaQuery.of(context).copyWith(textScaleFactor: textScaleFactor),
child: child,
);
},
title: 'Home Page',
);
This should do the trick~
username_3: Isn't making `textScaleFactor: 1.0` much easier than using `MediaQuery` stuff?
For example, the following will prohibit scaling based on system's accessibility settings:
```dart
Text(
'hello',
textScaleFactor: 1.0,
),
```
username_4: No, I think you should use `MediaQuery` Because of your method doesn't change textField label sizes and etc |
PIVX-Project/PIVX | 604230852 | Title: [GUI][Visual bug] Pull down menu / transactions disapears.
Question:
username_0: <!--- Remove this description and sections that do not apply -->
This issue tracker is only for technical issues related to PIVX Core.
General PIVX questions and/or support requests and are best directed to the [PIVX Discord](https://discord.pivx.org).
### Describe the issue
Please refer to the first screenshot below and note all 9 pull-down Transactions Types. Randomly select one Transactions Type, followed by another, and a third/forth/fifth. At some point the pull-down menu disappears (refer to the second screenshot below). The only way to recover (to my knowledge) is to close/open the wallet.
### Can you reliably reproduce the issue?
#### If so, please list the steps to reproduce below:
I have reproduced this issue twice today. Randomly select a Transaction Type in the pull-down menu and ensure the proper information is displayed. Repeat several times (4-6) with different Transaction Types. At some point the pull-down menu will disappear.
### Screenshots.
If the issue is related to the GUI, screenshots can be added to this issue via drag & drop.


### What version of PIVX Core are you using?
List the version number/commit ID, and if it is an official binary, self compiled or a distribution package. Self compiled 4.1.0
### Machine specs:
- OS: Windows 10
- CPU: i9
- RAM: 8GB
- Disk size: 512 GB
- Disk Type (HD/SDD): SDD
Answers:
username_1: possibly related to #1172
username_2: This was solved in #2073, closing.
Status: Issue closed
|
open-austin/iced-coffee | 138799743 | Title: [Code Across] Follow up emails with more info
Question:
username_0: - Google Slides
- Link to Slack
- Link to Github issues with Code Across tag
- Upcoming events
Answers:
username_1: Can I get a link to the slides? I want to post them to Meetup.
username_0: https://drive.google.com/open?id=1aWNB5wWr9ZIjwvMNEZR2uoGJcK9zDKuFw0_B1JIhFBk
username_2: Here is the view only link: https://docs.google.com/presentation/d/1aWNB5wWr9ZIjwvMNEZR2uoGJcK9zDKuFw0_B1JIhFBk/edit?usp=sharing
Status: Issue closed
username_0: <img width="475" alt="screen shot 2016-03-07 at 9 18 34 am" src="https://cloud.githubusercontent.com/assets/5697474/13573515/a83e3802-e445-11e5-988e-fc0501d72e83.png"> |
Redbility/paper-lightbox | 240404668 | Title: fix demo when doing polymer serve
Question:
username_0: minor thing but when doing `polymer serve` on this element to do some investigation it comes up 'not found'. I know that you can add demo to the path to get it to show up but still.
 |
kukko/web-wombat | 415075674 | Title: Is the repo open for contribution?
Question:
username_0: Hey!
Thanks for the presentation yesterday, well done!
Are you accepting PRs?
What is the style you are using for commits? I'm not yet familiar with the tags used: fix/enhancement/progress.
Cheers,
Balázs
Answers:
username_1: Hello!
Thanks for your attendance yesterday!
Yeah, I tottally accept PRs. In th beginning, I would merge pull requests, which is not conflicting with my future plans, about this repo. So If you want to make a feature, please contact with me, to prevent needless coding, which will not be merged into the main repo.
Keep coding!
Your sincerely, username_1.
Status: Issue closed
username_0: Thanks for the answer! I opened a PR for some minor typo fixes, I was guessing that can go without a separate issue beforhand. Let's see how that goes. :) Closing this issue. |
microsoft/CCF | 731242732 | Title: Upgrade Evercrypt to 0.3.0
Question:
username_0: Evercrypt has its first official release, we should upgrade to that: https://github.com/project-everest/hacl-star/releases/tag/v0.3.0
Thanks @wintersteiger for the heads up!
Answers:
username_0: #1903 revealed that the sha256 implementation we use from Evercrypt now is slower than mbedtls. @wintersteiger helped me take a look at why, and that's because we no longer get the fast assembly version, which is only enabled for chips that have SHA extensions (something no modern Intel CPUs feature, and no SGX-enabled CPUs ever supported).
One of the things we want to watch out for in the upgrade is that we do get a fast SHA implementation.
Status: Issue closed
|
tekHudson/DruidBarClassic | 502940918 | Title: "Bliz-Like" Text Display Inconsistencies and Suggested Improvements
Question:
username_0: DruidBarTextLeft displays with a slightly incorrect x axis offset. Changing Line 50 of `DruidBar.xml` from `<AbsDimension x="5" y="0"/>` to `<AbsDimension x="7" y="0"/>` fixes this.
`DruidBarTextRight` is set to the value of `ManaValues()` on Line 342 of `DruidBar.lua`, which returns both the current mana and maximum mana separated by a slash. On the Blizzard UI only the current mana is shown.
Changing Line 342 from `DruidBarTextRight:SetText(ManaValues());` to `DruidBarTextRight:SetText(floor(DruidBarKey.currentmana));` fixes this, suggest making a new way to handle the requests for mana values by the UI so that current mana can be retrieved separately, or separated from maximum mana.
Before Image: https://i.imgur.com/l3vynL8.png
After Image: https://i.imgur.com/QYafDzN.png
Answers:
username_1: Meged with #17
Status: Issue closed
|
uber/baseweb | 524971737 | Title: [Notification] Force line break in message
Question:
username_0: <!--- Provide a general summary of the issue in the Title above -->
<!--- Provide a codesandbox that reproduces your issue - you can fork one from https://baseweb.design/ -->
Example at https://codesandbox.io/s/notification-qiwp0
## Current Behavior
Inserting `\\n`in the string doesn't line break the message.
<!---
Describe what happens instead of the expected behavior.
-->
## Expected Behavior
Should give a line break at the position where I input `\\n`.
<!---
Describe what should happen.
-->
## Your Environment
<!--- Include as many relevant details about the environment with which you experienced the bug. -->
| Tech | Version |
| ------- | ------- |
| Base UI | v9.17.0 |
| React | v16.11.0 |
| browser | Chrome |
- [x] I have searched the [issues](https://github.com/uber/baseweb/issues) of this repository and believe that this is not a duplicate.
Answers:
username_1: You can do this https://codesandbox.io/s/notification-dls2y?fontsize=14&hidenavigation=1&theme=dark
Use it with caution though.
If the message is determined using user input, there's a possibility the user can inject arbitrary scripts that could compromise security.
username_0: Realised that you could do it like in this codesandbox: https://codesandbox.io/s/notification-jde02
No need for the `dangerouslySetInnerHTML`.
Status: Issue closed
username_2: I believe a more correct and cleaner way would be to use an override and use `white-space: pre-line` instead of `white-space: pre`.
Something like:
```
overrides: {
Body: {
style: {
whiteSpace: 'pre-line',
},
},
},
```
This will also work in a `toaster` where you cannot use a function with a child `div`. |
hbz/link-templates | 220933464 | Title: TH Köln OPAC link is broken
Question:
username_0: For example https://katalog.bibl.fh-koeln.de/webOPACClient/start.do?Query=0010=%22HT001045964%22
Answers:
username_0: Also Duisburg-Essen, see http://primo.ub.uni-due.de/primo_library/libweb/action/dlSearch.do?vid=UDE&institution=UDE&search_scope=localude&bulkSize=10&lang=ger&indx=1&onCampus=false&query=any,contains,HT002391361 . Other example: http://lobid.org/resources/HT002526560#
username_0: We should also test the other organizations.
username_0: TH Köln oviously switched to DigiBib, see https://www.th-koeln.de/hochschulbibliothek/katalog_24926.php. :-)
Status: Issue closed
|
jlippold/tweakCompatible | 556596714 | Title: `libcolorpicker` working on iOS 13.3.1
Question:
username_0: ```
{
"packageId": "org.thebigboss.libcolorpicker",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "org.thebigboss.libcolorpicker",
"deviceId": "iPhone10,5",
"url": "http://cydia.saurik.com/package/org.thebigboss.libcolorpicker/",
"iOSVersion": "13.3.1",
"packageVersionIndexed": true,
"packageName": "libcolorpicker",
"category": "Development",
"repository": "BigBoss",
"name": "libcolorpicker",
"installed": "1.6.7",
"packageIndexed": true,
"packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.",
"id": "org.thebigboss.libcolorpicker",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.1.5",
"shortDescription": "color picker library for developers",
"latest": "1.6.7",
"author": "PixelFire",
"packageStatus": "Unknown"
},
"base64": "<KEY>
"chosenStatus": "working",
"notes": ""
}
```<issue_closed>
Status: Issue closed |
neuronsimulator/nrn | 934890000 | Title: Hoc Object deletion not occurring with combination of deleted Section, expect_err, and PointProcess.
Question:
username_0: ## Context
The following code does not delete IClamp[0] on exit from foo.
```
from neuron import h
from neuron.expect_hocerr import expect_err, set_quiet
set_quiet(False)
s = h.Section()
def foo():
ic = h.IClamp(s(.5))
h.delete_section(sec=s)
expect_err('ic.amp = .1')
#del ic # problem goes away if this and following statements are uncommented
#print (locals())
foo()
h.allobjects()
```
prints:
```
CHECKING: ic.amp = .1
NEURON: point process not located in a section
near line 0
objref hoc_obj_[2]
^
No value
IClamp[0] with 1 refs
ParallelContext[0] with 1 refs
```
Note: ignore the fact that ParallelContext[0] exists. That is created by the expect_hocerr module. If the last two lines of def foo(): are uncommented then we get the correct output:
```
CHECKING: ic.amp = .1
NEURON: point process not located in a section
near line 0
objref hoc_obj_[2]
^
No value
{}
ParallelContext[0] with 1 refs
```
### Overview of the issue
I find expect_err to be very useful for testing error messages. However if, because of its use, objects are not properly deleted then that can cause problems during later tests with ```python -m pytest ...```
### Expected result/behavior
IClamp[0] should not exist whether or not the last two lines of foo are commented.
Answers:
username_0: When experiencing this problem, a general work around is to ```del``` all the offending local variables at the end of the function and then call ```locals()```. It remains a puzzle why ```del``` is needed but an even greater puzzle why the mere call to ```locals()``` is also needed. |
MicrosoftDocs/azure-docs | 610897969 | Title: Under Linux section - Configure agent communication typo
Question:
username_0: Command line should only be in the following format
azcmagent connect --resource-group "<resourceGroupName>" --tenant-id "<tenantID>" --location "<regionName>" --subscription-id "<subscriptionID>"
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6036c6de-a2e8-ffd7-50e9-44ae32bf1682
* Version Independent ID: f7ab9abc-abd0-09e1-4b64-9ec304495ee5
* Content: [Connect hybrid machines to Azure from the Azure portal](https://docs.microsoft.com/en-us/azure/azure-arc/servers/onboard-portal#feedback)
* Content Source: [articles/azure-arc/servers/onboard-portal.md](https://github.com/Microsoft/azure-docs/blob/master/articles/azure-arc/servers/onboard-portal.md)
* Service: **azure-arc**
* Sub-service: **azure-arc-servers**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
Answers:
username_1: @username_0
Thanks for the feedback! We are currently investigating and will update you shortly.
username_2: @username_0 Updated the referred document with the required changes. The changes will go live within 24 hours. Thanks #please-close
Status: Issue closed
|
lasso-js/lasso | 238003224 | Title: Mention Sass in the description
Question:
username_0: When I see this, I notice you don't mention Sass, which over 70% of web devs use (Less is down to sub-5%, and Stylus is still kicking it at around the 1% mark, and Marko is another internal eBay project that isn't well known).
I immediately assumed you didn't support Sass and realized Lasso is just some internal-use eBay thing, and they must be stuck using outdated Less, or have some weird affiliation with the creator of Stylus or something. Clearly not a product for general use.
I was just about to close the tab when I decided to Ctrl+F for `sass` and saw a plugin for it.
Anyways, the way you associate yourselves to other technologies effects the perception people have of your project. Namedropping obscure projects isn't a good way to obtain adopters. I would either remove the example all together, or replace them with more commonly used tools (like Sass in this instance). But in a broader light, just be more aware of the associations people will make and how it will cause them to make assumptions about your project. |
PHP-DI/PHP-DI | 355916971 | Title: Support for variadic arguments
Question:
username_0: PHP-DI does not currently support variadic arguments for autowiring. It will only inject the first element to the variadic parameter in a constructor, silently discarding the rest.
For example for this class:
```PHP
class UserProfile implements Timeline\Engine
{
public function __construct(DbConnection $dbConnection, TimelineModule ...$modules)
{
```
cannot be autowired by:
```PHP
Engine::class => \DI\autowire(UserProfile::class)
->constructor(
\DI\get(DbConnection::class),
\DI\get(Module\CasinoFollow::class),
\DI\get(Module\UserFollow::class),
\DI\get(Module\Review::class)
),
```
but must be instead autowired by:
```PHP
UserProfile::class => function(ContainerInterface $c) {
return new UserProfile(
$c->get(DbConnection::class),
$c->get(Module\CasinoFollow::class),
$c->get(Module\UserFollow::class),
$c->get(Module\Review::class)
);
},
```
which in turn reduces the efficiency of caching I believe and is harder to debug.
Answers:
username_1: Thanks for the detailed issue!
Can you confirm that:
```php
Engine::class => \DI\autowire(UserProfile::class)
->constructor(
\DI\get(DbConnection::class),
\DI\get(Module\CasinoFollow::class),
\DI\get(Module\UserFollow::class),
\DI\get(Module\Review::class)
),
```
doesn't work and that it injects only `$dbConnection` and the first module?
username_0: Indeed, it only injects the first module: https://i.imgur.com/BANmom5.png
username_0: I would like to start working on this, but I don't feel so accustomed to PHP-DI internals yet. Do you have any tips @username_1 for me? Where should I start?
I have written some preliminary tests for this behavior.
username_1: @username_0 that's great!
I suggest you have a look here: https://github.com/PHP-DI/PHP-DI/blob/master/src/Definition/Resolver/ParameterResolver.php#L51
You can see that we do `foreach` on parameters returned by PHP's reflection. So a variadic parameter will be interpreted as 1 parameter only.
Try to write unit tests but more importantly functional tests in here: https://github.com/PHP-DI/PHP-DI/blob/master/tests/IntegrationTest/Definitions/CreateDefinitionTest.php
username_1: @username_0 yes that is awesome thanks! I had a quick look last week but because of conferences I didn't have enough time to review it carefully (this PR is more complex that other simple PRs :) )
I'll try to have a look at it this week.
username_2: Need this feature too.
username_2: Is it possible to make it work the same way with constructorParameter() method too?
To be able to extend definition for one param only?
Example:
```
class MyClass {
public function __construct(Foo $foo, Bar $bar, Baz $baz, ...$args) {}
}
MyClass::class => DI\autowire()->constructorParameter('args', ...['foo', 'bar', 'baz', 'etc']),
```
username_2: Not quite what I would like to do (-:
Now I can avoid defining other constructor arguments and have them autowired only this way:
```
class MyClass {
public function __construct(Foo $foo, Bar $bar, Baz $baz, $args) {}
}
[
MyClass::class => DI\autowire()
->constructorParameter('args', ['a', 'b', 'c', 'etc'])
]
```
But I would like to type-hint $args:
`public function __construct(Foo $foo, Bar $bar, Baz $baz, string ...$args) {}`
And define it like this:
```
[
MyClass::class => DI\autowire()
->constructorParameter('args', ...['a', 'b', 'c', 'etc'])
]
```
username_0: You could do it this way (with the patch merged):
```PHP
class MyClass {
public function __construct(Foo $foo, Bar $bar, Baz $baz, string ...$args) {}
}
MyClass::class => DI\autowire()
->constructorParameter(3, 'a')
->constructorParameter(4, 'b')
->constructorParameter(5, 'b')
```
Parameters 0 to 2 (types of Foo, Bar and Baz) would be autowired automatically without defining them as you'd expect with PHP-DI.
I do not personally prefer this
```PHP
MyClass::class => DI\autowire()
->constructorParameter('args', ...['a', 'b', 'c', 'etc'])
```
to work as you described, because that would show the user in the constructorParameter method signature that multiple parameters are valid to any constructor parameter, when the situation with PHP is that only the last parameter can be variadic.
For example: user might think based on method signature and documentation that this would be valid:
```PHP
MyClass::class => DI\autowire()
->constructorParameter('bar', ...['a', 'b', 'c', 'etc'])
->constructorParameter('args', ...['a', 'b', 'c', 'etc'])
```
When obviously the part where bar is defined would not work. I understand that you'd perhaps want to use the parameter name instead of the argument number for clarity, but I don't think it's a good idea.
username_1: I had a quick look at this and I think it might be worth implementing the same rules as PHP 8's named parameters: https://wiki.php.net/rfc/named_params (over all the PHP-DI features)
The integration with variadics is clearly defined (and constrained) in there, so it could be both a good lead to follow, as well as minimize confusion and friction down the road. |
github-vet/rangeloop-pointer-findings | 774906043 | Title: yzs981130/oyashirosama2: controllers/testjob_controller.go; 24 LoC
Question:
username_0: [Click here to see the code in its original context.](https://github.com/yzs981130/oyashirosama2/blob/c0c27bdd26dccfbd67eafa87bcfada4a6e0bc7f7/controllers/testjob_controller.go#L121-L144)
<details>
<summary>Click here to show the 24 line(s) of Go which triggered the analyzer.</summary>
```go
for i, job := range childJobs.Items {
_, finishedType := isJobFinished(&job)
switch finishedType {
case "": // ongoing
activeJobs = append(activeJobs, &childJobs.Items[i])
case schedulev1.JobFailed:
failedJobs = append(failedJobs, &childJobs.Items[i])
case schedulev1.JobComplete:
successfulJobs = append(successfulJobs, &childJobs.Items[i])
}
scheduledTimeForJob, err := getScheduledTimeForJob(&job)
if err != nil {
log.Error(err, "unable to parse schedule time for child job", "job", &job)
continue
}
if scheduledTimeForJob != nil {
if mostRecentTime == nil {
mostRecentTime = scheduledTimeForJob
} else if mostRecentTime.Before(*scheduledTimeForJob) {
mostRecentTime = scheduledTimeForJob
}
}
}
```
</details>
<details>
<summary>Click here to show extra information the analyzer produced.</summary>
```
The following dot graph describes paths through the callgraph that could lead to a function calling a goroutine:
no paths found; call may have ended in third-party code; stay tuned for diagnostics
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: c0c27bdd26dccfbd67eafa87bcfada4a6e0bc7f7 |
Mangodream01/ISMI_project | 771965468 | Title: Dataset
Question:
username_0: Hello Michel,
I have walked through your code on segmenting liver, liver cancer and spleen and have found it to well presented and informative.
However, in trying to replicate the work, I have realized that, the data set is not provided neither is any link provided to the it. Could you please assist me by providing either the link to the data or the data set itself?
Thanks and looking forward to your rapid response.
Answers:
username_1: Hi @username_0,
We used data from the Medical Segmentation Decathlon (http://medicaldecathlon.com/).
You can download the data there, we used the data from task 3 (Liver) and task 9 (Spleen).
Hope all is clear and goodluck!
username_0: Hello @username_1 ,
Thanks for your rapid reply and assistance. It is indeed helpful.
Cheers!
Status: Issue closed
|
vklochan/python-logstash | 64301310 | Title: SSL Support?
Question:
username_0: Hi
I want to know if python-logstash has SSL support, because i cannot find any way to define custom CA certificate.
Thanks
Answers:
username_1: Same issue here, TLS/SSL support would be nice.
username_0: ping
username_2: I had a quick look at the code to see if this was possible. It looks to me like the easiest way to achieve this is to implement `makeSocket()` in the `TCPLogstashHandler` class as has been done in @klynch's [logstash_handler](https://github.com/klynch/python-logstash-handler/blob/master/logstash_handler/__init__.py#L17) library.
@vklochan would you be likely to accept a patch that does something like that?
username_3: Ping for @vklochan
The possibility to encrypt logs is a super useful feature!
I think the one proposed by @username_2 is an easy-winning patch... If you want I can help out
username_4: +1 please. |
rancher-sandbox/rancher-desktop | 1082974421 | Title: [macOS] Stuck in Waiting for Kubernetes after update to 0.7
Question:
username_0: ### Rancher Desktop Version
0.7
### Rancher Desktop K8s Version
1.21.7
### What operating system are you using?
macOS
### Operating System / Build Version
macOS Monterey
### What CPU architecture are you using?
arm64 (Apple Silicon)
### Windows User Only
_No response_
### Actual Behavior
After manually updated my app from 0.7.0-beta to 0.7.0 (replace /Applications/Rancher\ Desktop.app), it stuck on Waiting for Kubernetes API... forever.
I check the k3s.log and got `invalid node ip, unable to get global unicast ip from interface name: can't find ip for interface rd0.`
My active network interface is `en0`, and checkout that interface in file `~/Library/Application\ Support/rancher-desktop/lima/_config/networks.yaml` is `en0`, which means no need to change.
I stop and restart Rancher Desktop but nothing fixed.
### Steps to Reproduce
Update the app version to 0.7.0.
### Result
Stuck on `Waiting for Kubernetes API...`.
### Expected Behavior
Start K3S successful.
### Additional Information
Here I paste the file content:
```yaml
paths:
vdeSwitch: /opt/rancher-desktop/bin/vde_switch
vdeVMNet: /opt/rancher-desktop/bin/vde_vmnet
varRun: /private/var/run/rancher-desktop-lima
sudoers: /private/etc/sudoers.d/rancher-desktop-lima
group: staff
networks:
shared:
mode: shared
gateway: 192.168.205.1
dhcpEnd: 192.168.205.254
netmask: 255.255.255.0
bridged:
mode: bridged
interface: en0
host:
mode: host
gateway: 192.168.206.1
dhcpEnd: 192.168.206.254
netmask: 255.255.255.0
```
And log files
[log.tar.gz](https://github.com/rancher-sandbox/rancher-desktop/files/7733385/log.tar.gz)
Status: Issue closed
Answers:
username_0: Well that actually works, after I switch an wifi ssid.
username_1: For those who may run into this problem, like I did. I am not using wifi. I am using an ethernet adapter, en6. I ended up changing `rd0` to `en6` in
~/Library/Application Support/rancher-desktop/lima/_config/0.yaml
~/Library/Application Support/rancher-desktop/lima/0/lima.yaml |
briansmith/ring | 794560623 | Title: PBKDF2 and other APIs should be redesigned to avoid `out: &mut [u8]` parameters.
Question:
username_0: Here's the current PBKDF2 API:
```
pub fn derive(
algorithm: Algorithm,
iterations: NonZeroU32,
salt: &[u8],
secret: &[u8],
out: &mut [u8],
) {
```
This was one of the first APIs written for *ring*, about 5 years ago, and it should be improved. The API uses `out: &,mut [u8]` because it wants to support `alloc`-less environments and it doesn't want to depend on crates like `SmallVec` or similar. We should use what we've learned about Rust and how Rust has improved to fix this to be something roughly like this:
```
pub fn derive(
algorithm: Algorithm,
iterations: NonZeroU32,
salt: &[u8],
secret: &[u8],
) -> T where T: Default + Extend<u8> {
```
See also issue #987 about allowing `secret` to be encapsulated. |
tc39/proposal-temporal | 978510576 | Title: Use cases and sample code for multi-Temporal-type string parser
Question:
username_0: To better think through #1751, I wanted to understanding the long tail of ISO-string-parsing use cases better. I also wanted to understand how well `from` methods support those parsing use cases. Here's my sense of those cases:
* **Impose stricter control on string inputs.** e.g. require minutes to be specified in ISO time strings, or require the time to be included to be specified in a date/time string
* **Clarify user intent.** e.g. differentiating Z strings vs. local+offset strings in `ZonedDateTime.from`
* **Handle partially-invalid/partially-valid ISO strings.** Temporal doesn't support this case today, because AFAIK invalid ISO strings throw even if some parts are valid.
* **Generically parse ISO strings for use in non-Temporal code.** In this case, the user wants a parser to chop up an ISO string into parts so they can do whatever they want with those parts (e.g. custom parsing, display to users, etc.) Temporal doesn't support this case today.
Are there others?
To validate Temporal's `from` methods for those first two use cases, as an exercise I wrote a simple parser with the goal of relying solely on `from` methods of various Temporal types to extract information about the original string.
### Issues Found
This effort helped me discover some gaps in Temporal's support for the first two parsing use-cases above, where "gaps" means that `from` was insufficient and I had to look inside the string was required to get info I was interested in. #1765 seems like a clear bug in the polyfill and/or spec. Maybe #1766 too. The other two seem like use cases we could support with new `from` options in a V2 if there's enough demand.
* PlainTime will parse anything, e.g. 01-01 or 2020-01-01 or 2020-01 - #1765
* Can't distinguish default ISO calendar from user appending [u-ca=iso8601] - #1766
* Can't distinguish Z from +00:00 when parsing strings with TZ annotations - #1767
* Can't determine if the original string had minutes or smaller time units - #1769
### Parser Prototype
I figured it'd be useful to file this issue to provide sample code for others who may want to do something similar in the future.
Some quirks:
* I worked around the top 3 issues above using simple string parsing. The workarounds I added aren't bulletproof, but if we choose not to fix those issues then it may be worth using better regexes for those cases.
* I was too lazy to work around #1769. This could be an exercise for the reader. 😄
* Time zone IDs, calendar IDs, and offsets are normalized to their canonical format. There's no way using the output of `from` to now what the original input was. I didn't think this was important enough to make into the list of issues above.
* Date-only strings can be parsed by PlainDateTime and ZonedDateTime, but the output of the parser doesn't include time units because someone using a multi-type parser probably wants to know that the input was only a date.
* The output includes properties like `monthCode` and `era` even though those are never in an ISO string. This seemed to be the right thing to enable the resulting object to be used as a property bag to initialize a Temporal type like PlainMonthDay.
* Speaking of PlainMonthDay, it has no `month` property so the result of parsing `01-01` has no `month` property either, only `monthCode`, `day`, `plainMonthDay`, and (if calendar annotation is present) a `calendar` property.
* Even though ISO is the default calendar when parsing strings, no `calendar` property is returned unless an annotation is present in the string. This lets callers to know if the string has a calendar annotation or not.
* `offset` is returned as a string, which may be "Z" or a numeric offset.
* As discussed in #1751, if an instant string (no TZ annotation) uses a "Z" offset, my opinion is that the components are meaningless, for the same reason that we don't offer `year`, `month`, `day` etc. properties on Instant. The code below therefore limits the output for Instant Z string to just `{instant, offset: "Z"}`.
**Parser Code**
```js
/**
* Parse an ISO string into Temporal instances and their components: date units,
* time units, offset, calendar, and time zone
*
* Issues found:
* * #1765 - PlainTime will parse anything, e.g. 01-01 or 2020-01-01 or 2020-01
* * #1766 - Can't distinguish default ISO calendar from user appending [u-ca=iso8601]
* * #1767 - Can't distinguish Z from +00:00 when parsing strings with TZ annotations
* * #1769 - No way to know if minutes and smaller units were present in a time string
*
* @param s
* @returns An object with:
* * a property for each Temporal type parsed successfully, e.g. `zonedDateTime`,
* `plainDate`, `plainTime`
* * the following properties from those successfully-parsed instances:
* 'year`, `eraYear`, `era`, `month`, `monthCode`, `week`, `day`, `hour`,
* `minute`, `second`, `millisecond`, `microsecond`, `nanosecond'
* * a `calendar` property if a bracketed calendar annotation was in the string
* * a `timeZone` property if a bracketed time zone annotation was in the string
* * an `offset` property ("Z" or numeric offset string) if an offset was in the string
*/
function parse(s) {
const types = [
[Truncated]
"plainMonthDay": "01-01",
"year": 2020,
"month": 1,
"monthCode": "M01",
"day": 1
},
"2020-01": {
"plainYearMonth": "2020-01",
"year": 2020,
"month": 1,
"monthCode": "M01"
},
"01-01": {
"plainMonthDay": "01-01",
"monthCode": "M01",
"day": 1
},
"bogus": {}
}
``` |
OpenFn/Miracle-Feet | 1010823797 | Title: Turn on/off SMS alerts based on `Account.Status__c` field
Question:
username_0: ## Background, context, and business value
The clinics (Salesforce `Account`-s) MiracleFeets works with can be temporarily or permanently suspended, which affects if patients (SF `Contact`-s) should receive SMSs, and which ones. Clininc status is set in SF. Depending on the clinic's status in SF, we want to deactivate some/all SMS-s for patients associated with that clinic.
## The specific request, in as few words as possible
We have 2 different flows based on clinic status which can be handled in the same job:
Flow #1 ([data flow diagram](https://lucid.app/lucidchart/9454d9ca-7c35-482d-b9e9-0e41284d1281/edit?page=3kldohKkc3rC#))
1. Query Salesforce `Account` records where "Status" = "Previously Supported "
2. Find all `Contact` records for each of the `Account` records
3. Disable all SMS for these Contacts.
Flow #2
1. Query Salesforce `Account` records where "Status" = "Temporarily Suspended"
2. Find all `Contact` records for each of the `Account` records
3. Disable Alert [#17, 18, 19](https://docs.google.com/spreadsheets/d/1quhQJgQkVRC8oObDzkwgnnm-Rov5BGOW85I4YqcNV0I/edit?pli=1#gid=262234774&range=112:114) for these Contacts.
## state.json
Either provide state directly, or link to a file. If sensitive information
should be in state, redact it and provide instructions for where it can be
found.
```json
{
"configuration": { "username": "abc", "password": "<PASSWORD>" },
"data": { "a": 1 },
"cursor": "2020-01-19 00:00:00"
}
```
```json
{
"configuration": ["SEE LAST PASS: 'client cred'"],
"data": { "a": 1 },
"cursor": "2020-01-19 00:00:00"
}
```
## adaptor
List the adaptor to be used for this job. If changes must be made to the
adaptor, explain why existing functions dont work and specify the new API you'd
like from a helper function.
```md
There is no "upsert" in postgres. I'd like an API where I can provide the table,
the UUID, and some data to upsert. Like this:
upsert('some_table', 'some_column', state.data.records);
```
## expression.js
In pseudocode, either in the current job expression or in a new file, describe
as best you can what changes need to be made
```js
each(
[Truncated]
```
## output.json
Either provide the output you'd like, or describe it in terms of final state and
side effects.
### side effects
1. upsert new records to postgres
### output.json
```json
{
"configuration": {},
"data": { "statusCode": 200 },
"references": { "a": 1 }
}
```
Answers:
username_1: @username_0 For step 3 (Disable SMSs)... how will this be done in Infobip?
A SF query like this will return a list of Contacts... but is there more info we need from SF in order to find the corresponding contacts in Infobip? Do Infobip Contacts maybe have the CommCare `case_id` or some other externalId we need in order to build the `bulkId` to match on?
```
Select Id, Account from Contact WHERE Account.Status__c = "Previously Supported"
```
username_0: @username_1 We would need to cycle through all alert `[bulkId](https://docs.google.com/spreadsheets/d/1quhQJgQkVRC8oObDzkwgnnm-Rov5BGOW85I4YqcNV0I/edit?pli=1#gid=262234774&range=M:M)`-s constructed from the `case_id`. Same method as we use when a patient opts out of all SMS-s.
username_1: @username_0 2 test clinics and 3 test patients have been setup in sandbox.
@username_2 moving to the bottom of the backlog. When you get to this, keep us posted on questions... and Rita will also be around tomorrow afternoon for a quick chat if needed.
username_2: @username_0 I have tried to fetch the bulkIds above you scheduled but they are not found on postman
username_0: @username_2 I scheduled the following bulkIds:
`casting_intro-3-1234567` -> `case_id: 1234567`
`bracing_day-6-345678` -> `345678`
`bracing_night-6-7777777` -> `7777777`
`visitBefore-234567-2021-12-24`
username_2: I pushed draft for the 2 jobs in the meantime.
The blocker for now is to get the visit date for some bulkIds.
username_0: @username_2
I added a visit in Infobip and SF:
`visitBefore-345678-2021-12-24` for caseId `345678`
to access it:
```
Select Id, Next_Visit_Date__c from Visit_new__c where Patient__r.CommCare_Case_ID__c = ${case_id} AND Next_Visit_Date__c !=null
```
This should return a list of `Next_Visit_Date__c` values for a selected patient
username_2: @username_0 I made the update.
username_0: @username_2 getting a syntax(?) error with disableSMS.js, can you check?
```Line 1: fn(state => {
^^
Function not available.
```
see run https://www.openfn.org/projects/pdbznd/runs/0615c2fa-66f8-7cbc-adc8-33c501cce5ca
username_0: @username_2
Now we're using `language-http v3.1.11`. I tested with Infobip credentials and set up a new one too to test, but same error
https://www.openfn.org/projects/pdbznd/runs/0615c474-e7d8-72ca-aec3-8ad59448ceef
username_2: @username_0 feel free to re-run after update on the credentials. I don't have access to it
username_2: @username_0 @username_1 find the run here. I add more log to separate for more precision: https://www.openfn.org/projects/pdbznd/runs/0615d873-d4c3-77ea-930a-ade52bfd0428
username_1: @username_0 please close whenever you've tested this |
llimllib/limbo | 151907713 | Title: New user can't execute bot command
Question:
username_0: The bot is running and then a new user is being added to the team. When the new user is trying to execute the bot command he/she will trigger the KeyError because `server.slack.server.users` hasn't been updated when the user got added. -> https://github.com/username_1/limbo/blob/master/limbo/limbo.py#L131
Answers:
username_1: ooh good catch, I hadn't considered that case!
Thanks so much for reporting, fixing it will be a top priority when I get time to work on this.
username_2: I'd encountered this before but wasn't sire what caused it. I knew a
restart fixed it.
A similar issue happens when you add the bot to a new private group on the
first time its triggered.
Mind if I have a go at fixing this?
username_1: Not at all, have at it
username_2: Looking at the offending code, it appears to be because slackrtm pulls the list of users originally and as such the server object has a static list of users in `server.slack.server.users`
When a new user enters, slackrtm doesn't update the list of users, which is giving the key error.
The immediately obvious way to fix it would be to make slackrtm update the users first when this KeyError occurs, try again and then if it happens again perform the debugging.
Having said that, what purpose does this line serve? (from `limbo/limbo.py ln 131`)
`msguser = server.slack.server.users[event["user"]]`
Can this be replaced with just `msguser = event["user"]`? As far as I can see this would server the same purpose and solve the issue, or is msguser used elsewhere that I can't find?
username_1: I think this is the correct answer! We no longer need to check that the user is in the slackrtm roster, though we should fix that too because plugins could depend on that behavior.
username_2: As `msguser` isn't used at all I've changed the code to the following
```
if "user" not in event:
logger.debug("event {0} has no user".format(event))
return
```
I'll make the PR soon.
username_2: Ok, I deviated a bit and kept some of the code closer to what it was, and I've made slackrtm update at the same time as fixing this issue (and another issue)
username_1: Fixed in https://github.com/username_1/limbo/commit/6be28e4c973780922b90eb3b02b522c5b4f6e5ff
Status: Issue closed
|
NLog/NLog.Extensions.Logging | 1004617361 | Title: JSON configuration always throws when throwConfigExceptions is true.
Question:
username_0: **NLog version**: NLog 5 preview 1
**Platform**: .NET Core 5
**Current NLog config**
```json
{
"NLog": {
"throwConfigExceptions": true
}
}
```
It looks like `SetNLogElementSettings` attempts to validate the key/value pair produced by the configuration abstraction below.

The `name` key/value pair throws in `SetNLogElementSettings` as it is unknown.
I am adding NLog to a .net core generic host using `.UseNLog()` which may be the problem?
The resulting exception is:
```LOG
NLog.NLogConfigurationException: Unrecognized value 'name'='NLog' for element 'NLog'
at NLog.Config.LoggingConfigurationParser.SetNLogElementSettings(ILoggingConfigurationElement nlogConfig)
at NLog.Config.LoggingConfigurationParser.LoadConfig(ILoggingConfigurationElement nlogConfig, String basePath)
at NLog.Extensions.Logging.NLogLoggingConfiguration.LoadConfigurationSection(IConfigurationSection nlogConfig)
at NLog.Extensions.Logging.NLogLoggingConfiguration..ctor(IConfigurationSection nlogConfig, LogFactory logFactory)
at NLog.Extensions.Logging.RegisterNLogLoggingProvider.<>c__DisplayClass1_0.<TryLoadConfigurationFromSection>b__0(ISetupLoadConfigurationBuilder configBuilder)
at NLog.SetupBuilderExtensions.LoadConfiguration(ISetupBuilder setupBuilder, Action`1 configBuilder)
at NLog.Extensions.Logging.RegisterNLogLoggingProvider.TryLoadConfigurationFromSection(NLogLoggerProvider loggerProvider, IConfiguration configuration)
at NLog.Extensions.Hosting.ConfigureExtensions.CreateNLogLoggerProvider(IServiceProvider serviceProvider, IConfiguration configuration, NLogProviderOptions options)
at NLog.Extensions.Logging.RegisterNLogLoggingProvider.<>c__DisplayClass0_1.<TryAddNLogLoggingProvider>b__2(IServiceProvider provider, IConfiguration cfg, NLogProviderOptions opt)
at NLog.Extensions.Logging.RegisterNLogLoggingProvider.<>c__DisplayClass0_0.<TryAddNLogLoggingProvider>b__4(IServiceProvider serviceProvider)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitFactory(FactoryCallSite factoryCallSite, RuntimeResolverContext context)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSiteMain(ServiceCallSite callSite, TArgument argument)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitCache(ServiceCallSite callSite, RuntimeResolverContext context, ServiceProviderEngineScope serviceProviderEngine, RuntimeResolverLock lockType)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitRootCache(ServiceCallSite singletonCallSite, RuntimeResolverContext context)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSite(ServiceCallSite callSite, TArgument argument)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitConstructor(ConstructorCallSite constructorCallSite, RuntimeResolverContext context)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSiteMain(ServiceCallSite callSite, TArgument argument)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitCache(ServiceCallSite callSite, RuntimeResolverContext context, ServiceProviderEngineScope serviceProviderEngine, RuntimeResolverLock lockType)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitRootCache(ServiceCallSite singletonCallSite, RuntimeResolverContext context)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSite(ServiceCallSite callSite, TArgument argument)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitConstructor(ConstructorCallSite constructorCallSite, RuntimeResolverContext context)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSiteMain(ServiceCallSite callSite, TArgument argument)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitCache(ServiceCallSite callSite, RuntimeResolverContext context, ServiceProviderEngineScope serviceProviderEngine, RuntimeResolverLock lockType)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitRootCache(ServiceCallSite singletonCallSite, RuntimeResolverContext context)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSite(ServiceCallSite callSite, TArgument argument)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitConstructor(ConstructorCallSite constructorCallSite, RuntimeResolverContext context)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSiteMain(ServiceCallSite callSite, TArgument argument)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitCache(ServiceCallSite callSite, RuntimeResolverContext context, ServiceProviderEngineScope serviceProviderEngine, RuntimeResolverLock lockType)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.VisitRootCache(ServiceCallSite singletonCallSite, RuntimeResolverContext context)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteVisitor`2.VisitCallSite(ServiceCallSite callSite, TArgument argument)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteRuntimeResolver.Resolve(ServiceCallSite callSite, ServiceProviderEngineScope scope)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.DynamicServiceProviderEngine.<>c__DisplayClass1_0.<RealizeService>b__0(ServiceProviderEngineScope scope)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.ServiceProviderEngine.GetService(Type serviceType, ServiceProviderEngineScope serviceProviderEngineScope)
at Microsoft.Extensions.DependencyInjection.ServiceLookup.ServiceProviderEngine.GetService(Type serviceType)
at Microsoft.Extensions.DependencyInjection.ServiceProvider.GetService(Type serviceType)
at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService(IServiceProvider provider, Type serviceType)
at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService[T](IServiceProvider provider)
at Microsoft.Extensions.Hosting.HostBuilder.Build()
at Program.Main(String[] args)
```
Answers:
username_1: @username_0 Thank you for reporting this issue. The fix will be included in NLog 5.0 preview-2
See also https://github.com/NLog/NLog.Extensions.Logging/pull/531
Status: Issue closed
username_1: NLog.Extensions.Logging 5.0 preview 2 is now available: https://www.nuget.org/packages/NLog/5.0.0-preview.2 |
urbit/urbit | 641680134 | Title: Bridge Error
Question:
username_0: TypeError: Cannot create property 'message' on string 'Node error: {"code":-32000,"message":"already known"}'
at renderAdditionalInfo (https://bridge.urbit.org/static/js/main.b6a701ac.chunk.js:2322:385)
at PassportTransfer (https://bridge.urbit.org/static/js/main.b6a701ac.chunk.js:2322:1914)
at dh (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:159072:7)
at Jh (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:159548:7)
at mj (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:161821:86)
at jj (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:160809:11)
at Z (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:160666:15)
at $i (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:160525:16)
at https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:158251:17
at push.exports.unstable_runWithPriority (https://bridge.urbit.org/static/js/2.a3f8391b.chunk.js:162946:12) |
openshift/origin | 289784475 | Title: Strange error when getting logs for terminating pod
Question:
username_0: 3.9 master, using json-file driver with docker 1.12.6-68
I had just deleted this pod. I get this when I call logs briefly
```
$ oc logs -n kube-system sts/prometheus -c prometheus
failed to get container status {"" ""}: rpc error: code = OutOfRange desc = EOF
```
@openshift/sig-pod
Answers:
username_0: I'll move this to a bug if this isn't already known
username_0: @username_1 if you can have someone at your team look at this - not sure what we're doing wrong, but looks broken.
username_1: @username_2 PTAL
username_1: xref https://bugzilla.redhat.com/show_bug.cgi?id=1537237
username_0: I *think* I opened a bug, but can't find the tab anymore.
username_2: Upstream issued filed here: https://github.com/kubernetes/kubernetes/issues/59296
Upstream PR posted here: https://github.com/kubernetes/kubernetes/pull/59297 |
openshift/origin | 413071996 | Title: Is the Openshift API swagger.json exposed anywhere?
Question:
username_0: I know there are swagger endpoint docs accessible at `OPENSHIFT_URL/openapi/v2` and also `OPENSHIFT_URL/swagger.json` for the Kubernetes related API's.
Is there also an equivalent endpoint for the Openshift api in swagger format. I see in this repo, there is this document, https://raw.githubusercontent.com/openshift/origin/master/api/swagger-spec/openshift-openapi-spec.json, but i'm not really sure if that is exposed as an endpoint in my Openshift Cluster.
Answers:
username_1: @username_0, as near as I can tell, for release-3.11 that file is generated by this script: https://github.com/openshift/origin/blob/release-3.11/hack/update-generated-swagger-spec.sh
Out of curiosity, what were you planning on doing with spec file?
username_0: We have a Node.js module, [openshift-rest-client](https://github.com/nodeshift/openshift-rest-client), that we created and have been adding functionality "manually".
we are looking to refactor the library so it generates the api automatically, instead of us adding things in manually. There is a nice [kubernetes node module](https://github.com/godaddy/kubernetes-client), which does something similar. One of the ways it can load the api is querying the running clusters `/openapi/v2` or `/swagger.json` endpoint.
If there isn't a similar endpoint for the Openshift api's, i can just use this file, which i've started to do, https://github.com/nodeshift/openshift-rest-client/pull/113
username_1: @username_0, that looks quite interesting.
My own interest in this is using the OpenShift OpenAPI specs to create REST API docs. I'll let you know if I learn anything helpful.
It does look like a running cluster is queried to generate the spec for 3.11 here: https://github.com/openshift/origin/blob/b6db8e68d78723f943eba6710ab6adc8d4254a2f/hack/update-generated-swagger-spec.sh#L40 |
ATMartin/TIY-Homework | 59117632 | Title: 2015-02-26
Question:
username_0: # ember-crud
## Description
A CRUD app using Ember.js
## Objectives
### Learning Objectives
After completing this assignment, you should:
- Understand how to separate data/syncing logic from presentation/interaction
and application state using Ember.js Routes, Controllers, and Templates.
- Understand how to properly write CRUD functionality using Ember.js actions.
### Performance Objectives
After completing this assignment, you should be able to:
* Use ember-cli to quickly generate the files needed to set up an application
with multiple routes, controllers, and templates.
## Details
### Deliverables
* A repo containing an ember-cli project
### Requirements
* No JSHint warnings or errors
## Normal Mode
Write an Ember.js application that allows you to create, update, read, and
delete a resource. The assignment is open ended, but some examples of things you
could make:
- A blog
- A bookmarking application
- A contact list
You may use either the localStorage adapter or Parse.
### Parse (preferred)
- Use an [ember
initializer](http://emberjs.com/api/classes/Ember.Application.html#method_initializer)
to set up your API keys on $.ajaxSetup, which you can generate with ember-cli
- Use $.ajax in your route's `model` and `actions` to GET/POST/PUT/DELETE
### localStorage
- Set up the [localStorage
adapter](https://www.npmjs.com/package/ember-localstorage-adapter)
- Use `this.store.find(thing)`, `this.store.createRecord({})`, etc. as seen in
the Todo tutorial.
## Hard Mode
Implement a filtering route. For example, if you made a bookmarking app, create
a route that allows you to view the bookmarks with a specific tag.
## Additional Resources
* Read [the Ember guides](http://emberjs.com/guides/). Try to get through
"Enumerables", but at least read through "Routing".
Answers:
username_1: https://github.com/username_1/TIY-HW-21-EmberCRUD
Status: Issue closed
username_1: At the moment, this has all CRUD features working, and has a working route to filter by category. However, I'm trying to render category names out to the app index route and am having no luck due to what I think is an implicit controller. The majority of my logic is controller-based, but I'd also like to refactor that to the route so I can be more comfortable with the MVVM vs MVC paradigm here. I'm gonna keep hacking on it with fresh eyes this morning, and if I can get the category list working then I plan on adding unique images (all placeholders right now) and auth to it next.
Thanks!
username_0: I think the problem is that there is only ever one active controller for a template, and it's the child-most controller (if that makes sense). So allCategories would only be available at `/`. In order to make it available elsewhere, you could add the following to that particular controller:
```js
needs: ['rentit'],
allCategories: Ember.computed.alias('controllers.rentit.allCategories')
```
That might not be the best way, but it's the first way that comes to mind. |
hlminh2000/lifeBalance | 321420554 | Title: Fix com.facebook.react.common.JavascriptException in ExceptionsManagerModule.java line 56
Question:
username_0: ### Version 1.0(1525833228) ###
### Stacktrace ###
com.facebook.react.modules.core.ExceptionsManagerModule.showOrThrowError (ExceptionsManagerModule.java:56);
com.facebook.react.modules.core.ExceptionsManagerModule.reportFatalException (ExceptionsManagerModule.java:40);
com.facebook.react.bridge.JavaMethodWrapper.invoke (JavaMethodWrapper.java:374);
com.facebook.react.bridge.JavaModuleWrapper.invoke (JavaModuleWrapper.java:162);
com.facebook.react.bridge.queue.NativeRunnable.run (NativeRunnable.java);
com.facebook.react.bridge.queue.MessageQueueThreadHandler.dispatchMessage (MessageQueueThreadHandler.java:31);
com.facebook.react.bridge.queue.MessageQueueThreadImpl$3.run (MessageQueueThreadImpl.java:194);
### Reason ###
com.facebook.react.common.JavascriptException
### Link to App Center ###
* [https://appcenter.ms/users/username_0-pyo4/apps/lifeBalance/crashes/groups/c51b37d817263cebfbf6d06c18a78556eff56f6a](https://appcenter.ms/users/username_0-pyo4/apps/lifeBalance/crashes/groups/c51b37d817263cebfbf6d06c18a78556eff56f6a) |
juliema/aTRAM | 221874071 | Title: Don't assume we have write access to directories
Question:
username_0: Use case: If you're running data built by another user then you may not have access to the --work-dir directory. Change it to use a prefix for input only (trim the names) and make an --output-dir option for the temp files.<issue_closed>
Status: Issue closed |
didykivoice/ride-my-way | 443811633 | Title: Code Formatting
Question:
username_0: @didykivoice Do not use more than one blank line in your html code
```
<span id="slide-line"></span>
</nav>
<div id="box" class="box-1">
<div id="boxtext" class="boxtext-1">
<h1>Ride-MyWay</h1>
<p>fORLoop week-01 Challnge 02 UI Designing</p>
<img src="img/picha.jpg" width="100%" height="70%" alt="pic">
<p>Ping a driver to get your ride</p>
</div>
</div>
<footer class="fixed_footer">
```
Also make sure before commit your code is formatted with 2 or 4 spaces for code indent.
Good work |
openebs/openebs | 284908119 | Title: Resolve maya-nodebot dependency on admin privilege(sudo)
Question:
username_0: <!-- This form is for bug reports and feature requests ONLY! -->
<!-- Thanks for filing an issue! Before hitting the button, please answer these questions.-->
## Is this a BUG REPORT or FEATURE REQUEST?
Choose one: BUG REPORT or FEATURE REQUEST
FEATURE REQUEST
<!--
If this is a BUG REPORT, please:
- Fill in as much of the template below as you can. If you leave out information, we can't help you as well.
If this is a FEATURE REQUEST, please:
- Describe *in detail* the feature/behavior/change you'd like to see.
In both cases, be ready for followup questions, and please respond in a timely
manner. If we can't reproduce a bug or think a feature already exists, we
might close your issue. If we're wrong, PLEASE feel free to reopen it and
explain why.
-->
**What happened**:
kubeconfig file location needs to be passed as flag variable and that can't be handled
when maya-nodebot runs as daemonset.
**What you expected to happen**:
Use incluster config kubernetes api. If it fails then try to use kubeconfig flag.
**How to reproduce it (as minimally and precisely as possible)**:
Running maya-nodebot start should work, without kubeconfig flag.
<!--
**Environment**:
- kubectl get nodes
- kubectl get pods --all-namespaces
- kubectl get services
- kubectl get sc
- kubectl get pv
- kubectl get pvc
- OS (e.g. from /etc/os-release):
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
-->
Answers:
username_1: Add to the implementation details for https://github.com/openebs/node-disk-manager/pull/1 |
versionone/component-library | 471101096 | Title: Switch doesn't behave as expected when passing default state
Question:
username_0: While working within CTM @mickeyrogers and I observed some unexpected behavior. The switch itself worked fine unless we passed state to it for the `checked` prop. After spending a couple hours on it we tried a regular checkbox and it worked fine.
Answers:
username_1: Can you express what the expected behavior was? If you pass a checked prop then the switch should reflect exactly that state independent of further user interaction. Controlled vs. uncontrolled React components. Can you verify this is still an issue?
username_0: Can you clarify "switch should reflect exactly that state independent of further user interaction"?
**Expected behavior:** When I pass a default value of `true` to the checked prop then the switch should show that it's checked. When the user clicks the switch to turn it off then the switch should show that it's unchecked.
username_2: Talked with Matt, the source of his issue was not the switch. Closing
Status: Issue closed
|
ChalkyBrush/roshpit-bug-tracker | 298726443 | Title: Arkimus - Glyph of Mana Shield
Question:
username_0: Even if damage instance is absorbed, mana still consumes


Answers:
username_0: I know, mana shield effect comes before any mitigations, right?
username_1: yea so it needs to be moved to final operation so this is in fact a bug (not fixed yet just commenting)
Status: Issue closed
username_1: fixed next patch |
stripe/bonsai | 125048467 | Title: Consider removing implicit/default Layouts
Question:
username_0: Since an efficient layout is such a crucial part of successfully using Bonsai, it might be better to have folks create their own, rather than providing implicit instances. Look into what it would take to remove the implicits and just use builders.
Answers:
username_1: I definitely generally agree.
I think the future will likely have people using some sort of Generic-style builder that'll remove a lot of the complexity in constructing instances in the common case (eg `Layout.generic[MyLabelType]`).
Also, serialization (eg on-disk) is a thing we need to start thinking about. It's unclear to me if we want to conflate in-memory layout with on-disk, but there are some obvious benefits in load time if we do (can even imagine an on-disk, memory-mapped back `Vec`). However, in that case we *really* need to be sure we don't accidentally swap out the `Layout` being used, so there can be no generic implicits. |
sailfishos/sailfish-secrets | 311495134 | Title: Misleading package name: libsailfishcryptoplugin
Question:
username_0: You would think that `libsailfishcryptoplugin` contains an actual crypto plugin, but it's just the QML plugin. This issue is a reminder to myself to change the name of this package, and also `libsailfishsecretsplugin`. |
osmlab/osm-community-index | 316044208 | Title: OSM Peru 🇵🇪
Question:
username_0: <!--
This template can help you add your community resource to the index.
Fill in whatever sections you want, or feel free to ignore it!
-->
## Community Resource Name
Openstreetmap Peru
### What is it?
- Homepage
- Mail list
- Matrix
- Telegram
- Twitter
- Facebook
### URL link to the resource
URL links to the resources
- Homepage OSMPE: http://osmpe.ourproject.org/
- Mail list OSMPE: http://lists.openstreetmap.org/listinfo/talk-pe
- Matrix: https://matrix.to/#/#osmpe:matrix.org
- Telegram: https://telegram.me/osmPe
- Twitter: https://twitter.com/osmpe
- Facebook: https://www.facebook.com/Osmpe
### Points of contact
@osm-pe Team on GIthub
### Bounding polygon
- Relation: https://www.openstreetmap.org/relation/288247
- Geojson file: https://gist.github.com/username_0/d80c5396b0d8efd22be69e768806a6e7
### Description (optional)
<!--
example: "Join our group here: {url}"
-->
### Languages spoken (optional)
"es"
### Country codes (optional)
"pe"
### Add an emoji flag (optional)
🇵🇪<issue_closed>
Status: Issue closed |
jaegertracing/jaeger-idl | 1162225713 | Title: Unclear documentation for max, minDuration for traces endpoint
Question:
username_0: Must all spans in a trace be larger than `minDuration`, if set? Must only one span in a trace fit within the min and max duration parameters? It's not really obvious what the intention is here.
Answers:
username_1: The zipkin files are just a copy from the Zipkin project, we're not going to change them.
Status: Issue closed
username_0: OK - can you point me to the right place to file a ticket?
username_1: probably here https://github.com/openzipkin/zipkin-api |
nuclio/nuclio | 644188369 | Title: Blocked by github for misbehaving after restarting too many times
Question:
username_0: When nuclio dashboard starts it fetches function templates from github.
Restarted it too many times, and got blocked by github for misbehaving
Am using minikube setup from https://nuclio.io/docs/latest/setup/minikube/getting-started-minikube/
See logs below from dashboard pod that has status crashloopbackoff:
Running in parallel
Starting dashboard
Starting nginx
20.06.23 08:46:36.210 [37m dashboard.platform[0m [32m(D)[0m Using kubeconfig {"kubeconfigPath": ""}
20.06.23 08:46:36.212 [37mrd.platform.docker.runner[0m [32m(D)[0m Executing {"command": "docker version"}
20.06.23 08:46:36.241 [37mrd.platform.docker.runner[0m [32m(D)[0m Command executed successfully {"output": "Client: Docker Engine - Community\n Version: 18.09.6\n API version: 1.39\n Go version: go1.10.8\n Git commit: 481bc77\n Built: Sat May 4 02:33:34 2019\n OS/Arch: linux/amd64\n Experimental: false\n\nServer: Docker Engine - Community\n Engine:\n Version: 19.03.8\n API version: 1.40 (minimum version 1.12)\n Go version: go1.12.17\n Git commit: afacb8b7f0\n Built: Wed Mar 11 01:30:32 2020\n OS/Arch: linux/amd64\n Experimental: false\n containerd:\n Version: v1.2.13\n GitCommit: 7ad184331fa3e55e52b890ea95e65ba581ae3429\n runc:\n Version: 1.0.0-rc10\n GitCommit: dc9208a3303feef5b3839f4323d9beb36df0a9dd\n docker-init:\n Version: 0.18.0\n GitCommit: fec3683\n", "stderr": "", "exitCode": 0}
20.06.23 08:46:36.250 [37m dashboard[0m [32m(D)[0m Fetching function templates from git repository {"templatesGitRepository": "https://github.com/nuclio/nuclio-templates.git", "templatesGitRef": "refs/heads/master"}
20.06.23 08:46:36.273 [37mitFunctionTemplateFetcher[0m [32m(D)[0m Fetching templates from git {"ref": "refs/heads/master"}
Error - Get "https://github.com/nuclio/nuclio-templates.git/info/refs?service=git-upload-pack": dial tcp: lookup github.com on 10.96.0.10:53: server misbehaving
.../gitfunctiontemplatefetcher.go:80
Call stack:
Failed to initialize git repository
.../gitfunctiontemplatefetcher.go:80
Failed to clone repository
.../gitfunctiontemplatefetcher.go:58
Failed to fetch one of given templateFetchers
.../dashboard/functiontemplates/repository.go:35
Failed to create repository out of given fetchers
/nuclio/cmd/dashboard/app/dashboard.go:131
parallel: This job failed:
/runners/dashboard.sh
Exiting
Answers:
username_1: Hi @username_0,
Not a very common issue as far as I can tell :\
Why were your pod "misbehaving" - in other words, whats the cause of the "too many" (?!) restarts?
I'm guessing after you solved the root cause of the restarts, life went back to normal and all is well?
username_2: Hey @username_3 ,
You may use the following envs on nuclio-dashboard
- `NUCLIO_DASHBOARD_OFFLINE:true`
- `NUCLIO_TEMPLATES_ARCHIVE_ADDRESS:none`
- `NUCLIO_TEMPLATES_GIT_REPOSITORY:none`
to achieve offline working.
also look at https://github.com/nuclio/nuclio/issues/2021#issuecomment-757130215 to avoid pulling base images while building functions
username_3: Thanks for the quick response. I had already circumvented the issue by creating an offline git repo and mounting the private CA pem file in the Nuclio container.
Status: Issue closed
|
virtualstaticvoid/asdf-docker-compose | 1016018273 | Title: Download URL changed since version 2
Question:
username_0: When installing docker-compose 2.0.1 using `asdf install docker-compose 2.0.1` the download fails and `~/.asdf/installs/docker-compose/2.0.1/bin/docker-compose` contains "Not found".
In fact, the download URL has changed :
- v1.29.2 : https://github.com/docker/compose/releases/download/1.29.2/docker-compose-Linux-x86_64
- v2.0.1 : https://github.com/docker/compose/releases/download/v2.0.1/docker-compose-linux-x86_64
Answers:
username_1: See the [About update and backward compatibility][update] and [Where to get Docker Compose][install]
pages of the `docker-compose` repository for further details.
I will update this plugin to include this on the README and show a warning when running `asdf list-all docker-compose` or when a user attempts to install a version >= 2.0.
[install]: https://github.com/docker/compose#where-to-get-docker-compose
[update]: https://github.com/docker/compose#about-update-and-backward-compatibility |
mbert/kubeadm2ha | 301257824 | Title: kube component can not do `Watch` when apiserver is set to master IP?
Question:
username_0: hi (:
the default load balancing strategy of nginx is **rr**, so when a pod(sth like kube-proxy) do `Watch` action, it will print a lots of warning log message like
W0301 02:10:52.929987 1 reflector.go:341] k8s.io/kubernetes/pkg/client/informers/informers_generated/internalversion/factory.go:85: watch of *core.Service ended with: very short watch: k8s.io/kubernetes/pkg/client/informers/informers_generated/internalversion/factory.go:85: Unexpected watch close - watch lasted less than a second and no items received
How to deal with this issue? or just ignore it?
Answers:
username_1: The most obvious thing to try would be setting up load balancing with a strategy other than rr. As far as I understand, this is not supported by the free version NGINX (you'd have to upgrade to the expensive 'plus' edition). Hence something other than NGINX may be worth a try - at least in order to see whether the effect you describe disappears.
I might take a look into this at some time, but this may take a while. If you have something in your sleeve I'd happily take a look at a PR :)
username_1: Closing this because there does not seem to be a tweak to get this working with _nginx_. Hence if one wants to use the 'watch' commands, then _nginx_ load balancing cannot be used.
Status: Issue closed
|
leolorenzoluis/xyz.MonacoEditorLoader | 530887382 | Title: Custom path for loading monaco doesn't work
Question:
username_0: I spent some time trying to pass in a custom path for the *loadMonacoEditor directive and I realized that no matter what I changed the path to, the path the service was trying to load from would not change.
I noticed that the editor load logic actually happens in the constructor of the loader service:
https://github.com/username_1/xyz.MonacoEditorLoader/blob/55c9ad2b12e808eabaf39b3b88105352ced21c58/lib/monaco-editor-loader/monaco-editor-loader.service.ts#L15
Even though the directive has an @Input that passes in the custom path, that doesn't actually ever activate since the loading logic executes in the constructor. By the time the input is set, the LoaderService has already attempted to load the editor.
Additionally, the README suggests that it's the inclusion of the *loadMonacoEditor directive
that loads the editor, but this isn't the case.
```
<monaco-editor *loadMonacoEditor></monaco-editor>
```
Even without the directive, the editor will load since the logic will be execute on construction of the service when any component imports the MonacoEditorLoaderModule.
I propose that the editor service should have a load() method that is actually only called in the MonacoEditorLoaderDirective, after the @Inputs are set within the ngOnInit.
Answers:
username_1: How are you setting the custom path? I tested with custom path and it is working as expected. The life cycle is correct. When the input directive is set, the property is respected to whatever value you set.
username_0: @username_1 Thanks for looking into this. I set the directive as specified in the README
```
<monaco-editor *loadMonacoEditor="'test/monaco'"></monaco-editor>
```
Perhaps there is a race condition. In your test, is the AMD loader being loaded via loaderScript? If so, onGotAmdLoader() might be executing after the Input is set - but that wouldn't be the case if the window.require was provided and onGotAmdLoader() fires immediately
username_1: Ah makes sense now. I have my require as undefined. Is this what you propose?
```
import { Injectable, NgZone } from '@angular/core';
// tslint:disable-next-line:import-blacklist
import { BehaviorSubject } from 'rxjs';
@Injectable()
export class MonacoEditorLoaderService {
isMonacoLoaded: BehaviorSubject<boolean> = new BehaviorSubject<boolean>(false);
private _monacoPath = 'assets/monaco-editor/vs';
set monacoPath(value: any) {
if (value) {
this._monacoPath = value;
this.load();
}
}
load = () => {
// Load monaco
console.log(this._monacoPath);
(<any>window).require.config({ paths: { 'vs': this._monacoPath } });
(<any>window).require(['vs/editor/editor.main'], () => {
this.ngZone.run(() => this.isMonacoLoaded.next(true));
});
};
constructor(private ngZone: NgZone) {
// Load AMD loader if necessary
if (!(<any>window).require) {
const loaderScript = document.createElement('script');
loaderScript.type = 'text/javascript';
loaderScript.src = `${this._monacoPath}/loader.js`;
loaderScript.addEventListener('load', this.load);
document.body.appendChild(loaderScript);
}
}
}
```
username_0: Yes, that should do the job. Thanks!
username_1: @username_0 Published. Should be fixed on `8.0.13`
Status: Issue closed
|
jfrimmel/cargo-valgrind | 607027939 | Title: Not working on unit tests
Question:
username_0: Hi. Really awesome project. I love how much simpler it is to check for memory leaks etc with this tool. However I'm not able to run this on my integration tests on a Fedora Linux machine. I'm not sure if Cargo changed in some incompatible way or if I have an unsupported version of valgrind?
This is the software I'm running:
```
$ valgrind --version
valgrind-3.15.0
$ cargo valgrind --version
cargo-valgrind 1.3.0
$ cargo --version
cargo 1.43.0 (3532cf738 2020-03-17)
```
And this is the issue I'm getting:
```
$ cargo valgrind --test all_tests
...
error: Could not parse XML: custom: 'missing field `xwhat`'
```
Answers:
username_0: It seems to work when running under Rust 1.39. `cargo +1.39.0 valgrind --test all_tests` does work. But only after a `cargo clean`. The builds with newer Rust seem to leave some metadata that `cargo-valgrind` does not understand.
username_1: that would be great if I could run this on unit tests.
username_2: Hello @username_0,
can you provide me some test input for reproduction? Alternatively, you can try out the new version I've just updated. If you want to, you can try the new software using the following command:
```bash
cargo install --git https://github.com/username_2/cargo-valgrind --branch custom-runner
```
username_0: I currently can't reproduce the issue. I can't remember exactly how I hit the problem, but I remember it being easy to get back then. I now have valgrind-3.16.1. So that could be why. Or newer Rust.
I think I got the problem in this repo: https://github.com/username_0/loom-executor. It's the only one of my repos where I can find a `all_tests.rs`.
username_3: I ran into this issue when using Valgrind 3.15.0 as packaged in the Ubuntu 20.04 repositories. The problem went away when I compiled Valgrind 3.16.1 on my own.
username_4: Hello and *thank you very much* for this tool which is incredibly useful!
I am encountering the same bug locally with valgrind 3.16.1. Therefore, I installed cargo-valgrind from the malformed-xml branch using:
```
$ cargo install --git https://github.com/username_2/cargo-valgrind --branch malformed-xml -- cargo-valgrind
```
I rerun the test and I now have the failing XML printed out!
[valgrind.log](https://github.com/username_2/cargo-valgrind/files/6488984/valgrind.log)
I hope this helps!
username_2: Thank you for the log, that's helpful. I'll have a look and hope to fix it soon.
username_5: Hi! I wanted to share the environment in which I’m encountering the same issue today, including sample project and Dockerfile to reproduce. I hope this helps!
[valgrind_test.zip](https://github.com/username_2/cargo-valgrind/files/6577728/valgrind_test.zip) |
mapbox/mapbox-navigation-android | 338315228 | Title: Issue in OffRouteDetector
Question:
username_0: **Android API:** N/A
**Mapbox Navigation SDK version:** 0.15.0
### Steps to trigger behavior
1. Start navigation
2. Move slowly along the route (walking pace can be enough), such that the distance to maneuver stays basically the same
3. Wait for 4 GPS readings
### Expected behavior
- Navigation continues normally
### Actual behavior
- Rerouting is triggered
I believe this is caused by an error in OffRouteDetector.movingAwayFromManeuver(). it checks that:
distancesAwayFromManeuver.peekLast()
- distancesAwayFromManeuver.peekFirst() < MINIMUM_BACKUP_DISTANCE_FOR_OFF_ROUTE
I believe there is an error here.
- The queue will already be cleared (later in the code) if the distance is decreasing.
- Therefore the queue from oldest to newest must be increasing (or it could remain constant)
- peekLast() takes the oldest value, peekFirst() takes the newest, so peekLast() - peekFirst() will always be <= 0
- Which means this check will always be true
I assume it is meant to check if the user has gone "more than 50m backwards" before triggering, in which case the check should be:
peekFirst() - peekLast() > MINIMUM_BACKUP_DISTANCE_FOR_OFF_ROUTE
Note that because distancesAwayFromManeuver are integers, even slightly decreasing distances will still trigger this - as long as they round to the same integer.
e.g. distances of 210.40, 210.18, 210.34 and 209.94 will trigger this.
As will distances of 200, 210, 220, 230 (which is backwards, but as per the above description, I assume is meant to be filtered).
Also note the 4th distance seems to be pointless. It will always trigger then anyway because the "return true" happens before there is a chance to clear if it's back in the right direction. If that's intended, it's not clear why it doesn't just trigger on the previous call.
Answers:
username_1: Hey @username_0 👋 thanks for the great write-up here. We will take a look at this logic in-depth and get back to you here on this ticket. At first glance, all of your points look completely valid to me.
username_2: I have the same problem.
Thank you
username_3: We've the same problem.
username_4: I have same problem on v16. here is logcat:
`8-17 17:14:31.778 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false
08-17 17:14:33.744 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false
08-17 17:14:34.827 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false
08-17 17:14:36.764 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: true
08-17 17:14:37.815 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false
08-17 17:14:38.874 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false
08-17 17:14:40.853 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false
08-17 17:14:42.788 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false
08-17 17:14:44.785 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: true
08-17 17:14:45.800 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false
08-17 17:14:46.820 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false
08-17 17:14:48.788 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false
08-17 17:14:49.848 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: true
08-17 17:14:52.904 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false
08-17 17:14:54.819 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: isUserOffRoute: false
08-17 17:14:55.823 1978-2089/com.mapbox.services.android.navigation.testapp D/RouteProcessorHandlerCallback: `
after first offRoute it should print true for all logs but as you can see sometimes it detect user is not offroute and Rerouting is not triggered.
username_4: **UPDATE**
I think I found the reason and some added information for debuging:
the problem would occur in last leg. the `checkForArrivalEvent` method in `NavigationEventDispatcher` consider it as Arrival and delete the listener so when offroute occur there is no listener to invoke reroute request
username_1: @username_0 @username_4 do ya'll mind retesting with `0.20.0`?
username_0: @username_1 If I get time I'll have a look. But we've actually taken the opportunity to build our own off route engine, because we wanted extra functionality. I'll have to see how easy it is to go back to the original.
username_1: @username_0 okay great, thank you. @username_4 if you don't mind testing as well, I'd just like to confirm this is no longer and issue with the new logic.
username_5: Hey @username_4 👋 were you able to retest? It'd be 💯 if you could report back confirming that the issue is gone. Thanks a lot!
I also want to take the opportunity to let you know that we released version [`0.22.0`](https://github.com/mapbox/mapbox-navigation-android/releases/tag/v0.22.0) so I encourage you to upgrade because there are a bunch of improvements and new features plus lots of bug fixes since `0.16.0`.
username_4: Hi sorry I didn't see the thread. I will test in tomorrow and let you know 🙏🏻
username_5: Sweet! Thanks for the update @username_4 🙇
username_1: Hey @username_4 @username_0 I'm going to go ahead and close this as resolved. @username_0 once we unblock you from upgrading and if you re-test and still see issues, please feel free to cut a new ticket and we will continue to dig on this. Thanks!
Status: Issue closed
username_1: @username_0 this should be right behind our current push for offline routing - I can't give a specific timeline but this is definitely high priority. We will be sure to update your ticket regarding this once we start moving. Thanks as always for your patience 👍 |
studydash/cards | 936362956 | Title: Release `Server 0.0.5` & `GUI 1.8.5`
Question:
username_0: ## Saturday
Tonight, I released `Server 0.0.5` and `GUI 1.8.5` into `QA`. The new server build now adds a `Record` and increments `RecordCount` and `Streaks` on every new card insertion:

And for the new GUI, I reorganized the `Members Pane`; it now looks like this:

<br />
## Extra/Fun 🎈🎉:
This is July 4th weekend, so Shaza and I visited a new bakery this afternoon and also checked out the Art Climb steps at the Cincinnati Art Museum. The weather was thankfully still cool; it was a good time! 🥳
 |
Hirnbix/captain-holetooth | 183537477 | Title: Missing translations
Question:
username_0: There appears to be support for adding translations to texts and labels, however it has not been fully implemented.
This is just a reminder that it is an issue and should be implemented. :)
Answers:
username_1: Milestone goal: Add english and german
username_1: Pushing this to a new milestone / project as it is unclear how much translation work is needed at all. |
nss-evening-cohort-10/dinomate | 491738243 | Title: Create Wireframes
Question:
username_0: # User Story
As a developer, when I do stuff I would like to see some wireframes
# AC
**WHEN** I open this ticket
**THEN** I should see screenshots for each page
# Dev Notes
* Use Moquops to generate wireframes
* paste them into this ticket
Answers:
username_0: Home Page:

username_0: About Us:

username_0: Profile:

Status: Issue closed
|
dask/distributed | 282580628 | Title: Move worker_info onto WorkerState?
Question:
username_0: @username_1 I'm curious why you chose not to move worker metadata like this onto the WorkerState object?
Answers:
username_1: Just that it was not relevant for the immediate goal of trying to cythonize the scheduler, and did not present any obvious difficulties that deserved to be tackled up front :-) But, yes, we should probably do it anyway.
username_1: This is fixed, no?
Status: Issue closed
|
gbif/portal-feedback | 299466620 | Title: Link to description of data validation rules in the data validator
Question:
username_0: **Link to description of data validation rules in the data validator**
It would be helpful if the data validator (which by the way is a great tool) could provide links to the data validation rules / short description of what is expected of the different fields checked. - i.e. a short link to the proper documentation. This would help data publishers in faster find the out what is wrong in cases where this is not obvious .
-----
User provided contact info: <EMAIL>
System: Firefox 58.0.0 / Ubuntu 0.0.0
User: [See in registry](https://www.gbif.org/api/feedback/user/15a8d2e2150b6c740b608f88ace5269d:ee803319c81be99ceb64813a4b8c0226f0f69b9435fb3c4972cd094f767c6346342f3b684e4e97bfe4c5635c4d42af1f57517bacbf2c6d2398da9492a117dbf6)
Referer: https://www.gbif.org/tools/data-validator/1516700048547
Window size: width 1663 - height 895
[API log](http://elk.gbif.org:5601/app/kibana?#/discover?_g=(refreshInterval:(display:Off,pause:!f,value:0),time:(from:'2018-02-22T18:47:00.712Z',mode:absolute,to:'2018-02-22T18:53:00.712Z'))&_a=(columns:!(_source),index:'prod-varnish-*',interval:auto,query:(query_string:(analyze_wildcard:!t,query:'response:%3E499')),sort:!('@timestamp',desc)))
[Site log](http://elk.gbif.org:5601/app/kibana?#/discover?_g=(refreshInterval:(display:Off,pause:!f,value:0),time:(from:'2018-02-22T18:47:00.712Z',mode:absolute,to:'2018-02-22T18:53:00.712Z'))&_a=(columns:!(_source),index:'prod-portal-*',interval:auto,query:(query_string:(analyze_wildcard:!t,query:'response:%3E499')),sort:!('@timestamp',desc)))
System health at time of feedback: OPERATIONAL<issue_closed>
Status: Issue closed |
linkerd/linkerd | 296633206 | Title: namerd admin interface hangs
Question:
username_0: ```
There are **no** log entries neither in the consul nor in the namerd log.
**What you expected to happen**:
Not to hang? :)
**How to reproduce it (as minimally and precisely as possible)**:
Configure as I did, curl /.
**Anything else we need to know?**:
Until I got the consul configuration right, the interface *did* load with an error telling me that no namespaces were found (as a side note, the documentation around namerd and its integration with linkerd and other services was not quite up to the level of what I got used to from linkerd's docs – I needed additional guidance otherwise I’d still be stumbling).
**Environment**:
- linkerd/namerd version, config files:
namerd.yml:
```yaml
---
admin:
port: 9991
ip: 10.6.180.250
tls:
certPath: /etc/ssl/certs/XXX.crt
keyPath: /etc/ssl/private/XXX.pkcs8
namers:
- kind: io.l5d.consul
includeTag: true
useHealthCheck: true
storage:
kind: io.l5d.consul
token: XXX
readConsistencyMode: stale
pathPrefix: /namerd/dtabs
interfaces:
- kind: io.l5d.mesh
ip: 10.6.180.250
port: 4321
telemetry:
- kind: io.l5d.prometheus
path: /admin/metrics/prometheus
prefix: namerd_
```
namerd: 1.3.5
consul: 1.0.1
- Platform, version, and config files (Kubernetes, DC/OS, etc):
Ubuntu Xenial, since this is an isolation namerd issue, the rest shouldn’t matter.
- Cloud provider or hardware configuration:
LXC running on metal. More than 1 GB of RAM free.
Answers:
username_1: @username_0 thanks so much for filing this detailed report. These kinds of issues make debugging linkerd/namerd much easier. We will take a look at this issue as soon as we can. Also, if you have any suggestions on where you would love to see more detail in namerd's documentation, please do let us know!
username_0: JFTR, just tested with consul 1.0.6 and the problem persists.
username_2: @username_0 we're taking a look at this
username_3: Hey @username_0, thanks for such a detailed report! I've been trying to reproduce this issue locally, and haven't been able to exactly. But I did notice something in your setup that seems potentially problematic, based on the contents of your consul KV store:
```
[
{
"CreateIndex": 137788444,
"Flags": 0,
"Key": "namerd/dtabs/",
"LockIndex": 0,
"ModifyIndex": 137788444,
"Value": null
},
{
"CreateIndex": 137788454,
"Flags": 0,
"Key": "namerd/dtabs/default",
"LockIndex": 0,
"ModifyIndex": 137801542,
"Value": "<KEY>"
}
]
```
The first key in that list is "namerd/dtabs/", which is also the folder where the second key, "namerd/dtabs/default", is located. When testing locally, if I create the first key first and the second key second, then namerd's admin UI doesn't hang, but it does tell me that there are no namespaces found:

If I use the consul API to delete the first key:
```
$ curl -X DELETE http://localhost:8500/v1/kv/namerd/dtabs/
true
```
And then bounce namerd, its admin interface successfully loads the namespace from the second key:

Am not positive you're experiencing the same issue, but can you try applying the same fix in your setup (delete the `namerd/dtabs/` key and restart namerd)?
---
When playing around with this locally, I setup a docker-compose env that might be useful. You can find it here:
[consul-issue.tar.gz](https://github.com/linkerd/linkerd/files/1726287/consul-issue.tar.gz)
There are some instructions for reproducing the issue in the README.
username_4: Wow, @username_3, great catch! I tried to repeat your steps and was able to reproduce the issue exactly as @username_0 reported it! In my case, as well as in @username_0's case, Consul has ACLs enabled with default policy set to "deny-all" and ACL permissions set to
```
key "namerd/dtabs/" { policy = "write" }
```
which may explain the difference in behavior we see.
Let me see if I can figure out what causes it...
username_0: I can confirm that deleting everything and creating `namerd/dtabs/default` in one step fixes the problem for me. Thanks everyone and good luck fixing. :)
username_4: The cause of the issue as well as the fix turned out to be embarrassingly simple and _it looks like_ this issues has been in place since first implementation of the store that I pushed almost 2 years ago. 😞
username_3: @username_0 Thanks for confirming that the workaround works -- that's great! And huge thanks to @username_4 for putting together the fix. I tested the changes from #1816 in my local repro, and they do indeed fix the issue that I was seeing.
Status: Issue closed
username_2: W00t! |
quarkusio/quarkus | 860756670 | Title: Controlling Service Account roles when using the Kubernetes and Kubernetes Client extension
Question:
username_0: ## Description
It currently is impossible to disable the service account generation when using the Kubernetes en Kubernetes Client extension in a project. It gererates a service account and a rolebinding linked to view role. Sometimes this is not strict enough. It is possible to work around using Kustomize, but this way using the deployment feature of the Kubernetes extension is impossible.
Being able to disable the Service Account generation will solve this, but that is not optimal as a Service Account and RoleBinding would have to be applied outside of the extension control.
An other option might be to specify a role to bind to in stead of view. The role could either be provided using the kubernetes.yml in the source tree or be provided or it would need to be present on deployment already. Any check on the existence of the role should be configurable.
Answers:
username_1: cc @username_2
username_2: In theory people could specify the rbac configuration in src/main/kubernetes and make it as strict as they need to.
Haven't tried it though...
Now I can see a feature flag to disable the generation of rbac resources. Also we could point to an existing Role binding or even control the privileges. These are all valid approaches.
username_0: @username_2 specifying the rbac configuration src/main/kubernetes does add them to the generated descriptors, but if the name is not the same as would be generated it is just added and the *-view rolebinding is still generated. A bit more control over how it all is generated seems a good idea to me. |
google/filament | 687325242 | Title: ANDROID: How to eliminate lag and improve FPS.
Question:
username_0: I am referencing sample gltf viewer demo to load my model. everything is working fine except lagging. once model loads, the screen starts lagging so much. is there any way to tackle this issue?
My model size is around 150 MB.
Answers:
username_1: The size of the glTF doesn't matter much, what matters is going to be:
- How many triangles
- How many materials
- How large the textures
- etc.
You can optimize your glTF by trying to reduce all those values. A good rule of thumb would be to use ~40,000 triangles and textures that are 1024x1024 each.
You can also:
- Make sure the app doesn't render at native resolution (1080p or 720p is good enough on many devices)
- Enable dynamic resolution
- Turn off expensive features like bloom or SSAO
Status: Issue closed
username_2: @username_1 you suggested changing the render resolution in a couple of threads. Do you mean setting it via `this.surfaceView.holder.setFixedSize(760, 360)` ?
username_2: Answering the initial question, removing unnecessary lights from the scene will improve performance drastically. My app was lagging bad and removing a light from the scene basically fixed the issue for me. |
MicrosoftDocs/azure-docs | 665706493 | Title: Incorrect output
Question:
username_0: Hi - I've followed these steps exactly, but the output doesn't give you a CSV with the transformed data. Instead, it just prints out the text "iris_setosa.csv" into Cell A1 of a CSV in my output container - is there an error in the code or something that I'm missing?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 894562e1-2c07-2f75-b420-d843262a7f07
* Version Independent ID: 1b1fa915-19f0-653c-04d5-218ed4335398
* Content: [Run Python scripts through Data Factory - Azure Batch](https://docs.microsoft.com/en-us/azure/batch/tutorial-run-python-batch-azure-data-factory)
* Content Source: [articles/batch/tutorial-run-python-batch-azure-data-factory.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/batch/tutorial-run-python-batch-azure-data-factory.md)
* Service: **batch**
* GitHub Login: @mammask
* Microsoft Alias: **komammas**
Answers:
username_1: @username_0 Thank you for your feedback . We will investigate and update the thread further.
username_2: @username_0 Apologies for the delay in response. I was checking with our team internally on the issue and it seems like there are some changes made. Can you try again and confirm if the issue is resolved for you or not.
Thanks.
username_2: @username_0 Hope the provided information is helpful.
We will now close this issue. If there are further questions regarding this, please tag me in a comment. I will reopen it and we will continue the discussion.
Status: Issue closed
|
uspki/policies | 271756880 | Title: Section 5.4.1
Question:
username_0: **Organization / Program:** DoD <br>**Section:** 5.4.1 <br>**PDF Page:** 52<br>**PDF Line(s):** 1257<br>**Comment:** Items b and c are too broad and do not constitute clear policy. Revise to make them more concrete.<br><br>**Suggested Change:** Revise per comment
Answers:
username_0: Is it preferred to state this policy requirement in terms of 800-53 security controls inclusive of AC-6, AC-7, AU-2, AU-3 (etc)?
Status: Issue closed
|
Shopify/semian | 188339905 | Title: Support for error percentage?
Question:
username_0: Hi guys, love the library, but I wanted to see if you would be open/interested in adding support for error percent in addition to the current absolute error threshold? I find dealing with error rates in terms of percentages much more flexible than absolute values. If you're open to it I would be interested in trying to tackle it and send you a pull request. However, I don't want to maintain a fork of the project, so only want to go down this path if it is likely to get integrated back into the main code line.
Answers:
username_1: The percentage would have to be over some sliding window of requests, wouldn't you have to configure the size of that window based on time or absolute size?
username_0: Agreed, if I'm understanding the current code there's already a SlidingWindow used for tracking errors so I would imagine that we'd have to track the number of requests utilizing the same window so that we could do the proper comparison. Plenty of details to work out, but not work figuring out unless it would be considered generally useful.
username_1: Honestly, I'm not sure this is worth it. I actually think it's harder to reason about errors in a window than it is for absolute errors in a row, and in our experience this has worked well. If you'd like to contribute, a huge, huge improvement to Semian would be a way to share the circuit breakers between processes in an IPC namespace. A SysV implementation of the circuit breaker using shared memory would be a massive improvement.
username_0: Strange, I don't think it sounds that hard to reason about (assuming we expose the right information). However, I'll take that as a no for now. Let me know if you guys reconsider. I'll see how far we get utilizing absolute values. I'm just worried that we'll find ourselves having to do very different configurations for services that are called very different volumes.
Status: Issue closed
username_1: Why do you think the configuration would be different? We use the same between MySQL shards that take 10,000s of queries per second and services that do few requests per minute. Works great.
username_0: That's interesting Simon as that's exactly the type of scenario I was worried would require very different configurations when using absolute values (although at this point we're planning on only using this for various HTTP based services not the database connection itself). If you don't mind can I ask what your configuration looks like?
username_1: Absolutely. For MySQL shards we use this configuration:
```yaml
semian: &semian
tickets: <%= Resiliency.semian_ticket_count %>
timeout: <%= Resiliency.semian_ticket_timeout %>
success_threshold: 2
error_threshold: 3
error_timeout: 10
```
`Resiliency.semian_ticket_count` is defined as the number of Unicorn or job workers on the box divided by two. We generally run around 48 workers per box, that's around 3 per core.
`Resiliency.semian_ticket_timeout` is `2s` for job workers, and `0s` for web workers. The difference between a ticket count and a timeout is that timeouts will allow short bursts on a resource, whereas ticket counts is the absolute maximum concurrency. This is a problem for job workers with our workload, but not for web workers.
With this configuration we run over a dozen shards with 10,000+ workers without issues. A single shard outage never has an impact on the latency of the entire platform. MySQL, given our throughput to it, is _by_ far the biggest threat.
As for HTTP based services, we do very low throughput to them. This means that bulkheads are not really useful. Circuit breakers are quite useful though. For HTTP services our configuration looks like this:
```
SEMIAN_PARAMETERS = {
tickets: 1000,
success_threshold: 1,
error_threshold: 3,
error_timeout: 20
}.freeze
```
This effectively disabled bulkheads for them because we do low throughput to them (https://github.com/Shopify/semian/pull/89 allows just disabling them). It's fairly hard to reason about good ticket counts for them, and we don't see a scenario where it'd be a huge issue. It'll take a while for all the circuits to trigger, but given the throughput spending that capacity waiting for the timeouts is not really an issue as long as we converge towards failing fast.
So they're not actually exactly the same, but the circuit breaker thresholds don't matter too much in our experience for these low latency services. We're much more concerned about the high throughput resources (MySQL, basically) than anything else.
username_0: Thanks so much for sharing this! We will experiment with a variation of this for our candidate services. I appreciate your insight!
username_1: Absolutely! Let me know if you have other issues.
Another thing I'd love to see in core Semian is that HTTP can be circuit broken by default, intead of ad-hoc configuring every endpoint. The reason it's not is because if you talk to 10,000s of hosts like we do, you'd allocate way too many circuit breakers. The implementation needs to have a ring-buffer scoped by resource for that to not be a subtle memory leak. If you have similar issues, let me know. |
uraimo/SwiftyGPIO | 1084106996 | Title: Segmentation Fault
Question:
username_0: ### Board Type
RaspberryP4
### Operating System
raspIO Bullseye 64 bit
### Swift Version
5.5
### Description
When using `onRaising` to measure the rotational speed of a 5000 rpm fan, I'm getting segmentation faults |
FreshPorts/freshports | 842517091 | Title: Two FreshSource issues
Question:
username_0: (Is this repo good for FreshSource issues?)
Via <https://www.freshsource.org/>:
1. <http://docs.freebsd.org/mail/current/cvs-all.html> is obsolete, according to <https://lists.freebsd.org/mailman/listinfo>
2. <https://www.freshsource.org/news.php> does not appear to be RSS.
I would have ignored point (1) but then, I vaguely recalled (2) the RSS feed working two or three days ago. This might be a confused recollection of a feed elsewhere (!) but I'm reporting it, just in case.
Answers:
username_1: Yes it is, and FreshSource should have a link to here.
username_1: Fixed by adding adding different links, mentioning subversion, and git.
See https://dev.freshsource.org
username_1: This is what I get when I fetch it to a file: https://gist.github.com/username_1/c92807d0a1b9a1071a6cbdd25158dce4
Although the DTD is deprecated: https://validator.w3.org/feed/check.cgi?url=https%3A%2F%2Fwww.freshsource.org%2Fnews.php
username_1: For the MIME type:
`header("Content-Type: application/rss+xml; charset=UTF-8");`
Please try again. :)
username_0: I'm not getting it in Firefox.
Debatably not the best choice, since Firefox dropped support for RSS, but I typically use this extension:
* [Want My RSS](https://addons.mozilla.org/addon/want-my-rss/)
– and today I enabled this:
* [Smart RSS Reader](https://addons.mozilla.org/addon/smart-rss-reader/)
(Off the top of your head, can you think of any browser in ports that has integral support for RSS?)
username_1: Do you have any feeds which work in Firefox? If you do, tell me which ones and I'll compare to the FreshSource feed.
username_0: Thanks, try https://freebsdfoundation.org/feed/ (referred from <https://freebsdfoundation.org/>) seems to work with Want My RSS
username_1: @username_0 Please try https://dev.freshsource.org/news.php - I fear I was not clear about where the fix was located.
username_0: Ah, got it now:

* to the left, the one feed that's found at https://dev.freshports.org/
* to the right, https://dev.freshsource.org/news.php
username_1: Fixed in repo.
username_1: @username_0 how is https://www.freshsource.org/news.php now?
username_0: Sorry! I lost sight of this.

With that extension disabled:
 |
rundeck/rundeck | 418706201 | Title: LDAPS not working on Rundeck 3.0.16
Question:
username_0: **Describe the bug**
Secure LDAP connection (LDAPS) cannot be established following the how-to under:
https://docs.rundeck.com/docs/administration/security/authenticating-users.html#communicating-over-secure-ldap-ldaps
`Error message: PKIX path validation failed`
**My Rundeck detail**
* Rundeck version: 3.0.16
* install type: rpm,
* OS Name/version: rhel 7.5
* DB Type/version: postgres
**To Reproduce**
1. : Do the steps to configure LDAPS described under: https://docs.rundeck.com/docs/administration/security/authenticating-users.html#communicating-over-secure-ldap-ldaps
**Expected behavior**
LDAPS connection can be established based on the stores described in /etc/rundeck/ssl/ssl.properties
**Additional context**
I had to manually pass the keystore and truststore as JVM Options:
RDECK_JVM_OPTS="-Djavax.net.ssl.keyStore=${RDECK_CONFIG}/ssl/keystore -Djavax.net.ssl.keyStorePassword=${CERT_STORE_PASS} -Djavax.net.ssl.trustStore=${RDECK_CONFIG}/ssl/truststore -Djavax.net.ssl.trustStorePassword=${CERT_STORE_PASS}"
Setting RUNDECK_WITH_SSL=true is not sufficient, even though the stores are being used by Rundeck, because HTTPS is working!
Answers:
username_1: I also had problems setting up LDAPs, as Java's way of dealing keyStore is a bit hard. I recommend using a proxy like stunnel to setup your ldaps. |
edouardlicn/Ifxoss_ragemp | 362860465 | Title: Can not set Hospital Leader
Question:
username_0: When i try the command "/sethospitalleader" or "/sethospitalleader 1" then nothing happend the Database is not changed. And there is a error on the console.
`(node:5356) UnhandledPromiseRejectionWarning: Unhandled promise rejection (rejection id: 1): TypeError: s.getAdminLvl is not a function
(node:5356) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
(node:5356) UnhandledPromiseRejectionWarning: Unhandled promise rejection (rejection id: 2): TypeError: s.getAdminLvl is not a function`
Answers:
username_1: this is not originally create by me that i don't have the ability to fix that bug.may be you ask the original author in readme.md better.i can update here when he fix.
username_1: BTW,you make a barbershop but did not set the cord?
username_1: if you not add cord i will delete it.for it will not work and become bug.
Status: Issue closed
|
Unidata/thredds | 129265748 | Title: PointFeature.getLocation returns altitude in wrong units (Point Feature Types)
Question:
username_0: [Migrated from Jira](https://bugtracking.unidata.ucar.edu/browse/TDS-672) per Sean
This issue concerns the Point Feature Types. Read in the attached nc/ncml combination (below). The data should be CF compliant to my knowledge. John looked at it too. Call the `getLocation()` method. This will return the altitude in kilometers, but is should be in meters. According to the docs: "double getAltitude() altitude in meters; missing = NaN".
[28219.476.nc](https://files.zenhub.io/56a92d81e79663ef099dfceb)
[28219.476.ncml](https://files.zenhub.io/56a92d81e79663ef099dfcea)
Answers:
username_1: I can confirm this behavior on both 4.6 and 5.0. `PointFeature.getLocation()` returns an `EarthLocation`. `EarthLocation.getAltitude()` claims that it returns altitude in meters, but its implementation class–`EarthLocationImpl`–makes no attempt to convert values that aren't already in meters.
The altitude unit **is** read and made available via `DsgFeatureCollection.getAltUnits()`, but we're currently not doing a whole lot with it.
The lazy solution is to edit `EarthLocation.getAltitude()`'s Javadoc to indicate that the unit is not necessarily meters, and can be found by calling `DsgFeatureCollection.getAltUnits()`. A better solution is to store the units in each `EarthLocation`, but that's significantly more work as we'd have to change every piece of code where an instance is created.
username_2: I think we can force it to be in km using the udunits package. For example, we could first check if the unit is compatible with km, and if so, change it using the udunits package to get the correct scale factor and make it correct. I sort of get the feeling, especially in coordinate system classes, that we do not utilize the udunits package enough.
username_1: Yeah, but not all vertical coordinates are compatible with meters. Some are even dimensionless. It's just a bad idea to guarantee meters from that method.
username_2: Ok, here is what we came up with:
* change docs to reflect reality
* make sure we give users an idea of where to look for the units
* All of this needs a major API review
* Try to address this in 5.0
* Look at OGC geotoolkit thingy and investigate for THREDDS 6.
username_1: Accomplished the first two bullets in #412. I only intend to address this in 5.0.0. |
gaffneyc/heroku-buildpack-jemalloc | 989412066 | Title: Jemalloc no longer working
Question:
username_0: true
```

Answers:
username_1: @username_0 There isn't a good way to see if jemalloc is enabled through RbConfig or inside a process. The best way at the moment (which needs to be added to the readme) is this: https://github.com/username_1/heroku-buildpack-jemalloc/issues/5#issuecomment-499932026
Give that a try and let me know if it's not actually enabled. Jemalloc's allocation algorithm does a better job of allocating objects in already claimed free memory rather than requesting more from the system so it's possible there is a change in your app which is causing it to request and retain a lot more memory
Status: Issue closed
username_0: You are correct. It seems to be enabled using this way of testing. I will investigate alternative routes. Thank you.
P.S. Could you please document this: `MALLOC_CONF=stats_print:true ruby -e "exit"` |
docker/for-mac | 185979103 | Title: Building OpenNetworkLinux yields unresponsive container
Question:
username_0: ### Expected behavior
Container completes build.
### Actual behavior
Container becomes unresponsive, i.e.
* `docker ps` shows it is `Up`
* `docker exec -it <cmd> <container_id>` hangs
* `docker stop <container_id>` hangs (unable to stop it)
### Information
```
Docker for Mac: version: 1.12.3-beta29.2 (902414d)
OS X: version 10.11.6 (build: 15G1108)
logs: /tmp/BC71B46C-FACD-4D89-8C0D-B26FE75FBF4E/20161028-102715.tar.gz
[OK] vmnetd
[OK] dns
[OK] driver.amd64-linux
[OK] virtualization VT-X
[OK] app
[OK] moby
[OK] system
[OK] moby-syslog
[OK] db
[OK] env
[OK] virtualization kern.hv_support
[OK] slirp
[OK] osxfs
[OK] moby-console
[OK] logs
[OK] docker-cli
[OK] menubar
[OK] disk
```
### Steps to reproduce the behavior
- https://opennetlinux.org/docs/build
```
git clone https://github.com/opencomputeproject/OpenNetworkLinux
cd OpenNetworkLinux
make docker
```
Answers:
username_1: Does Diagnose & Feedback from the :whale: menu work? Could you please run and upload the diagnostics and post your diagnostic ID back here?
Status: Issue closed
username_2: This issue has been inactive for more than 14 days while marked as `status/0-more-info-needed`. It is being closed due to abandonment. Please feel free to re-open with more information about the problem.
MORE_INFO_EXPIRY_TIMEOUT |
convox/rack | 292313709 | Title: ERROR: timeout starting process
Question:
username_0: CLI and web console returning this on all attempts to build or deploy.
Rack version: 20180126132639
Health checks all 200 and nothing unusual in rack logs
Status in AWS CloudFormation all looks fine, as per Troubleshooting Docs.
Any idea what might be causing this?
Answers:
username_1: We see this all the time. Was it resolved for you somehow?
username_2: Would generally mean that you build instance is not large enough and AWS is timing out trying to schedule the build container.
username_1: Thank you @username_2!
ℹ️ @tapajos |
cloudmesh-community/hid-sp18-602 | 382637207 | Title: swagger example is incomplete
Question:
username_0: Your Docker Glance example is not completed.
Two things are missing:
a) what is the abstraction to GLance (e.g. are there other clouds that have a similar service)\
b) the service that glance provides whith just a string attribute is certainly incomplete. However the focus is (a)
https://github.com/cloudmesh-community/hid-sp18-602/blob/master/swagger-docker/openstack/swagger.yaml<issue_closed>
Status: Issue closed |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.