repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
electron/electron
160472873
Title: can't use npm module `markdown' Question: username_0: * Electron version:1.2.2 * Operating system:windows10, 64bit I want to use use `markdown' library in my own projects, which is published on https://github.com/evilstreak/markdown-js. I have already installed `markdown' via npm, with it's newest version. When I run the electron app, this code doesn't work: ``` var gg = require('markdown').markdown; ``` Anyone can help me? Status: Issue closed Answers: username_1: [We're trying to control the amount of Issues that we have to triage and track](http://blog.atom.io/2016/04/19/managing-the-deluge-of-atom-issues.html). For general questions, [the Electron message board](https://discuss.atom.io/c/electron) or the [Electron Slack channel](http://atom-slack.herokuapp.com/) is where you'll probably be able to find answers faster. Hope that helps!
jlippold/tweakCompatible
446607303
Title: `SilentScreenshot` working on iOS 11.3.1 Question: username_0: ``` { "packageId": "com.smokin1337.silentscreenshot", "action": "working", "userInfo": { "arch32": false, "packageId": "com.smokin1337.silentscreenshot", "deviceId": "iPhone8,2", "url": "http://cydia.saurik.com/package/com.smokin1337.silentscreenshot/", "iOSVersion": "11.3.1", "packageVersionIndexed": false, "packageName": "SilentScreenshot", "category": "Tweaks", "repository": "Packix", "name": "SilentScreenshot", "installed": "1.2", "packageIndexed": true, "packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.", "id": "com.smokin1337.silentscreenshot", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.1.5", "shortDescription": "Silent screenshots with haptic feedback !", "latest": "1.2", "author": "smokin1337", "packageStatus": "Unknown" }, "base64": "<KEY> "chosenStatus": "working", "notes": "" } ```<issue_closed> Status: Issue closed
FillZpp/sys-info-rs
342603226
Title: What unit is DiskInfo represented in? Question: username_0: What unit (byte, kilobyte, etc.) is returned when accessing `sys_info::disk_info().unwrap().total` or `sys_info::disk_info().unwrap().total` I can't seem to figure it out. Status: Issue closed Answers: username_1: @username_0 Both `sys_info::disk_info().unwrap().total` and `sys_info::disk_info().unwrap().free` return kilobyte(KB).
uclapi/uclapi
528828803
Title: Timetable doesn't appear Question: username_0: Never a timetable shown for my account. Just a message on every day saying nothing scheduled on *day* **To Reproduce** Steps to reproduce the behaviour: 1. Go to 'timetable' **Please provide the following information:** - Device: iPhone6S - OS: iOS12.4<issue_closed> Status: Issue closed
ravibpatel/AutoUpdater.NET
687940401
Title: 更新到1.6.1.0又出现新问题了 Question: username_0: 提示找不到路径 ![Uploading image.png…]() Answers: username_1: Which .NET Framework you are using and does the update adds a new folder with file? username_0: .net Framework 4.8 Configs里面包含许多文件 Status: Issue closed username_1: Please try new version from [here](https://ci.appveyor.com/api/buildjobs/wm5qy6fay5gu5etq/artifacts/AutoUpdater.NET-1.6.2.zip). Use "AutoUpdater.NET.dll" from "net45" folder. Let me know how it goes. username_0: 用了这个1.6.2版本,以上的未能找到路径问题解决了。感谢大佬 username_1: I updated the NuGet package with the fix. It will be available soon.
php-actions/phpunit
565943745
Title: Add testsuite option to action Question: username_0: Hi! It will be great to have the --testsuite parameter as an optional parameter to run a specific testsuite Answers: username_1: Great. I'll look into this tomorrow. Thanks for your feedback. username_1: Hi @username_0 , please can you check out the PR. In case I don't hear anything back, this will be merged in within 24 hours as it is such a simple change. https://github.com/php-actions/phpunit/pull/3 username_0: Thank you Greg, I just comment something in the pr Status: Issue closed
kevoree-modeling/mwDB
203644765
Title: OffHeapTimeTreeChunk Load error Question: username_0: ``` java.lang.RuntimeException: get: bad address 1*8 in null at org.mwg.memory.offheap.primary.OffHeapLongArray.get(OffHeapLongArray.java:77) at org.mwg.memory.offheap.OffHeapTimeTreeChunk.internal_insert(OffHeapTimeTreeChunk.java:776) at org.mwg.memory.offheap.OffHeapTimeTreeChunk.internal_load(OffHeapTimeTreeChunk.java:283) at org.mwg.memory.offheap.OffHeapTimeTreeChunk.load(OffHeapTimeTreeChunk.java:228) at org.mwg.WSServer$7.on(WSServer.java:272) at org.mwg.WSServer$7.on(WSServer.java:263) at org.mwg.memory.offheap.OffHeapChunkSpace$1.on(OffHeapChunkSpace.java:170) at org.mwg.memory.offheap.OffHeapChunkSpace$1.on(OffHeapChunkSpace.java:160) at org.mwg.LevelDBStorage.get(LevelDBStorage.java:63) at org.mwg.memory.offheap.OffHeapChunkSpace.getOrLoadAndMark(OffHeapChunkSpace.java:160) at org.mwg.WSServer.process_put(WSServer.java:263) ``` with the following buffer ``` #rQO4I#A;////////+;A;A#C|Y|A|BA:Y2l0aWVz:CiJIlv:bmFtZQ:ZvUW:aW5kZXg:L9lGk:ZGlzdGFuY2U:iYxfq:Ym91bmRfbWlu:CJHb/f:Ym91bmRfbWF4:CJHcG7:cG9zaXRpb25z:DLgRjU:bnVtYmVy:B7Tklv:cG9zaXRpb24:BZJTJS:<KEY>:NWSLi#G;////////9;////////+;C#U#A;A;////////9;EAAAAAC#G|E|ZvUW|QW1pZW5z|a|DLgRjU|C:EAAAAAG|c|CdKEt+|C:Xp4046vu1:EAAAAAI#C;A;////////+;EAAAAAC#C:////////9#C;A;////////9;EAAAAAC#C:////////9#E;A;A;EAAAAAC#C:A:////////9#A;A;////////9;////////8#C|U|L9lGk|C:CiJIlv:EAAAAAE#C;A;////////+;////////8#C:////////9#C;A;////////9;////////8#C:////////9#E;A;A;////////8#C:A:////////9#A;A;////////9;EAAAAAE#C|c|L9lGk|C:xOmUkPQp8:EAAAAAC#C;A;////////+;EAAAAAE#C:////////9#C;A;////////9;EAAAAAE#C:////////9#E;A;A;EAAAAAE#DcBchh|C:A:////////9#A;A;////////9;EAAAAAG#K|I|iYxfq|C|M|CJHb/f|E:wFGg:wGGg|M|CJHcG7|E:QFGg:QGGg|i|Cbi4/v|C$Q%G%A%C%G%E%A%G%C%A%M%G%E:wFGg:wGGg%M%I%E:QFGg:QGGg%e%K%m:QAA:QCA:P/A:QEI8vW9ZxFd:QACXI9DaL42:f/I:f/I:f/I:f/I:f/I:f/I:f/I:f/I:f/I:f/I:f/I:f/I:f/I:f/I%g%M%W:C:Q:C:EAAAAAI:D:D:D:D:D:D:D%G%A%C|I|DVH0bm|A#C;A;////////+;EAAAAAG#C:////////9#C;A;////////9;EAAAAAG#C:////////9#E;A;A;EAAAAAG#Dt/7MZ|C:A:////////9#A;A;////////9;EAAAAAI#O|E|B7Tklv|MDAwMw|a|BZJTJS|C:EAAAAAK|a|DbSr0L|C:EAAAAAM|a|CBQITF|C:EAAAAAO|a|BX0pd4|C:EAAAAAQ|a|BjLLX3|C:EAAAAAS|a|BqZGAd|C:EAAAAAU#C;A;////////+;EAAAAAI#C:////////9#C;A;////////9;EAAAAAI#E:////////9:q76Juyg#E;A;A;EAAAAAI#C:A:////////9#A;A;////////9;EAAAAAK#A#C;A;////////+;EAAAAAK#C:////////9#C;A;////////9;EAAAAAK#E:////////9:q76Juyg#E;A;A;EAAAAAK#C:A:////////9#A;A;////////9;EAAAAAM#A#C;A;////////+;EAAAAAM#C:////////9#C;A;////////9;EAAAAAM#E:////////9:q76Juyg#E;A;A;EAAAAAM#C:A:////////9#A;A;////////9;EAAAAAO#E|I|DWDUab|FQ|G|CRuC9+|BIGQgA#C;A;////////+;EAAAAAO#BIGQgA|C:////////9#C;A;////////9;EAAAAAO#E:////////9:q7mezgA#E;A;A;EAAAAAO#DyNara|C:A:////////9#A;A;////////9;EAAAAAQ#A#C;A;////////+;EAAAAAQ#C:////////9#C;A;////////9;EAAAAAQ#W:////////9:q76Juyg:q7/CVoA:q7/dx+Q:q8APAXQ:q8BNF/w:q8BNdbw:q8BSEWw:q8CrNbw:q8DTxIw:q8D3lng#E;A;A;EAAAAAQ#C:A:////////9#A;A;////////9;EAAAAAS#A#C;A;////////+;EAAAAAS#C:////////9#C;A;////////9;EAAAAAS#W:////////9:q76Juyg:q7/CVoA:q7/dx+Q:q8APAXQ:q8BNF/w:q8BNdbw:q8BSEWw:q8Cq3dg:q8DTxIw:q8D3lng#E;A;A;EAAAAAS#C:A:////////9#A;A;////////9;EAAAAAU#A#C;A;////////+;EAAAAAU#C:////////9#C;A;////////9;EAAAAAU#E:////////9:q76Juyg#E;A;A;EAAAAAU#C:A:////////9#A;A;q76Juyg;EAAAAAK#E|K|0M+|QEI8vW9ZxFd|K|0ZK|QACXI9DaL42#A;A;q76Juyg;EAAAAAQ#C|I|NWSLi|G#A;A;q76Juyg;EAAAAAS#C|I|NWSLi|i#A;A;q76Juyg;EAAAAAU#C|E|NWSLi|T1BFTg#A;A;q76Juyg;EAAAAAM#C|I|NWSLi|o ``` Answers: username_1: I struggle to reproduce this bug. There is some # inside this buffer, meaning that this is not only a TimeTree buffer can you reproduce this bug from your side and send me only time tree related Buffer Status: Issue closed username_1: fixed in GreyCat, many thanks for reporting
awslabs/dynamodb-document-js-sdk
48826217
Title: publish to bower.io Question: username_0: Please consider providing this lib there via http://bower.io/docs/creating-packages/#register I see its `aws-sdk-js` dependency is already there: http://bower.io/search/?q=aws-sdk Answers: username_1: Would be great to have this Status: Issue closed username_2: The DynamoDB [DocumentClient is already in the AWS SDK for JS](https://github.com/aws/aws-sdk-js/blob/master/lib/dynamodb/document_client.js#L321). Closing.
googleanalytics/ga-dev-tools
387877897
Title: Server-Side Authentication demo is not entirely clear about requirements Question: username_0: Hi! **Introduction** I'm starting to work with Python, and I'm working on a little project that's already using OAuth and getting data from Google Analytics, but I was asked to implement a solution without signing in, and that's where I thought the server side authentication demo could help me. **Problem** During step 3, you have to create a .py file to get the access token (I already tried the demo for using a service account so it wasn't painful). The thing is, you get the token and then the demo isn't explicit about what to do or where to "use" the token. Do you have to print it (which I did), and then paste it in a separate HTML file that you have to create along with the scripts (just like the other demos)? Do you have to store it somewhere, tying yourself to using a Python-dependant system (e.g. Django or another templating engine) to complete the tutorial? **Possible solution** If you only need this Python script to create the access token, it should be clearly stated that you have to create that script, and then you can work just like in the other demos, with a HTML file. If you must have a server-side infrastructure like the one that powers this site, it should be indicated too. Step 4 should be the proper place to do it. I hope this suggestion can be helpful. Answers: username_1: Thank you for the suggestion! I've run into a few of these issues myself, so your feedback is very helpful. I'll add this to the list of things to work on when I do a work sprint on the Developer tools. Status: Issue closed
facebookresearch/ParlAI
932224220
Title: Issue in running eval script for wizards on wikipedia pre-trained agent Question: username_0: The script stops running after running ~50%. Ran the following script after installing parlai module using pip parlai eval_model -bs 64 -t wizard_of_wikipedia:generator:topic_split -mf models:wizard_of_wikipedia/end2end_generator/model The colab pro environment has 27gb RAM and is GPU enabled. The following is the output given before abruptly stopping tcmalloc: large alloc 3227770880 bytes == 0x55d7efaf0000 @ 0x7fe264f85001 0x7fe206e8754f 0x7fe206ed7b58 0x7fe206edae83 0x7fe206edb07b 0x7fe206f7c761 0x55d2c8e4ecc0 0x55d2c8e4ea50 0x55d2c8ec2be0 0x55d2c8ebd7ad 0x55d2c8e503ea 0x55d2c8ebe60e 0x55d2c8ebd7ad 0x55d2c8e503ea 0x55d2c8ebe60e 0x55d2c8e5030a 0x55d2c8ebe60e 0x55d2c8e5030a 0x55d2c8ebe60e 0x55d2c8e5030a 0x55d2c8ebe3b5 0x55d2c8e5030a 0x55d2c8ebe3b5 0x55d2c8e5030a 0x55d2c8ebe60e 0x55d2c8e5030a 0x55d2c8ec27f0 0x55d2c8e50561 0x55d2c8ebe3b5 0x55d2c8e5030a 0x55d2c8ebe3b5 ^C Answers: username_1: Looks like you ran out of RAM. Try lowering the batchsize.
VadimDez/ng2-pdf-viewer
221949250
Title: Invoke URL retrieval manually Question: username_0: My pdfSrc variable changes asynchronously and the document isn't even tried to be fetch after the url is changed. How can I invoke the retrieval of the URL manually? Answers: username_1: Dynamic change of src should work. See here https://vadimdez.github.io/ng2-pdf-viewer/ Status: Issue closed username_1: Closing.
openml/OpenML
59505215
Title: Sense of row ID attributes Question: username_0: Hi, in [data set 377](http://openml.liacs.nl/d/377) (synthetic_control), there's a row ID attribute called "index" going from 1 to 600. [Task 3512](http://openml.liacs.nl/t/3512) uses this data set, but the corresponding data splits are 0-based, i.e., do not take the row ID attribute into account. Now I don't really get the sense of the attribute then, because I would expect the data splits to range from 1 to 600 as well. Answers: username_1: Hi dominik, this attribute was in the dataset uploaded to OpenML. However, OpenML simply disregards it and uses its own row numbers. This is exactly to avoid issues with user-defined row id's. They can be 0-based, 1-based, or whatever other identifier. Sometimes they are not even numbers. We keep the features because quite possibly they have some meaning in their domain, but we flag them because they should not be modelled. Does that help? Cheers, Joaquin username_0: All right. That was helpful, thank you! Status: Issue closed
kununu/nukleus
194308397
Title: Automate releases with travis Question: username_0: It should already be setup in the `travis.yml` file but it does not seem to be working. Answers: username_1: I guess we could avoid this by just using the same manual release publishing workflow that `create-react-app` uses (from`lerna`). It's quite nice.
zenhob/hcl
152777362
Title: Error reading man page Question: username_0: I just installed this gem successfully. I ran `gem man hcl` and get this response ``` fgets: Undefined error: 0 Error reading man page /Library/Ruby/Gems/2.0.0/gems/hcl-0.4.15/man/hcl.1 No manual entry for /Library/Ruby/Gems/2.0.0/gems/hcl-0.4.15/man/hcl.1 ``` Am I doing something wrong? Thank you. Answers: username_1: :joy: I was using `write.tap do ... end` instead of `write do ... end` so the file wasn't actually being closed before the end of the job. Good catch, thank you! This will be fixed in the next release. Status: Issue closed
trabucayre/openFPGALoader
1183542651
Title: [Request] Add support for ECP5 key programming Question: username_0: If there is a simple way to run STAPL file with openFPGALoader this way might work. Nevertheless we made experience that programming keys does work with simple SVF files as well. That´s why I think that enabling svf support for lattice devices would do the job. It seems to me like openFPGALoader already supports this for Intel FPGAs, is that right? Furthermore just some quick information about key programming on Lattice FPGA: There is no bitstream file o similar thing. It does exist only a xcf file (file format for Lattice Diamond Programmer) which contains the key and the operation string. But one can convert this into a svf file.
aws/aws-sdk-cpp
349547650
Title: TransferManager uploads failing to S3 Question: username_0: Windows 10, Visual Studio 2017 Notable lines from my log file: host: max-engage-dm.s3.amazonaws.com user-agent: aws-sdk-cpp/1.4.33 Windows/10.0.16299.15 AMD64 x-amz-content-sha256: UNSIGNED-PAYLOAD x-amz-date: 20180810T141259Z [ERROR] 2018-08-10 10:12:59 WinHttpSyncHttpClient [25840] Failed to add HTTP request headers with error code: 12150 (Header not found). See here: [https://support.microsoft.com/en-us/help/193625/info-wininet-error-codes-12001-through-12156] I was able to work around this error by making a modification to WinHttpSyncHttpClient.cpp, line 164. I removed the flag WINHTTP_ADDREQ_FLAG_REPLACE from WinHttpAddRequestHeaders. [dmserver_aws_2018-08-10-10.log](https://github.com/aws/aws-sdk-cpp/files/2278342/dmserver_aws_2018-08-10-10.log) Answers: username_1: Hi @username_0, Sorry for the late response, it seems from your comment that you got it to work though , from my side, I was unable to reproduce the error, but I can see that the flag is still there, so I don't think it was necessarily related. I'd need a code sample to figure out the root cause. Status: Issue closed
MusikAnimal/pageviews
152187919
Title: Date bug on zhwiki Question: username_0: Hi, this problem has existed since 2016/04/01. The date would show wrong value (such as 3/0) when using this tool on zhwiki. (For example, [this page](https://tools.wmflabs.org/pageviews/#project=zh.wikipedia.org&platform=all-access&agent=user&range=latest-90&pages=荒勝文策)). And the date also shows wrong if you choose specific date (you can try yourself at "日期範圍"). Can someone fix it? Thanks! Answers: username_0: The screenshot is like this: http://ppt.cc/yIuT1 username_1: @username_0 This should be fixed now with https://github.com/username_1/pageviews/commit/a867da76ef4410c89abe31a040678df2830b9a85, please confirm! Apologies for the bug username_0: Thanks! It works well now. Status: Issue closed username_1: Excellent. Thank you for the report!
tsolucio/corebos
355462019
Title: In some cases the click event doesn't register Question: username_0: https://github.com/tsolucio/corebos/blob/35370c5bb1527e3063986cf118adc0ff322448a2/include/js/Inventory.js#L1652 I have one user (this maybe a fluke) where the click event does not register so he can't select a result using his mouse. My mouse, operating via Teamviewer on the same PC however DID register. I'm wondering if this is a timing issue where the blur event outruns the click event. I will investigate this better en further but if anyone has the same issue, please reply here and leave as much as you can about the used devices, browsers and other info since this is **very** hard to debug. Answers: username_0: - No, none. That's we weird thing. I even put a console.log in the click function but that didn't trigger - Yes this is system-wide, but all kind of results. The click event is bound to the li of the rsultbox so no character parsing is done there. At least that should register - Firefox (latest), also tried it in incognito I should also mention that for now it seems the behaviour is gone. I'll keep an eye out for a couple of days and close the issue if it doesn't come back. username_0: Haven't had any complaints anymore so closing this. Status: Issue closed
googlemaps/android-maps-utils
119365275
Title: map.clear() & cManager.clearItems() are incompatible Question: username_0: ``` map.addPolygon(...); //add poly 1, poly 2, poly 3 cManager.additems(...); // add item 1, item 2, item 3 cManager.cluster(); // the 3 polygons are shown // the 3 markers are shown map.clear(); //clear polygons cManager.clearItems(); map.addPolygon(...); //re-add poly 1, add poly 4 cManager.additems(...); // re-add item 1, add item 4 cManager.cluster(); // poly 1, poly 4 shown //item 1 NOT SHOWN!, item 4 shown map.clear(); cManager.clearItems(); map.addPolygon(...); //re-add poly 1, re-add poly 4 cManager.addItems(...); // re-add item 1 and re-add item 4 cManager.cluster(); // poly 1, poly 4 shown //NONE items are shown ``` I can work around this by iterating through all the polygons and removing them 1 by 1, so i don't need to use map.clear(), but i believe it would be nice to have it fixed.
arrayfire/arrayfire
83234968
Title: add gitter Question: username_0: like... https://github.com/Yonaba/30log e.g. : https://gitter.im/Yonaba/30log?utm_source=badge&utm_medium=badge&utm_campaign=pr-badge&utm_content=badge Answers: username_1: It is done: https://gitter.im/arrayfire/arrayfire The badge is going to be pulled into master from hotfixes branch. Status: Issue closed
Hammerspoon/hammerspoon
395812560
Title: noises feature stopped working in mojave? Question: username_0: The noises feature doesn't work for me anymore in Mojave. This is the gist of what I'm doing in my code, it used to work in 10.11 but I upgraded to 10.14 and the callbacks no longer fire. local noiseCallback = { [1] = function() hs.alert.show("start!") end, [2] = function() hs.alert.show("finish!") end, [3] = function() hs.alert.show("you did it!") end, } noisesListener = hs.noises.new(function(noiseType) noiseCallback[noiseType]() end) local mod = {} function mod.start() noisesListener:start() end function mod.stop() noisesListener:stop() end hs.hotkey.bind(cah, "y", function() hs.alert.show("listening") noisesListener:start() end) Thanks Answers: username_1: Maybe this is something @username_2 can help with? username_2: Darn. I haven't upgraded to Mojave yet but I will soon and then I might be able to reproduce this and possibly fix. username_0: When I first tried this feature, I thought it wasn't very useful, I could only produce the 'sssss' sound, so how useful could it be if I could only trigger one thing with it? But then I realized since there's a start and end event, I could track the duration and do something useful with it. So I started building a morse code thingy where longer hissing meant a different event and you could associate a hissing pattern to an action. So I built most of the initial logic for my idea and then sadly upgraded to mojave so I never finished the actual thing I imagined. Recently I got that itch again, so I [gave it a UI](https://gfycat.com/imaginaryvibrantamericancrow) that I'm triggering with a modifier key. I'm patient, but I really want to try hissing at this UI. I've been dreaming about this for a year. username_3: Huh, I used to use that module some a couple of machines ago, but hadn't really used it recently... I can confirm that it no longer seems to work in Mojave. Taking a quick look at the Security and Privacy settings panel, I'm going to guess that Hammerspoon needs to be given access to the microphone, but I don't see a way to add it manually, so we'd need to find a way to trigger the request... @username_4 or @username_1, any thoughts? @username_2 did you ever get a chance to try it out in Mojave? username_1: I've never actually tried `hs.noises`, but... In theory, macOS should automatically prompt the user for access if Hammerspoon requires microphone access. It's possible that if you've updated your machine from High Sierra to Mojave, that something could have gone funky with permissions? You could try resetting your microphone permissions by triggering `tccutil reset Microphone` from Terminal. If you're running Hammerspoon from Xcode, I believe you must have the `NSMicrophoneUsageDescription` key in the `Info.plist` for the notification to actually trigger. I've added a pull request here: #2172 It might also be worth adding a function for [`requestAccessForMediaType:completionHandler:`](https://developer.apple.com/documentation/avfoundation/avcapturedevice/1624584-requestaccessformediatype?language=objc) to `hs.sound`? username_1: print(string.format("Result: %s", result)) end):start() 2019-09-08 20:32:03: -- Loading extension: noises 2019-09-08 20:33:04: Result: 1 2019-09-08 20:33:04: Result: 2 2019-09-08 20:33:39: Result: 1 2019-09-08 20:33:40: Result: 2 2019-09-08 20:33:40: Result: 1 2019-09-08 20:33:40: Result: 2 2019-09-08 20:33:41: Result: 1 2019-09-08 20:33:41: Result: 2 ``` username_2: You can also check out https://talonvoice.com/ which is a scriptable input system which I worked with to integrate my noise recognizers, and then the author of Talon optimized them to use Accelerate.framework SIMD functions so it uses even less CPU. username_0: @username_2 - I am actually very interested in the idea of voice coding. I recently remembered that [2013 pycon video](https://www.youtube.com/watch?v=8SkdfdXWYaI) and I just did a couple of days research about what's available these days. I really liked the idea of connecting it to HS and doing all the automation in lua, but it seems like all these systems have their own scripting engines. talonvoice looks cool, I love the idea of combining the hisses with regular voice commands. I'll check it out. I wonder how usable it is without dragon. Thanks username_0: I tried talonvoice, it's an amazing tool. I managed to send the hiss events back to hammerspoon from talonvoice. It's a bit awkward because it communicates via `hammerspoon://` urls which feels a bit slow, but it works. username_2: I recommend trying to do everything within Talon, it's very powerful and all the hot reloading and things will work better and quicker that way. Talon can do basically everything Hammerspoon can, although it may not be documented and you have to look at examples or ask in the Talon Slack. username_0: I am too invested in HS, it's not even funny how reliant I am on it. I can't imagine going back and rewriting everything in python. Most of the things I need are not time sensitive, so I can accept some slowness due to talon>HS communication. I wonder if an applescript API in HS would have been better than `hammerspoon://` urls for communication. Status: Issue closed username_0: can confirm that 0.9.76 fixes the problem, thanks!
gopherjs/gopherjs.github.io
252948133
Title: playground: Update for Go 1.9. Question: username_0: This should be done with GopherJS version 1.9-1. Answers: username_0: I'm investigating this right now. Everything seems to work okay, except that something causes `unsafe` package to get loaded, and that fails, because it's not one of the installed packages: https://github.com/gopherjs/gopherjs.github.io/blob/b7ffcc8bb375a70cc85b414ec55636f102fe68c8/playground/update.sh#L135-L138 I tried adding it to that list, and that fixes the problem: ```diff diff --git a/playground/update.sh b/playground/update.sh index 5d62c70..578a5fa 100755 --- a/playground/update.sh +++ b/playground/update.sh @@ -134,7 +135,8 @@ gopherjs install -m \ time \ unicode \ unicode/utf16 \ - unicode/utf8 + unicode/utf8 \ + unsafe cp -a "$GOROOT"/pkg/*_js_min/* pkg/ cp -a "$GOROOT"/pkg/*_amd64_js_min/* pkg/ ``` However, I want to look around first to see if there's a better fix, and whether this is indicative of a problem somewhere in GopherJS 1.9-1. username_0: The "unsafe" package is special-cased here: https://github.com/gopherjs/gopherjs/blob/95deb33d587c9f6e24b494ea9bdf9648c48f9a60/compiler/package.go#L105-L107 I want to see what changed related to that causing it to be included as part of a build, where previously it wasn't. username_0: This is the generated code for the `unsafe` package: ```JavaScript $packages["unsafe"] = (function() { var $pkg = {}, $init; $init = function() { $pkg.$init = function() {}; /* */ var $f, $c = false, $s = 0, $r; if (this !== undefined && this.$blk !== undefined) { $f = this; $c = true; $s = $f.$s; $r = $f.$r; } s: while (true) { switch ($s) { case 0: /* */ } return; } if ($f === undefined) { $f = { $blk: $init }; } $f.$s = $s; $f.$r = $r; return $f; }; $pkg.$init = $init; return $pkg; })(); ``` It's basically empty. It contains an empty `init` function, and that's the only thing that gets called by other packages. username_0: Figured it out. It's due to an API change in `go/types` package. [`"go/types".Package.Imports`](https://godoc.org/go/types#Package.Imports) used to exclude package `unsafe` in Go 1.8, but no longer does in 1.9. Compare https://gotools.org/go/types?rev=go1.8.3#package.go-L54-L55 vs https://gotools.org/go/types?rev=go1.9#package.go-L50-L51. Status: Issue closed username_0: This should be done with GopherJS version 1.9-1. username_0: I might need to make additional changes, so I've undone the deploy of 8bcc3f3a576ba326bbd26a1d27f5de696b8cf9dd for now. Status: Issue closed
EoRImaging/FHD
274387851
Title: Discontinuities in frequency with diffuse HERA simulations Question: username_0: There are vertical discontinuities at (apparently) even spacing in waterfall plots from visibility simulations of diffuse emission, using using HERA beams and the GSM. For all of these nfreq_avg = 1024 and there are 1024 channels. Snapshots are 11 time samples and the total simulation covers 5 minutes. So far I've determined: - The effect depends on baseline length, but less so on direction. - Shorter baselines have wider spacing between discontinuities - They seem to be there with beam\_model\_version=2 (<NAME>), but are much more apparent with beam\_model\_version=3 (<NAME>, on the _hera\_beam\_model\_three_ branch). I haven't fully tested the new beam model yet. The plots below are generated by taking 30min of simulated data, selecting the given baselines, averaging in time, taking the first derivative, normalizing each by its maximum, then shifting the curves up so they don't overlap. ![bm2_14m](https://user-images.githubusercontent.com/12849618/32872664-b4cae47e-ca55-11e7-9cbf-cc0b29613693.png) 14m baselines with beam\_model\_version=2, with different orientations. ![bm2_29m-ish](https://user-images.githubusercontent.com/12849618/32872586-7ec07146-ca55-11e7-825d-8ed1a094f208.png) Approx. 29.2m baselines with beam\_model\_version=2 ![bm2_ew-14m](https://user-images.githubusercontent.com/12849618/32872587-7ed7da8e-ca55-11e7-9875-2faa6ae1060a.png) 14m east-west baselines with beam\_model\_version=2 ![bm2_waterfall](https://user-images.githubusercontent.com/12849618/32872588-7eea8cb0-ca55-11e7-9b42-cf03253c08f0.png) Waterfall plot with beam\_model\_version=2 ![bm2_waterfall_uglier](https://user-images.githubusercontent.com/12849618/32873186-bda971d4-ca58-11e7-8f55-150965c57a81.png) Bonus waterfall plot with extra jumpiness. beam_model_version=2 Figures below are all with beam\_model\_version=3. ![bm3_14m](https://user-images.githubusercontent.com/12849618/32872589-7f035f60-ca55-11e7-82b9-d1471378a067.png) ![bm3_29m-ish](https://user-images.githubusercontent.com/12849618/32872590-7f15c858-ca55-11e7-8ec7-406130bcca38.png) ![bm3_ew-14m](https://user-images.githubusercontent.com/12849618/32872591-7f268e5e-ca55-11e7-8ef4-4f93af9a0fd4.png) ![bm3_waterfall](https://user-images.githubusercontent.com/12849618/32872592-7f3e6114-ca55-11e7-9513-69c31706dccf.png) Each snapshot is only 11 time samples, and there are clear discontinuities between them. Answers: username_0: An update --- These stripes do not appear when the instrument is set to MWA, so it seems to be caused at least in part by the beam model. username_0: I've successfully run <NAME>'s code for plotting the high-resolution psf, and found that there is a sharp cutoff visible on the edges. This is especially obvious when I set a lower psf_resolution. psf_resolution = 10: ![full_kernel_psfres-10](https://user-images.githubusercontent.com/12849618/32984252-1c828c4e-cc71-11e7-9a81-54e5171131de.png) psf_resolution=100: ![full_kernel_psfres-100](https://user-images.githubusercontent.com/12849618/32984253-1c8ebc9e-cc71-11e7-9aba-70117c689b6e.png) Both of these are with beam_model_version=3 with HERA, on the beam_model_version_three branch. I haven't seen any sharp cutoffs with beam_model_version=2, so this might not be responsible for _all_ of the discontinuities we're seeing.
perry-mitchell/webdav-client
395966536
Title: Force trailing slash on PROPFIND requests Question: username_0: by its name without a trailing slash, the server MAY handle the request as if the trailing slash were present. In this case, it SHOULD return a Content-Location header in the response, pointing to the URL ending with the "/". For example, if a client invokes a method on http://example.com/blah (no trailing slash), the server may respond as if the operation were invoked on http://example.com/blah (trailing slash), and should return a Content-Location header with the value http://example.com/blah. Wherever a server produces a URL referring to a collection, the server SHOULD include the trailing slash. In general, clients SHOULD use the trailing slash form of collection names. If clients do not use the trailing slash form the client needs to be prepared to see a redirect response. Clients will find the DAV:resourcetype property more reliable than the URL to find out if a resource is a collection. Relates to #121 Status: Issue closed Answers: username_0: Released in 2.2.1.
cfig/Android_boot_image_editor
750793830
Title: vbmeta pack FAILED Question: username_0: * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org BUILD FAILED in 16s 20 actionable tasks: 3 executed, 17 up-to-date Answers: username_1: Hi, Are you using an intermediate version? ` "flags-pVg5ArA" ` error is only observed when I bump kotlin version. Please try latest origin/master, it should work. Thanks. username_0: Modify "flags-pVg5ArA" to "flags-WZ4Q5Ns" username_1: Can you attach your vbmeta.img here? username_0: [vbmeta.zip](https://github.com/username_1/Android_boot_image_editor/files/5606473/vbmeta.zip) username_1: @username_0 , Should be fixed by https://github.com/username_1/Android_boot_image_editor/commit/9e30b56015d46422be0de5e1d1a1c9d3e72398e8 KernelCmdlineDescriptor is not used in latest AOSP code by default. I will add this into integration test case. Can you fetch code and try again? username_0: [main] WARN username_1.packable.PackableLauncher - [vbmeta_a.img] will be handled by [VBMetaParser] [main] WARN username_1.packable.PackableLauncher - 'pack' sequence initialized [main] INFO avb.blob.AuxBlob - encodePubKey(): size = 1032, algorithm key size: 1032 [main] INFO avb.blob.AuxBlob - Using the same key as original vbmeta [main] INFO avb.blob.AuxBlob - no pubkey metadata in auxBlob [main] INFO avb.blob.AuxBlob - encodePubKey(): size = 1032, algorithm key size: 1032 [main] INFO username_1.Avb - pkmd size: 0, pkmd offset : 4072 [main] INFO username_1.Avb - raw vbmeta size 4928, padding size 3264, total blob size 8192 [main] INFO username_1.Avb - Writing padded vbmeta to file: vbmeta_a.img.signed [main] WARN username_1.packable.PackableLauncher - 'pack' sequence completed BUILD SUCCESSFUL in 19s 20 actionable tasks: 3 executed, 17 up-to-date` Status: Issue closed username_1: Good to hear.
r-hub/r-minimal
1103461343
Title: readxl not building anymore Question: username_0: Hello, tidyverse/readxl recently updated its github repository and now it does not compile anymore with your example: https://github.com/r-hub/r-minimal/blob/master/examples/tidyverse/Dockerfile Do you have any idea how to fix the problem? 😅 <details> <summary>Crashlog</summary> ``` #5 193.8 ✖ Failed to build readxl 1.3.1.9000 #5 194.6 #5 194.6 Error: <callr_remote_error: Failed to build source package 'readxl'> #5 194.6 in process 47 #5 194.6 --> #5 194.6 Failed to build source package 'readxl', stdout + stderr: #5 194.6 #5 194.6 OE> * installing *source* package ‘readxl’ ... #5 194.6 OE> staged installation is only possible with locking #5 194.6 OE> ** using non-staged installation #5 194.6 OE> ** libs #5 194.6 OE> g++ -std=gnu++11 -I"/usr/local/lib/R/include" -DNDEBUG -Iunix -I. -I'/usr/local/lib/R/library/cpp11/include' -I'/usr/local/lib/R/library/progress/include' -I/usr/local/include -fvisibility=hidden -fpic -D__MUSL__ -Wall -pedantic -c cpp11.cpp -o cpp11.o #5 194.6 OE> g++ -std=gnu++11 -I"/usr/local/lib/R/include" -DNDEBUG -Iunix -I. -I'/usr/local/lib/R/library/cpp11/include' -I'/usr/local/lib/R/library/progress/include' -I/usr/local/include -fvisibility=hidden -fpic -D__MUSL__ -Wall -pedantic -c XlsWorkBook.cpp -o XlsWorkBook.o #5 194.6 OE> g++ -std=gnu++11 -I"/usr/local/lib/R/include" -DNDEBUG -Iunix -I. -I'/usr/local/lib/R/library/cpp11/include' -I'/usr/local/lib/R/library/progress/include' -I/usr/local/include -fvisibility=hidden -fpic -D__MUSL__ -Wall -pedantic -c XlsWorkSheet.cpp -o XlsWorkSheet.o #5 194.6 OE> In file included from /usr/local/lib/R/library/progress/include/RProgress.h:6, #5 194.6 OE> from Spinner.h:4, #5 194.6 OE> from XlsWorkSheet.h:5, #5 194.6 OE> from XlsWorkSheet.cpp:1: #5 194.6 OE> /usr/include/unistd.h:51:25: error: 'uid_t' has not been declared #5 194.6 OE> 51 | int chown(const char *, uid_t, gid_t); #5 194.6 OE> | ^~~~~ #5 194.6 OE> /usr/include/unistd.h:51:32: error: 'gid_t' has not been declared #5 194.6 OE> 51 | int chown(const char *, uid_t, gid_t); #5 194.6 OE> | ^~~~~ #5 194.6 OE> /usr/include/unistd.h:52:17: error: 'uid_t' has not been declared #5 194.6 OE> 52 | int fchown(int, uid_t, gid_t); #5 194.6 OE> | ^~~~~ #5 194.6 OE> /usr/include/unistd.h:52:24: error: 'gid_t' has not been declared #5 194.6 OE> 52 | int fchown(int, uid_t, gid_t); #5 194.6 OE> | ^~~~~ #5 194.6 OE> /usr/include/unistd.h:53:26: error: 'uid_t' has not been declared #5 194.6 OE> 53 | int lchown(const char *, uid_t, gid_t); #5 194.6 OE> | ^~~~~ #5 194.6 OE> /usr/include/unistd.h:53:33: error: 'gid_t' has not been declared #5 194.6 OE> 53 | int lchown(const char *, uid_t, gid_t); #5 194.6 OE> | ^~~~~ #5 194.6 OE> /usr/include/unistd.h:54:33: error: 'uid_t' has not been declared #5 194.6 OE> 54 | int fchownat(int, const char *, uid_t, gid_t, int); #5 194.6 OE> | ^~~~~ #5 194.6 OE> /usr/include/unistd.h:54:40: error: 'gid_t' has not been declared #5 194.6 OE> 54 | int fchownat(int, const char *, uid_t, gid_t, int); #5 194.6 OE> | ^~~~~ #5 194.6 OE> /usr/include/unistd.h:108:1: error: 'uid_t' does not name a type; did you mean 'pid_t'? #5 194.6 OE> 108 | uid_t getuid(void); #5 194.6 OE> | ^~~~~ #5 194.6 OE> | pid_t #5 194.6 OE> /usr/include/unistd.h:109:1: error: 'uid_t' does not name a type; did you mean 'pid_t'? #5 194.6 OE> 109 | uid_t geteuid(void); #5 194.6 OE> | ^~~~~ #5 194.6 OE> | pid_t [Truncated] #5 194.6 14. get("pkg_install_do_plan", asNamespace("pak"))(...) #5 194.6 15. pkgdepends::install_package_plan(plan = plan, lib = lib, num_workers = num_ ... #5 194.6 16. base:::withCallingHandlers({ ... #5 194.6 17. pkgdepends:::handle_events(state, events) #5 194.6 18. pkgdepends:::handle_event(state, i) #5 194.6 19. pkgdepends:::stop_task(state, worker) #5 194.6 20. pkgdepends:::stop_task_build(state, worker) #5 194.6 21. base:::throw(new_pkg_build_error("Failed to build source package {pkg}", ... #5 194.6 22. base:::signalCondition(cond) #5 194.6 23. (function (e) ... #5 194.6 24. base:::stop(e) #5 194.6 25. (function (e) ... #5 194.6 #5 194.6 x Failed to build source package 'readxl' #5 194.6 #5 194.6 Execution halted ------ executor failed running [/bin/sh -c installr -d -t "R-dev file automake autoconf linux-headers libxml2-dev" -a "libxml2 icu-libs" `# needed for xml2 and stringr` tidyverse/readxl `# CRAN version (<=1.3.1) does not compile on alpine` tidyverse]: exit code: 1 ``` </details> Answers: username_0: To answer myself, a quick fix is to use the last working git commit hash: `tidyverse/readxl@895ccb494846249a4e771c8087a53e90b5a65283` username_1: This was supposedly fixed, but maybe it crept back: https://github.com/tidyverse/readxl/issues/562 username_2: After digging around I found a fork of readxl with a fix that worked for me. ``` RUN installr -d \ -t "R-dev file automake autoconf linux-headers libxml2-dev libxml2 icu-libs" \ struckma/readxl tidyverse ``` username_1: I fixed this in the progress package, so this should work: ``` installr -d -t r-lib/progress readxl ``` I'll update the README. username_1: Hmmm, unfortunately this breaks vroom.... username_1: OK, this fixes it: https://github.com/tidyverse/readxl/pull/687 Until it is merged you can use this branch: ``` installr username_1/readxl@fix/alpine-linux ``` username_1: I updated the README and the tidyverse example as well.
matiasa123/DailyProgrammerChallenge
196418817
Title: [2016-12-19] Challenge #296 [Easy] The Twelve Days of... Question: username_0: https://www.reddit.com/r/dailyprogrammer/comments/5j6ggm/20161219_challenge_296_easy_the_twelve_days_of/ #Description Print out the lyrics of [The Twelve Days of Christmas](http://www.41051.com/xmaslyrics/twelvedays.html) #Formal Inputs &amp; Outputs ##Input description No input this time ##Output description On the first day of Christmas my true love sent to me: 1 Partridge in a Pear Tree On the second day of Christmas my true love sent to me: 2 Turtle Doves and 1 Partridge in a Pear Tree On the third day of Christmas my true love sent to me: 3 French Hens 2 Turtle Doves and 1 Partridge in a Pear Tree On the fourth day of Christmas my true love sent to me: 4 Calling Birds 3 French Hens 2 Turtle Doves and 1 Partridge in a Pear Tree On the fifth day of Christmas my true love sent to me: 5 Golden Rings 4 Calling Birds 3 French Hens 2 Turtle Doves and 1 Partridge in a Pear Tree On the sixth day of Christmas my true love sent to me: 6 Geese a Laying 5 Golden Rings 4 Calling Birds 3 French Hens 2 Turtle Doves and 1 Partridge in a Pear Tree On the seventh day of Christmas my true love sent to me: 7 Swans a Swimming 6 Geese a Laying 5 Golden Rings 4 Calling Birds 3 French Hens 2 Turtle Doves and 1 Partridge in a Pear Tree [Truncated] French Hens Calling Birds Golden Rings Geese a Laying Swans a Swimming Maids a Milking Ladies Dancing Lords a Leaping Pipers Piping Drummers Drumming ##Output The song described as above #Finally Have a good challenge idea? Consider submitting it to /r/dailyprogrammer_ideas
ELIFE-ASU/Neet
219736588
Title: Control Kernel Identification Question: username_0: [ "Ste9", "Rum1", "Wee1", "Cdc25" ] Answers: username_0: @username_1 Can you please add a reference to the original CK paper to my above description? username_1: [Scientific reports 2013 Kim-1.pdf](https://github.com/ELIFE-ASU/Neet/files/901457/Scientific.reports.2013.Kim-1.pdf) username_2: I/we think that there are two main distinct types of control: "pinning" and "intervention". username_3: We talked about different definitions of control kernels that we could implement, and the need to formalize these definitions in some kind of outline/document. username_4: @username_2 I prefer the name "override" to "intervention". :) username_2: The concept of a control kernel may be related to Stu's ideas of "frozen cores".
Savickiokas/VenusPro0
1091501079
Title: Kvieskite draugus prisijungti prie Pi Network Question: username_0: Pi - nauja skaitmeninė valiuta, sukurta Stanfordo universiteto mokslų daktaro (PhDs). Turinti daugiau nei 25 milijonų narių visame pasaulyje. Norėdami gauti Pi, spustelkite šią nuorodą https://minepi.com/extrymas ir naudokite vartotojo vardą (extrymas) kaip pakvietimo kodą.
hamelsmu/Issue-Label-Bot-Examples
432302606
Title: ARM64 Native Compilation support Question: username_0: Dear Team, Problem Description ARM64 native compilation support Proposed Solution Steps followed to compile on arm64 platform: Compile gn and ninja tools and add repective paths in PATH variable. Install clang, lld, default-jdk packages. Run gclient sync --nohooks Modify DEPS file and build.gn as nacl and binutils are not supported in arm64 platform. gclient runhooks Execute gn command as mentioned in electron development docs and add target_cpu, clang path in args.gn. Execute ninja command to generate binary in out folder. Approaches: Raise PR for above mentioned changes in Chromium package. This will create merge dependency over Chromium package Create a patch and make a condition to apply that patch when package checks for ARM64. This will make Electron independent of Chromium based changes Please let me know, which approach should be best way to get the things done for Electron on ARM64 Regards, From: https://github.com/electron/electron/issues/17740 Answers: username_0: @nornagon from https://github.com/electron/electron/issues/17740 Status: Issue closed
elgatito/script.elementum.burst
452686144
Title: RARBG not **always** getting results Question: username_0: This is really weird but sometimes provider won't get results from RARBG, but sometimes it will. I tried 3 times for the same episode, the first two attempts failed but the third returned results. Here's the [log](http://paste.kodi.tv/fuweliwayu) You can see my three requests. We can see in the logs that the behaviour between the failed attempts and the one that succeeded isn't the same. ## Working request Burst sent only two requests : * one with `good+omens+s01e01` which returned results * the other with `good+omens+01x01%7CS01` which didn't return any result. ## Failing request This time Burst sent three requests : * one with `good+omens+s01e01` which returned results * another with `good+omens+01x01%7CS01` which didn't return any result. * finally one with `good+omens+s1` which didn't return any result My guess is that in the failing request, the one query that returned results didn't return them quickly enough maybe ? I'm here if you need more info. Of course the three requets were made without changing the configuration of Burst Answers: username_1: I have to search twice the same episode/movie to get a result (from RARBG), I have not posted so far because it is not a big problem. username_2: Looks like Rarbg is responding with 429 error (too much requests). I will add a retry option for that error in the next version of Burst. username_2: Can you try with the latest version? It problem is still there - please, send new log file. username_3: @username_2 could you check this out too: https://github.com/username_2/script.elementum.burst/issues/132 Status: Issue closed
invertase/react-native-firebase
545803254
Title: :fire: Unable to find a specification for `RNFBApp` depended upon by `RNFBFirestore` Question: username_0: <!--- Hello there you awesome person; Please note that the issue list of this repo is exclusively for bug reports; 1) For feature requests please visit our [Feature Request Board](https://boards.invertase.io/react-native-firebase). 2) For questions and support please use our Discord chat: https://discord.gg/C9aK28N or Stack Overflow: https://stackoverflow.com/questions/tagged/react-native-firebase 3) If this is a setup issue then please make sure you've correctly followed the setup guides, most setup issues such as 'duplicate dex files', 'default app has not been initialized' etc are all down to an incorrect setup as the guides haven't been correctly followed. --> <!-- NOTE: You can change any of the `[ ]` to `[x]` to mark an option(s) as selected --> <!-- PLEASE DO NOT REMOVE ANY SECTIONS FROM THIS ISSUE TEMPLATE --> <!-- Leave them as they are even if they're irrelevant to your issue --> ## Issue <!-- Please describe your issue here --^ and provide as much detail as you can. --> <!-- Include code snippets that show your usages of the library in the context of your project. --> <!-- Snippets that also show how and where the library is imported in JS are useful to debug issues relating to importing or methods not found issues --> When trying to run `pod install` we get this: ``` [!] Unable to find a specification for `RNFBApp` depended upon by `RNFBFirestore` You have either: * out-of-date source repos which you can update with `pod repo update` or with `pod install --repo-update`. * mistyped the name or version. * not added the source repo that hosts the Podspec to your Podfile. ``` Which is not an unusual error, I've seen multiple issues here on Github, but none of the presented solutions worked. Usually the feedback is that the project is using an outdated version of `@react-native-community/cli`, but it doesn't seem to be the case for our app. We don't have the project as a dependency but it is a requirement for `react-native`. You can see here: https://github.com/facebook/react-native/blob/v0.60.5/package.json#L84 So the version seems to be updated enough but we're still getting the same error. Also: ``` warn Package @react-native-firebase/app has been ignored because it contains invalid configuration. Reason: Unknown option dependency.platforms.ios.scriptPhases with value "[{"name":"[RNFB] Core Configuration","path":"./ios_config.sh","execution_position":"after_compile"}]" was found. This is either a typing error or a user mistake. Fixing it will remove this message. ``` What else should we be trying to fix this? How can I make sure the we're using the correct RN CLI version? There's no CLI version installed globally and locally, so it's using the dependency brought by `react-native`, which I assume would be enough. Thanks! --- ## Project Files <!-- Provide the contents of key project files which will help to debug --> <!-- For Example: --> <!-- - iOS: `Podfile` contents. --> <!-- - Android: `android/build.gradle` contents. --> <!-- - Android: `android/app/build.gradle` contents. --> <!-- - Android: `AndroidManifest.xml` contents. --> <!-- ADD THE CONTENTS OF THE FILES IN THE PROVIDED CODE BLOCKS BELOW --> [Truncated] - **`Firebase` module(s) you're using that has the issue:** - `e.g. Instance ID` - **Are you using `TypeScript`?** - `Y` </p> </details> <!-- Thanks for reading this far down ❤️ --> <!-- High quality, detailed issues are much easier to triage for maintainers --> <!-- For bonus points, if you put a 🔥 (:fire:) emojii at the start of the issue title we'll know --> <!-- that you took the time to fill this out correctly, or, at least read this far --> --- Think `react-native-firebase` is great? Please consider supporting all of the project maintainers and contributors by donating via our [Open Collective](https://opencollective.com/react-native-firebase/donate) where all contributors can submit expenses. [[Learn More]](https://invertase.io/oss/react-native-firebase/contributing/donations-expenses) - 👉 Check out [`React Native Firebase`](https://twitter.com/rnfirebase) and [`Invertase`](https://twitter.com/invertaseio) on Twitter for updates on the library. Answers: username_1: Not sure why, but this is the problem `warn Package @react-native-firebase/app has been ignored because it contains invalid configuration. Reason: Unknown option dependency.platforms.ios.scriptPhases with value "[{"name":"[RNFB] Core Configuration","path":"./ios_config.sh","execution_position":"after_compile"}]" was found. This is either a typing error or a user mistake. Fixing it will remove this message. ` And it probably still is some CLI version that is not up to date. I've got 3.0.4 resolved for it now in my yarn.lock. Quick low-probability mention, while RN0.59->RN0.60 was extraordinarily painful, 0.60->0.61 is trivial (esp with upgrade-helper) if you don't have a specific regression stopping you from moving to 0.61 I'd recommend doing that then you can have a clean upgrade slate and maybe that resolves it at the same time Status: Issue closed username_0: Upgrading React Native to the latest version did it. Thanks!
apache/incubator-ponymail
258838575
Title: Bug: use actual envelope header if present Question: username_0: The From_ line is currently generated by mbox.lua from the Return-Path and Received headers. This is not always accurate. The method msg.as_bytes() can include the original envelope header as used in the From_ line. This should only be done if the envelope is present, as otherwise a default is added which changes the source. Also the default is not as good as the one generated by mbox.lua. Since existing messages won't have the From_ line, mbox.lua needs to check if the From_ line needs to be added or not. This also affects From_ prefixing as the code currently assumes the buffer does not contain the From_ header<issue_closed> Status: Issue closed
gatling/gatling
187356380
Title: Upgrade Scala 2.12 Question: username_0: * [x] akka * [x] scala-swing * [x] scala-xml * [x] scala-java8-compat * [x] fastring * [x] scopt * [x] scala-logging * [ ] jsonpath * [ ] boopickle * [ ] redisclient * [x] quicklens * [x] scalatest * [x] scalacheck Status: Issue closed Answers: username_1: Is there a build using 2.12? username_2: Hi @username_0 , I was using gatling 3.0.0 and all was working fine until socket issues with ALB. I couldnt get proper document in migrating from 3.0.0 to Latest. Could you help me on this. --Alex username_0: @username_2 See the "Upgrading" section in our doc: https://gatling.io/docs/gatling/ Then, questions are better asked on our [community](https://gatling.io/community/).
vinz486/fingerprint-r503-mqtt
892678987
Title: Socket error on client <unknown> Question: username_0: Hi, I have used you project to create a fingerprint access for my home assistant. I haven't received any error on compiling the code, the esp is correctly connected to my wifi, but in the mqtt broker hassio component I see an endless errors list like "New connection from 192.168.*.* on port 1883. Socket error on client <unknown>, disconnecting." Obviously the esp ip is correct. Answers: username_1: Seems like an issue HA side: are you sure that MQTT username and password are ok? username_1: Are you able to "sniff" traffic using Wireshark or tcpdump? username_1: What string do you put as gate name? username_1: Hi, sorry but the issue must be reproducible: try using tcpdump, mqtt is not TLS and you can see what happens in clear text.
MSP-Greg/ruby_on_windows
267707272
Title: Appveyor - new builds... Question: username_0: @username_1 First of all, congratulations. When I got the email, I really hoped that you were also nominated. I wondered that, while you were working on RI2, I might have been more visible to the core team because of all my 'communication' about trunk tests. I'm glad they're aware of your contributions. There was some kind of Ruby event this weekend, and there were a lot of commits. Early on, the AV mswin build stopped working (and is still broken), so anytime I was near my computer, I'd start a new build. At times, it was running continuously. The issues I brought up were addressed, and it's still running with (basically) no failures/errors. Some OpenSSL issues, but that's probably due to the fact that Travis ruby-head may not be using 1.1.0, night even be using 1.0.1... Anyway, re Appveyor. I wish you'd reconsider the idea of a fresh install. 1. I've been overwriting my daily ruby trunk for months, and I've don't ever recall seeing the default gem list get garbled when running `gem list`. I do `gem cleanup` from time to time, but we're talking weeks, as opposed to new builds are daily or better... 2. Some of the 'duplicate gems' are default gems, and having been in the Bundler and RubyGems repos, this stuff is still pretty new. FYI, Bundler has never tested with AV, but RubyGems is and is also using my trunk build. But, as I'm sure you know, Bundler is used a lot for CI test setup. 3. I've been working more with Travis (although much less than AV), and don't recall any of their versions having the gem `clutter`. 4. Taking OpenSSL or JSON default gems as an example. Both have two gemspec files, so gem list is showing them. But, I think only the most recent is installed. Hence, if someone has bundler set to a particular version (common, I've seen amazing things using trunk) which matches the older gemspec, things might get amiss... 5. I think the AV guys are probably spread a little thin... FYI, I've got a script that can probably do everything AV needs, output is [here](https://ci.appveyor.com/project/username_0/appveyor-testing/build/1.0.45). Updates RubyGems and adds Bundler. If you decide a fresh build maybe the way to go, I can pass it along to the AV guys. Thanks again, Greg Answers: username_1: Congratulations to you as well! Will you be there in Matsue at the conference? Regarding appveyor install: I don't have a strong opinion to this. Both versions of OpenSSL and json should be working, although only the newer one is used unless the older is explicitly specified. If not, it's a bug in RubyInstaller2 or rubygems. Did you encounter anything not working? username_0: I'm missing this. First of all, I'm not familiar with the 'Inno Setup' installer. Let's assume that AV is installing Ruby from the *.7z files, so an existing 2.4.1 install is in place, and 2.4.2 is extracted into it. For bundled gems, there are version specific folders that they install into, along with a few version specific files (*.gem, *.gemspec). For default gems, there are no version specific locations, and the only file that is version specific is the *.gemspec file. Hence, it seems to me that, for an 'in-place' upgrade, there will be only one set of files for OpenSSL/JSON. But, there may be more than one gemspec. Which may or may not confuse RubyGems / Bundler... What am I missing? username_1: OK, now I got it! I thought about bundled gems first. Current RubyInstaller2.exe makes an upgrade by simply overwriting all installed files and adding new files. This was how RubyInstaller1 worked and obviously works for most people. It's known to leave some artifacts in the start menu (see #78). It obviously also leaves the old gemspecs for default gems. I already planned to switch to a smarter upgrade procedure in RubyInstaller2.exe. It will deinstall the previous package before installing the new one. The deinstallation will not touch any bundled gems nor gems installed per `gem install`. username_0: A good idea, but potentially unstable... But, a very good idea. Since I use trunk and don't package with the installer, I don't normally trip over that kind of issue. I can test against an RI2 build, so let me know when you move forward... Back to Appveyor. Maybe you could let them know there are issues, and we're working on it? I've got the start of a script that should work for them (AV), it downloads the two builds (32 & 64) from GitHub, puts them in temp folders, adds bundler, updates rubygems, then renames old, renames temp to old, then deletes old (new location) & downloads, etc. Main thing holding me up right now is files added in the bin folder by gems. I've got code working in my trunk (ruby-loco) system that works, but, under the assumption that they're called from the same folder that PATH points to. It also affects the 'reproducible build' issue (I consider a proper 'reproducible build' to contain no hard coded paths.) AYK, some of the files will be installed with a hard coded path included with `ruby.exe`. My current code replaces that with just `ruby.exe`. It should really replace it with `%~dp0ruby.exe`. But I'm having some issues with works local and in AV ruby-loco, but not in my current script on AV. When I get that sorted out... Said another way, the issue is that my preference is that, for two ruby installs A & B, with A_bin_path and B_bin_path: 1. no paths should be hard coded in any files in the bin folder 2. If `ENV['PATH']` contains A_bin_path, and from a cmd prompt one does ``` B_bin_path\bundle env ``` One should see the info from the B install. I hope this makes sense, and if you have any thoughts on it, I'll file an issue in RubyGems about replacing the hard coded paths with `%~dp0` on windows platforms... Thanks.
teamcapybara/capybara
262651683
Title: Calling #fill_in on a fillable_field Capybara::Node::Element doesn't work Question: username_0: ## Meta - Capybara Version: - 2.15.1 - Driver Information (and browser if relevant): - selenium-webdriver 3.6.0 - Chrome 61.0.3163.100 (Official Build) (64-bit) - ChromeDriver 2.32.498513 ## Expected Behavior Field should be filled. ## Actual Behavior Raises `Capybara::ElementNotFound Exception: Unable to find visible field nil that is not disabled` ## Steps to reproduce ```ruby When(/\AI populate "([^"]*)" with "([^"]*)"\z/) do |field_locator, value| field = find_field(field_locator) if field.matches_selector?(:fillable_field) field.fill_in(:with => value) elsif field.matches_selector?(:checkbox) if [ '0', '', 'nil', 'false', 'unchecked', 'no' ].include?(value.to_s.downcase) field.uncheck else field.check end elsif field.matches_selector?(:radio_button) field.choose(value) elsif field.matches_selector?(:select) field.select(value) elsif field.matches_selector?(:file_field) field.attach_file(File.expand_path(value)) else warn "Unknown field type for #{field.inspect}." field.set(value) end end When(/\AI populate "([^"]*)" with the following:\z/) do |form, fields| within(form) do fields.rows_hash.each do |name, value| step %{I populate "#{name}" with "#{value}"} end end end ``` ```gherkin Given I populate "#new_user_session" with the following: | user_session_email | <EMAIL> | | user_session_password | <PASSWORD> | ``` ## Suspected problem in Capybara This error occurs to be occurring because the `#fill_in` method runs `find(:fillable_field, locator, options)` in the context of the element and `find` appears to only look at descendants of the current element (of which there are obviously none). Intuitively, however, `#fill_in` should performed on the current object given the way I invoke it. *NB: I imagine similar issues also exist in the other form fill methods invoked in my (not yet fully tested) step definition.* Answers: username_0: Looking into this more, one could argue I should be using the `#set` method for everything, but this is rather unintuitive considering I can't use it for `<select>`s, and `#check`, `#uncheck` and `#choose` have added functionality with regard to the `:allow_label_click` option that I would have to duplicate, should I need it. Status: Issue closed username_1: @username_0 You're just using it wrong. As you stated, if you have already located the element you want to interact with , you call `set` on it (which you can call on checkboxes, etc with true or false). `fill_in` as documented - http://www.rubydoc.info/gems/capybara/Capybara/Node/Actions#fill_in-instance_method - finds a field a fills it in, it doesn't act on the current scope element Why you would ever write a cucumber step as Given I populate "some_check_box_id" with "unchecked" is beyond me though. username_0: Haha. I wouldn't write ```gherkin Given I populate "some_check_box_id" with "unchecked" ``` in long form, but I wrote the step to work with checkboxes too so I could fill in a whole web form (checkboxes included) if I supply the inputs in the tabular format. Anywho, I can accept I'm not using it as it was intended. It just seemed a bit unintuitive because you'd think calling `#fill_in` on an object of a class literally named `Element` would let you fill in said element. Not a huge deal though. username_2: The documentation must've been updated since this was written, because now it shows two examples: ``` username_2: Looks like 3.7.0 introduced support for calling fill_in on the current element. username_1: @username_2 You're using 2.18 and looking at the docs for 3.13.x, there are big differences between those two versions. The ability to call `fill_in` on the element you want filled in was added in 3.7 username_1: Nevermind - you apparently figured that out --- 2.18 was released a year ago -- it's pretty obsolete at this point
idnow/de.idnow.android
653221064
Title: App getting crash after open Idnow (My app migrated to AndroidX) Question: username_0: java.lang.NoClassDefFoundError: Failed resolution of: Landroid/support/v4/content/LocalBroadcastManager; at de.idnow.sdk.util.IDnowExternalLog.logExternally(IDnowExternalLog.java:21) at de.idnow.sdk.IDnowSDK.getApplicationName(IDnowSDK.java:568) at de.idnow.sdk.IDnowSDK.initialize(IDnowSDK.java:444) Answers: username_1: Yeah IDNow is incompatibilie with MaterialTheme, You app styles have to iherit from Material whereas IdNow from AppCompact <style name="AppTheme" parent="Theme.MaterialComponents.Light.NoActionBar"> ... </style> <style name="IdnowSdkTheme" parent="Theme.AppCompat.Light.NoActionBar"> ... </style> This will save You from crash, but IDNow styles will be destoryed - U will see old fasion dialog , progress bar like in KitKat Status: Issue closed username_2: I apologise your issue wasn't addressed, however, we've also migrated to androidx and this should no longer be the case. Closing.
se1exin/Vultr-Dynamic-DNS
150500812
Title: Attempting to setup error in Vultr DNS admin Question: username_0: When I login and attempt to add my domain with your suggested IP of 111.222.333.444 I'm given an error about entering valid info. Any tips? Answers: username_1: Try 192.168.1.1. Stupid me, 111.222.333.444 isn't even a valid IP address Status: Issue closed
westurner/opengov
718558261
Title: ENH: Finding the latest version of government resources Question: username_0: ## User Stories STORY1: Users can search for a document by {title, identifier, file hash, [...]} and find the latest version (1) in a format that works with their device; (2) with metadata that their device recognizes. ## Use Cases ### Find the DSCA Handbook - Search for "DSCA Handbook" - Search for file_hash(document_in_my_Download_folder): - Which is the latest version of this document? - Can I read this on my phone? - Can I read this on my tablet? - Can I read this on my eBook reader? ## Data Model - PDF metadata - DOC, DOCX, ODF document metadata - OAI-PMH metadata - **Schema.org metadata** - This is what search engines now index ### Schema.org data elements Attributes of these government documents (in terms of schema.org classes and properties) #### Classes https://schema.org/CreativeWork https://schema.org/Book https://schema.org/ScholarlyArticle https://schema.org/CreativeWorkSeries https://schema.org/BookSeries https://schema.org/MovieSeries #### Properties https://schema.org/identifier https://schema.org/name https://schema.org/description https://schema.org/abstract https://schema.org/about https://schema.org/url https://schema.org/datePublished ### Gaps - [ ] How to link to newer and older versions of a document?
ProyectoIntegrador2018/makers
503659308
Title: Lab Space Admins do not see the equipment of the spaces they administrate Question: username_0: In the admin dashboard, Lab Space Admins should be able to see the equipment of the spaces they administrate. ## Steps for reproducing bug 1. Sign in as a Lab Space Admin 2. Go to Admin dashboard 3. Click on Equipment tab ## Expected behavior Equipment the Lab Space Admin administrate should be listed ## Current behavior Equipment the Lab Space Admin administrates are not listed (nothing is listed) Answers: username_1: As I mentioned on #135 this wasn't a bug, because the only problem was that the Lab Space Admin wasn't assigned to any Lab Space, but with the fix, it is now possible to assign a Lab Space to a Lab Space Admin and with this he/she is able to see their Equipments. Status: Issue closed
spcowboy/Mynote
171037597
Title: win7安装CENTOS7 Question: username_0: 1.下载EASYBCD EXT2FSD 分区助手 2.使用分区助手分出2个逻辑分区(一定要逻辑分区,不然后面安装时无法分区),1个分区6G左右,装安装光盘,有卷标;1个分区作为centos的工作空间,无卷标,都为EXT3 3.使用EXT2FSD配置6G的分区为win7可识别,然后将安装光盘拷入6G的分区(安装盘小于4g这一步省略);解压centos镜像中isolinux文件夹下的vmlinuz和initrd.img(如果只有vmlinuz0则使用vmlinuz0,下面指令也需对应修改)到6G分区盘根目录,解压images文件夹到6g分区根目录下,也就是说跟centos镜像在同一目录下 4.使用EASYBCD,添加新条目-neogrub-安装-配置,在弹出的文本文件里填入: title centos root (hd0,2) 2表示你的6G的分区,可能是其他数字,可以自己尝试 kernel /vmlinuz linux repo=hd:/sda3:/ 3可能是其他数字,需要尝试 initrd /initrd.img 五、其他信息 安装,启动,登录后,发现上不了网,需要做如下配置更改(假设你是以root权限登录的) 1. vim /etc/sysconfig/network-scripts/ifcfg-eth0 ONBOOT=yes #系统启动时是否设置此网络接口,设置为yes时,系统启动时激活此设备,默认值是no 2. 如果你是公司内的域账号,那么还需要设置一下:打开system->preferences->network connections 在centos里添加win7启动项: 1.首先下载安装ntfs-3g包,这样linux才能识别ntfs盘; 1. 添加EPEL源(ntfs-3g在该源中): $ yum search epel 只找到一个epel-release.noarch, 就装它了。 # yum install epel-release 2. 安装ntfs-3g用于识别ntfs分区: # yum install ntfs-3g 3. 重新生成引导项: # grub2-mkconfig -o /boot/grub2/grub.cfg 配置无线网络: 1.# ip addr 找到自己的无线网接口 (ps:本人的是wlp3s0) 2.# ip link set wlp3s0 up 打开无线网的驱动 3.# ip link show wlp3s0 查看该网络接口的状态 括号中出现UP 说明驱动已经开启 网络名字可以通过其他在本网的设备查看一下 注意:引号不要省略 执行wpa_passphrase "<PASSWORD>" “PSK“ 注意:引号不要省略 修改/ETC/WPA_PASSPHASE.CONF 文件,把上一部执行的结果拷贝进去 4.执行wpa_supplicant -B -i wlp3s0 -c /ETC/WPA_PASSPHASE.CONF 5.dhclient wlp3s0 dhcp分配ip 6. ip addr show wlp3s0
FACG5/reuters-news
346741390
Title: don't repeat your self Question: username_0: https://github.com/FACG5/reuters-news/blob/3a9c4a8efc89d405070a8a85b06458df420e81f3/src/test.js#L30-L40 what is the different between this two checks ? ` .expect(200) ` ` t.equal(res.statusCode, 200, 'Should return 200'); `
ray-project/ray
589941288
Title: Incorrect unreconstructable error message and raise different exception. Question: username_0: is not actually correct because the error is caused by a node failure, not by LRU eviction or deletion by the user. `UnreconstructableError` doesn't seem like the right error because the actor isn't configured to be reconstructed. The expected behavior is to raise an exception of some sort. In contrast, the following script (in which the actor dies while executing the task) raises `RayActorError: The actor died unexpectedly before finishing this task.`. Perhaps something like that would be more appropriate. ```python import ray import time ray.init() @ray.remote class Foo: def method(self): time.sleep(100) f = Foo.remote() x_id = f.method.remote() time.sleep(1) ray.kill(f) time.sleep(1) ray.get(x_id) ```
marstr/envelopes
633890862
Title: Panic trying to print balance Question: username_0: <details><summary>System Information</summary> <pre> Version: v0.2.0-alpha-modified System: windows/amd64 Go: go1.12.5 Source Revision: 36f7f1924ae315fe7fbd03b95cd5e08037ed514b </pre> </details> ## What did you do? After transferring both cash and stock out of a budget, I executed the following command: `baronial bal .\budget\queue\` The balance file contains several different 0.000 line items. All seem well-formed. ## What did you expect to happen? It should have printed a zero balance. ## What actually happened? A recovered panic is generated while figuring out the balance string. <details><summary>Stack Trace</summary> <pre> Total: [redacted] Balance: %!v(PANIC=String method: bytes.Buffer: truncation out of range) Children: espp: [redacted] payable: USD 0.000 </pre> </details> Answers: username_0: The trouble here seems to be that when the balance function simplifies the multi-term balance, everything gets removed, and then it tries to truncate an empty buffer. Status: Issue closed
mitodl/mitxpro
432631845
Title: Can't logout of open edX Question: username_0: ### Steps to Reproduce 1. Login to xpro 2. Follow a catalog link from xpro to open edX, to ensure that your browser is logged in to open edX 3. Using the user menu in the top right, choose Logout ### Expected Behavior Ideally, logging out should: 1. Log the user out of open edX 2. Log the user out of xpro 3. redirect the user to the /login page of xpro ### Actual Behavior In the browser, the page just reloads. In the background, I think this is what happens: 1. User is logged out of open edX 2. User is directed back to the current page 3. the current page in open edX is proected, so the user is directed to login to xpro 4. Since an xpro session is already open, the user is directed back to open edX Answers: username_1: This should probably cover the reverse as well - that logging out of xPro also logs you out of open edX Status: Issue closed
SeleniumHQ/selenium-ide
343865462
Title: I got a warning message when I closed a selenium-ide, although I didn't anything. Question: username_0: ## Meta - OS: Window 10 <!-- Windows 10? OSX? --> Selenium IDE Version: v3.2.0-beta.4 ![ezgif com-video-to-gif](https://user-images.githubusercontent.com/33743343/43114090-2efa422e-8f38-11e8-9859-057eb7be3f6f.gif) <!-- e.g.: 3.0.3 --> Selenium SIDE Runner Version: <!-- e.g.: 0.7.0 Please note that selenium-side-runner supports Node.js 8 - 10 --> Node version: <!-- e.g.: 10.4 Only required if using selenium-side-runner --> Browser: Chrome Browser Version: 67.0.3396.99 (Official Build) (64-bit) <!-- e.g.: 49.0.2623.87 (64-bit) --> ## Expected Behavior - If I didn't anything, the window of ide will be closed without a warning message. ## Actual Behavior - I got a warning message when I closed a selenium-ide, although I didn't anything. ## Steps to reproduce - <!-- Please be sure to include an SSCCE (Short, Self Contained, Correct [compilable] example) http://sscce.org/ --> I didn't modify my IDE, I just opened. I don't want the message to appear when I didn't anything.<issue_closed> Status: Issue closed
MicrosoftDocs/azure-docs
332375326
Title: Example Currently Broken due to Power BI bug Question: username_0: See https://community.powerbi.com/t5/Desktop/PowerBI-changes-data-table-name-to-RealTimeData-and-creates/td-p/152268 - basically, unless the table name is "RealTimeData" in the streaming analytics output, this doesn't work today. Also, I was unaware that PowerBi didn't like nested data structures and just imported them as Text... Might be worth adding a note to clarify that you must use a flat data structure if you are using a real device. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 97ce2614-0e74-0f67-bf3f-8a477bd5f329 * Version Independent ID: e29a2f4e-63a3-2f28-35bf-6d990f9f3b84 * Content: [Real-time data visualization of sensor data from Azure IoT Hub – Power BI](https://docs.microsoft.com/en-us/azure/iot-hub/iot-hub-live-data-visualization-in-power-bi) * Content Source: [articles/iot-hub/iot-hub-live-data-visualization-in-power-bi.md](https://github.com/Microsoft/azure-docs/blob/master/articles/iot-hub/iot-hub-live-data-visualization-in-power-bi.md) * Service: **iot-hub** * GitHub Login: @rangv * Microsoft Alias: **rangv** Answers: username_1: @username_2 Hi Dominic, I think this issue is for one of your articles. username_2: @rangv - Hi Ranga - this issue is for one of your articles. @username_10 - can you reassign this please? username_3: in this step https://docs.microsoft.com/en-us/azure/iot-hub/iot-hub-live-data-visualization-in-power-bi#add-an-output-to-the-stream-analytics-job after authorization, the save button is still greyed out so I can't save the output. username_3: however, I tried to use the different type of output, like Blob storage or Table storage, it worked fine. username_4: Dataset is not showing in the PowerBi dataset list. Dataset was created in the the step which creates Output for Streaming Analytics. This has been checked for login credentials and for the path to My workspace. Any Suggestion please! username_5: Same here. Streaming dataset not showing in PowerBI username_6: @delrid1, @username_3, @username_4, @RDSimmons01, @username_5 -- Thanks for reporting these issues. We've updated the topic and the product has undergone some upgrades as well. The tutorial should work as-advertised now. Please have a look and let us know if we can close this issue. @delrid1 -- thanks for your suggestion about adding a note wrt flat data structures. We may not add this at this time, but will monitor the topic and make a note in case of future revisions. username_7: #reassign:jimacoms3 username_6: @username_0, @username_3, @username_4, @RDSimmons01, @username_5 Thanks again for reporting these issues. Since we haven't heard back from you, we are going to close this issue. Please feel free to re-open it if you feel the problem you reported hasn't been adequately addressed. username_6: #please-close Status: Issue closed username_9: Same Error. every thing is okay on IoT hub and job side but Streaming dataset is not showing in PowerBI username_6: Hi @username_9, I'm not having any trouble running this. I'm using a dataset named **ContosoSampleHubTelemetry** and a data table named **TelemetryData**. Everything seems to work fine. ![Capture](https://user-images.githubusercontent.com/47835911/61091037-75ca0680-a3f5-11e9-87e1-fd7f4adaeaad.PNG) I think we need to figure out what is different in our scenarios. Can you tell me: 1. Are you using Power BI online or the desktop app? (I'm using online) 2. What device you are using to send data? (I would suggest trying the [Raspberry Pi Online Simulator](https://docs.microsoft.com/en-us/azure/iot-hub/iot-hub-raspberry-pi-web-simulator-get-started) as a sanity check). 3. Have you verified that data is actually being arriving at your hub? You can use the [az iot hub monitor-events](https://docs.microsoft.com/en-us/cli/azure/ext/azure-cli-iot-ext/iot/hub?view=azure-cli-latest#ext-azure-cli-iot-ext-az-iot-hub-monitor-events) CLI command for this. 4. When you say everything is OK on IoT hub and job side -- do you mean that the job starts and gets to the running state with no errors? Or? 5. When you say that the Streaming dataset is not showing in PowerBI. Do you mean that data is not showing up or that you cannot select the dataset to configure a line chart? I'm not really sure that these are going to make a difference. I'm just trying to narrow down the differences between our scenarios. Thanks, Jimaco username_9: What subscription model you are using? I am using free tier subscription, I think that may be difference. 1. I tried both desktop app and Power BI online. 2. I am using IIOT gateway but that is sending data on IoT hub. 3. NO I am not able to see in CLI monitoring. 4. Yes 5. Nothing change on Power BI, I couldn't find any data set or any other thing from Iot hub side. username_6: I'm using a standard (S1) tier hub. The Free tier could be an issue if you have exceeded the 8k/day message limit: https://azure.microsoft.com/en-us/pricing/details/iot-hub/ ; otherwise, I don't think that is the problem. Your answer to #3 points to a potential issue. How are you verifying that you are actually getting messages from your gateway to your IoT hub? You should be able to see these with the **az iot hub monitor-events** command. Specifically, in the tutorial you create a consumer group so you should be able to do the following: ```azure-cli az iot hub monitor-events --hub-name {yourHubname} --cg {yourConsumerGroup} ``` **Note**: You may have to add the `-g` parameter to specify your hub's resource group. You can also use other parameters to format output and/or filter for a device. Another potential issue is if you have custom routes created on your IoT hub -- for example to route every message received to storage. If that is the case, the custom route will suck the data before it ever hits the default endpoint, so you need to add a compensatory route to make sure the data also gets sent to the default endpoint. Can you also try using the **Raspberry Pi online simulator** that I pointed you to in #2 above? This might help narrow down the problem. username_7: I only use S1 tier hubs. You can’t use more than one endpoint on a free hub, so you can’t do the routing tutorial with a free hub, as noted by the big-ass note that’s in the article. So that’s fine. You might add a note as to why you don’t recommend they use a free hub. These hubs are expensive, and people will try the free hub unless you tell them not to. I don’t have any idea what is this “gateway” that you’re talking about?? For testing the routing, I need specific content in the messages, so there’s an app that sends the messages and randomizes the content so I get what I need. It’s the one for iot-hub in the c# samples iot samples repo. The messages are routed to different destinations, and I can check the destinations to see if the whole thing is working. I don’t have any reason to use anything else, and anything else won’t result in the messages being routed correctly, so it will just confuse the customer. This is ok. It’s a routing tutorial, the messages are supposed to be routed, and this is how Ashita wanted them to be. As this came through in a github comment e-mail, I don’t know what problem you’re talking about here. I thought the main q was about PowerBI. Robin username_6: HI @username_9, let me know if any of the above helped. username_10: @username_9 please feel free to open a new issue and reference this one if you are still blocked. Thanks!
Wistoft2410/Digital-design---Spilprojekt
699608444
Title: Grøn og rød farve på dinosaurussen Question: username_0: # Grøn og rød farve på dinosaurussen ## Task tilhører følgende user story: #7 ### Forventet tid: 20 minutter Dinosaurussen skal kunne skifte farve mellem grøn og rød. Vi arbejder med 10 sekunders intervaller (10 sekunder er 100% af den tid vi arbejder med). Man er grøn i 9 sekunder (90%), og rød i et sekund (10%) til at starte med. Efter et tids interval øger vi tiden man er rød med et sekund og reducerer tiden man er grøn med et sekund. Answers: username_1: En ny person skal **assignes** til denne issue, da den skal testes af en ny person! username_0: **Test af farver på timern** Timern skal have en grøn farve for 10 sekunder, og i de 10 sekunder, så skal der kun komme dejlige æg ned fra himlen, som giver points. Når timern efter 10 sekunder så bliver rød, skal der komme metatorer ned, som spillern skal undvige, ellers mister han liv. Timern viser den røde og grønne farve godt, og de er en god kontrask til hinanden. I iteration 2, skal vi gøre sådan at timern bliver skiftevis mere og mere rød og mindre og mindre grøn. I interation 3, kan vi også lave en gul farve, hvor der både er æg og meteorer Status: Issue closed username_1: **Test nr. 2 af farver på timern** Det ser ud til at det stadig virker. Der er dog lavet en afvigelse fra denne task, da hvergang man modtager 10 point så bliver sværhedsgraden højere/svære på den måde at tidsmængden bliver øget med 1 sekund i den røde del, og den formindskes med 1 sekund i den grønne del!
glasklart/hd
12637409
Title: Little Things® Forever Question: username_0: **App Name:** Little Things® Forever **Bundle ID:** com.klicktock.littlethings2 **iTunes ID:** <a target="_blank" href="http://getart.username_1.at?id=520762327">520762327</a> **iTunes URL:** <a target="_blank" href="https://itunes.apple.com/us/app/little-things-forever/id520762327?mt=8&uo=4">https://itunes.apple.com/us/app/little-things-forever/id520762327?mt=8&uo=4</a> **App Version:** 1.7.0 **Seller:** KlickTock Pty Ltd **Developer:** <a target="_blank" href="https://itunes.apple.com/us/developer/klicktock/id348840966?uo=4">© KLICKTOCK</a> **Supported Devices:** iPad2Wifi-iPad2Wifi, iPad23G-iPad23G, iPhone4S-iPhone4S, iPadThirdGen-iPadThirdGen, iPadThirdGen4G-iPadThirdGen4G, iPhone5-iPhone5, iPodTouchFifthGen-iPodTouchFifthGen, iPadFourthGen-iPadFourthGen, iPadFourthGen4G-iPadFourthGen4G, iPadMini-iPadMini, iPadMini4G-iPadMini4G, iPhone5c-iPhone5c, iPhone5s-iPhone5s, iPadAir-iPadAir, iPadAirCellular-iPadAirCellular, iPadMiniRetina-iPadMiniRetina, iPadMiniRetinaCellular-iPadMiniRetinaCellular, iPhone6-iPhone6, iPhone6Plus-iPhone6Plus, iPadAir2-iPadAir2, iPadAir2Cellular-iPadAir2Cellular, iPadMini3-iPadMini3, iPadMini3Cellular-iPadMini3Cellular, iPodTouchSixthGen-iPodTouchSixthGen, iPhone6s-iPhone6s, iPhone6sPlus-iPhone6sPlus, iPadMini4-iPadMini4, iPadMini4Cellular-iPadMini4Cellular, iPadPro-iPadPro, iPadProCellular-iPadProCellular, iPadPro97-iPadPro97, iPadPro97Cellular-iPadPro97Cellular, iPhoneSE-iPhoneSE, iPhone7-iPhone7, iPhone7Plus-iPhone7Plus, iPad611-iPad611, iPad612-iPad612, iPad71-iPad71, iPad72-iPad72, iPad73-iPad73, iPad74-iPad74, iPhone8-iPhone8, iPhone8Plus-iPhone8Plus, iPhoneX-iPhoneX **Original Artwork:** <img src="http://is3.mzstatic.com/image/thumb/Purple127/v4/62/ff/f8/62fff85a-1618-3b40-264e-08062783a61c/source/1024x1024bb.png" width="180" height="180" /> **Accepted Artwork:** \#\#\# THIS IS FOR GLASKLART MAINTAINERS DO NOT MODIFY THIS LINE OR WRITE BELOW IT. CONTRIBUTIONS AND COMMENTS SHOULD BE IN A SEPARATE COMMENT. \#\#\# Answers: username_1: **App Name:** Little Things® Forever **Bundle ID:** com.klicktock.littlethings2 **iTunes ID:** <a target="_blank" href="http://getart.username_1.at?id=520762327">520762327</a> **iTunes URL:** <a target="_blank" href="https://itunes.apple.com/us/app/little-things-forever/id520762327?mt=8&uo=4">https://itunes.apple.com/us/app/little-things-forever/id520762327?mt=8&uo=4</a> **App Version:** 1.7.0 **Seller:** KlickTock Pty Ltd **Developer:** <a target="_blank" href="https://itunes.apple.com/us/developer/klicktock/id348840966?uo=4">© KLICKTOCK</a> **Supported Devices:** iPad2Wifi-iPad2Wifi, iPad23G-iPad23G, iPhone4S-iPhone4S, iPadThirdGen-iPadThirdGen, iPadThirdGen4G-iPadThirdGen4G, iPhone5-iPhone5, iPodTouchFifthGen-iPodTouchFifthGen, iPadFourthGen-iPadFourthGen, iPadFourthGen4G-iPadFourthGen4G, iPadMini-iPadMini, iPadMini4G-iPadMini4G, iPhone5c-iPhone5c, iPhone5s-iPhone5s, iPadAir-iPadAir, iPadAirCellular-iPadAirCellular, iPadMiniRetina-iPadMiniRetina, iPadMiniRetinaCellular-iPadMiniRetinaCellular, iPhone6-iPhone6, iPhone6Plus-iPhone6Plus, iPadAir2-iPadAir2, iPadAir2Cellular-iPadAir2Cellular, iPadMini3-iPadMini3, iPadMini3Cellular-iPadMini3Cellular, iPodTouchSixthGen-iPodTouchSixthGen, iPhone6s-iPhone6s, iPhone6sPlus-iPhone6sPlus, iPadMini4-iPadMini4, iPadMini4Cellular-iPadMini4Cellular, iPadPro-iPadPro, iPadProCellular-iPadProCellular, iPadPro97-iPadPro97, iPadPro97Cellular-iPadPro97Cellular, iPhoneSE-iPhoneSE, iPhone7-iPhone7, iPhone7Plus-iPhone7Plus, iPad611-iPad611, iPad612-iPad612, iPad71-iPad71, iPad72-iPad72, iPad73-iPad73, iPad74-iPad74, iPhone8-iPhone8, iPhone8Plus-iPhone8Plus, iPhoneX-iPhoneX **Original Artwork:** <img src="http://is3.mzstatic.com/image/thumb/Purple127/v4/62/ff/f8/62fff85a-1618-3b40-264e-08062783a61c/source/1024x1024bb.png" width="180" height="180" /> **Accepted Artwork:** \#\#\# THIS IS FOR GLASKLART MAINTAINERS DO NOT MODIFY THIS LINE OR WRITE BELOW IT. CONTRIBUTIONS AND COMMENTS SHOULD BE IN A SEPARATE COMMENT. \#\#\# username_1: ![com klicktock littlethings2-large](http://preview.username_1.at?image=https://user-images.githubusercontent.com/2068130/38596613-d55d23e4-3d52-11e8-9be3-fef750baf0d7.png) https://user-images.githubusercontent.com/2068130/38596613-d55d23e4-3d52-11e8-9be3-fef750baf0d7.png --- --- Source: https://user-images.githubusercontent.com/2068130/38596631-e584e00e-3d52-11e8-868f-7e5ab5ab3ff1.png Status: Issue closed
INTECS-ITFAC/intecs_webapp_front-end
611350159
Title: Re Locate About Us section Question: username_0: - Currently About Us is in footer section. - Recent discussion, the committee decided to relocate about us section to new navbar tab - Task 01 : Remove current about us section - Task 02 : Add new tab to navbar after the "gallery" tab stating "About Us" (Content of about us is not addressed in this issue, here only consider the relocating) Contact INTECS Open Source community for more information. (You can find a link to join the community in Read me file.) Answers: username_0: - Currently About Us is in footer section. - Recent discussion, the committee decided to relocate about us section to new navbar tab - Task 01 : Remove current about us section - Task 02 : Add new tab to navbar after the "gallery" tab stating "About Us" (Content of about us is not addressed in this issue, here only consider the relocating) Contact INTECS Open Source community for more information. (You can find a link to join the community in Read me file.) username_1: May I work on this? username_0: OK Sure, Please Proceed Status: Issue closed
pelagios/recogito2
378256253
Title: 'New' button in Shared With Me view Question: username_0: We should do something with the __New__ button in the __Shared with me__ view. The issue is that new uploads will always go to users' own documents (obviously), so uploading from the shared docs view might be confusing. (Alternatively, we could force-switch to My Docs. But might be just as confusing.)
Altinn/altinn-studio
465112129
Title: Choose version of runtime for deploy Question: username_0: ## Considerations Look at work in #472 ### Questions - Are there other ways to solve this scenario e.g. back-wards compatibility for runtime? - Can update of runtime be mandatory e.g. security patch? - Should the developer also choose version of altinn studio? ## Acceptance criteria - What is allowed/not allowed - Validations - Error messages and warnings - ... ## Tasks - [ ] Update developer tasks - [ ] Design review ## Specification tasks - [ ] Answer the open questions - [ ] Update acceptance criteria - [ ] Sketches - [ ] QA ## Development tasks - [ ] Documentation (if relevant) - [ ] Manual test (if needed) - [ ] Automated test (if needed) Answers: username_1: Can already be done by changing the version that is loaded from CDN. Default is that major version 1 is loaded, so any security or minor fixes that we do will be used automatically. If we introduce breaking changes, we will bump that to version 2 etc. Here's the line of code to change in the app. https://github.com/Altinn/altinn-studio/blob/master/src/Altinn.Apps/AppTemplates/AspNet/App/views/Home/Index.cshtml#L29 Here are the available versions: https://github.com/Altinn/altinn-cdn/tree/master/toolkits/altinn-app-frontend Status: Issue closed username_2: Already available through referring to version numbers, as per commented by @username_1.
Zrips/CMI
830264858
Title: Command doubles in /ic Question: username_0: When a player clicks on an entity, the command written in interactiveCommands is doubled, but with blocks everything is ok. Watch the video. https://youtu.be/EYEqFr_ETrg --- interactiveCommands: shop: UUID: - 001ee404-c581-4671-96ab-1112e20495cd Commands: - give [playerName] diamond ReqPerm: false --- Cmi Version 172.16.58.3 Server Type Spigot Server Version 1.16.5 Answers: username_1: Should be fixed with latest builds
Andu2/FEH-Mass-Simulator
289058754
Title: Blade Tomes Question: username_0: Hi, I was doing some simulations with Nino, and saw that in the summary popup windows, it showed this: ![bug in duel simulator](https://user-images.githubusercontent.com/25864406/35012601-ebdde20e-facf-11e7-80ac-41e4befba3c8.png) That is incorrect because it shows Ayra getting bonus attack from the blade tome, based on her buffs, which should not happen, since she does not have the blade tome equipped. I'm not sure if this is effecting the actual calculations of the simulation, or if it is purely a display bug, but it seems like it would be worth addressing either way. Answers: username_1: I think it's a display issue, let me quickly fix it. username_1: Testing if bladetome buffs affect enemy Heavy Blade effects. eg. Ike attacks Nino (with buffs), will Ike gain an extra charge? username_0: That is an interesting question. According to the wiki it effects stuff like bonfire if she has the equipped, so I think it is safe to assume it would also effect whether or not an enemy's heavy blade is triggered, but I'm not sure. I'm also not sure whether or not the mass simulator handles that properly. username_1: Yeah, bladetome and blizzard wording are a bit different and confusing, I'll be confirming this when I get a chance. username_1: [https://imgur.com/a/1swus](https://imgur.com/a/1swus) Seems to apply on defense and affects Heavy Blade. The combat tooltip will show the bonus on both attack and defense phases because of how effective attack is calculated, it is not added twice though. May fix later. Status: Issue closed
scala-exercises/scala-exercises
142404466
Title: Some sample code snippets are not extracted correctly Question: username_0: We have noticed that there are some code snippets that are not correctly extracted from the comments. See this example in the "Tuples" section: ![blank](https://cloud.githubusercontent.com/assets/409039/13926288/ee2ad8ba-ef8b-11e5-82f7-79b8591fc30b.png) The code from the exercise looks like the following, i can't see anything wrong with it: ```scala object Tuples extends FlatSpec with Matchers with exercise.Section { /** Scala tuple combines a fixed number of items together so that they can be passed around as a whole. They are one indexed. Unlike an array or list, a tuple can hold objects with different types but they are also immutable. Here is an example of a tuple holding an integer, a string, and the console: * * {{{ * val t = (1, "hello", Console) * }}} * * Which is syntactic sugar (short cut) for the following: * * {{{ * val t = new Tuple3(1, "hello", Console) * }}} * As you can see, tuples can be created easily: */ def oneIndexedTuples(res0: String, res1: String) { val tuple = ("apple", "dog") val fruit = tuple._1 val animal = tuple._2 fruit should be(res0) animal should be(res1) } // ... } ``` Answers: username_1: You can inspect the generated source in the content project at `content/target/scala-2.11/src_managed/...`. Below is the Functor exercise from the Cats lib. An empty code block, `<pre class=\"scala\"><code class=\"scala\"></code></pre>`, is emitted-- so something is definitely wrong. ```scala // ... object Exercise_usingFunctor$1 extends Exercise { override val name = "usingFunctor"; override val description = scala.Some("<h3> Using Functor </h3><h4> map </h4><p><code>List</code> is a functor which applies the function to each element of the list:</p><pre class=\"scala\"><code class=\"scala\"></code></pre><p><code>Option</code> is a functor which only applies the function when the <code>Option</code> value\nis a <code>Some</code>:</p>"); override val code = "Functor[Option].map(Option(\"Hello\"))(_.length) should be(res0)\nFunctor[Option].map(None: Option[String])(_.length) should be(res1)"; override val qualifiedMethod = "catslib.FunctorSection.usingFunctor"; override val imports = scala.collection.immutable.List("import cats.implicits._", "import cats._", "import org.scalatest._", "import cats.syntax.cartesian._", "import cats.syntax.apply._", "import cats.std.all._", "import cats._", "import ApplyHelpers._", "import org.scalatest._", "import cats.std.list._", "import cats.std.option._", "import cats._", "import org.scalatest._", "import ValidatedHelpers._", "import cats.data.Xor", "import cats.data.NonEmptyList", "import cats.data.Validated", "import org.scalatest._", "import org.scalatest._", "import cats._", "import cats.std.all._", "import cats._", "import org.scalatest._", "import cats.data.Xor", "import org.scalatest._"); override val explanation = scala.None } //... ``` It might be worth inserting some debugging statements in the comment parsing code. See the [renderBlock method in comments.scala](https://github.com/scala-exercises/scala-exercises/blob/11df0c8abf4b4f94c7ca82d5b57e0eae335a36ab/core/compiler/src/main/scala/com/fortysevendeg/exercises/compiler/comments.scala#L289). username_1: I've finished investigating this and it was introduced by the unwrap method introduced as part of #264. Fix (and test) coming soon. Status: Issue closed
latex3/babel
1145131232
Title: PL lang \cc double colons Question: username_0: For polish language when I try to add \cc{} to \documentclass{letter} I've got double colons. `\cc{Some CC}` -> Kopie:: Some CC It should be change in **locale/pl/babel-pl.ini** 46 line: `cc = Kopie:` to `cc = Kopie` 69 line `cc = Kopie:` to `cc = Kopie`<issue_closed> Status: Issue closed
purescript/spago
858082531
Title: spago reports module `Data.Foreign` was not found but package `foreign` is installed Question: username_0: I am trying to migrate a project from purescript 0.11.7 to 0.13.8 and to spago. I have successfully done this before with other projects. But for one project I am getting errors that the following modules are not found, even though I have installed `foreign `and `foreign-generics`: `npx spago build` reports (yes, I use nmp / npx because of a local install of purs and spago, to be able to run more than one version of purescript) ``` Module Data.Foreign was not found. Module Data.Foreign.Class was not found. Module Data.Foreign.Generic was not found. Module Data.Foreign.Generic.EnumEncoding was not found. Module Data.Foreign.Index was not found. ``` But the packages are installed: ``` $ npx spago ls packages | grep foreign ffi-foreign v0.0.2 Remote "https://github.com/markfarrell/purescript-ffi-foreign.git" foreign v5.0.0 Remote "https://github.com/purescript/purescript-foreign.git" foreign-generic v10.0.0 Remote "https://github.com/paf31/purescript-foreign-generic.git" foreign-object v2.0.3 Remote "https://github.com/purescript/purescript-foreign-object.git" parsing-foreign v0.0.2 Remote "https://github.com/markfarrell/purescript-parsing-foreign.git" ``` This is my `spago.dhall` file: ``` {- Welcome to a Spago project! You can edit this file as you like. -} { name = "zzab" , dependencies = [ "assert" , "console" , "effect" , "foreign" , "foreign-generic" , "node-fs" , "prelude" , "psci-support" ] , packages = ./packages.dhall , sources = [ "src/**/*.purs", "test/**/*.purs" ] } ``` For migrating, I cloned my project, ran `spago init`, then ran `spago install` for the packages I need, then, because I got the problems above but many more than those, I created a new directory and copied only the files and directories really needed by spago (basically leaving out all old bower and node files and directories), ran npm install purescript@~0.13 and then tried to run spago build, and this time getting less errors but still the ones concerning foreign and foreign-generics above. I don't see why spago lists the packages but does not find them when trying to build with purs. Any ideas? Am I doing something wrong? Here some more info if it helps: ``` $ npx purs --version 0.13.8 ``` This is my package.set as written to `packages.dhall`: ``` https://github.com/purescript/package-sets/releases/download/psc-0.13.8-20210226/packages.dhall ``` This is my project directory (my user name is x-ed out): ``` drwxr-xr-x+ 1 xxxx Domain Users 0 Apr 13 15:29 node_modules -rwxr-xr-x+ 1 xxxxx Domain Users 45066 Apr 13 15:29 package-lock.json -rwxr-xr-x+ 1 xxxxx Domain Users 3012 Apr 12 20:43 packages.dhall -rwxr-xr-x+ 1 xxxxx Domain Users 330 Apr 12 20:43 spago.dhall drwxr-xr-x+ 1 xxxxx Domain Users 0 Apr 13 13:11 src drwxr-xr-x+ 1 xxxxx Domain Users 0 Apr 13 13:11 test ``` Answers: username_1: I don't think you're doing anything wrong with `spago` itself - the dependencies should be installed correctly. In preparation for 0.12, those modules you are referring to (`Data.Foreign`, etc) were changed - they are now under the `Foreign` namespace, see https://github.com/purescript/purescript-foreign/releases/tag/v5.0.0. Since you upgraded `purs` and are now using the `0.13.8` package set, your dependencies got upgraded and there were breaking changes. I haven't exhaustively checked, but I think changing all of your imports to be `Foreign`, `Foreign.Generic`, etc will fix these particular errors - but you might run into more errors down the line if any other breaking changes were made username_0: Ah, I overlooked that. I already caught that for `Effect`, for example (and the breaking changes there). Indeed, there they are: ``` $ find . -name Foreign ./.spago/foreign/v5.0.0/src/Foreign ./.spago/foreign-generic/v10.0.0/generated-docs/Foreign ./.spago/foreign-generic/v10.0.0/src/Foreign ./.spago/foreign-object/v2.0.3/bench/Bench/Foreign ./.spago/foreign-object/v2.0.3/src/Foreign ./.spago/foreign-object/v2.0.3/test/Test/Foreign ``` I'm good to go now, and am now running into the (expected) breaking changes and will deal with those. Thanks for the quick help! Status: Issue closed
Stephan-S/FS19_AutoDrive
587583016
Title: [BUG] [MP] Waypoint sort Question: username_0: AutoDrive Version: 1.1.0.2 Release or selbsterstelltes zip: Release Karte (falls relevant): Oberthal Fahrzeug(art) (falls relevant): **Beschreibe den Bug** Zielpunkte lassen sich nicht in Ordner verschieben. Rückfragen gern per Discord Crystep / Manu#7882 kann es euch auch im Livestream zeigen **Falls anwendbar - Schritte mit denen ich den Bug nachstellen kann** **Screenshots** Falls relevant, bitte Screenshots beifügen Answers: username_1: Soviel ich weiss geht das im Multiplayer nicht username_2: -> https://github.com/Stephan-S/FS19_AutoDrive/issues/1238 Status: Issue closed
pyscaffold/pyscaffold
1013371387
Title: Prefer `ConfigUpdater` instead of string interpolation in `setup_cfg.template`. Question: username_0: _Originally posted by @FlorianWilhelm in https://github.com/pyscaffold/pyscaffold/issues/509#issuecomment-932203438_ We use the template just as a structure for the sections, default values and comments, but not do any string substitution on it. Instead, the best would be to use `ConfigUpdater` to fill out all the values given as PyScaffold parameters.
dotnet/aspnetcore
890561889
Title: JsonPatch without Newtonsoft Question: username_0: So people have been asking for years now to allow JsonPatch without requiring Newtonsoft. I've completely moved away from Newtonsoft (as Microsoft has for pretty much everything else) and don't want yet another dependency in my project, so now I'm faced with the decision of including Newtonsoft and potentially having other developers accidentally use that library vs. having to implement my own version of JsonPatch. There are so many requests for this, is there any chance of it finally happening in .NET 6? Answers: username_1: @username_0 thanks for contacting us. Unfortunately that's not something we have planned for .NET 6.0 and its not a small work item for us to tackle. Something like that has a lots of potential to break existing JSON patch customers (in fact, just changing the library will break their deps and libraries) and we want to make sure those customers don't get disrupted by switching libraries. If you are concerned about the usage of JSON.NET types in other projects/assemblies, it should be doable to write an analyzer to prevent such usage in other areas/parts of your product and include it by default in a `Directory.Build.Props` so that it gets automatically applied to new projects. That would effectively make it impossible to reference a type in Newtonsoft.Json by accident.
databricks/spark-avro
58421086
Title: Possible loss of precision when converting from SQL to Avro Question: username_0: According to this, TimestampType is converted to long: ``` TimestampType -> long ``` https://github.com/databricks/spark-avro/blob/master/README.md However, java.sql.Timestamp has a nanos portion that is not fitted into a long. Would there be a loss of precision then? Should TimestampType -> string? http://docs.oracle.com/javase/7/docs/api/java/sql/Timestamp.html Answers: username_0: AvroSaver.scala is calling Timestamp.getTime which returns the milliseconds but not the nanos value (see getNanos) username_1: True. Avro doesn't have a timestamp type, so the two options are either using string (but then when you want to use it, you would probably want to convert string back to something meaningful), or to get a long representation, which is meaningful, but looses precision. We decided to go with the second option. It's unlikely that we can change it at this point, as someone may be relying on the current behavior. If for you purposes, you need the extra accuracy, you can preprocess your RDD to replace the timestamp column with strings that contain the necessary information. That way you won't lose precision and you're free to choose representation that suits your needs best. username_0: I understand. Is there a way to have a config switch to map timestamp -> string? I'd be happy to contribute. username_0: Also, framework in general have very good support for string -> timestamp-like types. Impala, for example, can handle that automatically: http://www.cloudera.com/content/cloudera/en/documentation/core/v5-2-x/topics/impala_timestamp.html#timestamp " Conversions: Impala automatically converts STRING literals of the correct format into TIMESTAMP values. Timestamp values are accepted in the format YYYY-MM-DD HH:MM:SS.sssssssss, and can consist of just the date, or just the time, with or without the fractional second portion. For example, you can specify TIMESTAMP values such as '1966-07-30', '08:30:00', or '1985-09-25 17:45:30.005'. You can cast an integer or floating-point value N to TIMESTAMP, producing a value that is N seconds past the start of the epoch date (January 1, 1970). " username_2: Hi folks, it looks like this issue explains the following behavior. I'm using Spark 1.6.2 with spark-avro 2.0.1: ``` $ spark-shell --packages "com.databricks:spark-avro_2.10:2.0.1" scala> import com.databricks.spark.avro._ scala> import java.sql.Timestamp scala> import java.time.Instant scala> case class Foo(ts: Timestamp) scala> val foos = Seq.fill(100)(Foo(Timestamp.from(Instant.now()))) scala> val df = sc.parallelize(foos).toDF() scala> df.write.orc("hdfs:///user/vagrant/tmp.orc") scala> df.write.avro("hdfs:///user/vagrant/tmp.avro") scala> val orcFoos = sqlContext.read.orc("hdfs:///user/vagrant/tmp.orc").as[Foo] scala> val avroFoos = sqlContext.read.avro("hdfs:///user/vagrant/tmp.avro").as[Foo] scala> orcFoos.take(10) res4: Array[Foo] = Array(Foo(2016-09-13 17:59:18.207), Foo(2016-09-13 17:59:18.207), Foo(2016-09-13 17:59:18.207), Foo(2016-09-13 17:59:18.207), Foo(2016-09-13 17:59:18.207), Foo(2016-09-13 17:59:18.207), Foo(2016-09-13 17:59:18.207), Foo(2016-09-13 17:59:18.207), Foo(2016-09-13 17:59:18.207), Foo(2016-09-13 17:59:18.207)) scala> avroFoos.take(10) res5: Array[Foo] = Array(Foo(48672-07-13 12:23:27.0), Foo(48672-07-13 12:23:27.0), Foo(48672-07-13 12:23:27.0), Foo(48672-07-13 12:23:27.0), Foo(48672-07-13 12:23:27.0), Foo(48672-07-13 12:23:27.0), Foo(48672-07-13 12:23:27.0), Foo(48672-07-13 12:23:27.0), Foo(48672-07-13 12:23:27.0), Foo(48672-07-13 12:23:27.0)) ``` Note that the AVRO "Foos" contain incorrectly deserialized timestamps. How can I work around this issue?
Canadensys/vascan-data
67616601
Title: English vernacular Question: username_0: [Originally posted on GoogleCode (id 2344) on 2014-10-30] Marilyn, you may want to check this. <b>What is the URL of the page where the problem occurs?</b> Astragalus beckwithii var. weiserensis <b>What data are incorrect or missing?</b> Weiser's milk-vetch <b>What data are you expecting instead?</b> Weiser milk-vetch <b>If applicable, please provide an authoritative source.</b> 1. Name used in Ill. Flora of BC, vol. 3 2. Type locality is Weiser, Idaho Answers: username_1: [Originally posted on GoogleCode on 2014-10-30 22:15Z] fixed, many thanks Status: Issue closed
mobileappdevhm/cie-server-team-2
327220225
Title: Course management interface Question: username_0: The management interface (angular app) should have the possibility to manage courses: - Add course (Should have the same information as in the pdfs and the table on hm.edu) - Delete course - Edit course<issue_closed> Status: Issue closed
linkerd/linkerd-examples
243592265
Title: Need dnsPolicy: ClusterFirstWithHostNet when deploying for CNI + zipkin Question: username_0: At least for kubernetes 1.7 one has to specifically request `dnsPolicy: ClusterFirstWithHostNet` for linkerd container in deployment config [linkerd-cni.yml](https://github.com/linkerd/linkerd-examples/blob/master/k8s-daemonset/k8s/linkerd-cni.yml) if one wants to use zipkin telemetry ```yaml ... telemetry: - kind: io.l5d.zipkin host: zipkin-collector.default.svc.cluster.local port: 9410 sampleRate: 1.0 ... ``` Otherwise due to `hostNetwork: true` Pod specifier (for CNI deployment) the `dnsPolicy: default` is applied so that "cluster" addresses (`....svc.cluster.local`) are not resolved because `default` dns policy is to use node's host /etc/resolv.conf which may or may not (as it is in my case) refer to kube-dns for name resolution, at least this kind (the latter) of resolv.conf setup is done by kubeadm. Anyways this clarification must be stated somewhere (https://discourse.linkerd.io/t/flavors-of-kubernetes/53 ?) P.S. I do not know since what kubernetes version dns policy `ClusterFirstWithHostNet` is available. Answers: username_1: First available in 1.6 Should we just add `dnsPolicy: ClusterFirstWithHostNet` to the linkerd-cni.yml? thoughts @username_2? username_2: @username_1 Yeah, that sounds like the right fix to me, but am also worried it will make the configs backwards incompatible with earlier Kubernetes versions. Hmm, maybe we need a separate linkerd-cni.yml config file for 1.6+. username_1: OK let's do that. As soon as linkerd-cni.yml is split into linkerd-cni.yml and linkerd-cni-legacy.yml, I'll update the k8s flavor page to reflect it username_3: OMG thank you for this. I couldn't get the new service_mesh.yml to work with CNI and Zipkin and this solved it!!! username_1: oh snap! we should def add this to the config then! Status: Issue closed
STAT545-UBC-hw-2019-20/stat545-hw-armetcal
510905612
Title: Peer Review for Assignment 5 Question: username_0: Topic | Excellent | Satisfactory | Needs Work -- | -- | -- | -- Coding style | ✔️ |   |   Coding strategy | ✔️ |   |   Presentation: graphs | ✔️  | |   Presentation: tables | ✔️ |   |   Achievement, creativity | ✔️ |   |   Ease of access/reading | ✔️ |   |   - I really like your use of the table contents/index on the left-hand side. - I also liked that you included an example of how to use the here package within Exercise 1. - I never thought to create a comparison table for Exercise 2, but it made it easier to compare levels and rows between the original and modified dataset. - It was useful to compare the 'written' and 'read' datasets to each other in Exercise 3 and I learned how to modify classes from your work. - I appreciated how you played with the options in Exercise 5 and was able to show the reader how to use different parameters. Awesome last assignment :) @armetcal
fullcontent/sistemacelic
1051471416
Title: [BUG] VINCULAR PENDÊNCIAS Question: username_0: Bruno, boa noite Preciso vincular 2 pendências mas o sistema não mostra o número da O.S. que eu preciso. Por exemplo: Tenho a O.S. AC0106 que refere-se ao Alvará de Funcionamento. Possuo dois outros serviços que só poderão ser iniciados após finalização do serviço AC0106, são eles: AC0108 e AC0658. O correto seria: Entrar na pendência das duas O.S's citadas acima e vincular ambas à O.S. AC0106. Mas na listagem suspensa não puxa. Vejamos abaixo: 1- O.S. AC0106 - Serviço "Alvará de Funcionamento" que está tramitando. ![image](https://user-images.githubusercontent.com/60766925/141388215-859bca4b-a727-4330-9227-8076b78f4739.png) 2- O.S. AC0108 - Serviço "Alvará de Pubicidade" - Só posso iniciar após finalização da AC0106 ![image](https://user-images.githubusercontent.com/60766925/141388246-ac164710-660e-46ef-9f6e-724bec5e626f.png) 3- O.S. AC0658 - Serviço "Licença Ambiental" - Só posso iniciar após finalização da AC0106 ![image](https://user-images.githubusercontent.com/60766925/141388261-e8ab4549-c934-4274-a334-7dd13fd71b10.png) Abaixo, coloco as duas telas onde mostra a lista suspensa para vincular e NÃO CONSTAM a AC0106 que eu necessito. 1- O.S. AC0108 - Serviço "Alvará de Pubicidade" - Lista suspensa/vínculo de pendências ![image](https://user-images.githubusercontent.com/60766925/141388372-1667889f-c85e-4c83-8cb5-4cd399351a0f.png) 2- O.S. AC0658 - Serviço "Licença Ambiental" - Lista suspensa/vínculo de pendências ![image](https://user-images.githubusercontent.com/60766925/141388351-37a7686d-d50a-43a1-8a46-5eb7993ddc60.png) Consegue verificar, por gentileza? Obrigada! Status: Issue closed Answers: username_1: @username_0 Corrigido
GothenburgBitFactory/taskwarrior
297224146
Title: [TW-1850] Adding a task with \' between \" breaks sync with taskserver Question: username_0: _<NAME> on 2016-09-06T16:27:56Z says:_ I added a task. My desired description was: Blah blah "quoted text that says Don't" So I run: task add "Blah blah \"quoted text that says Don\'t panic and carry a towel\"" +tag1 +tag2 and every sync after that failed with error: The Taskserver returned error: 500 Unrecognized username_0 file format. otherwise, username_0 seemed to be OK. STEPS I WENT THROUGH: This was one of 629 tasks that I added one after the other with a script. I did not check how that "don't" looked like. I did a: task export to save all those new tasks (I didn't need it later), then went into the .taskrc directory and removed all the .data files. Then a sync worked and brought all my tasks from before that crazy script, from taskserver into username_0. Then I ran the first 50 adds, and a sync (succesful) then 50 more adds and another sync. Until one sync gave the same error. Deleted the *.data files again. sync to get all the tasks from the server (including the first few batches of 50). And did 10 adds, sync, 10 adds, sync until it crashed again. Deleted *.data, sync to get everything back except the last batch of 10, and went one by one, until it crashed on the "don't" task. That time it gave me a different error, but it was too late at night and I didn't save it. Removing the "\" before the " ' " fixed the problem, and the rest of the adds went OK, and now it's syncing with all and the 629 tasks. SUGGESTIONS: Something is not getting escaped as it should. If that \ wasn't necessary, I should have got it as part of the description (perhaps escaped as \\ in the file...), but the sync shouldn't break. This way of deleting all the .data files is a bit violent, I lost the undo and perhaps something else (I don't care), but I think it could be refined a bit and made into a standard procedure recommended somewhere in the help as a way of getting over these sync problems that seem to be relatively common... perhaps with an export with some format before... but it solves the problem, and helps isolate the cause. Answers: username_0: Migrated metadata: ``` Created: 2016-09-06T16:27:56Z Modified: 2017-10-12T03:02:20Z ``` username_1: This might be fixed by #2588 - I cannot really reproduce, since it seems that the backslashes in the description were not merged to github. Status: Issue closed username_2: I agree, I think this is actually a duplicate. The description matches the underlying problem exactly (and in the previous versions, this would corrupt the data files). username_2: Duplicate of #2189.
minbrowser/min
389235743
Title: Issue when using Ctrl-F to search Question: username_0: First, I use ctrl+f and type in the word I want to search for. After that I want to scroll up and down the page to see where the highlighted comments are. As of now when I start to scroll using the **scroll bar** on the right the search goes away.I think the preferred outcome would be for the text to stay highighlighted until you cleared the search or decided to search for something else.Thoughts?? Answers: username_1: Right, that probably would be better. If you start interacting with the page itself, I think we do want to hide the find bar, but we need some way to detect clicks on the scrollbar vs the actual page, and ignore the scrollbar ones. username_2: How about an invisible full-width and full-height container to catch the page clicks? This way it will catch all clicks except for the scrollbar. username_1: We can't stack anything on top of the BrowserView, but we could probably just attach an event listener to the document from within the preload script, and send an IPC message to the UI process. If we do that, we'd also need to figure out what other scenarios the find bar should be hidden in as well - right now, we just hide it as soon as the text input loses focus, but if we do this we'll need to also hide it whenever you switch tabs, open the task overlay, etc. Status: Issue closed
dotnet/core
591587207
Title: /usr/share/dotnet/sdk/3.1.101/NuGet.targets(123,5): error : 'vxxxx.xx.xx' is not a valid version string. (Parameter 'value') Question: username_0: I am getting This error /usr/share/dotnet/sdk/3.1.101/NuGet.targets(123,5): error : 'v2020.03.16' is not a valid version string. (Parameter 'value') "unset version" didn't work as well, per _Originally posted by @jbenguira in https://github.com/dotnet/core/issues/3909#issuecomment-558856208_ Answers: username_1: Can you share your project files and whatever command line you're running that shows this? Thanks. username_0: ~> dotnet new console --output sample1 The template "Console Application" was created successfully. Processing post-creation actions... Running 'dotnet restore' on sample1/sample1.csproj... /usr/share/dotnet/sdk/3.1.101/NuGet.targets(123,5): error : 'v2020.03.16' is not a valid version string. (Parameter 'value') [~/sample1/sample1.csproj] Restore failed. Post action failed. Description: Restore NuGet packages required by this project. Manual instructions: Run 'dotnet restore' ~> cd sample1 ~/sample1> dotnet restore /usr/share/dotnet/sdk/3.1.101/NuGet.targets(123,5): error : 'v2020.03.16' is not a valid version string. (Parameter 'value') [~/sample1/sample1.csproj] username_2: @wli3 @username_3 would you be the best people to help with this? username_3: @username_0 That is very strange. What do the contents of sample1.csproj look like? Could you also run `dotnet restore -bl` to create a binary log and share that? Be aware that this will include information from your machine such as environment variables, paths, etc. See here for more info: http://aka.ms/binlog username_0: [sample1.zip](https://github.com/dotnet/core/files/4457774/sample1.zip) username_2: ping @username_3 username_3: @username_0 There's nothing special in the project file, so this is probably something in your environment. I think I'd need a [binlog](http://aka.ms/binlog) to be able to investigate what's going on. username_4: Closing due to the lack of response after two weeks or more. Status: Issue closed username_5: The same thing happened to me, and it was due to me setting up the Environment Variable "VERSION". Solved with with `set VERSION=`
FabianIAM12/memory-game
343642437
Title: Higher Threshold Question: username_0: https://github.com/username_1/memory-game/blob/80df08ecb7818cd272535babc141e79ad2b6e810/js/app.js#L45 You should really consider to raise the threshold needed to get three stars. 11 moves might be a good start. Answers: username_1: To prevent high frustration with bad skilled noobs, i did take 13. Thanks for your contribution. I'm looking forward for your first pull request. Status: Issue closed
CNMAT/CNMAT-odot
107776849
Title: interleave(?) crash when dealing ridiculously large lists Question: username_0: this came up in some homework that I just received. not sure if it's interleave or not. and, no, I'm not saying this is a good idea : ) but probably we shouldn't crash. <pre><code> ----------begin_max5_patcher---------- <KEY> -----------end_max5_patcher----------- </code></pre> Answers: username_1: Doesn't crash for me. We can't predictively determine when a list or something will exhaust the memory. { "version" : "Version 7.0.4 (cbb83ff) (32-bit mac)", "platform" : "mac", "arch" : "x86", "osversion" : "Mac OS X Version 10.10.4 x86_64", "samplerate" : 44100, "iovs" : 1024, "sigvs" : 1024, "scheduler_in_audio_interrupt" : "off", "audio_drivername" : "Core Audio", "audio_driver_subname" : "", "license" : "permanent full", "machine_id" : "9e39a8f2dd16c55c866911287cd55401", "eventinterval" : 1, "overdrive" : "off", "mixerparallel" : "off", "mixercrossfade" : 0, "mixerlatency" : 30.0, "mixerramptime" : 10.0, "videoengine" : "avf", "packages" : { "Beap" : { "name" : "BEAP", "type" : "package", "author" : "<NAME>, <NAME>", "description" : "BEAP (Berklee Electro Acoustic Pedagogy) Modular is a synthesis pedagogical tool.", "version" : "Master from GitHub", "major_version" : "0", "minor_version" : "0", "min_max_version" : "612", "min_osx_version" : "None", "min_win_version" : "None", "website" : "https://github.com/stretta/BEAP/wiki", "link_mac32" : "https://github.com/stretta/BEAP/archive/master.zip", "link_mac64" : "https://github.com/stretta/BEAP/archive/master.zip", "link_win32" : "https://github.com/stretta/BEAP/archive/master.zip", "link_win64" : "https://github.com/stretta/BEAP/archive/master.zip", "relative_path" : "None", "path" : "Macintosh HD:/Applications/Max.app/Contents/Resources/C74/packages/Beap", "vol" : -682 } , "Gen" : { "name" : "Gen", "path" : "Macintosh HD:/Applications/Max.app/Contents/Resources/C74/packages/Gen", "vol" : -683 } , "Jitter" : { "name" : "Jitter", "path" : "Macintosh HD:/Applications/Max.app/Contents/Resources/C74/packages/Jitter", "vol" : -684 } , "Max for Live" : { "name" : "Max for Live", "path" : "Macintosh HD:/Applications/Max.app/Contents/Resources/C74/packages/Max for Live", "vol" : -685 [Truncated] "vol" : -689 } , "vizzie" : { "name" : "vizzie", "path" : "Macintosh HD:/Applications/Max.app/Contents/Resources/C74/packages/vizzie", "vol" : -690 } , "Max" : { "name" : "Max" } , "MSP" : { "name" : "MSP" } } } Status: Issue closed username_0: ok, not crashing here now too.
kumarrk21/S1QuotePDF
537058198
Title: Code Coverage Issue Question: username_0: I am trying to implement this package in production and am only getting 72% coverage.... can I propose a controller test class? I have little experience but my understanding is that if we write a test class for the controller we will get better coverage...I am using this package to deliver pdf's via mobile.. I already modified it to be a global action by eliminating the standard controller and getting the quote id from page referrence... Answers: username_0: updated controller `public with sharing class S1QuotePDFController { public Quote Quote{get;set;} public List<Quote> Quote2; public S1QuotePDFController() { String qId = ApexPages.currentPage().getParameters().get('id'); if (qId == null) {Quote = new Quote();} else {Quote = [Select Id FROM Quote Where Id = :qId];} } @AuraEnabled public static String getPDFData(String quoteId){ returnMessage ret = new returnMessage(); ret.success = false; List<S1QuotePDFSettings__mdt> tsList = [SELECT Value__c FROM S1QuotePDFSettings__mdt WHERE Label ='Template ID']; S1QuotePDFSettings__mdt ts = new S1QuotePDFSettings__mdt(); if(tsList.size()>0){ ts = tsList[0]; } try{ String pdfURL = '/quote/quoteTemplateDataViewer.apexp?id='+quoteId+'&summlid='+ts.Value__c; PageReference pdf = new PageReference(pdfURL); String b64 = EncodingUtil.base64Encode(pdf.getContent()); ret.success=true; ret.message = b64; }catch(Exception e){ ret.message = JSON.serialize(e.getMessage()); } return JSON.serialize(ret); } @AuraEnabled public static String emailPDF(String quoteId, String emailId, String pdfData){ returnMessage ret = new returnMessage(); ret.success = false; //Implement send email logic here.. ret.success=true; ret.message = 'Email Sent Successfully'; return JSON.serialize(ret); } @AuraEnabled public static String savePDF(String quoteId, String pdfData){ returnMessage ret = new returnMessage(); ret.success = false; try{ Quote q = [Select Name from Quote where id=:quoteId]; QuoteDocument qd = new QuoteDocument(); qd.document = EncodingUtil.base64Decode(pdfData); qd.quoteId = quoteId; insert qd; [Truncated] } } update apex:page `<apex:page docType="html-5.0" showHeader="false" sidebar="false" standardStylesheets="false" applyHtmlTag="false" applyBodyTag="false" controller="S1QuotePDFController" >` with these adjustments got it deployed in production but the header sits on top of other parts of template Status: Issue closed
MicrosoftDocs/OfficeDocs-Enterprise
614733772
Title: Need a page like this for Teams Question: username_0: It is not easy to find the documentation on how to connect to Teams with PowerShell for an account with MFA enabled. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 0f318af5-7da8-daec-d038-7bca3d4cabf8 * Version Independent ID: 7cb9bf9d-5b01-7cbd-fa54-cb9847b61fc2 * Content: [Manage Skype for Business Online with Office 365 PowerShell](https://docs.microsoft.com/en-us/office365/enterprise/powershell/manage-skype-for-business-online-with-office-365-powershell#feedback) * Content Source: [Enterprise/powershell/manage-skype-for-business-online-with-office-365-powershell.md](https://github.com/MicrosoftDocs/OfficeDocs-Enterprise/blob/live/Enterprise/powershell/manage-skype-for-business-online-with-office-365-powershell.md) * Service: **o365-administration** * GitHub Login: @username_1 * Microsoft Alias: **josephd** Answers: username_1: @username_0 A new page like this one has been created at https://docs.microsoft.com/en-us/office365/enterprise/powershell/manage-microsoft-teams-with-office-365-powershell. Thanks for the suggestion. Status: Issue closed
schmunk42/yii2-giiant
317269379
Title: View not Found in Question: username_0: Go to web gii and click "Giiant Test **Start**" Controller /gii/default/view?id=giiant-test I got error: The view file does not exist: *******\vendor\username_1\yii2-giiant\src\generators\test/form.php Status: Issue closed Answers: username_1: I added a very basic view, improvements and PRs welcome!
gaois/terminologue
473976047
Title: Teachtaireacht earráide mhíthreorach Question: username_0: Má aighnítear seoladh ríomhphoist atá in úsáid [anseo](https://www.terminologue.org/signup/), faightear an teachtaireacht earráide 'Seoladh ríomhphoist mícheart.'/'Incorrect e-mail address.' Ba cheart an teachtaireacht earráide 'Seoladh ríomhphoist in úsáid.'/'E-mail address in use.' a fháil. Answers: username_1: Tá sé seo ceartaithe anois. Status: Issue closed
preactjs/preact
913729162
Title: Cannot find name 'h' with ts-loader Question: username_0: - [ ] Check if updating to the latest Preact version resolves the issue I already have a website on React and would like to migrate to Preact. My website is bundling with Webpack and ts-loader. And that's a problem, because I could find how to migrate from React to Preact if I don't use babel. For example, [here](https://preactjs.com/guide/v10/getting-started#typescript-preact/compat-configuration) are the steps for migrating using babel. But editting `.babelrc` will not affect in my case. So, I decided to search the Internet and as I understand such configuration should have helped: **tsconfig.json** ``` { "compilerOptions": { ... "jsx": "react", "jsxFactory": "h", "jsxFragmentFactory": "Fragment", }, ... } ``` **webpack.config.js** ``` module.exports = (env = {}, argv = {}) => { return { ... resolve: { extensions: ['.tsx', '.ts', '.js'], alias: { ... 'react': 'preact/compat', 'react-dom': 'preact/compat', }, }, } }; ``` But it didn't work. The building process ends up with en error: `TS2304: Cannot find name 'h'.` <details> <summary><b>tsconfig.json</b> полностью</summary> <pre> { "compilerOptions": { "module": "es2020", "target": "es5", "skipLibCheck": true, "jsx": "react", "jsxFactory": "h", "jsxFragmentFactory": "Fragment", "allowJs": true, "sourceMap": true, "moduleResolution": "node", "experimentalDecorators": true, "allowSyntheticDefaultImports": true, "resolveJsonModule": true, "baseUrl": "./", "paths": { "@components": ["components"], "@components/icons": ["components/icons"], "@utils": ["utils"], "@pages/*": ["views/pages/*"], "@assets/*": ["assets/*"], "@data/*": ["data/*"] } }, "include": ["./**/*"], "exclude": [ "node_modules/**/*", "app/**/*" ] } </pre> </details> Answers: username_1: The classic JSX mode (=`"react"`) doesn't add imports for the jsx pragma function automatically. Those need to be imported by hand if the newer runtime mode isn't an option. The new runtime mode which adds those imports automatically can be enabled by by setting `jsx` to either `react-jsx` or `react-jsx-dev`. ```json { "compilerOptions": { "jsx": "react-jsx", "jsxImportSource": "preact", } } ``` Status: Issue closed
auth0/auth0-PHP
291995560
Title: Double Login / Logout Question: username_0: I've come across a strange issue that seem to allow a user to login twice and then requires logging out twice. This happens when: • a user logs in to the secured area of the app • then navigates back with the browser back button. • This prompts the 'Last time you logged in with' login screen which allows you to then login using the remembered account • This returns you to the secured area. • If you then click logout (which uses the logout url endpoint + sdk logout method) you are logged out • clicking the browser back button at this point still returns you to the secured area as if you are logged in. • Logging out a second time then means the back button no longer returns you to the secured area and instead prompts the login screen. After logging out the first time and navigating back I can see that auth0__access_token stored in the session is now different. Status: Issue closed Answers: username_1: @username_0 - We added state validation in [5.1.0](https://github.com/auth0/auth0-PHP/releases/tag/5.1.0) and when I test with that version, I get a fatal `Invalid state` error, as expected. This might solve your issue, though maybe not in the way you expect. To summarize what I'm doing/seeing on my end: 1. Load my local app and clear out all my cookies for that domain 1. Click **Login** which calls `Auth0()->login()` 1. I land on `https://my-domain.auth0.com/login?&response_type=code&state=RANDOM_STRING&etc...` with a prompt to login 1. Enter credentials and submit, get redirected back to my local app here `http://php.localhost.test/?code=AUTH_CODE&state=RANDOM_STRING` 1. Now I'm logged in so I click **Back** and get that "Last time you logged in with..." message 1. Click my credential and get redirected back to my local app here `http://php.localhost.test/?code=DIFFERENT_AUTH_CODE&state=RANDOM_STRING` (note that the state is the same in the second URL because the same one was in the URL) 1. Click **Logout** which calls `Auth0()->logout()` 1. Now I'm logged out so I click **Back** and I get the fatal state error because it's trying to check the same state value that was already validated and deleted All that to say ... 5.1.0 gets this to work as expected. I'm going to close this but please feel free to report back if you're having the same/similar issue or need help moving to 5.1.0. username_0: Ok thanks @username_1 I'm unable to test right now but the sequence / outcome you described above certainly sounds like it should be fixed. Will report back if theres any related issues though.
highsource/jaxb2-basics
261036154
Title: simplify: Bad namespace URL http://jaxb2-commons.dev.java.net/basic/simplify Question: username_0: java.net has been shutdown. This seems to cause issues as the namespace url is no longer valid. http://jaxb2-commons.dev.java.net/basic/simplify I cannot find that file. Makes anything using it, not able to be usable via xjc/jaxb2. Thanks! Answers: username_1: `http://jaxb2-commons.dev.java.net/basic/simplify` *is* the valid namespace URI. Namespaces are URIs ("identifiers"), *not* URLs ("resource locators"). A namespace URI does not need to point to some existing resource. It *may* but that is completely optional. As the matter of fact, there have never been any resource under `http://jaxb2-commons.dev.java.net/basic/simplify`, there is no (and never was) a "file" there. So the shutdown of java.net is completely irrelevant and has no influence on functionality of the plugin in xjc/jaxb2. Now you seem to have a problem that you thought have been caused by "invalid namespace URL". Why don't you start with describing that problem - like, how you execute the plugin (ideally a build script/Maven pom), which error messages you get (complete log) etc. username_0: Ok. Not to get into semantics but a URL is a URI. The fact that it has http makes it a URL, http://. I was not sure if the missing namespace mattered or not. I guess it does not matter. I am looking into building the plugins and using them. Seems they reside in the [javaee/jaxb2-commons](https://github.com/javaee/jaxb2-commons) repository. Once I build and package them, I can see about using the Simplify plugin. Maybe then the namespace will not matter. I think that is the first time I have come across a namespace/schema that has no source. username_1: Or you can get the JAR here: http://repo1.maven.org/maven2/org/jvnet/jaxb2_commons/jaxb2-basics/0.11.1/ As it seems there was no actual problem with the plugin, I'm closing this issues as invalid. username_0: I am packaging this stuff on Gentoo, building jars, not using them. This is actually something Hibernate is using. I was not sure what this repo relationship was to the plugin. I did not see it on maven.org, at least version wise etc. What is the relation between this one and the one under javaee? username_1: Historically JAXB2-Basics was a subproject of JAXB2 Commons (hosted on java.net). Later on I've moved my projects from java.net to GitHub, but some older repos probably remained on java.net. Status: Issue closed
MicrosoftDocs/sql-docs
486092783
Title: Error in Continue Scripting on Error documentation Question: username_0: "Continue scripting on error - When True, scripting stops when an error occurs. When False, scripting continues. The default is False." I think the True and False need to be switched around for this description. That or the option name is horribly, HORRIBLY named. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 222643c9-2ac1-11f6-4913-5e2e5216dfe6 * Version Independent ID: 7bb94693-57be-7775-0f40-ce7681a8b655 * Content: [Generate and Publish Scripts Wizard - SQL Server](https://docs.microsoft.com/en-us/sql/ssms/scripting/generate-and-publish-scripts-wizard?view=sql-server-2017#feedback) * Content Source: [docs/ssms/scripting/generate-and-publish-scripts-wizard.md](https://github.com/MicrosoftDocs/sql-docs/blob/live/docs/ssms/scripting/generate-and-publish-scripts-wizard.md) * Product: **sql** * Technology: **scripting** * GitHub Login: @markingmyname * Microsoft Alias: **maghan**
genicam/harvesters
347355771
Title: Support Bayer16 formats Question: username_0: **Is your feature request related to a problem? Please describe.** Support the following Bayer16 formats: * ``BayerGR16`` * ``BayerRG16`` * ``BayerGB16`` * ``BayerBG16`` **Describe the solution you'd like** Harvester Core should support those formats. **Describe alternatives you've considered** None. **Additional context** None. Answers: username_1: @username_0 I'm only able to test `BayerRG16` with our camera, but it works. :+1: Status: Issue closed username_0: For people who those confirmed any of pixel formats didn't work: Please reopen this issue again once you noticed it.
hoelzro/lua-repl
59660764
Title: pretty_print plugin blows up on tables with functions as keys Question: username_0: lua5.1: ./repl/plugins/pretty_print.lua:114: attempt to compare two function values stack traceback: ./repl/plugins/pretty_print.lua:114: in function <./repl/plugins/pretty_print.lua:109> [C]: in function 'tsort' ./repl/plugins/pretty_print.lua:134: in function 'sortedpairs' ./repl/plugins/pretty_print.lua:204: in function 'dump' ./repl/plugins/pretty_print.lua:251: in function 'displayresults' repl/init.lua:103: in function 'old_value' repl/init.lua:227: in function 'handleline' ./repl/sync.lua:33: in function 'run' rep.lua:51: in main chunk [C]: ? ``` This is because Lua only defines less than for numbers, strings, or anything with an `__lt` metamethod. We could check for this explicitly, or we could wrap the check in a `pcall`; I don't know how poorly the latter would perform. Answers: username_1: Same thing for tables.
DiscipleTools/disciple-tools-mobile-app
512205804
Title: Inconsistent icons Question: username_0: Have the same icons that represent the same fields in the app that the website uses. This is what the website uses... <img width="544" alt="Screenshot 2019-10-24 17 07 00" src="https://user-images.githubusercontent.com/43966676/67529121-dae1f900-f680-11e9-8c9c-43827be49a26.png"> This is what the app uses at present... ![Screenshot_20191024-170450](https://user-images.githubusercontent.com/43966676/67529135-ecc39c00-f680-11e9-8028-79f783c01679.png)<issue_closed> Status: Issue closed
iamswain25/political-life
475571553
Title: 博主为什么维护这么奇怪的项目? Question: username_0: 今天闲来无事在网上乱逛,莫名其妙的的进入了这个repo,看了几篇感觉还蛮有意思,但是谈不上多少有意思的内容(毕竟只是日记)。不知道博主为什么要把这本书做成一个网页还翻译成各种语言?果然世界这么大,什么样奇怪的人都有啊 Answers: username_1: @username_0 书已绝版了,不能正常翻译出版。你去问问王沪宁为什么将它绝版。 username_0: 书绝版很正常,但你为什么要把他放到网上呢?感觉维护这个还是要耗不少精力的啊 username_1: @username_0 初次做好以后,维护不耗精力。github免费,我就闲时间没事翻译翻译。 username_2: 牛逼了 username_2: 这个书讲了啥? username_1: 讲了当代上海人的精神境界 Status: Issue closed
dotnet/efcore
554174340
Title: Type-Generic Where Clause unable to be translated when using projections (EF Core 3.1.1) Question: username_0: <!-- Describe what isn't working as expected --> When trying to access a defined property on a generic, projected type, the object in the `Where` clause lambda is unable to be translated. ```C# public virtual async Task<IEnumerable<T>> GetAsync(IEnumerable<Guid> ids) => await this.repository.Read().Where(r => r.Id.ToString().Contains("0")).ToListAsync(); ``` Outputs: ``` System.InvalidOperationException HResult=0x80131509 Message=The LINQ expression 'DbSet<Person> .Where(p => new Person{ ContactNumber = p.ContactNumber, Email = p.Email, FullName = p.FullName, Id = p.Id } .Id.ToString().Contains("0"))' could not be translated. Either rewrite the query in a form that can be translated, or switch to client evaluation explicitly by inserting a call to either AsEnumerable(), AsAsyncEnumerable(), ToList(), or ToListAsync(). See https://go.microsoft.com/fwlink/?linkid=2101038 for more information. Source=Microsoft.EntityFrameworkCore ``` ### Steps to reproduce ``` C# //service class, either method fails public abstract class Service<T> : IService<T> where T : class, ICommonModel { protected readonly IRepository<T> repository; public Service(IRepository<T> repository) => this.repository = repository; public virtual async Task<IEnumerable<T>> GetAsync(IEnumerable<Guid> ids) => await this.repository.Read().Where(r => ids.Contains(r.Id)).ToListAsync(); public virtual async Task<IDictionary<Guid, T>> GetDictionaryAsync(IEnumerable<Guid> ids) => await this.repository.Read().Where(r => r.Id.ToString().Contains("0")).ToDictionaryAsync(i => i.Id, i => i); } } //repo class public abstract class Repository<TCommon, TData> : IRepository<TCommon> where TCommon : class, ICommonModel where TData : class, IDataModel { protected DenaliContext Context { get; } protected IMapper Mapper { get; } public Repository(AppContext context, IMapper mapper) { this.Context = context; this.Mapper = mapper; } public virtual IQueryable<TCommon> Read() => this.Context.Set<TData>().ProjectTo<TCommon>(this.Mapper.ConfigurationProvider); [Truncated] Microsoft.EntityFrameworkCore.Query.QueryableMethodTranslatingExpressionVisitor.VisitMethodCall.__CheckTranslated|8_0(Microsoft.EntityFrameworkCore.Query.ShapedQueryExpression, ref Microsoft.EntityFrameworkCore.Query.QueryableMethodTranslatingExpressionVisitor.<>c__DisplayClass8_0) Microsoft.EntityFrameworkCore.Query.QueryableMethodTranslatingExpressionVisitor.VisitMethodCall(System.Linq.Expressions.MethodCallExpression) Microsoft.EntityFrameworkCore.Query.RelationalQueryableMethodTranslatingExpressionVisitor.VisitMethodCall(System.Linq.Expressions.MethodCallExpression) System.Linq.Expressions.MethodCallExpression.Accept(System.Linq.Expressions.ExpressionVisitor) Microsoft.EntityFrameworkCore.Query.QueryableMethodTranslatingExpressionVisitor.VisitMethodCall(System.Linq.Expressions.MethodCallExpression) Microsoft.EntityFrameworkCore.Query.RelationalQueryableMethodTranslatingExpressionVisitor.VisitMethodCall(System.Linq.Expressions.MethodCallExpression) System.Linq.Expressions.MethodCallExpression.Accept(System.Linq.Expressions.ExpressionVisitor) ... [Call Stack Truncated] ``` ### Further technical details EF Core version: 3.1.1 Database provider: Microsoft.EntityFrameworkCore.SqlServer Target framework: .NET Core 3.1 Operating system: Windows IDE: Visual Studio 2019 16.4.3 Answers: username_1: @username_2 to take a look. username_1: Pinging @username_2 username_2: Seems to be working in current master. I'm using the following repro: ```cs [ConditionalFact] public void Repro19679() { using var ctx = new Context19679(); ctx.Database.EnsureDeleted(); ctx.Database.EnsureCreated(); var query = ctx.People.Where(p => new Person19679 { ContactNumber = p.ContactNumber, Email = p.Email, FullName = p.FullName, Id = p.Id }.Id.ToString().Contains("0")).ToList(); } public class Context19679 : DbContext { public DbSet<Person19679> People { get; set; } protected override void OnConfiguring(DbContextOptionsBuilder optionsBuilder) { optionsBuilder.UseSqlServer(@"Server=.;Database=Repro19679;Trusted_Connection=True;MultipleActiveResultSets=True"); } } public class Person19679 { public int Id { get; set; } public string ContactNumber { get; set; } public string Email { get; set; } public string FullName { get; set; } } ``` which yields the following sql: ```sql SELECT [p].[Id], [p].[ContactNumber], [p].[Email], [p].[FullName] FROM [People] AS [p] WHERE CHARINDEX(N'0', CONVERT(VARCHAR(11), [p].[Id])) > 0 ``` username_2: @username_0 is there more to your query? Translation for ToString should have been present in 3.1.1, so likely something else is wrong. Ideally, please submit full runnable repro, so it's easier for us to pin it down. Status: Issue closed username_1: **EF Team Triage:** Closing this issue as the requested additional details have not been provided and we have been unable to reproduce it. *BTW this is a canned response and may have info or details that do not directly apply to this particular issue. While we'd like to spend the time to uniquely address every incoming issue, we get a lot traffic on the EF projects and that is not practical. To ensure we maximize the time we have to work on fixing bugs, implementing new features, etc. we use canned responses for common triage decisions.* username_3: WTF? This is one critical issue and it is super easy reproduced in ef 3.1.1. I am fighting this all the time: ---> System.InvalidOperationException: The LINQ expression 'DbSet<OBTOBDocument> .Where(o => !(o.BIsDeleted) == True) .Where(o => new Document{ Created = o.DCreateDate, Description = o.SDescription, FileType = o.SMimeType, Id = o.PkOBDocumentID, Identity = o.GIdentity, IsDeleted = o.BIsDeleted, IsVirtual = o.BIsVirtual, Name = o.SOrigFilename, Size = o.ISizeInkBytes } .Identity == __key_0)' could not be translated. Either rewrite the query in a form that can be translated, or switch to client evaluation explicitly by inserting a call to either AsEnumerable(), AsAsyncEnumerable(), ToList(), or ToListAsync(). See https://go.microsoft.com/fwlink/?linkid=2101038 for more information. Saying "Seems to be working in current master." is an insult to everyone who has to ship something in the timeframe BEFORE 5.0 some point in half a year. This is a regression from 2.2 down into 3.1. it is possibly FIXED in 5.0, but that fix is half a year down the road. username_4: May be every issue critical issue. username_3: The funny thing with EfCore is that the moment you actually try to use it you find out that it is so full of bugs it is not even an alpha. Odata was bad for a time, but the EfCore puts up a new fight on making sure that everyone realizes how much behind even Ef efcore is not - what is it? 10 years? Have you see n the amazing number of bugs to be fixed next release? Hint: NOTHING. The team is happil working on the next major release. 2.1: Wait . 2.2: Wait 3.0: Wait 3.1: Wait 5.0? Will it finally handle simple LINQ cases? All the while Ef works and works and works and chugs along. And people get told "hey, we fixed that, it does not happen in the branch to be released in more than half a year, you know, go and play with dolls". A case for HR, it seems - there is somethign VERY odd with the way the EfCore teams hadnels this. Paid by Oracle to make MS look bad? username_0: @username_2 your example is using a defined type. If you abstract your person object to a generic the query will fail. And I'm starting to agree with @username_3, we are really starting to regret using EF Core for the ORM for a product that is supposed to be shipping this summer. We will spend more time fighting EF than writing functional code on a lot of our development tasks, and it seems there is no hope to getting any issues addressed. username_3: @username_0 I have to disagree with you here. There is hope. Hope in the name of 5.0. As someone from the financial markets let me tell you why there is hope. HOPE DIES LAST. When you have moved off to a better product and EfCore is not fixed in 10 years, there STILL is hope left. And this is the ONLY reason why there is still hope. username_4: Duplicate of #19087 username_3: And the amazing thing here is: This is a one line change with unit tests that was merged in in December and is not in any bug fix. Maybe there would be less bug reports if bug fixes would be deployed? Oh, common sense, you are sooooo lost.
vuejs/vetur
363877268
Title: Autocomplete not work with update 0.12.7 Question: username_0: - [x ] I have searched through existing issues - [x ] I have read through [docs](https://vuejs.github.io/vetur) - [ x] I have read [FAQ](https://github.com/vuejs/vetur/blob/master/docs/FAQ.md) ## Info - Platform: <!-- Win/macOS/Linux --> - Vetur version: 0.12.7 - VS Code version: 1.27.2 ## Problem Vue output langugague serve ``` Vetur initialized `` I can autocomplete in .vue file with update 0.12.7. With 0.12.1 before every normal Vetur is package very bad :( ! Issue everywhere Answers: username_1: +1 username_2: 0.12.6 is ok username_3: removing node_modules and clearing npm cache fixed it for me. username_0: I try, but it don't work. I installed v0.12.6, it work ok ! Update 0.12.7 is very bad ... :( username_4: For me the same, have to Install the 0.12.6 VSIX and disable automatic updates to make it work as expected. username_5: @username_0 Since you never mention any details of what goes wrong, I can't help much. Status: Issue closed
wolfika/scirocco
115678340
Title: Create watch tasks Question: username_0: Add watch tasks to Gulp to make development easier. It should be able to transpile the scripts, copy the HTML templates, and compile SCSS. Linting could be left out of this task to make it faster. The renderer process (browser window) should be reloaded (?) when changes are made to the files.<issue_closed> Status: Issue closed
openebs/openebs-docs
509407017
Title: `python-software-properties` has been replaced with `software-properties-common` Question: username_0: The command ``` sudo apt-get install python-software-properties ``` in README.md wouldn't work as the package has now been replaced with `software-properties-common`. A good first issue for beginners.<issue_closed> Status: Issue closed
marcone/teslausb
976471851
Title: Random missing time from recordings Question: username_0: I've noticed this happen a number of times. I will go look for a recording and find a missing block of time. Here is one example. I was definitely driving for most of that missing time. In fact I took a photo of something interesting, and then decided to go look for the video. So I have a photo timestamped 16:34 but that block of time is missing from the recordings. This isn't the only time, but it's one where I had complete assurance that I had indeed been driving at the time. ![image](https://user-images.githubusercontent.com/39809/130370194-0b865a14-ec80-498b-805e-570fcc6cbff7.png) Answers: username_1: The script that the car uses to save recordings to the USB storage (https://twitter.com/greentheonly/status/1214669344933896192) simply drops the recordings when there's any kind of error. This seems consistent with that, and it's not something teslausb can do anything about. Status: Issue closed
realm/realm-java
118713010
Title: New Gson 2.5.0: Do we still need to `setExclusionStrategies`? Question: username_0: Hi I have noticed that Gson has moved from 2.3.1 to 2.4.x and now 2.5. I was wondering if we still need to do the GsonBuilder().setExclusionStrategies() thing? Thanks! Answers: username_1: Hi @username_0 I just tested with GSON 2.5 and you still need to set the exclusion strategy. I updated our example to reflect his as well. Status: Issue closed username_0: Thanks!
wallabyjs/public
1090895200
Title: Option to toggle the path in the Output Inspector Question: username_0: ### Issue description or question Option to toggle the path in the Output Inspector. Ideally I'd like the filename only to be displayed by default with a toggle to display the whole path when needed. Or even some better styling ala how Jest does it: ![CleanShot-2021-12-30-at-14 01 18](https://user-images.githubusercontent.com/14298160/147725906-16e583f2-8966-4faf-8f56-af643f0b14b6.png) It can be difficult to identify the included test files when starting a exclusive test run. Here's a screenshot for your reference: ![CleanShot-2021-12-29-at-13 08 43@2x](https://user-images.githubusercontent.com/14298160/147725527-216b955b-68c0-4e76-bd78-ebad2128ff60.png) Artem has recommended using the Wallaby App for now and this is what I'm currently doing. I'd like to reduce context switching though, so implementing this option would be a nice-to-have. Thanks for the amazing product and support guys. :)
FoodOntology/foodon
537770424
Title: NTR: Consider adding specific diets Question: username_0: Hi All, We are currently adding diet and supplement restrictions, but some terms may be more suited for FOODON. I am listing a few that are higher on our list. I don't have an easy answer to where in FOODON these should go. There is a term diet meal, liquid (FOODON_03302872) but that is a child of 'food product', so I don't think that is a natural placement. Thanks. Term | in Maxo | Definition | Definition source | Synonym | Foodon Term -- | -- | -- | -- | -- | -- vegetarian diet | no/will not add currently |   |   |   |   vegan diet | no/will not add currently |   |   |   |   Low-carbohydrate, high-protien, high fat diet | no | The low-carbohydrate, non–restricted-calorie diet aimed to provide 20 g of carbohydrates per day for the 2-month induction phase and immediately after religious holidays, with a gradual increase to a maximum of 120 g per day to maintain the weight loss. | PMID:18635428 | Low carbohydrate diet Meditteranean diet | no | The moderate-fat, restricted-calorie, Mediterranean diet was rich in vegetables and low in red meat, with poultry and fish replacing beef and lamb. | PMID:18635428 |   |   low fat diet | no | A restricted-energy diet with less than 30 % of energy from fat. | PMID:23651522 |   | low fat food FOODON_03510039 intermittent fasting | no | Intermittent fasting is a form of time restricted eating, typically 16 h fasting and 8 h eating. | PMID: 30897855 |   |   DASH diet | no | Dietary Approaches to Stop Hypertension (DASH) trial, which demonstrated that consuming a diet rich in fruits, vegetables, and low-fat dairy products and reduced in saturated fat and cholesterol | PMID: 29141784 |   |   ketogenic diet | ketogenic diet intake MAXO_0030010 | The ketogenic diet is a high-fat, low-carbohydrate, adequate-protein diet that mimics a fasting state, resulting in the substitution of ketone bodies as a source of energy. The ketogenic diet can be used to reduce recurrent epileptic seizures in individuals with intractable epilepsy. Intractable epilepsy is defined as epilepsy that failed to respond to three or more antiepileptic drugs. | PMID:23419562, PMID:27759811, PMID:28701250 | Modified Atkins diet fasting | fasting intervention MAXO_0000098 | Going without food or drink for a certain period of time. | PMID: 30897855 |   |   Low calorie diet | calorie restrictive diet intake MAXO_0000089 |   |   |   | low calorie food FOODON_03315839 texture modified diet | texture modified diet intake MAXO_0030011 |   |   |   |   https://github.com/monarch-initiative/MAxO/issues/66 https://github.com/monarch-initiative/MAxO/issues/82 https://github.com/monarch-initiative/MAxO/issues/94 @leechuck aurenechan Answers: username_1: Indeed diet meal is a material entity reference, whereas diet seems to speak to a pattern of behaviour. We've had some past discussion about this (https://github.com/EcologicalSemantics/ecocore/issues/108 @pbuttigieg @cmungall @diatomsRcool) and explored whether a diet is a material aggregate of sorts, or qualities, but actually, now I'm thinking it is best considered a type of protocol, essentially a (food) inclusion criteria (http://purl.obolibrary.org/obo/OBI_0500027) that pertains to food consumption processes. Whether the protocol is adhered to strictly is a different question! username_1: We'd probably use "eligibility criterion" (http://purl.obolibrary.org/obo/OBI_0500026)- both inclusion and exclusion - to define diet components. Its definition is nicely general: "an eligibility criterion (rule) is_a selection criterion which defines and states the requirements (positive or negative) for an entity to be considered as suitable for a given task or participation in a process." username_2: Agreed, this is more of a protocol! @username_0 Just to make sure I understand, are you recommending to add all of these eating patterns/therapies into FOODON rather than add them into MAxO? I am primarily wondering if we are then limiting representation of any nutrition therapies in MAxO should we choose to not include things such as a ketogenic diet, which is a prescribed medical nutrition therapy. There are some separations between medically prescribed nutrition recommendations and more of a lifestyle eating pattern, but therapeutic diets should likely still be represented in MAxO I would think. username_1: Tried to solve this more as a result of cmungall's and diatomsRcool examples on EcologicalSemantics/ecocore#108 . New term proposals are there that fit into BFO / IAO / OBI paradigm. I'll repeat them here for feedback: This one has broad synonym of "diet" because it is a more specific sense of diet: label: '**ingested food set**': subClassOf: 'material aggregate' definition: 'A comprehensive set of food items which an organism has ingested over some span of time.' hasBroadSynonym: diet elucidation: 'This names various food items an organism has ingested. The quantity or mass of the items is not indicated.' This could be derived from: label: '**ingested food log**' subClassOf: IAO "data set". definition: 'A set of food consumption events detailing food ingested and the time of ingestion.' elucidation: 'This may include the mass or quantity of a food item, and other details such as its preparation process.' label: '**ingested food**' subClassOf: "food material". definition: 'A food material consumed by an organism at some point or duration in time.' ... one could add terms for other statistics derived from the ingested food log Another sense of diet is as a defined set of food proportions, so not referencing material entities directly. The proportions are with respect to some open-ended "norm", or with respect more particular food groups. Outright exclusions are needed, e.g. "A vegan diet contains no animal products." Technically this sense of diet sounds a lot like a "dataset of features" (http://purl.obolibrary.org/obo/OBI_0000649) about an organism's ingested food: _A data set that is produced as the output of a descriptive statistical calculation data transformation and consists of producing a data set that represents one or more features of interest about the input data set._ . If we allow a looser interpretation of "descriptive statistical calculation", then this would cover both intentional plan/goal and descriptive perspectives, and could include constraints on timing of ingestion, so could cover fasting regimes. I'll check with OBI folks on monday about this. label: "**diet**" subClassOf: OBI 'dataset of features' and 'is about' some 'ingested food' definition: 'A set of generalizations about an organism's ingested food' label: "**mediterranean diet**" subClassOf: diet definition: 'A diet emphasizing vegetables, fruits, herbs, nuts, beans and whole grains, and allowing moderate amounts of dairy, poultry, eggs, and seafood, and a lower intake of red meat.' etc. One could add "objective specification" parts that contain diet constraints. username_1: OBI call was informative. Call reinforced not to mint a basic "diet" term. All the more particular senses of diet live in different places in BFO etc. class/concept hierarchy. Instead, to connect the word "diet" as people might use it in a search engine to these varied term entities, we can use the hasBroadSynonym annotation. '**ingested food log**' survived critique in the call as a clear case of a data set. Should we label it as "... log" or "... record" or "... data set"? And I just found "**ingested food**" already exists as an UBERON term so we will reuse that. The '**ingested food set**' was agreeable but it does sound problematic to have it be an "object aggregate" as the aggregate is intended to reference a collection of object instances at some time t, whereas the proposed set is only meant to reference the **classes** or types of food eaten. So I would revise ingested food set to be a type of dataset too, derived from the food log dataset. May only need to add this term in the future. The above "diet" was discouraged as the "dataset of features" parent was too tied to parameters of an actual dataset. People preferred "dietary regimen" as a subclass of protocol, so - and here I welcome definition edits - something like: label: "**dietary regimen**" subClassOf: protocol hasBroadSynonym: diet definition: "A set of restrictions on food type ingestion and/or timing of food ingestion" Another suggestion was to define classes of dietary food - a "vegetarian food", a "carnivorous food" - here one can define in an equivalency exactly what is permitted or not in the class by reference to other food types. I really like this. Under "FoodOn product type, I would add a convenience class "food product by dietary group" class to gather the various diet categories. label: "**food product by dietary group**" subClassOf: "FoodOn product type" definition: "A food item categorized by a dietary group of food" Then subclasses can have equivalency axioms that spell out the dietary group: label: "**pescatarian food**" subClassOf: "food product by dietary grouping" definition: "A dietary group of food that does not include land-based animals or animal products, but does allow freshwater- or saltwater-based animal products." equivalentTo: "'food material' and not ('meat food product' or 'avian food product' or 'dairy food product') " The https://www.dietaryguidelines.gov/sites/default/files/2019-05/2015-2020_Dietary_Guidelines.pdf document conveys in its "eating patterns" not just the set of foods to eat, but also amounts, so I avoided using "eating pattern" in label as it carries this extra level of detail. We can then simply say some organism is 'consumer of' [some | only] 'pescatarian food'. (It isn't defined in RO at moment but 'consumer of' is reverse of 'has consumer'.) An experimental process can [paraphrasing] have an associated 'dietary regimen' protocol with objective (objective specification) to provide 'oral administration' (a planned process) of certain food (a 'food product by dietary group') according to some schedule. username_3: @username_1 are there notes from the OBI call? @username_2 is implementing our dietary interventions in Maxo (she is a registered dietitian) and it would be great for her to participate in the future. I do think we need to think more about dietary regiments- there are similar to a protocol but also have dynamic and behavioral aspects. These are in fact what we need for interventions and disease causal relationships (e.g. exacerbates or causes). We probably need greater community coordination and discussion before any design patterns are implemented - where should we have these discussions? username_1: I didn't take any notes, but can recount that <NAME>. was advocating for real study data to use as a basis to describe diet related terms from (sounds like you could provide some!); while Bjorn was happy to see curation work proceeding on diet as it related to protocol, with the analogy of protocols related to drug administration to patients in certain amounts/times of day. We talked about diet and poked holes in above "diet" definition, leading to my response above. I would be happy to engage with Lauren and others on this more. If she wanted to present to the OBI group in January on the dietary regime/nt and explain protocol parts and beyond, and clinical study needs, I think that would be welcome - I can verify with the OBI group. I foresee FoodOn housing diet related food groups but would accept other curation teams taking the lead there too. The ECOCORE folks certainly have a parallel interest in capturing food system related diet. Alternately, outside of the OBI call, in January we could organize a telecon on the subject? username_1: BTW I'd like to see these diets added too: Lacto-vegetarian, Ovo-vegetarian, Lacto-ovo vegetarian diet, and a semivegetarian diet — also called a flexitarian diet — which is primarily a plant-based diet but includes meat, dairy, eggs, poultry and fish on occasion or in small quantities. (source: https://www.mayoclinic.org/healthy-lifestyle/nutrition-and-healthy-eating/in-depth/vegetarian-diet/art-20046446) username_0: I support this for terms, but for MAxO...we mainly get recommendations for 'Supplement X if vegetarian/Vegan', but I think more granular terms could be useful. username_1: Ok. For FoodOn application to websites like https://fdc.nal.usda.gov/ (which we are working on) I'm thinking here of being able to select a diet (or indirectly, a list of foods), and then be able to calculate which nutrients may need special attention (supplementation or particular food sourcing) as a result. Understanding of course that food preparation and individual nutrient uptake are wildcards. username_1: A diagram that in my view disambiguates types of dietary / eating pattern term. Feedback welcome: ![image](https://user-images.githubusercontent.com/4000582/79642012-dee26580-814f-11ea-83c9-c4ad616e39ea.png) And a list of fasting and food-specific diets - not complete but potential organizing scheme ![image](https://user-images.githubusercontent.com/4000582/79642031-02a5ab80-8150-11ea-82fd-30132db88598.png) username_2: @username_1 Is there a particular reason we are using proscribed as opposed to prescribed? Also, looking at the 'diet by type of food' --> 'composition' sections, what are you proposing we guide the "high" or "low" values off of for indicating composition? Would we use percentages of acceptable macronutrient distribution ranges or some other guideline? Further, I think we may reconsider "ketogenic" as a term for composition as you will get a huge number of foods that fall into that category. Ketogenic simply describes production of ketones in the body, but ketones are produced by all humans throughout the day with lipid breakdown even in someone who is not observing a 'classic ketogenic diet'. As any substrate can become lipid and lipid catabolism can happen in a human fed any substrate, this could arguably encompass a huge amount of foods. Additionally, achievement of ketosis (the goal for a ketogenic diet) also relies on timing of substrate consumption, i.e. if you have 30g carb throughout the day in small quantities you are more likely to maintain ketosis, whereas if you had all 30g at one time, some human bodies will not maintain ketosis. While I understand the means to try and include this as a composition factor, we may consider just focusing on macronutrient content labels as opposed to functional labels to ensure we can clearly define the classes. username_0: My 2 cents: Halal diet is unlikely to be prescribed, but it may be proscribed. Conversely, vegetarian may be proscribed or prescribed or just a preference. Not sure if there is a term that encapsulates all 3. I don't think we should necessarily divide them. I don't think we need to have terms that breakdown composition too specifically. Higher and lower might be subjective, but I think that's preferable because there are different situations. Maybe textual defines could be 'higher than normal' or 'higher than baseline' or something that could fit a number of circumstances. I have no opinion on ketogenic-- Maybe textual definitions and comment field can cover some of the information? username_1: Oy - that was my spelling mistake. I meant "prescribed"! Fixed above. But good point Leigh, there seems to be a middle space as well for cultural diets - "practiced" might be better than "prescribed" in a number of cases. And then advice not to eat a certain way. So possibly proscribed/prescribed/practiced diet are their own stand-alone terms, and we don't place other diet categories under them - but this suggests building a bit of a model to capture who is doing the prescribing / proscribing. Something that could be done later. And I would be fine relabeling "high fat / low fat" to "higher than normal" and "lower than normal" fat intake. We could discuss having diet terms defined so they can at least be used in text mining without necessarily expecting to link them to food categories. username_0: Our biggest concern for text mining is to include synonyms of how they might be stated in the literature. So, if 'higher than normal fat diet' is the term label, we might have 'high fat diet' as a synonym. We currently do use annotation properties for exact, narrow, and broad synonyms to try bridge what things are commonly referred to with the exact concept. Does FoodOn use those annotation properties? username_1: Yes FoodOn uses the oboInOwl broad and narrow and exact synonym. And we try to keep them well populated, especially the exact one. We also use the language attribute with all labels and synonyms. username_1: Diet vocabulary is now on its way to being incorporated in ONS. Status: Issue closed
icon-project/loopchain
736716269
Title: Too much memory network usage Question: username_0: I use docker-compose to start, which takes up too much network and memory Answers: username_0: this is my docker-compose version: "3" services: prep: image: iconloop/prep-node:latest container_name: "prep-mainnet" network_mode: host restart: "on-failure" environment: NETWORK_ENV: "mainnet" # mainnet, testnet, PREP-TestNet (zicon) CERT_PATH: "/cert" LOOPCHAIN_LOG_LEVEL: "DEBUG" ICON_LOG_LEVEL: "DEBUG" IS_AUTOGEN_CERT: "true" # 自动生成证书 PRIVATE_PASSWORD: "<PASSWORD>" # 证书密码 FASTEST_START: "yes" # 加载快照 TZ: "Asia/Shanghai" # 时区 cap_add: - SYS_TIME volumes: - ./data:/data # mount a data volumes - ./cert:/cert # Automatically generate cert key files here ports: - 9000:9000 - 7100:7100 username_0: Is it normal for the bandwidth to occupy more than 2M? Status: Issue closed username_1: I left comment on icondev forum. https://www.icondev.io/discuss/5fa3b4da7f21b60051a560e5
magda-io/magda
341767682
Title: When I click on an organisation page from the search results, weird stuff happens Question: username_0: ### Problem description - You see another organisation when the page first loads - The page loads halfway through the content - Recent search keeps on appearing as documented here: https://github.com/magda-io/magda/issues/1464 ### Problem reproduction steps Tested on Android Google Pixel ### Screenshot / Design / File reference [Org search_small.zip](https://github.com/magda-io/magda/files/2200287/Org.search_small.zip)<issue_closed> Status: Issue closed
ogham/exa
301640078
Title: exa -F flag Question: username_0: I was disappointed to see that exa has no listed flag that emulates the behavior of the `-F` flag in ls...only to be even more surprised when I idly tested it to discover that the functionality existed, just without documentation. Any reason for this, or am I somehow missing somethign? Answers: username_1: Nope, I just forgot to add it to the documentation! Fixed in 9c77a953a31fd4198d3a96e2f73cc9dd57afe0e1 and will be part of the next release. Status: Issue closed
simpleidserver/SimpleIdServer
1161196880
Title: Is it possible to integrate SQL Server Question: username_0: Hello. I could not find the proper starting point of SQL integration. Is it possible to use this project to provision the users into SQL Database? Best regards, Answers: username_1: Hello, There are different methods to provision users into an SQL Server database : * **Just In Time Provisioning** : Automatically create new accounts in the right systems for new people when they authenticate to an external identity provider. * **Automate provisioning** : Automatically create new accounts in the right systems for new people when they join your team or organization. Both methods and more are supported by SimpleIdServer. # Pre-requisite Install SimpleIdServer template : ``` dotnet new --install SimpleIdServer.Templates ``` # Just In Time Provisioning An OPENID server with an SQL Connection must be configured : A SimpleIdServer template exists to create OPENID server with EF support. * Execute the command line `dotnet new openidef -n OpenId`. * Create the migration scripts `dotnet ef migrations add Init`. * Update the `CONNECTIONSTRING` parameter in the files `OpenIDMigration.cs` and `Startup.cs` before running the solution. Example project : https://github.com/username_1/SimpleIdServer/tree/master/src/OpenID/SimpleIdServer.OpenID.SqlServer.Startup # Automate provisioning A SCIM2.0 endpoint with an SQL Connection must be configured : A SimpleIdServer template exists to create SCIM server with EF support. * Execute the command line `dotnet new scimef -n ScimHost`. * Create the migration scripts `dotnet ef migrations add Init`. * Update the `CONNECTIONSTRING` parameter in the files `ScimMigration.cs` and `Startup.cs` before running the solution. Example project : https://github.com/username_1/SimpleIdServer/tree/master/samples/UseSCIMSqlServer username_1: Hello, Indeed the automatic provisioning can be configured between Azure AD and our SCIM server. To do so, please follow the steps below : * A SimpleIdServer SCIM server must be deployed. * Configure Microsoft Azure Active Directory to enable automatic provisioning of users between SCIM and Azure AD. For more information please refer to this documentation : https://docs.microsoft.com/en-us/azure/active-directory/app-provisioning/use-scim-to-provision-users-and-groups. * When a representation is added in SCIM (for example : User, Group etc...). A client can be developed to subscribe to the messages received from a queue. When a message like "representation-added" or "representation-updated" is received, an API operation can be called to create or update user in your preferred OPENID server. SimpleIdServer offers some APIS to facilitate the user management. The schema below shows the interactions between the different components : ![image](https://user-images.githubusercontent.com/10213388/157119278-ee6e1cde-0ab7-487e-8d77-485a0638a38a.png) username_0: Thank you for this great explanation. I followed the instructions till "When a representation is added in SCIM (for example : User, Group etc...). A client can be developed to subscribe to the messages received from a queue. When a message like "representation-added" or "representation-updated" is received, an API operation can be called to create or update user in your preferred OPENID server. SimpleIdServer offers some APIS to facilitate the user management." Could you please give more detail about subscribing to the messages, and the ongoing process? What do you mean by 'OpenID Server'? Do I have to combine 'OpenID' system to secure the application by following this part of the documentation? https://username_1.github.io/SimpleIdServer/documentation/openid/installation.html [cid:8022c46a-3f4a-432c-9d5d-27d5031b1a7f] Thank you, Best regards. username_1: Hello, When a message is received from the message broker (for examples "RepresentationAdded" or "RepresentationUpdated"). You can decide to create a user in an Identity server like OPENID, SAML IDP or softwares like Slack, gitlab etc... It's not mandatory to create a user in an OPENID server, it depends on your use case :). If the scenario you want to support is to add a new user in the OPENID server then you can download the ZIP file and run the projects : * Download the ZIP file : [QuickStart.zip](https://github.com/username_1/SimpleIdServer/files/8207384/QuickStart.zip) * Download the POSTMAN collection & import it : [Provisioning.postman_collection.zip](https://github.com/username_1/SimpleIdServer/files/8207395/Provisioning.postman_collection.zip) * Update the `CONNECTIONSTRING` in both projects. * Open a command prompt and execute this command under the OPENID folder: ``` dotnet run --urls=http://localhost:5001 ``` * Open a command prompt and execute this command under the SCIM folder : ``` dotnet run --urls=http://localhost:5002 ``` * When everything is up & running you can execute the request present in the POSTMAN collection. It contains two requests: ** First request : Get an access token. ** Second request : Add a user. At the end, a new user should be added in the OPENID database :) Kind Regards, SimpleIdServer
JosefNemec/Playnite
283487807
Title: Can't auto update Question: username_0: You said: Every time you start Playnite it checks for new version, downloads it in background and offers you update installation. Once you started Playnite it periodically checks for update every 4 hours. But my playnite never checks for new version, I had to download new version and install it manually . I don't know if this is a bug or something . Answers: username_1: Can you go to "About" menu, create diagnostics package and upload it here? Thank you username_0: [1.zip](https://github.com/username_1/Playnite/files/1575593/1.zip) username_1: Looks like Playnite can't reach update server. Can you check if [this URL](http://www.playnite.link/update/update.json) is available via browser? And also check if firewall is blocking Playnite? username_0: The URL is available via my browser ,and I added Playnite to the firewall list ,waiting for the next update to check if this will work. username_1: You can also download 3.0 and check if it updates to 3.2 automatically. username_0: 3.2 didn't auto update to 3.3. I downloaded 3.0 and it didn't auto update to the latest version as well.... It would be better if you add: 1. a update button to update manually 2. a download process bar for update download username_2: I can confirm that sometimes this bug occurs. This happened for me when I was waiting for an update from 2.9 to 3.0. The update button would be cool. username_1: The update button will not solve this, the problem is that Playnite can't access update server for some reason. username_1: Released in 3.4, there are now fallback addresses when connection to the main one fails. Please feel free to reopen this if the issue still occurs when (not)updating from 3.4 onward. Thanks Status: Issue closed
libgit2/libgit2sharp
153268148
Title: Test failure: NoPublicTypesUnderLibGit2SharpCoreNamespace Question: username_0: This test fails in the master branch (as of 1fb7fc5b1d26c893e085c2782d5ee483d56c5a18) ``` Test Name: LibGit2Sharp.Tests.MetaFixture.NoPublicTypesUnderLibGit2SharpCoreNamespace Test FullName: LibGit2Sharp.Tests.MetaFixture.NoPublicTypesUnderLibGit2SharpCoreNamespace Test Source: C:\Users\andarno\git\libgit2sharp\LibGit2Sharp.Tests\MetaFixture.cs : line 298 Test Outcome: Failed Test Duration: 0:00:00.003 Result StackTrace: at LibGit2Sharp.Tests.MetaFixture.NoPublicTypesUnderLibGit2SharpCoreNamespace() in C:\Users\andarno\git\libgit2sharp\LibGit2Sharp.Tests\MetaFixture.cs:line 324 Result Message: Public type 'LibGit2Sharp.Core.LeaksContainer' under the 'LibGit2Sharp.Core' namespace. ``` Answers: username_1: This happens because the compilation defines `LEAKS_IDENTIFYING` which makes us create this flag, but the tests got compiled by VS itself which did not, thus not making us include `LeaksContainer` in the list of classes that are OK to export. We should probably get rid of this long-term as we'd actually want to expose things in core. username_1: Does #1317 fix it for you? username_2: I think #1317 should fix this - @username_0 please feel free to reopen if it doesn't. Status: Issue closed
gerwitz/sitewriter
322290197
Title: Video post variables Question: username_0: video-specific posts need location, e.g. with OwnYourGram: location-latitude location-longitude but we must also add "video" variables to the base post to match the existing "photo" values N.b. that we might use dotted hierarchy for Liquid after #23
Azure/azure-sdk-tools
1006651361
Title: Make consistent the criteria for daily pipeline jobs Question: username_0: For example the criteria for JS' `PublishPackages` job which publishes dev packages is: ``` or(eq(variables['SetDevVersion'], 'true'), and(eq(variables['SetDevVersion'], ''), eq(variables['Build.Reason'],'Schedule'), eq(variables['System.TeamProject'], 'internal'))) ``` While the .NET criteria for the similar `PublishPackages` job is: ``` or(eq(variables['SetDevVersion'], 'true'), and(eq(variables['Build.Reason'],'Schedule'), eq(variables['System.TeamProject'], 'internal'))) ``` It's unclear whether these criteria are different. They should be made as consistent as possible to smooth the process of editing these pipelines. Answers: username_1: I don't believe `eq(variables['SetDevVersion'], '')` is interesting in the second part of this condition. It probably doesn't hurt the execution in our scenarios because we don't set that variable as a global variable for our scheduled builds, but it would be nice to remove it here as it just makes the conditions more complex then it needs to be. username_0: * Indeed it is no longer relevant -- https://github.com/Azure/azure-sdk-for-js/pull/11476#issuecomment-699395920 * The expression in .NET, Java, and Python are all the same, only JS is different username_2: Triaging this... we're pretty sure that you (@username_0) have already fixed the place(s) where this happens. If there is broader tools work to do here, please reassign to @chidozieononiwu. username_0: Oh yes, this is merged here and it only impacts JS -- https://github.com/Azure/azure-sdk-for-net/pull/24199 Closing. Status: Issue closed