repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
rust-lang/rust | 153161641 | Title: Missing context from Error Handling section of The Book
Question:
username_0: In the [Adding Functionality section](https://doc.rust-lang.org/book/error-handling.html#adding-functionality) of the Error Handling guide, a line with `args.flag_quiet` is referenced. This is not a standard method and after doing some googling I found [this post](http://blog.burntsushi.net/rust-error-handling/). It would seem a few sections regarding the creation of a custom `Args` struct were removed or left out by mistake. Either way, I'm not sure if this those sections should be re-added or if this section needs to be reworked without that struct.<issue_closed>
Status: Issue closed |
MisaelGC/invie-GitHub | 338132885 | Title: Feature Request = precios
Question:
username_0: ## ¿Cómo puedo replicar el problema?
no hay pagina de precios, me encantaría saber el precio de la guitarra acústica
## ¿En que versión ocurre de invie ocurre?
Ahora mismo no existe una pagina de precios |
solgenomics/sgn | 926277417 | Title: queries involving locations don't work in the wizard
Question:
username_0: Expected Behavior <!-- Describe the desired or expected behavour here. -->
--------------------------------------------------------------------------
For Bugs:
---------
### Environment
<!-- Where did you encounter the error. -->
#### Steps to Reproduce
<!-- Provide an example, or an unambiguous set of steps to reproduce -->
<!-- this bug. Include code to reproduce, if relevant. --><issue_closed>
Status: Issue closed |
cocos2d/cocos2d-x | 175694762 | Title: schedule::update dt is no accurate
Question:
username_0: - cocos2d-x version:3.9
- devices test on:android huawei
when fps is low than 60, the dt is 0.016, in fact it maybe 0.02 or more
```
self.nScheduleID = schedule:scheduleScriptFunc(function(dt)
self:onTick(dt)
end, 1/60, false)
```
Answers:
username_1: could you post a complete snippet... something that I can copy & paste and reproduce the issue? Thanks. |
theopolis/uefi-firmware-parser | 41663757 | Title: More powerful auto-checking
Question:
username_0: A script that uses the firmware-type detector. When no type is found and brute-forcing is an option, have the type-checker auto-suggest to the script invoker (via stdout/stderr) they should try bruteforce/scanning.
Answers:
username_0: This occurs by default for the `AutoChecker` class when a volume is found. Since volume content usually involves 'stacked' volumes.
Status: Issue closed
|
pcbuilderapp/pcbuilder | 203743643 | Title: Price history date filtering
Question:
username_0: 6.8 Price history can be filtered by using a date range
6.8.1 Quality attribute
Performance
6.8.2 Description
The backend server should provide a date range, on which price history can be filtered in order to reduce performance strain on the server, when large sets of pricing history are accumulated.
6.8.3 Indicator
When retrieving the pricing history of a given product, it should be limited by a start and an end date.
6.8.4 Measurement
When showing pricing history data screen loading times are measured and deemed acceptable if below 1 second.
6.8.5 Norm
Performance can be guaranteed even when large sets of pricing data are accumulated.
Status: Issue closed
Answers:
username_1: backend supports selecting pricehistories by dateranges, frontend gets pricehistory of last month by default |
lomik/zapwriter | 357251875 | Title: What does this do and why does it exist?
Question:
username_0: $subject.
The answers to these questions is not very clear from the readme or the git history.
Answers:
username_1: PRs with readme are welcome ;)
username_1: This is just reusable helper for my own public and private projects. I have no purpose and no desire to document and sell it
username_0: That's fine. Maybe I can phrase my question differently. What was missing from zap that made you write this? I can see _what_ this thing is doing, but I can only guess at _why_ it's doing it.
username_1: Primary goals was:
1. File output with rotation support
2. Custom encoder ("mixed"). At the time, it was necessary to maintain backward compatibility in the private projects
3. A flexible output manager who could solve the following problems:
* Write full log to file with level "info" and to stderr with level "error"
* Ability to record a part of the logs into a separate file. Sample, access log of http server. And also to solve feature requests like this https://github.com/username_1/go-carbon/issues/154
Zap could not do any of the above at the time of writing this library
username_0: Thanks for taking the time to explain this to me.
Status: Issue closed
|
rancher/rancher | 186672400 | Title: kubernetes & other environments are showing cow icon instead of their own
Question:
username_0: **Rancher Version:** master 11/01
**Steps to Reproduce:**
1. Create a kubernetes, mesos and cattle environments
2. go to cattle environment and add a host
3. Go to manage environments and add environment
**Results:** All icons are cows until I refresh screen

**Expected:** Should not be cows
Answers:
username_1: Switching environments was the key piece here..
username_0: Version - master 11/10
Verified fixed
Status: Issue closed
|
dubocr/homebridge-tahoma | 740663019 | Title: Capteur de consommation éclectique effet joule
Question:
username_0: Bonjour, je suis équipé d'un compteur a effet joule somfy et celui-ci n'est pas reconnue dans mes accessoires, l'avez-vous rendu compatible?
Merci
Answers:
username_1: Thank you for using Homebridge TaHoma plugin.
Please follow these step to obtain better support :
1. Execute failling operations from official app (TaHoma/Cozytouch/etc.) then execute same operation from Homekit
2. Report your config by browsing [https://dev.duboc.pro/tahoma](https://dev.duboc.pro/tahoma)
3. Search issues with title corresponding to your device widget name (see picture below). If no opened issue, rename your issue with this widget name.
4. Provide your bridge last 4 digits (number visible as SETUP-XXXX-XXXX-XXXX at step 2.)

Thank you.
username_1: Bonjour,
Ce compteur n'est pas encore compatible mais je pourrais me pencher sur le sujet avec le votre. Pourrez-vous vous connecter sur https://dev.duboc.pro/tahoma et me transmettre les derniers chiffre de votre bridge pour que je puisse consulter les paramètre de cet équipement.
MErci
username_2: Idem j'ai aussi ce module, je pensais le supprimer définitivement du tableau électrique mais si vous allez l'intégrer je veux bien le garder et vous aider dans son développement...
username_0: Bonjour,
Merci mais je ne peut me connecter sur https://dev.duboc.pro/tahoma je n'ai aucun username qui fonctionne.
Avez vous une solution,
Merci
username_1: Bonjour @username_0,
Vous n'avez pas de box TaHoma (ou équivalent) ?
username_0: Si la box energeasyconnect, j'ai essayer de me connecter avec les mêmes identifiants mais le message d'erreur "Bad credentials" s'affiche.
username_1: As-tu essayé en sélectionnant le Endpoint "Rexel" ?
Quel service utilises tu sur le plugin Homebridge ?
Merci
username_2: De mon coté avec ma nouvelle Tahoma Switch le capteur à effet joule n'est pas encore reconnu/supporté par contre c'est bizarre je vois des traces dans Homebridge.... soit un device IO (1234#7) dans les logos ou sinon j'avais trois icône GMDE_Zone1, GMDE_Zone2 et Boost que j'ai exclu dans la config.
Je pense qu'il faut attendre pour la Switch que le capteur soit déjà supporté.
username_0: Maintenant j'ai une erreur 500 server error
username_1: Merci, j'ai corrigé mais j'ai pu récupéré la config.
username_0: C’est bon la connexion est OK, bridge 4055
username_2: @username_0 vous avez une Tahoma ? Avec ma nouvelle Switch le capteur à effet joule n'est pas reconnu, d'après Somfy il devrait arriver pour la fin de l'année !
username_0: @username_2 non , j’ai une Energeasy connect rail Din de chez rexel |
JuliaImages/Images.jl | 313934688 | Title: convert Image to Array{Float64}?
Question:
username_0: MethodError: Cannot `convert` an object of type ColorTypes.RGB{FixedPointNumbers.Normed{UInt8,8}} to an object of type Float64
This may have arisen from a call to the constructor Float64(...),
since type constructors fall back to convert methods.
Stacktrace:
[1] _collect(::Array{ColorTypes.RGB{FixedPointNumbers.Normed{UInt8,8}},2}, ::Base.Generator{Array{ColorTypes.RGB{FixedPointNumbers.Normed{UInt8,8}},2},Type{Float64}}, ::Base.EltypeUnknown, ::Base.HasShape) at ./array.jl:488
[2] map(::Type{T} where T, ::Array{ColorTypes.RGB{FixedPointNumbers.Normed{UInt8,8}},2}) at ./abstractarray.jl:1868
[3] include_string(::String, ::String) at ./loading.jl:522
```
I also tried `map(Float64, img)` and `Float64.(img)` with the same error. Also, if there could be an example of how to convert an array to an Image that would be great!
Many thanks for the help!
Answers:
username_1: One way to create a multidimensional array of `Float64` where the third dimension corresponds to the red, green and blue channel is as follows:
```julia
using TestImages, Colors
img = testimage("mandrill")
I = Float64.(cat(3,red(img),green(img),blue(img)))
```
Another option is to create a view using `channelview`.
```julia
using TestImages, Colors, Images
img = testimage("mandrill")
mat = channelview(img)
```
However, this creates a view with dimensions `3 x 512 x 512` where the first dimension is used to index into the red, green and blue channels.
Status: Issue closed
username_0: Awesome thanks! My code to read in a video is now working and you really helped my understanding, thanks so much!
```julia
function readVideo(videoPath, nframes=100)
vid = VideoIO.openvideo(videoPath)
img = read(vid)
frames = zeros(ColorTypes.RGB{FixedPointNumbers.Normed{UInt8,8}},nframes, size(img)...)
i = 1
while !eof(vid)
read!(vid,img)
frames[i,:,:] = img
if size(frames, 1) >= nframes
break
end
i+=1
end
return frames
end
frames = readVideo(videopath);
```
username_2: as a side note. The main upside of `channelview` is that in general it does not need to copy/change your data; it simple reinterprets the memory in most practical cases. This is why it makes sense to have the colorchannel in the first dimension, because then RGB are right next to each other in memory and can simply be grouped. check out this article if you'd like to know more https://evizero.github.io/Augmentor.jl/images/ |
ooade/NextSimpleStarter | 428591924 | Title: Change to using Material Web Component
Question:
username_0: @username_3 Feel free to send in a PR :grinning:
Thanks :)
Answers:
username_1: Add a description naw 😥
username_2: When?
username_3: @username_0 do you need any help porting? I know Material very well and would be happy to give back
username_0: @username_3 Feel free to send in a PR :grinning:
Thanks :)
username_3: @username_0 great - I'm exploring the repo - I'll make a small draft PR and update 1 component to the latest material UI and we can discuss the results
username_0: Sure. Good to know 👌
username_3: @username_0 so when I said I know Material well - I actually know React Material [https://material-ui.com](https://material-ui.com) not Material Components Web - I briefly looked into that project and they look similar but I'd more comfortable with the react version.
So can we go with React Material UI instead? It does seem like it's more suited for NextJS...
username_0: Cool @username_3. I love Material UI, as well. The latest releases make a whole lot of sense.
Status: Issue closed
|
intel-analytics/analytics-zoo | 803110251 | Title: jenkins: UT failed on mac
Question:
username_0: PEP8_SCRIPT_PATH /private/var/jenkins_home/workspace/ZOO-NB-UnitTests-2.1-PYTHON-MAC/pyzoo/dev/../dev/pep8-1.7.0.py
PEP8 checks failed.
./zoo/examples/orca/learn/horovod/pytorch_estimator.py:120:101: E501 line too long (101 > 100 characters)
./zoo/examples/orca/learn/horovod/pytorch_estimator.py:131:101: E501 line too long (111 > 100 characters)
./zoo/examples/orca/learn/horovod/pytorch_estimator.py:133:101: E501 line too long (117 > 100 characters)
./zoo/examples/orca/learn/horovod/pytorch_estimator.py:135:101: E501 line too long (115 > 100 characters)
./zoo/examples/orca/learn/horovod/pytorch_estimator.py:137:101: E501 line too long (106 > 100 characters)
./zoo/examples/orca/learn/horovod/pytorch_estimator.py:139:19: E128 continuation line under-indented for visual indent
./zoo/examples/orca/learn/horovod/pytorch_estimator.py:140:19: E128 continuation line under-indented for visual indent
./zoo/examples/orca/learn/horovod/pytorch_estimator.py:141:19: E128 continuation line under-indented for visual indent
./zoo/examples/orca/learn/horovod/pytorch_estimator.py:142:19: E128 continuation line under-indented for visual indent
jenkins link
http://10.239.47.210:18888/view/ZOO-NB/job/ZOO-NB-UnitTests-2.1-PYTHON-MAC/211/console
Answers:
username_0: @username_1 [3458](https://github.com/intel-analytics/analytics-zoo/pull/3458)
username_1: It has been fixed in https://github.com/intel-analytics/analytics-zoo/pull/3474
Status: Issue closed
|
bcherny/frontend-interview-questions | 242432266 | Title: isBalanced: blog post issue in examples
Question:
username_0: [Source test](https://github.com/username_1/frontend-interview-questions/blob/master/coding-easy/isBalanced.js) on github is ok, but [blog post](https://performancejs.com/post/hde6d32/The-Best-List-of-Frontend-JavaScript-Interview-Questions) has issue:
```javascript
isBalanced('{}{}') // false
```
Should be `true`.
Status: Issue closed
Answers:
username_1: Updated - thanks @username_0. |
rust-lang/rfcs | 54966123 | Title: Switch rustdoc to Common Markdown
Question:
username_0: <a href="https://github.com/brson"><img src="https://avatars.githubusercontent.com/u/147214?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [brson](https://github.com/brson)**
_Wednesday Sep 03, 2014 at 22:27 GMT_
_For earlier discussion, see https://github.com/rust-lang/rust/issues/16978_
_This issue was labelled with: A-rustdoc, E-easy, I-wishlist in the Rust repository_
----
Assuming this takes off it's probably best to be on the right side of history: http://standardmarkdown.com/. Probably easiest to just make hoedown conformant.
Status: Issue closed
Answers:
username_1: This has been implemented behind a flag, and rust-lang/rust#44229 is tracking the plan to enable it by default. |
mojombo/chronic | 104117091 | Title: project URL 404s
Question:
username_0: ```
$ curl --write-out %{http_code} --silent --output /dev/null http://injekt.github.com/chronic
404
```
Answers:
username_1: This no longer seems to happen. Might have been an issue with github at that time, as they were dealing with redirects to the aliased project name. "injekt" was renamed/aliased to "leejarvis" at some point.
If you want to use the `curl` command now, you'll have to follow redirects, using the `-L` flag, like so:
curl -L --write-out %{http_code} --silent --output /dev/null http://injekt.github.com/chronic |
monarch-initiative/mondo | 638067781 | Title: [Merge] MONDO:0011543 BRCA3
Question:
username_0: **Mondo term (ID and Label)**
MONDO:0011543 BRCA3
**Reason for deprecation**
OMIM merged these
**Term to be merged with**
MONDO:0016419 hereditary breast carcinoma
Answers:
username_1: Don't merge these, they are different concepts
OMIM has a description of the putatitive BRCA3 concept:
https://omim.org/entry/114480
From my reading of this, BRCA3 should indeed by obsoleted. Cite 114480 as rationale. Use a consider link to point to MONDO:0016419
Status: Issue closed
|
ropensci/refsplitr | 849904789 | Title: Error in `$<-.data.frame`(`*tmp*`, "EM", value = c(NA_character_, NA_character_ : replacement has 2 rows, data has 1
Question:
username_0: I installed refsplitr package as follows:
install.packages("remotes")
remotes::install_github("username_1/refnet2")
---
For the sake of simplicity, I've given my data the same name as in your example.
I runed the example code and because of errors I cleanend NA's as follows.
dat1 <- dat1 %>% filter(!is.na(AU))
dat1 <- dat1 %>% filter(!is.na(AF))
dat1 <- dat1 %>% filter(!is.na(EM))
However, when I run the code below:
dat2 <- authors_clean(references = dat1)
I got the following error.
Error in `$<-.data.frame`(`*tmp*`, "EM", value = c(NA_character_, NA_character_ : replacement has 2 rows, data has 1
Do you have an idea for the solution.
Best regards.
Answers:
username_0: Possibly I found the problem.
I am trying it know.
I changed all ; to , in EM collumn.
The code is still runnig and did not give an error.
Because my data is bir, it will possibly take an hour.
I will write the result.
username_1: From an author email to EMB: " I changed all semicolons (;) to comma (,) in EM collumn, error is disappeared"
Status: Issue closed
|
mschlenstedt/Loxberry | 981797765 | Title: Unterstützung Raspberry Compute Modul
Question:
username_0: Wird aktuell nicht unterstützt. Neuer Kernel und diverse Änderungen in /boot notwendig.
https://www.loxforum.com/forum/projektforen/loxberry/allgemeines-aa/316021-loxberry-webseite-funktioniert-nicht-auf-dem-raspberry-pi-compute-module-4 |
gchq/CyberChef | 509315222 | Title: Inventory notification
Question:
username_0: Your tool/software has been inventoried on [*Rawsec's CyberSecurity Inventory*](https://inventory.rawsec.ml/).
https://inventory.rawsec.ml/tools.html#CyberChef
### What is Rawsec's CyberSecurity Inventory?
An inventory of tools and resources about CyberSecurity. This inventory aims to help people to find everything related to CyberSecurity.

More details about features [here](https://inventory.rawsec.ml/features.html).
Note: the inventory is a FLOSS (Free, Libre and Open-Source Software) project.
### Why should you care about being inventoried?
Mainly because this is giving visibility to your tool and improve its referencing.
### Badges
The badge shows to your community that your are inventoried. It looks good but also shows you care about your project, that your tool is referenced.
Feel free to claim your badge here: http://inventory.rawsec.ml/features.html#badges, it looks like that [](https://inventory.rawsec.ml/), but there are several styles available.
### Want to thank us?
If you want to thank us, you can help make our open project better known by tweeting about it! For example: [](https://twitter.com/intent/tweet?text=We%20have%20been%20inventoried%20on%20%23Rawsec%20CyberSecurity%20Inventory%21%20Check%20their%20awesome%20inventory%20of%20tools%20and%20resources%20about%20CyberSecurity%3A%20https%3A//inventory.rawsec.ml/%21%20Thx%20@rawsec_cyber)
### So what?
That's all, this message is just to notify you if you care. Else you can close this issue.<issue_closed>
Status: Issue closed |
nervetattoo/banner-card | 450075048 | Title: Enhancement - Tap to execute script entities
Question:
username_0: I have a script to activate the nightlight hue scene and have added the script as an entity to the banner-card. It would be nice if I could execute the script by tapping that part of the card; currently it opens more info which has an execute button.
Alternatively, is it possible to add tap_action customization options?
Answers:
username_1: I want config to be simple and lightweight with sane defaults over deep customisation options, thus I don't like the action/tap_action config support I'm seeing elsewhere. For scripts I guess one wouldn't add a script to this card if running it was dangerous so there shouldn't be much harm in just supporting that with a button somehow. Will a single button be enough? I mean, is it any meta data about the script that is needed to make this usable?
username_0: I can't speak for others, but mine is just a service call to turn on the
light with a specific Hue scene so a button would be fine.
Status: Issue closed
username_0: ```
- type: custom:banner-card
heading: "\U0001F6CF Bedroom"
entities:
- light.bedroom
- entity: sensor.aqara_temperature
name: Temp
- entity: binary_sensor.aqara_motion
name: Motion
- light.closet
- entity: sensor.lumi_lumi_sensor_motion_aq2_02f39828_1_1024
name: Lux
- entity: script.bedroom_nightlight_hue_scene
name: Nightlight
value: mdi:lamp
action:
service: homeassistant.turn_on
```
Renders

Clicking on Nightlight renders this pop-up instead of executing the service. I can manually execute homeassistant.turn_on for script.bedroom_nightlight_hue_scene and it executes the script, as does the execute button in the pop-up.

username_1: How strange. I configured a similar thing with a script for myself to test and it works like a charm.
And I assume all of the toggles etc work on the card?
username_1: Hang on; Are you clicking the `Nightlight` heading, or the icon? The heading is supposed to open the popover, only the icon button triggers the action
username_0: I tried both, and all other toggles are functioning.
username_0: FYI, issue was because I'd migrated to HACS And was unable to use the new functionality until Banner-Card was installed via HACS after their updates this past weekend. Clicking the icon executes the script successfully. |
department-of-veterans-affairs/va.gov-team | 830775069 | Title: User transitions to Cerner
Question:
username_0: ### **Background**
As VAMCs transition from VISTA to Cerner, Veterans within those facilities will no longer have access to MHV features, including secure messaging. When it comes to the mobile app, this means that there will be Veterans who one day have access to secure messaging, and do not the next once their facility flips the Cerner switch.
### **User Story**
As a Veteran, I want to see an error message when my facility transitions to Cerner so that I know I need to go to a different portal to manage my health.
### **Acceptance Criteria**
# When a user's facility switches to the Cerner portal, the mobile app displays an error state giving the Veteran context and directing them to use the Cerner portal.
Answers:
username_1: A use case we may need to consider here is if a user is registered at multiple facilities, they might have the switch flipped off for some facilities but still be able to use messaging for other facilities.
username_2: @username_1 I've started to brainstorm around this in your [User Flows and Unhappy Paths mural](https://app.mural.co/t/adhoccorporateworkspace2583/m/adhoccorporateworkspace2583/1614017071435/8061f2093dc4679b4b1da3ab3f20423c8f1c8f66?wid=0-1617728321726). Have on the agenda for our next check in. |
MicrosoftDocs/azure-docs | 724613250 | Title: Link to A3 download is incorrect
Question:
username_0: Takes to wrong URL
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 71485a19-f1d1-c034-f41f-5aa91d7e3e71
* Version Independent ID: 8b441303-97d5-0c87-1d8f-b5c017fc6da8
* Content: [Azure Cosmos DB PDF query cheat sheets](https://docs.microsoft.com/en-us/azure/cosmos-db/query-cheat-sheet)
* Content Source: [articles/cosmos-db/query-cheat-sheet.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/cosmos-db/query-cheat-sheet.md)
* Service: **cosmos-db**
* GitHub Login: @SnehaGunda
* Microsoft Alias: **sngun**
Answers:
username_1: @username_0 Thank you for the feedback. This is being assigned to the content author to evaluate and update as appropriate.
username_2: @username_0 The link downloads the A3 pages to your download folder. Please elaborate regarding the wrong URL.

username_0: The link here:
https://docs.microsoft.com/en-us/azure/cosmos-db/query-cheat-sheet#oversized-cheat-sheets
Specifically this one:

Has this href:
https://go.microsoft.com/fwlink/?linkid=870413
Which in turn takes me to:

username_0: Just checked with some of my colleagues, they're getting the correct PDF in Chrome. Looks like this is somehow my issue?
Status: Issue closed
|
RevenueCat/purchases-ios | 745468046 | Title: Offerings/purchaserInfo no longer returned from cache via main thread
Question:
username_0: **Describe the bug**
A small change in 3.7.4 now has `purchaserInfo` and `offerings` returning from cache on DispatchQueue.main async instead of the main thread. I understand apps shouldn't depend on a main thread return here, but it'd be nice if it could when possible, and it seems like it can without disrupting current behavior.
1. Environment
1. Platform: iOS
2. SDK version: 3.7.4
3. OS version: 14.1
4. Xcode version: 12.1
5. How widespread is the issue. Percentage of devices affected. 100%
3. Steps to reproduce, with a description of expected vs. actual behavior
Call `Purchases.shared.purchaserInfo` when a cached value is present, and now it returns just a bit slower on async DispatchQueue.main
4. **Other information** (e.g. stacktraces, related issues, suggestions how to fix, links for us to have context, eg. stackoverflow, etc.)
Relevant change was in this commit: https://github.com/RevenueCat/purchases-ios/commit/38d16aa9d59b5204f22f7cbdac627d3699b3f623
Specifically, `purchaserInfoWithCompletionBlock` and `offeringsWithCompletionBlock` now asynchronously request application background state before returning a cached value. Previously, this application state was requested after cache return. Since application state seems to only be necessary when updating a stale cache or no cache, proposed fix would be to shift the background state request after cache return and to the else-block when no cache is present.
Answers:
username_1: I'm not sure I understand the issue here: `DispatchQueue.main` is a queue that's directly associated with the main thread by the OS (as documented here https://developer.apple.com/documentation/dispatch/dispatchqueue/1781006-main), so there return will still be on the main thread.
As for the slowness, yes, it does unfortunately return slightly slower than before. This is because of some new app state checks, that unfortunately need to run on the main thread (because they're done on `UIApplication`), so there's a bit more back and forth between a worker thread and the main thread.
Those checks are tied to the app state, because the state is also used to determine whether a refresh is needed, so there's no clean way of disassociating them.
username_0: Sorry I probably didn't make it too clear - yup I was primarily highlighting how previously it was returned synchronously on the main thread, you are right that right now it's dispatched asynchronously to the main thread as well. I understand you have the app state checks in place, but it doesn't seem the app state check is strictly necessary for the cache return (it seems like it is for a cache refresh).
For example, [L425](https://github.com/RevenueCat/purchases-ios/blob/4190aceade7ea7aaaecf6dd3cb54bd57f43f2c5e/Purchases/Public/RCPurchases.m#L425) for `purchaserInfoWithCompletionBlock` seems like it could be after L429 instead of where it is right now (and I guess before L436 since it is necessary there too).
Anyway this is a small change that we observed in our UI animations - not critical, but assuming I'm reading this code correctly, could be an easy perf improvement with no change to the existing functionality.
username_1: @username_0 great catch, you're correct and we could improve those. I'll add it to the backlog. Thanks for pointing this out!
username_1: addressed in https://github.com/RevenueCat/purchases-ios/pull/433
Status: Issue closed
|
digital-asset/daml | 1053690557 | Title: Check types on exercise/fetch by template for interface choices
Question:
username_0: Both in Daml and over the ledger API you can exercise a choice from an interface by template id,e.g.,
```
exercise (cid : ContractId InterfaceImpl) InterfaceChoice
```
However, currently we only check that `cid` corresponds to a template that implements the interface, we do not check that it is `InterfaceImpl`. So someone can pass a cid of a different template that implements the interface and things will succeed.
This is at best confusing and at worse a potential attack vector so we need to fix this.
This is slightly tricky since we desugar such a choice as
```
instance IsToken t => HasExercise t GetRich (ContractId Token) where
exercise cid = GHC.Types.primitive @"UExerciseInterface" (toTokenContractId cid)
```
This means that we generate the same call regardless of whether we call by interface or by template.
To fix this, we should be able to change the primitive for execriseInterface and fetchInterface to accept an optional typerep and validate against that.
Another option would be to accept a boolean predicate on the template payload that we evaluate between the fetch & the exercise. However, if we want to make failures of that catchable we have to record a fetch in the case of failure to track transaction inputs which is somewhat tricky.
so for now, the typerep solution seems like the better option.
Answers:
username_1: This is done! Thanks to @akrmn for merging the final PR into ghc :)
Status: Issue closed
username_1: There's even a test: https://github.com/digital-asset/daml/blob/main/compiler/damlc/tests/daml-test-files/InterfaceTypeRepCheck.daml |
File-New-Project/EarTrumpet | 155603967 | Title: VMware Workstation machines only have a single entry
Question:
username_0: Ear Thrumpet can only control only one of the running VMware Workstation virtual machines.
Answers:
username_1: Do multiple machines show up in the classic sound mixer? VMware might be piping all audio through a single broker process
username_0: It did show up as multiple entries on Windows 7. I don't control the volume of all machines either, just one of them.
username_1: Can you repro this on Windows 10 with the classic sound mixer?
username_0: No, they are separate entries there.
Status: Issue closed
|
kyrylo/airbrake-rust | 230488411 | Title: Notification retry feature idea
Question:
username_0: Hi @kyrylo
I'm using Airbrake-Rust in an environment with really bad network connection, so I'm in a desperate need of a new feature to reliably send error notifications. I'm thinking about adding a "retry" feature with some configurable timeouts and number of maximum retries.
Before implementing something this big I would like to get your feedback on the idea and discuss the implementation details.
A: We will definitely need better error handling in the lib to determine if a notice was sent successfully or not. I've tried to add a simple boilerplate Error type with the error_chain! macro, and replace every unwrap() in SyncSender::send() with a try!(). It works fine, except for the serde_json::from_str call, which has some extra requirements which I couldn't handle, but I will look into it a bit more.
Do you mind changing the notify_sync function to return a Result<Value> instead a Value? I hope it's not a problem, after all if a developer wants to block the app while the notification is beeing sent, he would probably want to know if it was successfull or not (and also don't want his program to die in case of an unexpected response of course)
B: Where should I put the retry logic? I think AsyncSender could have the same retry functionality, because it basically includes a SyncSender in a separate thread, so that thread can also do the retry thing, even if we discard the final results after a given number of failures. So maybe we can have a simple wrapper type for SyncSender which contains the retry logic and the related params. Then the AsyncSender worker should decide if a simple SyncSender or a new Retry Sender should be created based on the config params. In this case the Notifier would have 3 senders, Sync, RetrySync and Async.
The other option is to not separate this, and make the SyncSender smarter by adding the retry logic to the send() function. I mean it gets the whole config in the constructor, so it can easily decide itself if any retries are needed. This way the whole structure can be the same, and we only have to add some new optional config params.
Both methods are backward compatible with the API.
Answers:
username_0: I've managed to get the error_chain feature working, so you can check out a working prototype in my repo on branch: retry_prototype
I went with the second option, making the send() function in SyncSender smarter to handle retries, so AsyncSender doesn't have to know anything about the new retry functionality, it just passes the config.
If it's looking good on your side I will make this change into separate, fully fledged pull requests with updating the Readme and everything, just wanted to get your opinion first. |
kyverno/kyverno | 1133805110 | Title: [VerifyImage] Support more certificate-extensions from cosign
Question:
username_0: **Is your feature request related to a problem? Please describe.**
Currently when using the "verifyImages" check it is possible to check the "subject" and "issuer" of the certificate. When the keyless cosign-signature was created within Github Actions, the "subject" is typically the value from `job_workflow_ref`. As already described in sigstore/fulcio#305 this is insufficient, when the sign-operation is placed in a reusable Github workflow and it is "only" called from a specific repository. The PR sigstore/fulcio#306 added more extracted claims (https://github.com/sigstore/fulcio/blob/main/docs/oid-info.md) to the certificate which can be verified afterwards.
**Describe the solution you'd like**
With the "verifyImages" check from Kyverno it should be possible to verify additional claims from the cosign-certificate as described above. Maybe something like this would be possible:
```yaml
verifyImages:
- image: "ghcr.io/username_0/my-awesome-product:*"
subject: "..."
issuer: "https://token.actions.githubusercontent.com"
additionalClaims:
1.3.6.1.4.1.57264.1.4: release
1.3.6.1.4.1.57264.1.5: username_0/my-awesome-product
roots: |-
...
```
Maybe the internal extension-names should be replaced with more user-friendly names.
**Describe alternatives you've considered**
**Additional context**
Without this, it would not be possible to verify that the image is really created by the repo-owner, as an attacker could also use the public reusable-workflow and would create a valid signature-cert with the exact same issuer and subject.
Answers:
username_1: Thanks, @username_0 - for opening this issue.
I'm just wondering how can we extract these new extensions in kyverno, from the certs? For the subject and issuer we are using the cosign's `signature` pkg - https://github.com/kyverno/kyverno/blob/main/pkg/cosign/cosign.go#L366-L367
username_0: You are right, I missed the fact, that this is the place where the area of cosign starts. 😆
I think we should cross-post the issue at cosign, that the extraction logic there could be expanded.
@username_1 Should I do that?
username_1: Yes, that would be nice, thanks :)
username_0: @username_1 @username_2 I've updated my proposal above for the additionalClaims validation. If its ok for you I will try to start a PR in a few days.
username_1: Thanks, @username_0, for your help! Please go ahead.
username_2: Thanks @username_0! I saw Cosign 1.6.0 is out so would be great to update and get this in.
username_0: I'm working on this issue in the last few days and had problems to debug my code locally (via VS Code).
1. I executed the `deploy-controller-debug.sh` against a kind-cluster with my local IP and Port 8000 as `--serverIP`.
2. I started the main.go from VS Code in debug-mode with the `--kubeconfig` and `--serverIP` arguments. (Kyverno logs are fine and the webhook is registered)
3. The webhook can not be accessed locally or from the Kubernetes-API-Server at the `serverIP` endpoint with e.g. `/mutate` path. With HTTP I always get 404, with HTTPS, the server seems to send a HTTP-Message anyway.
Is there any step missing to debug Kyverno locally?
username_3: @username_0 since you already have kind cluster running i would suggest to build the kyverno image locally using `make docker-build-kyverno-local` and deploy using the `config/install.md` operator yaml file as a pod.
username_0: @username_3 I've tried that too, but than I can't debug right? The only way to get infos from my code would be then to insert logs?
username_3: Right , then we have to check the pod logs to see the debug statements.
username_0: Thanks @realshuting and @username_3. I will give it a try.
Status: Issue closed
|
stanvanrooy/instauto | 704081138 | Title: API No longer logging in
Question:
username_0: I get the following error when trying to run the login code flow:
`Traceback (most recent call last):
File "insta.py", line 158, in <module>
main()
File "insta.py", line 48, in main
getsubnums()
File "insta.py", line 55, in getsubnums
client.login()
File "/home/user/anaconda3/lib/python3.8/site-packages/instauto/api/actions/authentication.py", line 64, in login
self.state.logged_in_account_data = LoggedInAccountData(**resp.json()['logged_in_user'])
KeyError: 'logged_in_user'
`
Answers:
username_0: And it seems I have discovered what 100+ requests in a second will do
Status: Issue closed
username_1: This happens sometimes when logging in from a saved file. Could you share the code you use to initiate the client?
username_0: It's all good! I had a while loop with the login flow and it sent over 100 requests to instagram to login. it works fine after waiting 2 minutes |
fluentmigrator/fluentmigrator | 560943603 | Title: Obsolete way to run migration on demand
Question:
username_0: Currently FluentMigrator scanning migrations on startup when using it in aspnet core project.
I found way to run migration on demand by calling for example action
Now I do that with RunnerContext which is obsolete
Any ways to do that not only at startup when configuring services ?
Answers:
username_1: @username_0 Thanks for your patience - was on vacation last week.
Artur, can you please post a code sample of what you're currently doing? My understanding is you want to remove all Obsolete compiler warnings from your solution. The best way, IMHO, for me to help you with that is to see a code example representative of what your solution is doing. The other option is to follow the example in FluentMigrator samples directory and build an "In-Process Runner".
username_0: @username_1 Yep, sure
```C#
internal virtual void MigrateToLatestDatabaseVersion(string connectionString)
{
var assembly = Assembly.GetExecutingAssembly();
var logAnnouncer = new TextWriterAnnouncer(s => _logger.Debug(s));
var migrationContext = new RunnerContext(logAnnouncer)
{
Namespace = "Logic.Dal.Migrations"
};
var options = new MigrationOptions();
var factory = new SqlServer2012ProcessorFactory();
using (var processor = factory.Create(connectionString, logAnnouncer, options))
{
var runner = new MigrationRunner(assembly, migrationContext, processor);
runner.MigrateUp();
}
}
```
`TextWriterAnnouncer` marked as `Obsolete` and other like `RunnerContext`, `SqlServer2012ProcessorFactory`, `MigrationRunner` with specified constructor
username_1: Got it. I am releasing 3.2.2. hopefully tonight which contains so better warnings than just Obsolete.
TextWriterAnnouncer still exists, but the general Announcer abstraction has been replaced with the ability to log arbitrarily anything you want, and each logger can thus have their own output streams (so you can log things to a database, windows event log, file, cloud service like datadog, whatever).
Please look at: https://github.com/fluentmigrator/fluentmigrator/blob/e82aafa20e6dbe3cefa221303fe23cf8bf59fffd/samples/FluentMigrator.Example.Migrator/Program.DependencyInjection.cs
In particular, https://github.com/fluentmigrator/fluentmigrator/blob/e82aafa20e6dbe3cefa221303fe23cf8bf59fffd/samples/FluentMigrator.Example.Migrator/Program.DependencyInjection.cs#L31-L32
username_1: @username_0 See this push just now: https://github.com/fluentmigrator/fluentmigrator/commit/69e1a267b7278d02f794d0a94d32054243f6f56c
username_0: Thank you very much, that was helpful 😃
Status: Issue closed
|
ant-design/pro-components | 1158270017 | Title: 🐛[BUG] ProFormList 表单项无法基于label右对齐
Question:
username_0: 提问前先看看:
https://github.com/ryanhanwu/How-To-Ask-Questions-The-Smart-Way/blob/main/README-zh_CN.md
### 🐛 bug 描述
使用ProFormList包裹的表单项无法label右对齐
<img width="605" alt="image" src="https://user-images.githubusercontent.com/8169783/156552164-7356e88b-b1e4-4a9e-afde-6b6d816017d4.png">
<!--
详细地描述 bug,让大家都能理解
-->
### 📷 复现步骤
https://codesandbox.io/s/jin-e-forked-knvchd?file=/App.tsx
https://codesandbox.io/s/biao-qian-yu-biao-dan-xiang-bu-ju-forked-juh3kt?file=/App.tsx:443-698
### 🏞 期望结果
ProFormList可以不影响表单对齐结果
<!--
描述你原本期望看到的结果
-->
### 💻 复现代码
<!--
提供可复现的代码,仓库,或线上示例
-->
### © 版本信息
- ProComponents 版本: [e.g. 4.0.0]
- umi 版本
- 浏览器环境
- 开发环境 [e.g. mac OS]
### 🚑 其他信息
<!--
如截图等其他信息可以贴在这里
-->
Answers:
username_1: 这个实在是没法实现,你 ProFormList 里面包个 Row 手动实现一把。
ProFormList 这个块要和 form 做切分的。
Status: Issue closed
username_0: 最后通过给每一个ProFormText ProFormCheckBox 都传入layout配置实现了,虽然有点怪,但可用
https://codesandbox.io/s/jin-e-forked-xdk03o?file=/App.tsx |
MicrosoftDocs/azure-devops-docs | 599402292 | Title: Which role is needed to manage approvals / checks
Question:
username_0: There is no documentation on which role is needed for managing approvals and checks on the environment. I assume this should be administrator?
Also I hope that we can see a new role over time that does allow setting approvers and checks but not grant setting security permissions.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77d95db6-9983-7346-d0eb-4b7443e4e252
* Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087
* Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops#feedback)
* Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @username_6
* Microsoft Alias: **jukullam**
Answers:
username_1: Thanks for your question. Here are a couple of options you might consider:
- [Azure DevOps Support Bot](https://azuredevopsvirtualagent.azurewebsites.net/)
- [Azure DevOps on Stack Overflow](https://stackoverflow.com/questions/tagged/azure-devops)
Also, thanks for the suggestion! The product teams aren’t always watching this repository, so to make sure it gets in front of them for triage, please [submit your idea]( https://developercommunity.visualstudio.com/content/idea/post.html?space=21) here.
Status: Issue closed
username_0: This is not a support request, this is feedback that your description of the roles is lacking. You do not describe which role on an environment is needed to manage approvals and checks.
username_2: Reopening for testing PRMerger priority label assignment
username_2: There is no documentation on which role is needed for managing approvals and checks on the environment. I assume this should be administrator?
Also I hope that we can see a new role over time that does allow setting approvers and checks but not grant setting security permissions.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77d95db6-9983-7346-d0eb-4b7443e4e252
* Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087
* Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops#feedback)
* Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @username_6
* Microsoft Alias: **jukullam**
username_1: @username_0 -- thank you for clarifying.
@username_2 -- this issue is alive again; please don't close it.
username_2: #pri3
Status: Issue closed
username_2: There is no documentation on which role is needed for managing approvals and checks on the environment. I assume this should be administrator?
Also I hope that we can see a new role over time that does allow setting approvers and checks but not grant setting security permissions.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77d95db6-9983-7346-d0eb-4b7443e4e252
* Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087
* Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops#feedback)
* Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @username_6
* Microsoft Alias: **jukullam**
username_3: @username_2 I think this is to remain open. Is that correct, @username_1?
username_3: Woops, page was stale when I last commented. Pls ignore.
username_4: This issue hasn't been updated in more than 180 days, so we've closed it. If you feel the issue is still relevant and needs fixed, please reopen it and we'll take another look. We appreciate your feedback and apologize for any inconvenience.
Status: Issue closed
username_0: Nevermind I guess..
username_5: @username_6 Can you take a look at this one?
username_5: There is no documentation on which role is needed for managing approvals and checks on the environment. I assume this should be administrator?
Also I hope that we can see a new role over time that does allow setting approvers and checks but not grant setting security permissions.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77d95db6-9983-7346-d0eb-4b7443e4e252
* Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087
* Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops#feedback)
* Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @username_6
* Microsoft Alias: **jukullam**
username_6: @username_0 You can manage approvals and checks with the creator, administrator and user roles. The reader role cannot manage approvals and checks. Updating the docs now.
Status: Issue closed
|
ioBroker/ioBroker.iot | 506171827 | Title: [GHOME] Crashing after Thermostat setting Value / License Bug ?
Question:
username_0: Here is my log output:
`
iot.0 | 2019-10-12 12:51:17.556 | info | (12173) Terminated (NO_ERROR): Without reason
-- | -- | -- | --
iot.0 | 2019-10-12 12:51:17.555 | info | (12173) terminating
iot.0 | 2019-10-12 12:51:17.549 | error | at processImmediate [as _immediateCallback] (timers.js:745:5)
iot.0 | 2019-10-12 12:51:17.549 | error | at tryOnImmediate (timers.js:768:5)
iot.0 | 2019-10-12 12:51:17.549 | error | at runCallback (timers.js:810:20)
iot.0 | 2019-10-12 12:51:17.549 | error | at Immediate.setImmediate [as _onImmediate] (/opt/iobroker/node_modules/iobroker.iot/lib/GoogleHome.js:2432:41)
iot.0 | 2019-10-12 12:51:17.549 | error | at GoogleHome.getStates (/opt/iobroker/node_modules/iobroker.iot/lib/GoogleHome.js:2430:26)
iot.0 | 2019-10-12 12:51:17.549 | error | at Adapter.getForeignState (/opt/iobroker/node_modules/iobroker.js-controller/lib/adapter.js:5306:24)
iot.0 | 2019-10-12 12:51:17.549 | error | (12173) TypeError: Cannot read property 'startsWith' of undefined
iot.0 | 2019-10-12 12:51:17.549 | error | (12173) uncaught exception: Cannot read property 'startsWith' of undefined
`
After setting the temperature of a thermostat, I got the feedback that this mode would not be supported by the device
here my settings:

After adpater restarting, I got this:
`[GHOME] Invalid URL key. Status update is disabled you can set states but not receive state updates: {"error":"Email not found or URL key not found"}`
I have a valid .pro license, how could this be fixed?
Answers:
username_1: Which js-Controller?
username_1: If it is js-controller 2.0 this could happen if id is not a string ... I will prevent this error in js-controller, but question stays: why id is no string ?
username_0: js controller is 2.0.25
Which id string do you mean?
username_2: Can you reinstall from my github and restart
https://github.com/username_2/ioBroker.iot
Please upload your devices in the forum
https://forum.iobroker.net/topic/24061/google-home-assistant-iobroker-einrichten-nutzen/
username_0: I allready installed the latest 1.1.9 and this issue is still present.
This is right after trying to set the temp:
`
| | |
-- | -- | -- | --
iot.0 | 2019-10-12 18:33:32.668 | info | (31639) Terminated (NO_ERROR): Without reason
iot.0 | 2019-10-12 18:33:32.668 | info | (31639) terminating
iot.0 | 2019-10-12 18:33:32.623 | error | (31639) TypeError: Cannot read property 'startsWith' of undefined at Adapter.getForeignState (/opt/iobroker/node_modules/iobroker.js-controller/lib/adapter.js:5306:24) at GoogleHome.getStates
iot.0 | 2019-10-12 18:33:32.623 | error | (31639) uncaught exception: Cannot read property 'startsWith' of undefined
iot.0 | 2019-10-12 18:33:09.494 | warn | (31639) [GHOME] Invalid URL key. Status update is disabled you can set states but not receive state updates: {"error":"Email not found or URL key not found"}
iot.0 | 2019-10-12 18:33:09.494 | warn | (31639) [GHOME] Invalid URL key. Status update is disabled you can set states but not receive state updates: {"error":"Email not found or URL key not found"}
iot.0 | 2019-10-12 18:33:09.493 | error | (31639) [GHOME] Url Key error. Google Request and Response are working. But device states are not reported automatically: {"error":"Email not found or URL key not found"}
`
What do you mean with uploading my devices? Simply post a list?
username_1: Is this all log? Please provide also some Logs before the error
username_2: Please reinstall the latest version from github.
And put the instance in debug (Instanzen ->Expertmode-> Log Level ->debug)
It is described in post #1 you can press the export arrow in objects (arrow down) to export a json of your object
username_0: @username_1 this is a log with a triggered settpoint @19:19 Uhr:
`iot.0 2019-10-12 19:20:27.277 warn (19608) [GHOME] Invalid URL key. Status update is disabled you can set states but not receive state updates: {"error":"Email not found or URL key not found"}
iot.0 2019-10-12 19:20:27.277 warn (19608) [GHOME] Invalid URL key. Status update is disabled you can set states but not receive state updates: {"error":"Email not found or URL key not found"}
iot.0 2019-10-12 19:20:27.276 error (19608) [GHOME] Url Key error. Google Request and Response are working. But device states are not reported automatically: {"error":"Email not found or URL key not found"}
iot.0 2019-10-12 19:20:27.130 info (19608) Connection changed: connect
iot.0 2019-10-12 19:20:26.952 info (19608) hm-rpc.2.INT0000001.2 is auto added with type info.
iot.0 2019-10-12 19:20:26.947 info (19608) hm-rpc.2.INT0000001.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.942 info (19608) hm-rpc.2.INT0000011.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.938 info (19608) hm-rpc.2.INT0000004.2 is auto added with type info.
iot.0 2019-10-12 19:20:26.934 info (19608) hm-rpc.2.INT0000004.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.930 info (19608) hm-rpc.2.INT0000002.2 is auto added with type info.
iot.0 2019-10-12 19:20:26.925 info (19608) hm-rpc.2.INT0000002.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.921 info (19608) hm-rpc.2.INT0000007.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.917 info (19608) hm-rpc.2.INT0000010.2 is auto added with type info.
iot.0 2019-10-12 19:20:26.913 info (19608) hm-rpc.2.INT0000010.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.908 info (19608) hm-rpc.2.INT0000008.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.904 info (19608) hm-rpc.2.INT0000006.2 is auto added with type info.
iot.0 2019-10-12 19:20:26.900 info (19608) hm-rpc.2.INT0000006.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.896 info (19608) hm-rpc.2.INT0000005.2 is auto added with type info.
iot.0 2019-10-12 19:20:26.891 info (19608) hm-rpc.2.INT0000005.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.887 info (19608) hm-rpc.0.XXXXX is auto added with type socket.
iot.0 2019-10-12 19:20:26.883 info (19608) hm-rpc.0.XXXXX is auto added with type socket.
iot.0 2019-10-12 19:20:26.879 info (19608) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 19:20:26.859 info (19608) hm-rpc.0.XXXXX is auto added with type dimmer.
iot.0 2019-10-12 19:20:26.854 info (19608) hm-rpc.0.XXXXX is auto added with type dimmer.
iot.0 2019-10-12 19:20:26.850 info (19608) hm-rpc.0.XXXXX is auto added with type dimmer.
iot.0 2019-10-12 19:20:26.846 info (19608) hm-rpc.0.XXXXX is auto added with type dimmer.
iot.0 2019-10-12 19:20:26.840 info (19608) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 19:20:26.821 info (19608) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 19:20:26.817 info (19608) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 19:20:26.813 info (19608) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 19:20:26.791 info (19608) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 19:20:26.786 info (19608) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 19:20:26.782 info (19608) hm-rpc.0.XXXXX is auto added with type dimmer.
iot.0 2019-10-12 19:20:26.778 info (19608) hm-rpc.0.XXXXX is auto added with type dimmer.
iot.0 2019-10-12 19:20:26.773 info (19608) hm-rpc.0.XXXXX is auto added with type dimmer.
iot.0 2019-10-12 19:20:26.769 info (19608) Cannot auto convert hm-rpc.0.XXXXX. Type motion is not available, yet. If you need the state please add him manually
iot.0 2019-10-12 19:20:26.765 info (19608) Cannot auto convert hm-rpc.0.XXXXX. Type motion is not available, yet. If you need the state please add him manually
iot.0 2019-10-12 19:20:26.761 info (19608) Cannot auto convert hm-rpc.0.XXXXX. Type motion is not available, yet. If you need the state please add him manually
iot.0 2019-10-12 19:20:26.757 info (19608) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 19:20:26.752 info (19608) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 19:20:26.748 info (19608) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 19:20:26.565 info (19608) hm-rpc.2.INT0000001.2 is auto added with type info.
iot.0 2019-10-12 19:20:26.558 info (19608) hm-rpc.2.INT0000001.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.534 info (19608) hm-rpc.2.INT0000011.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.511 info (19608) hm-rpc.2.INT0000004.2 is auto added with type info.
iot.0 2019-10-12 19:20:26.504 info (19608) hm-rpc.2.INT0000004.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.481 info (19608) hm-rpc.2.INT0000002.2 is auto added with type info.
iot.0 2019-10-12 19:20:26.473 info (19608) hm-rpc.2.INT0000002.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.450 info (19608) hm-rpc.2.INT0000007.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.423 info (19608) hm-rpc.2.INT0000010.2 is auto added with type info.
iot.0 2019-10-12 19:20:26.415 info (19608) hm-rpc.2.INT0000010.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.392 info (19608) hm-rpc.2.INT0000008.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.369 info (19608) hm-rpc.2.INT0000006.2 is auto added with type info.
iot.0 2019-10-12 19:20:26.362 info (19608) hm-rpc.2.INT0000006.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.333 info (19608) hm-rpc.2.INT0000005.2 is auto added with type info.
iot.0 2019-10-12 19:20:26.324 info (19608) hm-rpc.2.INT0000005.1 is auto added with type thermostat.
iot.0 2019-10-12 19:20:26.290 info (19608) hm-rpc.0.XXXXX is auto added with type socket.
iot.0 2019-10-12 19:20:26.275 info (19608) hm-rpc.0.XXXXX is auto added with type socket.
iot.0 2019-10-12 19:20:26.259 info (19608) hm-rpc.0.XXXXX is auto added with type light.
[Truncated]
iot.0 2019-10-12 18:45:38.383 info (4714) Cannot auto convert hm-rpc.0.XXXXX. Type motion is not available, yet. If you need the state please add him manually
iot.0 2019-10-12 18:45:38.369 info (4714) Cannot auto convert hm-rpc.0.XXXXX. Type motion is not available, yet. If you need the state please add him manually
iot.0 2019-10-12 18:45:38.351 info (4714) Cannot auto convert hm-rpc.0.XXXXX. Type motion is not available, yet. If you need the state please add him manually
iot.0 2019-10-12 18:45:38.324 info (4714) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 18:45:38.305 info (4714) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 18:45:38.283 info (4714) hm-rpc.0.XXXXX is auto added with type light.
iot.0 2019-10-12 18:45:37.452 info (4714) Connecting with a18wym7vjdl22g.iot.eu-west-1.amazonaws.com
iot.0 2019-10-12 18:45:37.425 info (4714) starting. Version 1.1.9 in /opt/iobroker/node_modules/iobroker.iot, node: v8.16.1
iot.0 2019-10-12 18:45:04.775 info (32050) Connection lost
iot.0 2019-10-12 18:45:04.775 info (32050) Connection changed: disconnect
iot.0 2019-10-12 18:45:04.772 info (32050) Terminated (NO_ERROR): Without reason
iot.0 2019-10-12 18:45:04.771 info (32050) terminating
iot.0 2019-10-12 18:45:04.726 error (32050) TypeError: Cannot read property 'startsWith' of undefined at Adapter.getForeignState (/opt/iobroker/node_modules/iobroker.js-controller/lib/adapter.js:5306:24) at GoogleHome.getStates
iot.0 2019-10-12 18:45:04.726 error (32050) uncaught exception: Cannot read property 'startsWith' of undefined
iot.0 2019-10-12 18:44:19.838 debug (32050) system.adapter.admin.0: logging false
iot.0 2019-10-12 18:34:07.809 warn (32050) [GHOME] Invalid URL key. Status update is disabled you can set states but not receive state updates: {"error":"Email not found or URL key not found"}
iot.0 2019-10-12 18:34:07.808 warn (32050) [GHOME] Invalid URL key. Status update is disabled you can set states but not receive state updates: {"error":"Email not found or URL key not found"}
iot.0 2019-10-12 18:34:07.807 error (32050) [GHOME] Url Key error. Google Request and Response are working. But device states are not reported automatically: {"error":"Email not found or URL key not found"}
iot.0 2019-10-12 18:34:07.678 info (32050) Connection changed: connect
`
username_0: The adapter is already in debug mode. I just reinstalled it as you want. I dont get you with the export thing. The linked post does not describe anything.
username_2: thanks for the log but I added addtional debug messages in the latest github version can you check you really reinstalled from github
 Press the down arrow and upload the json in the forum
username_0: Maybe this will help :
`
iot.0 | 2019-10-12 19:47:47.971 | info | (31337) Connecting with a18wym7vjdl22g.iot.eu-west-1.amazonaws.com
-- | -- | -- | --
iot.0 | 2019-10-12 19:47:47.945 | info | (31337) starting. Version 1.1.9 in /opt/iobroker/node_modules/iobroker.iot, node: v8.16.1
host.ioBroker | 2019-10-12 19:47:45.308 | info | instance system.adapter.iot.0 started with pid 31337
chromecast.0 | 2019-10-12 19:47:28.459 | info | (17271) Reconnecting Küche to 192.168.178.33:42229
chromecast.0 | 2019-10-12 19:47:28.459 | info | (17271) Received keep alive for disconnected device Küche
host.ioBroker | 2019-10-12 19:47:15.287 | info | Restart adapter system.adapter.iot.0 because enabled
host.ioBroker | 2019-10-12 19:47:15.287 | info | instance system.adapter.iot.0 terminated with code 0 (NO_ERROR)
host.ioBroker | 2019-10-12 19:47:15.287 | error | Caught by controller[0]: at processImmediate [as _immediateCallback] (timers.js:745:5)
host.ioBroker | 2019-10-12 19:47:15.287 | error | Caught by controller[0]: at tryOnImmediate (timers.js:768:5)
host.ioBroker | 2019-10-12 19:47:15.286 | error | Caught by controller[0]: at runCallback (timers.js:810:20)
host.ioBroker | 2019-10-12 19:47:15.286 | error | Caught by controller[0]: at Immediate.setImmediate [as _onImmediate] (/opt/iobroker/node_modules/iobroker.iot/lib/GoogleHome.js:2432:41)
host.ioBroker | 2019-10-12 19:47:15.286 | error | Caught by controller[0]: at GoogleHome.getStates (/opt/iobroker/node_modules/iobroker.iot/lib/GoogleHome.js:2430:26)
host.ioBroker | 2019-10-12 19:47:15.286 | error | Caught by controller[0]: at Adapter.getForeignState (/opt/iobroker/node_modules/iobroker.js-controller/lib/adapter.js:5306:24)
host.ioBroker | 2019-10-12 19:47:15.285 | error | Caught by controller[0]: TypeError: Cannot read property 'startsWith' of undefined
iot.0 | 2019-10-12 19:47:15.266 | info | (19608) Terminated (NO_ERROR): Without reason
iot.0 | 2019-10-12 19:47:15.265 | info | (19608) terminating
iot.0 | 2019-10-12 19:47:15.259 | error | at processImmediate [as _immediateCallback] (timers.js:745:5)
iot.0 | 2019-10-12 19:47:15.259 | error | at tryOnImmediate (timers.js:768:5)
iot.0 | 2019-10-12 19:47:15.259 | error | at runCallback (timers.js:810:20)
iot.0 | 2019-10-12 19:47:15.259 | error | at Immediate.setImmediate [as _onImmediate] (/opt/iobroker/node_modules/iobroker.iot/lib/GoogleHome.js:2432:41)
iot.0 | 2019-10-12 19:47:15.259 | error | at GoogleHome.getStates (/opt/iobroker/node_modules/iobroker.iot/lib/GoogleHome.js:2430:26)
iot.0 | 2019-10-12 19:47:15.259 | error | at Adapter.getForeignState (/opt/iobroker/node_modules/iobroker.js-controller/lib/adapter.js:5306:24)
iot.0 | 2019-10-12 19:47:15.259 | error | (19608) TypeError: Cannot read property 'startsWith' of undefined
iot.0 | 2019-10-12 19:47:15.259 | error | (19608) uncaught exception: Cannot read property 'startsWith' of undefined
`
I'll send you the json as a PM in forum
username_2: The issue is GoogleHome lib is calling a getForeignState on undefined
We will fix this on both sides.
username_1: js-controller "id sanitizing" will be fixed in 2.0.26 propably Sunday night or monday
username_3: hi, ist this Error fixed already?
I'm getting the same error (js-controller 2.1.1 iot v.1.1.9) if i try to use en Thermostat over Google Home.
username_4: Must be fixed with 1.1.11
username_0: I still have the issue that the license wont be recognised.
`GHOME] Url Key error. Google Request and Response are working. But device states are not reported automatically: {"error":"Email not found or URL key not found"} `
The Thermostats are currently not recognised and not controllable. I changed it to the virtual groups maybe thos is the issue?
I got many of these:
`
Missing name, type or treat for: hm-rpc.2.INT0000002.2.STATE. Not added to GoogleHome
`
username_2: Try to delete iot.0.certs
username_2: Remove INT0000002 manually and let it recreate from hm rpc Adapter
username_0: Seems to be an issue with my grouping, all devices must have some room and group not like mine only some funktions I wanted to be filtered. 
username_2: The channel you want to add needs room and func. Not only the state
username_0: The License error is still persistant
`
iot.0 | 2020-01-04 09:28:14.982 | warn | (28867) [GHOME] Invalid URL key. Status update is disabled you can set states but not receive state updates: {"error":"Invalid URL-KEY"}
-- | -- | -- | --
iot.0 | 2020-01-04 09:28:14.981 | warn | (28867) [GHOME] Invalid URL key. Status update is disabled you can set states but not receive state updates: {"error":"Invalid URL-KEY"}
iot.0 | 2020-01-04 09:28:14.980 | error | (28867) [GHOME] Url Key error. Google Request and Response are working. But device states are not reported automatically: {"error":"Invalid URL-KEY"}
`
username_2: Reset credentials in iot options and delete iot.0.certs
username_0: `
st.ioBroker | 2020-01-04 10:10:53.530 | info | Restart adapter system.adapter.iot.0 because enabled
-- | -- | -- | --
host.ioBroker | 2020-01-04 10:10:53.530 | info | instance system.adapter.iot.0 terminated with code 0 (NO_ERROR)
host.ioBroker | 2020-01-04 10:10:53.530 | error | Caught by controller[0]: at processImmediate (timers.js:658:5)
host.ioBroker | 2020-01-04 10:10:53.529 | error | Caught by controller[0]: at tryOnImmediate (timers.js:676:5)
host.ioBroker | 2020-01-04 10:10:53.529 | error | Caught by controller[0]: at runCallback (timers.js:705:18)
host.ioBroker | 2020-01-04 10:10:53.529 | error | Caught by controller[0]: at Immediate.setImmediate (/opt/iobroker/node_modules/iobroker.iot/lib/GoogleHome.js:2063:41)
host.ioBroker | 2020-01-04 10:10:53.529 | error | Caught by controller[0]: at GoogleHome.getStates (/opt/iobroker/node_modules/iobroker.iot/lib/GoogleHome.js:2061:26)
host.ioBroker | 2020-01-04 10:10:53.529 | error | Caught by controller[0]: at Adapter.getForeignState (/opt/iobroker/node_modules/iobroker.js-controller/lib/adapter.js:5561:24)
host.ioBroker | 2020-01-04 10:10:53.528 | error | Caught by controller[0]: TypeError: Cannot read property 'startsWith' of undefined
iot.0 | 2020-01-04 10:10:53.013 | info | (1146) Connection lost
iot.0 | 2020-01-04 10:10:53.012 | info | (1146) Connection changed: disconnect
iot.0 | 2020-01-04 10:10:53.009 | info | (1146) Terminated (NO_ERROR): Without reason
iot.0 | 2020-01-04 10:10:53.008 | info | (1146) terminating
iot.0 | 2020-01-04 10:10:52.963 | error | (1146) TypeError: Cannot read property 'startsWith' of undefined at Adapter.getForeignState (/opt/iobroker/node_modules/iobroker.js-controller/lib/adapter.js:5561:24) at GoogleHome.getStates (
iot.0 | 2020-01-04 10:10:52.962 | error | (1146) uncaught exception: Cannot read property 'startsWith' of undefined
iot.0 | 2020-01-04 10:10:45.498 | error | (1146) [GHOME] invalid protocol version: {"error":"Unsupported version"}
`
username_0: Okay got it, somehow it was related to my old subscription (3year plan for pro) after purchasing the extra year of cloud remote access its working now. |
mrdoob/three.js | 101401235 | Title: MorphAnimMesh and MorphBlendMesh
Question:
username_0: I found 2 similar classes for playing morph animations. MorphBlendMesh allows to run multiple animations, MorphAnimMesh only one(probably).
I'm not found any docs. I'm not any example with MorphAnimMesh (Is it works?). They have different interfaces(update/updateAnimation, (setAnimationFPS + playAnimation)/(playAnimation))...
I confused. MorphAnimMesh is deprecated?
Answers:
username_1: I've close problems with MorphAnimMesh. My app. with R72 no loger works with R73 and R74 (THREE.MorphAnimMesh is not a constructor???)
username_2: THREE.MorphAnimMesh is deprecated??? i think we need to describe it on docs.
username_1: I strongly advice to choose between the following strategies:
1. you may still use THREE.MorphAnimMesh. Since it is no longer in the core, you have to include the suitable file located here: https://github.com/username_4/three.js/tree/dev/examples/js/THREE.MorphAnimMesh.js (please download the very last THREE release: R79)
2. You throw overboard THREE.MorphAnimMesh and you adopt the new approach for animation in THREE. Unfortunately, your require a lot of re-work in your app. The new support for animation in THREE is poorly illustrated. For basic understanding, I advice to look at this app: morphtargets horse (Browse examples in R79)
3. (I personnaly choose this one): Drop off THREE.MorphAnimMesh and reuse http://www.createjs.com/tweenjs. From experience, TWEEN.JS is easy to understand and can be integrated with THREE without difficulty.
Hope this helps
username_3: @username_0 can this issue be closed?
Status: Issue closed
username_5: I've run into this issue when porting my app to ES6 and using modules. I need to use `SceneLoader` as there is no replacement for this yet, even though it was removed from `three` entirely.
I've got `SceneLoader` ported to ES6 (simply doing `import * as THREE from 'three'` and then export { SceneLoader }`) but it's complaining that `MorphAnimMesh`, which `SceneLoader` uses, is not present in `three` anymore.
The block of code in question is
```
else if ( objJSON.morph ) {
object = new THREE.MorphAnimMesh( geometry, material );
if ( objJSON.duration !== undefined ) {
object.duration = objJSON.duration;
}
if ( objJSON.time !== undefined ) {
object.time = objJSON.time;
}
if ( objJSON.mirroredLoop !== undefined ) {
object.mirroredLoop = objJSON.mirroredLoop;
}
if ( material.morphNormals ) {
geometry.computeMorphNormals();
}
```
I'm tempted to simply comment it out and see what happens but if I could get a solution that would be great.
username_4: Oh? `ObjectLoader` is supposed to be the replacement of `SceneLoader`.
username_5: It doesn't work the same way though, I tried it during my upgrade and just couldn't get it to work the same way :( I've made a stackoverflow post about it somewhere. It's fine though, `SceneLoader` still works, just had to do these little changes for my ES6 migration. |
HandyFriends/HandyFriends | 747883550 | Title: Code log in & sign up screen
Question:
username_0: Time commitment: 1-2 hours
Required user story:
- User logs in (or signs up) to access menu of various services they can get. Use Google Firebase
Optional:
- Code the following functionality: App remembers the user for certain period of time (automatic logout only after certain period of time)<issue_closed>
Status: Issue closed |
wireservice/csvkit | 197514566 | Title: Test the tutorial
Question:
username_0: So much has changed
Answers:
username_1: * `csvformat -M "\r" examples/dummy.csv` prints `a,b,c\r1,2,3\r`. Did not expect literal `\r`.
* In `csvsql.rst` multiple commands error with `sqlalchemy.exc.ArgumentError: Column must be constructed with a non-blank name or assign a non-blank .name before adding to a Table.`. Some others error with `Row 0 has 4 values, but Table only has 3 columns.`
* `csvstat examples/realdata/FY09_EDU_Recipients_by_State.csv` has lots of agate warnings.
* Other `csvstat` commands don't have expected output. See #714
* `in2csv examples/test.xls` errors with `UnicodeDecodeError: 'ascii' codec can't decode byte 0xce in position 9: ordinal not in range(128)`
* `curl https://api.github.com/repos/wireservice/csvkit/issues\?state\=open | in2csv -f json -v` fails with `UnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position 1249: ordinal not in range(128)`
* `curl "http://oakland.crimespotting.org/crime-data?format=json&type=robbery&count=10" | in2csv -f geojson` needs to be cut/replaced, since that API no longer returns a useful response.
username_1: Looked at the tutorial now. Besides #711, there's:
* `csvsql -i sqlite joined.csv` and `csvsql --db sqlite:///leso.db --insert joined.csv` and `csvsql --query "select county,item_name from joined where quantity > 5;" joined.csv` warn `DuplicateColumnWarning: Column name "fips" already exists in Table. Column will be renamed to "fips_2".`
username_0: Is that last issue you reported meant to be checked off? If so, what was the fix?
username_0: I can't replicate your csvsql issue. Can you confirm with latest agate?
username_0: Can not confirm `csvstat` issue either.
username_0: #714 I'm not going to fix for 1.0. That can be resolved in the next minor release.
username_0: Nevermind, scratch my comments from a moment ago. I've replicated your bugs.
username_0: Okay I've broken out the issues I can replicate into separate tickets. (Some for 1.0, some not.)
Status: Issue closed
username_1: * `curl "http://oakland.crimespotting.org/crime-data?format=json&type=robbery&count=10" | in2csv -f geojson` I just removed. I don't think we absolutely need a GeoJSON example. The API wasn't returning any features to transform.
* If you mean the`DuplicateColumnWarning` line, that occurred because I had two `fips` columns at some point while testing, but that was before you merged a csvjoin fix.
username_0: Got it, great! |
nuxodin/ie11CustomProperties | 658105116 | Title: !important not working
Question:
username_0: On version 4.1.0, I encountered the next behaviour:

My code looks like: color: var(--white) !important; where var(--white) is --white: #ffffff
Answers:
username_1: Is this library being maintained?
username_0: Don't think so. No response whatsoever.
username_2: Sorry, I have no free resources at the moment.
username_3: @username_0 @username_1 - can you please test the proposed fix in https://github.com/username_2/ie11CustomProperties/pull/63/files
If you could post back here with any feedback on that it'd be appreciated! |
lukasroegner/homebridge-dyson-pure-cool | 764425620 | Title: API error with TP02
Question:
username_0: Getting these error messages:
[12/12/2020, 13:25:56] [DysonPureCoolPlatform] Error while signing in. Status Code: 429
[12/12/2020, 13:25:56] [DysonPureCoolPlatform] API could not be reached. Retry is disabled.
I've triple-checked the email/username/countrycode/serial.
Any response on what 429 represents?
Answers:
username_1: Same problem....TP02 not work with the plugin...but my DP04 work fine
username_2: As stated several times in the issues:
- Make sure you have installed the latest version of the plugin
- Make sure the country code is correct
- Make sure your credentials are correct
- Check whether the name of the device in the Dyson app contains any special characters
If you provide more information (sanitized config), I can take a deeper look.
username_1: Its ok now, the name of the device work fine with other home bridge plugin but with this I must delete the DYSON- before and the -475 after the serial.
Status: Issue closed
username_3: Can you please show me a screenshot how you setted everything up.
I'm not able to get my TP02 working.
username_3: Looks like the plugin is not supporting an special characters in the password.
I now use a password without any special characters, only letters and numbers.
now the login is possible. |
dresden-elektronik/deconz-rest-plugin | 723508854 | Title: deCONZ 2.05.85 crashes with Segmentation fault
Question:
username_0: <!--
- Use this issue template to report a bug in the deCONZ REST-API.
- If you want to report a bug for the Phoscon App, please head over to: https://github.com/dresden-elektronik/phoscon-app-beta
- If you're unsure if the bug fits into this issue tracker, please ask for advise in our Discord chat: https://discord.gg/QFhTxqN
- Please make sure sure you're running the latest version of deCONZ: https://github.com/dresden-elektronik/deconz-rest-plugin/releases
-->
## Describe the bug
<!--
deCONZ 2.05.85 running in Docker
ConBee II Firmware: 26580700
deCONZ crashes after some 20 seconds with an error message: "Segmentation fault"
Version 2.05. 79 and Version 2.05.84 runs just fine.
-->
## Steps to reproduce the behavior
<!--
If the problem is reproducable, list the steps here:
1. Environment settings:
```
DEBIAN_FRONTEND noninteractive
DEBUG_APS 0
DEBUG_INFO 1
DEBUG_OTAU 0
DEBUG_ZCL 0
DEBUG_ZDP 0
DECONZ_DEVICE /dev/ttyACM0
DECONZ_UPNP 1
DECONZ_VERSION 2.05.85
DECONZ_VNC_DISPLAY 1
DECONZ_VNC_MODE 1
DECONZ_VNC_PASSWORD <PASSWORD>
DECONZ_VNC_PORT 5900
DECONZ_WEB_PORT 6080
DECONZ_WS_PORT 6443
LC_ALL C.UTF-8
PATH /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
TZ Europe/Oslo
```
Network Mode = HOST
The Container runs in Privileged Mode
Every time the container is started, it connects properly with the ConBee II and both VNC as well as Phoscon can be launched. but after some 20 seconds, deCONZ crashes with the following console output:
```
21:54:05:216 [INFO] - No button handler for: Super TR endpoint: 0x01 cluster: 0x0201 command: 0x0A payload[0]: 015
21:54:05:216 ZCL attribute report 0x000D6F000E9DC646 for cluster: 0x0201, ep: 0x01, frame control: 0x08, mfcode: 0x0000
21:54:22:129 rule event /groups/1/state/any_on: 0 -> 1
21:54:23:031 Bind response success for 0xccccccfffec2e48a ep: 0x01 cluster: 0x0000
21:54:23:118 ZCL configure reporting rsp seq: 35 0xCCCCCCFFFEC2E48A for ep: 0x01 cluster: 0x0000 attr: 0x4000 status: 0x00
21:54:23:219 Bind response success for 0xccccccfffec2e48a ep: 0x01 cluster: 0x0006
21:54:23:305 ZCL configure reporting rsp seq: 36 0xCCCCCCFFFEC2E48A for ep: 0x01 cluster: 0x0006 attr: 0x0000 status: 0x00
21:54:23:405 Bind response success for 0xccccccfffec2e48a ep: 0x01 cluster: 0x0008
21:54:23:405 skip configure report for cluster: 0x0008 attr: 0x0000 of node 0xCCCCCCFFFEC2E48A (wait reading or unsupported)
21:54:24:531 Bind response success for 0xccccccfffec2e48a ep: 0x01 cluster: 0x0300
[Truncated]
- Host system: QNAP 453Be NAS
- Running method: Marthoc Docker container
- Firmware version: 26580700
- deCONZ version: 2.05.85
- Device: ConBee II
- Do you use an USB extension cable: yes
### deCONZ Logs
<!--
Debug logs can help to investigate certain problems like device pairing and communication issues.
Logging can be enabled on the command line as described in:
https://github.com/dresden-elektronik/deconz-rest-plugin/wiki/deCONZ-debug-switches
-->
## Additional context
<!--
If relevant, add any other context about the problem here, like network size, number of routers and end-devices
and what kind of devices/brands are in the network.
-->
Answers:
username_1: I have the same behaviour. Do you have an Elko Super TR as well?
## Describe the bug
<!--
I just upgraded to 2.05.85. deCONZ crashes after few minutes. I suspect the bug is related to the following PR: https://github.com/dresden-elektronik/deconz-rest-plugin/pull/3329
-->
## Steps to reproduce the behavior
<!--
The bug is reproduced at each launch. If the Elko Super TR is removed from the network, deCONZ is running fine.
-->
## Environment
- Host system: Raspberry Pi 3B+
- Running method: Raspbian
- Firmware version: 26580700
- deCONZ version: 2.05.85
- Device: Raspbee I
### deCONZ Logs
<!--
20:48:53:923 ZCL configure reporting rsp seq: 51 0x000B57FFFE8CC7F2 for ep: 0x01 cluster: 0x0008 attr: 0x0000 status: 0x00
20:48:54:078 Bind response success for 0xd0cf5efffe11cfcd ep: 0x01 cluster: 0x0006
20:48:54:079 configure reporting rq seq 52 for 0xD0CF5EFFFE11CFCD, attribute 0x0006/0x0000
20:48:54:126 Incr. ZDP retry count 2 on item 7
20:48:54:209 ZCL configure reporting rsp seq: 52 0xD0CF5EFFFE11CFCD for ep: 0x01 cluster: 0x0006 attr: 0x0000 status: 0x00
20:48:54:379 rule event /config/localtime: 20:48:53.377 -> 20:48:54.377 (1000ms)
20:48:54:807 poll node fc00:e968:6179::de52:7100-01-0201
**20:48:54:807 Poll ZHAThermostat sensor node Super TR
Segmentation fault**
Full log:
[deconz_log.txt](https://github.com/dresden-elektronik/deconz-rest-plugin/files/5394270/deconz_log.txt)
-->
username_2: @username_1 You save this one, otherwise I would have closed it. So far, I've spotted two errors (usage of wrong function), but that shouldn't lead to a segfault. For another candidate related to polling, I'm not too sure. Could you make a core dump as described in the wiki? That would presumably help speeding up the search.
username_1: Here is a core dump:
[core-deCONZ-sig11-user1000-group1000-pid2192-time1602915330.gz](https://github.com/dresden-elektronik/deconz-rest-plugin/files/5395080/core-deCONZ-sig11-user1000-group1000-pid2192-time1602915330.gz)
username_0: I discovered that by cmpletely deleting all content in the _config_ folder (/root/.local/share/dresden-elektronik/deCONZ) instead of carrying it over from my previous version, and then reinstalled deCONZ, I no longer got the segfault.
But of course, instead - I lost all my devices and the prospect of re-paring isn't fun because it breaks all my automations and scripts in Home Assistant as well since all entity names are renewed (changed).
username_1: @username_0 Does the Segmentation Fault reappear when adding the Super TR?
username_0: The funny thing is that the Super TR appeared in deConz without me doing anything. But with a new (0xYYYY) code. I havent' tried to add it to PhosCon - is that what you mean?
username_1: Is the Super TR available in the Rest API?
After deleting my Super TR nodes, deCONZ .85 is running fine. Immediately after adding this through Phoscon I get Segfault and also after new startup attempts.
username_0: Yes, you are right. At first it was not available in the API. After adding it through Phoscon, I too immediately got Segfault.
username_2: @username_1 I asked @manup to have a look, but apparently he also couldn't spot anything causing the segfault. You may want to give compiling my latest PR a try, but I have little hope that it solves it.
username_1: @username_2 Thanks! I compiled your code. I had to add VENDOR_EMBER in two of the function calls. Unfortunately, the segmentation fault is still there.
By mistake I copied the .84 libde_rest_plugin.so. This is running flawlessly with deconz .85 (this maybe obviously for you but I thought I could mention it)
username_2: Crap. Where did you need to add that and why?
username_1: In the two calls: addTaskThermostatReadWriteAttribute
The function call was not in agreement with definition, so compilation failed. I thought there was a missing manufacturer code. Do you agree?
username_2: Ah, in my memory, the mfc was optional and defaulting to 0...
It should have been 0 or 0x0000 btw since you mentioned that the attributes are not manufacturer specific. Anyway, that shouldn't have made any difference...
username_3: I have never used that, but there is some tool to read the core dump to have more informations ? Or it s something platform dependant too ?
username_1: Since the rest_plugin from .84 works, maybe I could compile try to compile different commits from .84 and .85 and see for which version the seg fault is appearing? Can this contribute to debugging? I could also provide new core dump to see if it has changed (not sure if previous dump gave indication on which call that caused the seg fault?)
username_3: Ha yep, this error message can be usefull yes. We know the command now.
username_1: Sorry, I am new with gdb but this backtrack should be consistent with @username_2 fixes branch.
0 0x72ce2af8 in Resource::item(char const*) const () from /usr/share/deCONZ/plugins/libde_rest_plugin.so
#1 0x72cd354c in LightNode::manufacturer() const () from /usr/share/deCONZ/plugins/libde_rest_plugin.so
#2 0x72cd8434 in PollManager::pollTimerFired() () from /usr/share/deCONZ/plugins/libde_rest_plugin.so
#3 0x760cabfc in QMetaObject::activate(QObject*, int, int, void**) ()
from /usr/lib/arm-linux-gnueabihf/libQt5Core.so.5
#4 0x760d7d04 in QTimer::timerEvent(QTimerEvent*) () from /usr/lib/arm-linux-gnueabihf/libQt5Core.so.5
#5 0x760cb9c4 in QObject::event(QEvent*) () from /usr/lib/arm-linux-gnueabihf/libQt5Core.so.5
#6 0x76930408 in QApplicationPrivate::notify_helper(QObject*, QEvent*) ()
from /usr/lib/arm-linux-gnueabihf/libQt5Widgets.so.5
#7 0x7693860c in QApplication::notify(QObject*, QEvent*) () from /usr/lib/arm-linux-gnueabihf/libQt5Widgets.so.5
#8 0x02047c80 in ?? ()
username_2: Cool, that should help. I'll check it out.
username_2: Could you please try replacing the current code in `thermostat.cpp` with this one here?
```
case 0x0403: // Temperature measurement
{
if (zclFrame.manufacturerCode() == VENDOR_EMBER && sensor->modelId().startsWith(QLatin1String("Super TR"))) // ELKO
{
quint8 mode = attr.numericValue().u8;
QString mode_set;
if ( mode == 0x00 ) { mode_set = QString("air sensor"); }
else if ( mode == 0x01 ) { mode_set = QString("floor sensor"); }
else if ( mode == 0x03 ) { mode_set = QString("floor protection"); }
else { mode_set = QString("unknown"); }
item = sensor->item(RConfigTemperatureMeasurement);
if (item && item->toString() != mode_set)
{
item->setValue(mode_set);
enqueueEvent(Event(RSensors, RConfigTemperatureMeasurement, sensor->id(), item));
configUpdated = true;
}
}
sensor->setZclValue(updateType, ind.srcEndpoint(), THERMOSTAT_CLUSTER_ID, attrId, attr.numericValue());
}
break;
```
username_1: Sorry! Still same Segmentation fault!
I made a dirty trick:
- in void PollManager::poll(RestNodeBase *restNode, const QDateTime &tStart) I disabled the ZHAThermostat sensors by
else if (r->prefix() == RSensors)
{
sensor = dynamic_cast<Sensor*>(restNode);
DBG_Assert(sensor);
if (!sensor || sensor->deletedState() != Sensor::StateNormal)
{
return;
}
**if (qPrintable(sensor->type()) == QLatin1String("ZHAThermostat") )
{
return;
}**
I no longer have Segmentation Fault so at least some progress.
I expect this has something to do with LightNode manufacture name. I notice that the Manufacturer name from the Rest API is Heiman.
username_2: Thanks. The Heiman thing is interesting, as it shouldn't be. Is a light resource created for it?
username_1: No light resource created. Maybe I should remove and add again in the new version?
{
"config": {
"heatsetpoint": 2500,
"locked": null,
"offset": 0,
"on": true,
"reachable": true,
"schedule": {},
"schedule_on": null,
"temperaturemeasurement": null
},
"ep": 1,
"etag": "918b64f252e1a2de0bb4d4bdba14cc9c",
"lastseen": "2020-10-18T16:49Z",
"manufacturername": "Heiman",
"modelid": "Super TR",
"name": "Super TR (2)",
"state": {
"floortemperature": 0,
"heating": false,
"lastupdated": "2020-10-18T16:49:47.500",
"on": false,
"temperature": 2000
},
"type": "ZHAThermostat",
"uniqueid": "00:0d:6f:00:15:61:31:2d-01-0201"
}
username_1: Let me double check that Heiman thing. I might have changed some Vendor ID in the debugging
username_1: Just checked some more. The "Heiman" manufacturer name is still present with the username_2:fixes branch + https://github.com/dresden-elektronik/deconz-rest-plugin/issues/3427#issuecomment-711164306 + disabling polling for ZHAThermostats.
Should I try to remove node and add again in 2.05.85?
username_2: I just added a protection to change the manufacturer to Heiman in the PR so that shouldn't happen again. I still don't see the root cause for the segfault to occur.
username_0: I follow this thread with great interest. Regretfully I'm not able to contribute much, but I depend on you guys to solve this.
For what it's worth, "Heiman" is the name of the German manufacturer of the ELKO Thermostat.
ELKO doesn't make any of this themselves. They just buy OEM from Heiman in Germany. But ELKO have defined part of the feature spec and they contribute somewhat to the firmware in the device. So the name Heiman should be there, and has always been there.
username_1: Ok, but in Deconz when reading the Manufacturer name attribute is ELKO. I was just asking if this was an indication of a logical error in the API.
I will continue to add more debug statements in order to pin-point where the segmentation fault is occurring.
username_2: Hi. I've done some debugging with @manup and looks like we've found a possible candidate for this one. Seems that you have a certain combination of values, not too common, that's causing it.
@username_1 Would be cool if you could include the upcoming changes by manup locally and test it.
username_1: Thanks for the update. I can now confirm with the .86 release that this solved the segmentation fault!
Great work :)
username_0: Congratulation!! Any chance for us to see .86 on Docker HUB ?
username_2: Great!
That's up to marthoc, but usually doesn't take long
username_4: That Heiman as manufacturer is a thing, as it also happened to me with Tuya/Moes TRV. It was correct in deCONZ, but REST API calls said Heiman. After @username_3 made changes to his Tuya branch it's correct in API as well. It's issue #3440
No light resource created for me on 2.05.85, but on earlier versions that didn't support that valve it was.
username_0: I can confirm now that .86 runs in Docker without any segmentation fault.
I'm not sure what exactly to expect from .86 but even if the Thermostat is "successfully" paired with Phoscon, it still doesn't show up in Phoscon. The REST-API shows the ELKO (and the name "Heiman" is gone).
There is no visible to differences in the deCONZ attribute-list (as far as I can see).
So, what now?
username_1: I have the new attributes in the list but the new attributes are not updating. It is also interesting that one Super TR has manufacturer Elko while the other Super TR has Heiman:
"config": {
"heatsetpoint": 500,
"locked": true,
"offset": 0,
"on": true,
"reachable": true,
"schedule": {},
"schedule_on": null,
"temperaturemeasurement": null
},
"ep": 1,
"etag": "356ccc27c349d7ed14547df0097cbac4",
"lastseen": "2020-10-20T18:39Z",
"manufacturername": "Heiman",
"modelid": "Super TR",
"name": "Super TR (2)",
"state": {
"floortemperature": 0,
"heating": false,
"lastupdated": "2020-10-20T18:39:19.666",
"on": false,
"temperature": 1900
},
"type": "ZHAThermostat",
"uniqueid": "00:0d:6f:00:15:61:31:2d-01-0201"
}
{
"config": {
"heatsetpoint": 500,
"locked": null,
"offset": 0,
"on": true,
"reachable": true,
"schedule": {},
"schedule_on": null,
"temperaturemeasurement": null
},
"ep": 1,
"etag": "36b219c30bae5783bd90dbdac72690e2",
"lastseen": "2020-10-20T20:36Z",
"manufacturername": "ELKO",
"modelid": "Super TR",
"name": "<NAME>",
"state": {
"floortemperature": 0,
"heating": false,
"lastupdated": "2020-10-20T17:01:01.534",
"on": false,
"temperature": 2020
},
"type": "ZHAThermostat",
"uniqueid": "00:0d:6f:00:15:55:29:4f-01-0201"
}
username_3: Lol, I m realy bored by this heiman ^^.
If you want I can remove the line in the fonction, username_2 (that we have talked the last time) ?
Status: Issue closed
username_2: First of all, it is no surprise that the device is not visible in Phoscon, as it is just another REST API client. Also take note, that we're here in the REST API respository. So in that matter, please raise that in the Phoscon beta repository.
Secondly, you should reset and re-pair the devices. This deletes the current sensors (or marks them as such) and creates a new one with all current capabilities. Having no updates in terms of value updates could have 2 reasons: 1. bindings have been missed. Should also be resolved by reset/re-pair. 2. The device does not support attribte reporting.
Lastly, please note that we're driving here on the highway in the night without any lights. No dev has the device at hand to sniff and see what's over the air, so all is based on best effort.
Anyway, please raise a new issue if anything's not working and we can see what to do. As is is not topic related anymore, I'm closing this. |
reactphp/socket | 1169742740 | Title: How to handle socket connection on Server
Question:
username_0: I am building Socket TCP Server for laravel .
I want to keep all the connections alive till i don't do connection->close(). But when i do connectionWrite->write ("Hello world") the message is depicted on client screen but connection is also closed. I don't want the connection to die. I want to reuse the connection in future transaction with server. I don't want to close the connection as it does on its own. I am using Postman like Nightingale for testing( as a client).

Answers:
username_1: Hey @username_0, thanks for bringing this up 👍
It looks like what you're seeing is normal HTTP behavior. After the client requests something and the server responds there's no need for the client to leave the connection open. After receiving its answer the client may close the connection.
If you're using postman for testing purposes you're most likely speaking HTTP. You can take a look at https://github.com/reactphp/http, this may help!
username_2: I believe this has been answered, so I'm closing this for now. Please come back with more details if this problem persists and we can always reopen this :+1:
Status: Issue closed
|
ChalkyBrush/roshpit-bug-tracker | 691080074 | Title: [Oracle UI] Putting additional Reanimation Stone over your stash limit (5 or 20) destroys it
Question:
username_0: I just did it accidently, and probably will never happen normally if you pay attention, but stone just got devoured
I think other keys are impossible to put into stash in a first place if you have too many of them
Answers:
username_0: ok that's now interesting because I don't know how it works yet |
asabaylus/react-command-palette | 1182637294 | Title: UNSAFE_componentWillReceiveProps in strict mode is not recommended
Question:
username_0: **Describe the bug**
Getting this warning when running the code in `React.StrictMode`:
```tsx
<React.StrictMode>
<CommandPalette ... />
</React.StrictMode>
```
```
Warning: Using UNSAFE_componentWillReceiveProps in strict mode is not recommended and may indicate bugs in your code. See https://reactjs.org/link/unsafe-component-lifecycles for details.
```
**Desktop (please complete the following information):**
- OS: Windows
- Browser chrome
Status: Issue closed
Answers:
username_0: **Describe the bug**
Getting this warning when running the code in `React.StrictMode`:
```tsx
<React.StrictMode>
<CommandPalette ... />
</React.StrictMode>
```
```
Warning: Using UNSAFE_componentWillReceiveProps in strict mode is not recommended and may indicate bugs in your code. See https://reactjs.org/link/unsafe-component-lifecycles for details.
```
**Desktop (please complete the following information):**
- OS: Windows
- Browser chrome
username_0: Same issue as #804
username_0: https://github.com/cyberuni/react-command-paletter-strict-mode
Here is the repro. 🌷 |
gpuweb/gpuweb | 239567240 | Title: Binding Inheritance
Question:
username_0: Now that we agreed that compute and graphics work should be started and ended explicitly, an open question is whether resource bindings can be bound outside of "compute / render passes" and inherited inside these passes.
D3D12
-------
D3D12 doesn't have explicit bounds for graphics / compute work so be default root signature tables/descriptors/values are persistent. However there are two root signatures, one for graphics and one for compute. Also the root signature is invalidated if on a call to SetGraphicsRootSignature that changes the layout.
Overall when setting resource bindings outside of passes, we don't know if they should be put in the compute and graphics root signature, and they will in most cases be invalidated by a change to the root signature layout at the beginning of the pass.
Metal
------
Bindings are set inside of the MTLComputeCommandEncoder and MTLRenderCommandEncoder and do not propagate outside of them.
Vulkan
------
Supports inheritance from outside render passes to inside.
Proposal
-------
Given that both D3D12 and Metal do not support inheritance, and given that vkCmdBindDescriptorSet is usually multiple order of magnitudes cheaper than vkCmdBeginRenderPass, it sounds ok to not have inheritance and ask the app to all the bindings inside the passes.
Answers:
username_1: This is true, although it's not clear if that sacrifice needs to happen. If we do inheritance of resource bindings, we'll see that the overhead is moved to `MTLRenderCommandEncoder` of Metal backend, where the same sort of justification may apply (e.g. binding resources is cheaper than starting a pass).
Another concern here is usability. Allowing an application to set descriptor sets outside of render passes would simply be more flexible, and thus allow richer set of abstractions to be build. For example, one may bind the layout/root signature followed by some global resources being set (e.g. weather parameters), only to proceed with a number of passes (e.g. shadow pass per light source) that only set their own specific resources.
TL;DR:
D3D12 model is a lot closer to Vulkan than it is to Metal, and believe providing the resource inheritance in the way of D3D12/Vulkan makes more sense.
username_0: This is a good point, the resolution of this looks like it will be based on ergonomics: I think we agree that the performance impact of allowing or disallowing this inheritance is basically zero. The flip side if that no inheritance would mean the "BindDescriptorSet" function would know what "pipelineBindPoint" is, based on the wether we are in a compute or render pass. I don't have a strong opinion at this point and think we can defer this unless someone has a portability of performance concern.
username_1: Consensus has been reached to not have binding inheritance.
Status: Issue closed
|
kubernetes/kubernetes | 165960015 | Title: kubectl says that PetSet is alpha in master
Question:
username_0: I assume that PetSet will not be alpha in 1.4.
```console
$ cluster/kubectl.sh get
You must specify the type of resource to get. Valid resource types include:
* componentstatuses (aka 'cs')
* configmaps
* daemonsets (aka 'ds')
* deployments (aka 'deploy')
* events (aka 'ev')
* endpoints (aka 'ep')
* horizontalpodautoscalers (aka 'hpa')
* ingress (aka 'ing')
* jobs
* limitranges (aka 'limits')
* nodes (aka 'no')
* namespaces (aka 'ns')
* petsets (alpha feature, may be unstable)
```
Petsets should not be listed as alpha in master.
Obligatory version
```console
$ cluster/kubectl.sh version
Client Version: version.Info{Major:"1", Minor:"4+", GitVersion:"v1.4.0-alpha.1.309+7eeec7aefeec8d-dirty", GitCommit:"<PASSWORD>", GitTreeState:"dirty", BuildDate:"2016-07-17T04:18:10Z", GoVersion:"go1.6.2", Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"4+", GitVersion:"v1.4.0-alpha.1.309+7eeec7aefeec8d-dirty", GitCommit:"<PASSWORD>", GitTreeState:"dirty", BuildDate:"2016-07-17T04:15:47Z", GoVersion:"go1.6.2", Compiler:"gc", Platform:"linux/amd64"}
```
Answers:
username_1: FYI, we actually aren't committed to anything yet. We hope to get enough miles on it to bring PetSet to beta, but it's a huge feature, so we want to be deliberate.
username_2: It's going to say alpha at HEAD, till it's beta at HEAD, regardless of release
Status: Issue closed
username_3: It is still going to be alpha in 1.4.
username_4: @username_3 do we have any idea when it'll be out of alpha? the dynamic pv claims per pod instance solves so many problems for us but we're on GCE who don't support alpha features in anything other than 30day test environments
username_2: GCE supports it, GKE doesn't. I think the target is beta in 1.5 (track https://github.com/kubernetes/kubernetes/issues/28718, not all the issues mentioned there are blockers).
username_4: Apologies I meant GKE. Busy day and frantically typing ! I will follow the issue.
Many thanks |
tensorflow/tfjs | 712721879 | Title: Save tensors from react-native cameraWithTensors api to jpeg
Question:
username_0: <em>Please make sure that this is a feature request. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:feature_template</em>
**System information**
- TensorFlow.js version (you are using):
"@tensorflow/tfjs": "2.0.0",
"@tensorflow/tfjs-backend-webgl": "2.0.0",
"@tensorflow/tfjs-core": "2.0.0",
"@tensorflow/tfjs-react-native": "0.3.0",
- Are you willing to contribute it (Yes/No): Yes
**Describe the feature and the current behavior/state.**
Currently when using the `cameraWithTensor` HOC one can handle directly the underlying tensor to make e.g. real-time prediction. It would be nice to be able to store the stream of tensors as jpeg or video for instance for later use or visualization.
I have created a repo to showcase my currently unsuccessful attempts: https://github.com/username_0/tfjs-expo-example
**Will this change the current api? How?**
**Who will benefit with this feature?**
**Any Other info.**
Answers:
username_1: @username_0 Thanks for your request. Quick question about the code in the repo you posted. What is the goal of this line? https://github.com/username_0/tfjs-expo-example/blob/13b552720d11d7a0be02eef2a26d7253cefe968a/utils.js#L9 What shape tensor to you get from that operation?
I guess it would be nice to know if the issue is with the tensor data or with the jpeg encoding/file writing. Are you able to write a synthetic image to disk and have it appear as expected (e.g. a gradient or something similar that can be created programatically)?
username_0: Ok so there were two things:
- actually the `encodeJpeg(tensor)` was wrong because
- I mixed channels and pixels order
- I was adding a 4th channel in case tensor only had 3 but not stacking a channel full of 255 but copying the red channel of the input tensor (the link you mentionned)
- I was using `Dimensions.height` and `Dimensions.width` for the `cameraTexture(Height|Width)` and it cropped the image. Indeed I had to use your magic number from https://github.com/tensorflow/tfjs/blob/master/tfjs-react-native/integration_rn59/components/webcam/realtime_demo.tsx to have this work.
Things that are still unclear to me:
- where does these numbers (1920 x 1080 for iOs) come from?
- why the `GLView.takeSnapshotAsync(gl, {rect: {x: 0, y: 0, width: 1080, height: 1920}})` still crops the image a bit?
And eventually but not related to this encodeJpeg feature: wether to use the `autorender` props or call the `updatePreview` and `gl.endFrameEXP`.
username_2: Hi @username_0 Have you managed to get it done?
I'm facing the same problem, trying to save a jpeg from the stream.
username_0: yes! actually if you go to https://github.com/username_0/tfjs-expo-example/blob/master/utils.js the function does the job. I will try to push it as a contrib to tfjs-react-native very soon. However, as I have said above, I don't really understand the magic numbers used for width and height in the `cameraWithTensor` module.
username_0: But is it just iOs VS Android or does it depend on the device as well? Because I have noticed that if they are too small for instance, then the image is cropped. Maybe it is the reason why you get black padding, because they are too big for your device
username_2: I'm assuming it depends on the device. you can actually get the device available picture sizes. take a look at [expo-camera docs](https://docs.expo.io/versions/latest/sdk/camera/#getavailablepicturesizesasync).
when I'm changing the dimensions according to the results, I'm getting better capture proportions. |
BlueWallet/BlueWallet | 638223126 | Title: Teach GlaDOS to post e2e failed tests report in PR with artefacts download link
Question:
username_0: files to look at:
`./tests/e2e/upload-artifacts.sh`
`./appcenter-post-build.sh` (posts to github pr)
`tests/e2e/bluewallet.spec.js` (in `specDone` - get access to result of run)
careful, e2e tests can run retry several times in a run
PS. GlaDOS sarcastic comment is a must |
MicrosoftDocs/mixed-reality | 338289146 | Title: .NET Scripting Backend deprecated
Question:
username_0: Unity is apparently deprecating the .NET Scripting Backend, as per this page: https://docs.unity3d.com/2018.2/Documentation/Manual/windowsstore-dotnet.html
This means that the section on the Scripting Backend will soon be out-of-date, especially if the newer releases of Unity are recommended. As it stands, the .NET Scripting Backend also doesn't properly support .NET Standard 2.0, which is giving me a lot of issues.
What is the best course of action on this? I tried searching what Microsoft recommends, but aside a singular reference in the porting guide (which says to switch to IL2CPP), I couldn't find any information related to the deprecation. Is Microsoft working on their own solution to the deprecation, or do I have to simply suck it up and switch to IL2CPP?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 8969b329-9061-f1c0-70c7-b99aa0aa1b94
* Version Independent ID: a80974ef-fbc2-7559-7711-b899e349c010
* Content: [MR Basics 100 - Getting started with Unity - Mixed Reality](https://docs.microsoft.com/en-us/windows/mixed-reality/holograms-100)
* Content Source: [mixed-reality-docs/holograms-100.md](https://github.com/MicrosoftDocs/mixed-reality/blob/master/mixed-reality-docs/holograms-100.md)
* Product: **windows**
* GitHub Login: @username_2
* Microsoft Alias: **kurtie**
Answers:
username_1: That sounds funny, Microsoft has always deprecated (or better removed) stuff not caring about 3rd part projects maintainability and now they are getting kicked out from the Unity environment, without which the Hololens is a pretty useless device..
thats how karma works i think
username_2: We'll need to revisit some of our documentation as we move into 2018.2. The deprecation of the .NET scripting backend is planned alongside some of its benefits finally making their way to IL2CPP, such as [managed code debugging](https://blogs.unity3d.com/2018/07/11/scripting-runtime-improvements-in-unity-2018-2/). If you run into any issues with IL2CPP and our course content, please raise an issue on it! I've run a few courses without issue.
username_3: Hi,
I'm following this tutorial to learn Basic Hololens development with Unity. In my console, I'm seeing this warning (Virtual Reality SDK Windows Mixed Reality failed to initialize. Will attempt to enable None instead) and my cube (from Chapter 4) is not appearing in the device. Can anyone help?
username_4: @username_2, Raising the issue now -> I cant deploy to my hololens, i have cited what is going on on the thread above with the label : unable to deploy, there seems to be lots of pdb files that are not being resolved once you build in unity and open in visual studio, can you please give us a step by step deployment that does not give errors and also does not reference pdb files that cannot be found? for a $3,000 device, i would expect that there would be more activities on the forum, in trying to resolve this from Microsoft??
Status: Issue closed
|
styled-components/styled-components | 1119856744 | Title: Element visually loses font when I interact with unrelated element (Example 2)
Question:
username_0: <!--
🚨HOLD UP PLEASE, BEFORE DELETING THIS TEMPLATE!🚨😁
Thank you for contributing and please follow this guide before creating an issue! 🙌
1. Do you have a usage question?
=============================
- Look for prior or closed issues (but please avoid replying to them if they're old)
- Check the docs: https://www.styled-components.com/docs
- Look for/ask questions on stack overflow: https://stackoverflow.com/questions/ask?tags=styled-components
- Start a thread on our Spectrum help channel: https://spectrum.chat/styled-components/help
- Is this about Typescript? TS types are provided by DefinitelyTyped, all issues opened about TS will be closed and referred there.
2. Think you found a bug?
=============================
- Consider submitting a PR with a failing test instead
- Use the "BUG TEMPLATE" below to report a test
- Don't forget to 👉PROVIDE A REPRODUCTION PLEASE 😉 (Codesandbox links are provided and most useful, repositories are ok)
- If you can't provide a reproduction, snippets of code can help, but are incomplete reports
3. Do you have a feature request?
=============================
- Look for old & closed issues (replying might be ok if they're not too old or have no conclusion)
- Otherwise: Remove the template below and provide thoughtful commentary *and code samples* on what this feature means for your product
What will it allow you to do that you can't do today?
How will it make current work-arounds straightforward?
What potential bugs and edge cases does it help to avoid?
Please keep it product-centric.
-->
<!-- BUG TEMPLATE -->
## System:
- OS: Windows 10 10.0.19043
- CPU: (12) x64 AMD Ryzen 5 5600H with Radeon Graphics
- Memory: 9.93 GB / 15.40 GB
## Binaries:
- Node: 16.13.2 - C:\Program Files\nodejs\node.EXE
- Yarn: 1.22.15 - ~\AppData\Roaming\npm\yarn.CMD
- npm: 8.3.1 - C:\Program Files\nodejs\npm.CMD
## Reproduction
[Demo](https://codesandbox.io/s/strange-orla-xm7lz?file=/src/Test.js).
1. You can see initially "Header" text has "Fraunces" font
2. Inside `Test `component, uncomment `font-weight`:
```
let SectionLink = styled.a`
font-weight: 700;
`;
```
3. Refresh. Now you will see the "Header" text from parent component visually lost "Fraunces" font. Why?
React: 17
Styled-components: 5.3.3
## Expected Behavior
I suppose the "Fraunces" should not disappear visually from "Header" text, when I interact with other element.
Status: Issue closed
Answers:
username_1: It's unfortunately not possible to add `@import` statements for external CSS files to random StyledComponent definitions. It also wouldn't scale that well
You can try to add it using a `createGlobalStyle` component, but you're better off just using a regular `link` tag for external CSS files.
The reason for this is that `@import` statements need to be arlt the top of style sheets. As other rules get added to your app, it may end up not being at the top anymore, which breaks your styles
username_0: @username_1 is there some link to docs on best practice for importing fonts?
Reason I don't want link tag in this case, is following. I have demo project, which contains different components (say component A and B). Now, for component A, I want to have different fonts. For, component B different fonts. What is the way to go in such case?
My App.js always renders either A, or B component, never both. |
nightwatchjs/nightwatch | 92013367 | Title: The globals in test-settings doesn't work.
Question:
username_0: I create a file ci.js:
module.exports = {
adTitle: 'hello'
};
then I add to the globals of test-settings in nightwatch.js like this:
"test_settings" : {
"default" : {
"globals" : {
"myGlobal" : "./data/ci.js"
}
}
}
Then I use in the test.js like following, but I can't get the value for data below. i don't know where the problem is. Could you please help to take a lokk at? Thanks so much.
module.exports = {
'Test': function (browser) {
var data = browser.globals;
}
}
Answers:
username_1: I think what you're looking for is `globalsPath` which gets set at the top level. See http://nightwatchjs.org/guide#basic-settings
Status: Issue closed
|
influxdata/flux | 730200558 | Title: Dispatcher panic when pivoting data for monitoring
Question:
username_0: I don't have time te delve deeper into these issues but I would at least like to point out they happen. I don't think a syntax or semantic error in flux should result in a panic or throw on the backend
```
import "math"
import "strings"
import "date"
import "influxdata/influxdb/monitor"
my_aggregate = (tables=<-) =>
tables
|> range(start: v.timeRangeStart, stop: v.timeRangeStop)
|> filter(fn: (r) => r["_measurement"] == "wqqweqwe")
|> filter(fn: (r) => r["_field"] == "asdasd")
|> aggregateWindow(every: 1h, fn: mean, createEmpty: false)
my_map = (tables=<-) =>
tables
|> map(fn: (r) => ({ r with wd: string(v: date.weekDay(t: r._time) + 4), h: string(v: date.hour(t: r._time)) }))
|> map(fn: (r) => ({ r with mapped_time: r.wd + " " + r.h }))
|> map(fn: (r) => ({ r with h: if strings.strlen(v: r.h) == 1 then "0" + r.h else r.h }))
|> map(fn: (r) => ({ r with wd: if strings.strlen(v: r.wd) == 1 then "0" + r.wd else r.wd }))
my_group = (tables=<-) =>
tables
|> group(columns: ["h", "wd"], mode: "by")
|> mean()
|> group()
tab1 = from(bucket: "asd")
|> my_aggregate()
|> my_map()
|> my_group()
as_day = tab1
|> map(fn: (r) => ({ r with _time: time(v: "2020-10-" + r.wd + "T" + r.h + ":43:50.1004002Z") }))
|> sort(columns: ["_time"])
// |> yield()
tab2 = from(bucket: "asd")
|> my_aggregate()
|> my_map()
tab3 = tab2
|> filter(fn: (r) => r["_measurement"] == "qweqwe")
|> filter(fn: (r) => r["_field"] == "asdasd")
|> aggregateWindow(every: 1h, fn: mean, createEmpty: false)
joined = join(tables: {averages: tab1, actual: tab2}, on: ["h", "wd"], method: "inner")
|> map(fn: (r) => ({ r with diff: math.abs(x: r._value_averages - r._value_actual) / (r._value_averages + r._value_actual)}))
|> map(fn: (r) => ({ r with diff_percent: r.diff * 100.0 }))
|> map(fn: (r) => ({ r with _result: r.diff * 100.0 }))
|> sort(columns: ["diff"], desc: true)
// |> yield()
ok2 = (r) => (r.diff < 0.3)
crit2 = (r) => (r.diff > 0.4)
[Truncated]
github.com/influxdata/flux/execute.PanicUnknownType(...)
/go/pkg/mod/github.com/influxdata/[email protected]/execute/table.go:1216
github.com/influxdata/flux/execute.(*ColListTableBuilder).AddCol(0xc002043c20, 0xc001905000, 0x5, 0x0, 0x5d4c000, 0x0, 0x0)
/go/pkg/mod/github.com/influxdata/[email protected]/execute/table.go:696 +0xd5c
github.com/influxdata/flux/stdlib/universe.(*pivotTransformation).Process.func1(0x3832900, 0xc002043860, 0xc0000000025635e0, 0xc000f89a80)
/go/pkg/mod/github.com/influxdata/[email protected]/stdlib/universe/pivot.go:312 +0xa52
github.com/influxdata/flux/execute.(*ColListTable).Do(0xc002043860, 0xc001417080, 0xc00186a200, 0x63)
/go/pkg/mod/github.com/influxdata/[email protected]/execute/table.go:1410 +0x74
github.com/influxdata/flux/stdlib/universe.(*pivotTransformation).Process(0xc003ecf7a0, 0x945cf10071762b45, 0x44f39a117ad5c86, 0x3816d40, 0xc002043860, 0x7f808d1ce101, 0x0)
/go/pkg/mod/github.com/influxdata/[email protected]/stdlib/universe/pivot.go:283 +0xa68
github.com/influxdata/flux/execute.processMessage(0x380f900, 0xc003ca13e0, 0x3818280, 0xc003ecf7a0, 0x37e7200, 0xc0039c2c40, 0xc0040a3020, 0x0, 0x0)
/go/pkg/mod/github.com/influxdata/[email protected]/execute/transport.go:239 +0x302
github.com/influxdata/flux/execute.(*consecutiveTransport).processMessages(0xc001d5b420, 0x380f900, 0xc003ca13e0, 0xa)
/go/pkg/mod/github.com/influxdata/[email protected]/execute/transport.go:192 +0xb8
github.com/influxdata/flux/execute.(*poolDispatcher).run(0xc0040a3140, 0x380f900, 0xc003ca13e0)
/go/pkg/mod/github.com/influxdata/[email protected]/execute/dispatcher.go:132 +0x65
github.com/influxdata/flux/execute.(*poolDispatcher).Start.func1(0xc0040a3140, 0x380f900, 0xc003ca13e0)
/go/pkg/mod/github.com/influxdata/[email protected]/execute/dispatcher.go:86 +0x87
created by github.com/influxdata/flux/execute.(*poolDispatcher).Start
/go/pkg/mod/github.com/influxdata/[email protected]/execute/dispatcher.go:63 +0x7e
Answers:
username_1: @username_0 Thanks for the report we will take a look.
username_2: We're going to fix the panic to give a meaningful error message after confirming this is the case, but it appears that this error might be because you called `pivot()` with the `valueColumn` as `_value` and that column doesn't exist.
Status: Issue closed
|
electron/electron | 158824286 | Title: Electron <webview> elements do not retain input focus between BrowserWindow blur/focus
Question:
username_0: * Electron version: 1.1.1
* Operating system: 10.11.3
This is a bug @felixrieseberg and I noticed while attempting to [fix an issue](https://github.com/TryGhost/Ghost-Desktop/issues/116) on Ghost Desktop.
When a `BrowserWindow` instance loses and regains focus (a `blur`, followed by a `focus`), it automatically refocuses, and gives input focus to any input-like elements within the application. This is expected.
When a `BrowserWIndow` instance contains a `<webview>`, which itself contains an input-like element, a `blur`, followed by a `focus`, does not retain input focus on the input elements. This is unexpected.
[Here's a video in gif form of what I'm trying to demonstrate is an issue.](https://gist.githubusercontent.com/username_0/818785874346f7b6775df4fcc1d4cd5c/raw/9264bafb2258b7300606d652e5f0bcda25d9114d/electron-focus.gif)
[And here's a code sample of this issue.](https://gist.github.com/username_0/818785874346f7b6775df4fcc1d4cd5c)
Answers:
username_1: Yeah, we fix this up in the Slack app by-hand, it's a bummer
username_0: @username_1 can you share how Slack gets around this? I've tried a few things, but none appear to work definitively.
username_1: Basically it's just:
```js
window.addEventListener('focus', () => theWebViewElement.focus())
```
username_0: @username_1 thank you! I'd somehow ended up taking the long route around getting the webview to focus properly.
username_2: Interestingly we can replicate the same issue with a simple `<textarea>` element under Windows 10 x64 but it happens just 30% of the time. I hope this will also resolve our issue.
username_3: Any progress on this? Following soultion does not work with textarea.
``` javascript
window.addEventListener('focus', () => theWebViewElement.focus())
```
You need to click twice to get the focus on textarea.
username_4: Thanks for contributing!
This is an old version that we are no longer releasing bug fixes for. We prefer to only keep issues that are on our immediate radar.
If you are still having the same issue in an up-to-date version, feel free to reach out and we'll reopen.
Status: Issue closed
username_5: @username_4 This is still an issue in Electron 2.0. Can this be reopened, or is there another issue tracking this?
username_4: - Electron version: 1.1.1
- Operating system: 10.11.3
**Background**:
This is a bug @felixrieseberg and I noticed while attempting to [fix an issue](https://github.com/TryGhost/Ghost-Desktop/issues/116) on Ghost Desktop.
**Working case:**
When a `BrowserWindow` instance loses and regains focus (a `blur`, followed by a `focus`), it automatically refocuses, and gives input focus to any input-like elements within the application. This is expected.
**Not working case:**
When a `BrowserWindow` instance contains a `<webview>`, which itself contains an input-like element, a `blur`, followed by a `focus`, does not retain input focus on the input elements. This is not expected.
**Video/Code**:
- [Here's a video in gif form of what I'm trying to demonstrate is an issue.](https://gist.githubusercontent.com/username_0/818785874346f7b6775df4fcc1d4cd5c/raw/9264bafb2258b7300606d652e5f0bcda25d9114d/electron-focus.gif)
- [And here's the code used in the video above.](https://gist.github.com/username_0/818785874346f7b6775df4fcc1d4cd5c)
username_6: I have similar issue but I am using BrowserView instead of web view and I think BrowserView does not have focus function so How I can set focus on BrowserView when main BrowserWindow get the focus?
username_6: Found the focus function which I can use with BrowserWindow.
webView.webContents.focus();
username_5: still an issue in Electron 4.0.4. Also broken when using BrowserViews instead of webviews.
username_5: @username_4 Hi, this is still a bug in Electron 8.0.1. Thanks
Status: Issue closed
username_7: The Electron version reported on this issue is no longer supported. See our [supported versions documentation](https://www.electronjs.org/docs/tutorial/support#supported-versions).
If this is still reproducible on a supported version, please open a new issue with any other new information that a maintainer should know.
Thank you for taking the time to report this issue and helping to make Electron better! Your help is appreciated.
username_5: Hm, I actually can't repro with webview anymore in Electron 12.0.1:

I can however with a BrowserView. Will create a separate issue for that and update this comment with it. |
EnvironmentOntology/envo | 745932853 | Title: NASA is using ENVO and other OBO foundry terms for definitions of terms
Question:
username_0: Primarily in documents about their United Metadata Model.
Answers:
username_1: Great!
Is the action to add this to the website?
I see this on E-54
"MeasurementContextMedium": "atmosphere", "MeasurementContextMediumURI": "http://www.ontobee.org/ontology/ENVO?iri=http://purl.obolibrary.org/obo/ENVO_01000267", "MeasurementObject": "cloud", "MeasurementObjectURI": "http://www.ontobee.org/ontology/ENVO?iri=http://purl.obolibrary.org/obo/ENVO_01000760", "MeasurementQuantities/Value": "albedo ", "MeasurementQuantitiesMeasurementQuantityURI": "http://www.ontobee.org/ontology/ENVO?iri=http://purl.obolibrary.org/obo/PO_0009087".
They should just use the PURLs, e.g. http://purl.obolibrary.org/obo/ENVO_01000760
username_2: I see this pattern all the time.
ChEBI references are particularly bad (and the ChEBI HTML pages don't help, as you can't find the PURL anywhere.) Or am I mistaken in thinking that http://purl.obolibrary.org/obo/CHEBI_29351 is the proper PURL for Nitrogen?
username_1: This is the correct PURL, many resources are not used to thinking in semantic web terms
username_0: OK, so I submitted a comment on their system....
username_3: This is concerning for ethical and other reasons.
1) this appears to be misconstrued. Based on the document, it looks like it is an example, not a use of envo.
2) in any case, the more important point is that is should not use envo for a few reasons, some of which are ethical.
- Envo currently uses other ontologies, and those should not be forced or imposed on others projects.
- in doing so it would disadvantage other similar ontologies in the global community, as well as other users.
etc.
If there are any vested interest in pushing thsoe or other ontologies, that would not be proper.
We should be endeavoring for neutrality and fair practices.
username_4: Please feel free to recommend other appropriate ontology projects to the relevant agencies/bodies, I think in the end of the day they probably just want something that works. My understanding is that OBO is trying to do it's best to provide comprehensive representation of many domains as possibly. I don't think that's the same as trying to monopolize everything. AFAIK (which is why I spend my time contributing to them) OBO ontologies are completely open source and open access, so anyone should be able to contribute as desired.
username_3: @username_4 I believe that is incorrect. Any vested interests in obo should direct you or others to push that or their products.
username_4: @username_0 was just pointing out that this particular working document/working group at NASA was already using ENVO. We (the envo team) didn't even know about this let alone push or ask them to use ENVO/PATO.
Like I said @username_3 I'm certainly not trying to ignore distinctions/ontological commitments, I'm all ears to suggestions about other vocabs which don't have the issues you raise. If there are some other vocabularies you think that NASA group should be considering reach out to them and suggest it to them. |
dotnet/docs | 934350578 | Title: PowerShell script to determine version
Question:
username_0: Please provide a small PowerShell script to determine the .Net version on a Windows OS quickly. This is especially helpful when we need to check the version for Exchange Server installations and upgrades.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 2115da38-e95e-46fc-1e2d-da88ef92c6cf
* Version Independent ID: a3790e8a-0654-851b-c991-ab7382025406
* Content: [Determine which .NET Framework versions are installed](https://docs.microsoft.com/en-us/dotnet/framework/migration-guide/how-to-determine-which-versions-are-installed)
* Content Source: [docs/framework/migration-guide/how-to-determine-which-versions-are-installed.md](https://github.com/dotnet/docs/blob/main/docs/framework/migration-guide/how-to-determine-which-versions-are-installed.md)
* Product: **dotnet-framework**
* Technology: **dotnet-appcompat**
* GitHub Login: @gewarren
* Microsoft Alias: **gewarren**
Answers:
username_1: @username_0 Hi. There is a small powershell oneliner on the article at the [Use PowerShell to check for a minimum version](https://docs.microsoft.com/en-us/dotnet/framework/migration-guide/how-to-determine-which-versions-are-installed#use-powershell-to-check-for-a-minimum-version) heading. Hopefully this helps. I'll close this issue.
Status: Issue closed
|
ChromatixAU/phpcs-config-chromatix | 283048434 | Title: Take some cues from Neutron
Question:
username_0: https://github.com/Automattic/phpcs-neutron-standard/
fyi: https://github.com/Automattic/phpcs-neutron-standard/issues/9
- [ ] Add a directory above the Sniffs directory so we can have a standard name which is not the name of the package (eg. `<rule ref="Chromatix"/>`)
- [ ] Set the type field of `composer.json` to `phpcodesniffer-standard` so that it can be used without having to run the `phpcs --config-set installed_paths` stuff.
- [ ] Consider whether we should adopt any of the other rules and sniffs Neutron are using
- [ ] Look at adding some tests, see how Neutron are doing it |
rigetti/pyquil | 508809940 | Title: macOS Catalina ForestSDK - Unable to install
Question:
username_0: Issue Description
-----------------
I am attempting to install the ForestSDK on MacOS Catalina 10.15, but I am getting an error saying it is incompatible.
How to Reproduce
----------------
1. Open Forest SDK 2.12.1.pkg on MacOS.
2. Follow the installation steps.
3. Get an error.
### Error Output
This package is incompatible with this version of macOS.
The package is trying to install content to the system volume. Contact the software manufacturer for assistance.
Environment Context
-------------------
Operating System: macOS Catalina 10.15
Python Version (`python -V`): Python 3.6.5:: Anaconda, Inc.
Answers:
username_1: @username_0 thanks for opening an issue! we're working on a fix
username_2: Facing the same issue on macOS Catalina
username_3: I am facing this exact issue as well with macOS Catalina 10.15.
username_4: Facing the same issue , please advise
username_5: While we work on a fix for this (ETA undetermined), I would recommend using the docker-ized version of the Forest SDK. Get docker here https://www.docker.com/products/docker-desktop
With that installed, run the following commands in your terminal:
```
$ docker run -p 5555:5555 -d rigetti/quilc -S
$ docker run -p 5000:5000 -d rigetti/qvm -S
```
Within pyquil, you should then be able to run `get_qc("8q-qvm")` etc. as normal.
username_5: The latest package (2.15.1) should work on Catalina: https://qcs.rigetti.com/sdk-downloads
Unfortunately I have not fixed the security issue that macOS will complain about.
To circumvent that, you will need to `Right click > Open` on the .pkg file.
Let me know if you need more help. |
kiali/kiali | 858863815 | Title: clean up remnant of old ServiceMeshPolicies
Question:
username_0: We completely removed everything to do with the old ServiceMeshPolicies and ServiceMeshRBACConfigs, but there is this one remaining remnant that should probably be removed:
https://github.com/kiali/kiali/blob/master/models/istio_validation.go#L95<issue_closed>
Status: Issue closed |
thelastpickle/cassandra-reaper | 195629531 | Title: Support for resume
Question:
username_0: Repairs don't intelligently resume. They should.
Answers:
username_1: Currently, resume is done at the segment level : Each failed/interrupter segment will get fully retried.
Can you be more specific about the behavior you would like to have on resume ?
Status: Issue closed
|
KappaAnalysis/Kappa | 122536081 | Title: Fill jet correction factor
Question:
username_0: Jets in miniAOD are already corrected. For offline corrections (as possible in Artus using the CondFormats offline corrections), this member should be filled with the proper call of `getCorrectionFactor(...)`.
Answers:
username_1: @cheidecker @thoberger
braucht ihr das noch oder ist das mittlerweile erledigt?
Status: Issue closed
|
alexdebril/feed-io | 260672375 | Title: valid RSS feed doesnt get parsed
Question:
username_0: https://validator.w3.org/feed/check.cgi?url=https%3A%2F%2Fwww.bafin.de%2FDE%2FService%2FTopNavigation%2FRSS%2F_function%2FRSS_Aufsicht.xml
Answers:
username_0: its the user-agent
username_1: Please provide the following information:
I love that one. You seem to know that the user-agent is involved, how do you know that ?
Thank you for your feedback,
username_1: I made a test with a tweaked user-agent, it works. So you're right, it is actually the user-agent that mislead the server and provoke an error.
I'll release a hotfix in the next few hours, thank you very much
Status: Issue closed
|
citusdata/citus_docs | 311360827 | Title: Using datadog with Cloud
Question:
username_0: ## Why are we implementing it? (sales eng)
### What are the typical use cases?
## Communication goals (e.g. detailed howto vs orientation)
### Good locations for content in docs structure
## How does this work? (devs)
### Example sql
### Corner cases, gotchas
### Are there relevant blog posts or outside documentation about the concept/feature?
### Link to relevant commits and regression tests if applicable<issue_closed>
Status: Issue closed |
laminas/laminas-form | 544284131 | Title: Nested forms each with input filter, sub form input filter not added
Question:
username_0: If you nest forms and the main form uses an input filter then the sub form is completely ignored, no matter if the sub form uses an input filter or not. You have to explicitly add the input filter of the sub form to the main form's input filter.
```
// main form with one element
$mainForm = new \Zend\Form\Form('mainForm');
$mainElem = new \Zend\Form\Element\Text('mainElem');
$mainForm->add($mainElem);
// sub form with one element
$subForm = new \Zend\Form\Form('subForm');
$subElem = new \Zend\Form\Element\Text('subElem');
$subForm->add($subElem);
$mainForm->add($subForm);
$data = ['mainElem' => 'foo', 'subForm' => ['subElem' => 'foobar']];
$mainForm->setData($data);
if ($mainForm->isValid()) {
echo '<pre>', var_dump($mainForm->getData()), '</pre>';
}
```
Output is as expected
```
array(2) {
["mainElem"]=>
NULL
["subForm"]=>
array(1) {
["subElem"]=>
NULL
}
}
```
Now if you add input filter to both forms:
```
$mainForm->getInputFilter()->add([ 'required' => false ], $mainElem->getName());
// ...
$subForm->getInputFilter()->add([ 'required' => false ], $subElem->getName());
```
The output is not as expected:
```
array(1) {
["mainElem"]=>
NULL
}
```
While this makes sense, because the main form is checked and its input filter does not know anything about the sub form or its input filter. On the other hand this is highly annoying because a working form (sub form) - either with input filter or without - is working standalone but not working while used as a sub form.
Fieldsets are no alternative to build nested forms, because they have no input filter assigned.
Workaround / "Solution":
```
$mainForm->getInputFilter()->add($subForm->getInputFilter(), $subForm->getName());
```
I would suggest a better / automatic solution within the `add()` method of `\Zend\Form\Form` but I'm not sure if this is a good solution. But maybe someone has a better idea?
(This is not a question, more a feature request) (I'm talking about ZF 2.4.10)
-Cheers
---
Originally posted by @antiphp at https://github.com/zendframework/zend-form/issues/111
Answers:
username_0: nvm my previous comment. This was just valid for defaults.
You could call
```
$mainForm->getInputFilter()->merge($subForm->getInputFilter());
```
that should do the trick.
If your Forms both implemented `InputFilterProviderInterface` you wouldn't need to do this manually, as merging is done during Form's `getInputFilter()` method, when calling `attachInputFilterDefaults()`
---
Originally posted by @demichl68 at https://github.com/zendframework/zend-form/issues/111#issuecomment-236346406
username_0: In my case the forms should be nested (with its own name), so merge wouldn't create a correct input filter for the nested forms AFAIK.
If there is a method adding a form to another it should work as expected. Ignoring or discarding input filters is not "work as expected" IMHO. I don't see the point of creating a sub-form class for the only purpose of implementing InputFilterProviderInterface? Zend\Form\Form has already a possibility to set/get input filters? Or am I misunderstanding you?
---
Originally posted by @antiphp at https://github.com/zendframework/zend-form/issues/111#issuecomment-236547682 |
OpenEmu/OpenEmu | 413791536 | Title: Cheat Codes on Pokemon White 2
Question:
username_0: I installed a cheat to get a pokemon named bronzor. The cheat worked, then I deselected the cheat to stop the cheat from occuring. But when I went to find more pokemon the Bronzor cheat was stil in affect even though the cheat was deselected. I tried reseting my game four times and tried saving and getting out, but that didn't work. So my question is how I get the cheat to stop working?
Here is the cheat: 94000130 fffb0000
b2000024 00000000
c0000000 0000002f
10036b04 000001b4
dc000000 00000004
d2000000 00000000
And I am doing this on a macOS Mojave Version 10.14.2
Answers:
username_1: DeSmuME has bugs with Action Replay cheat codes as well as issues handling the Pokemon games in the first place, which they note are unsupported. You can do nothing about this. Perhaps some of the cheat bugs will be resolved if their next version is ever released.
Also:
[Contributing guidelines](https://github.com/OpenEmu/OpenEmu/blob/master/CONTRIBUTING.md) page is pretty clear on issues with NDS games.
Status: Issue closed
username_0: What is DeSmuMe?
username_1: ...the reason you are able to play an NDS pokemon game in the first place: the emulator core plugin. OpenEmu is not the emulator. |
definelicht/hlslib | 1080505781 | Title: `hlslib::Stream` on the interface
Question:
username_0: I'm connecting my kernel to an AXI4 stream of the platform.
This works with the plain `hls::stream` class, but I would prefer to use the superior `hlslib::Stream`.
Since it is on the interface, I would like to specify something like
```
#pragma HLS INTERFACE axis port=to_fft.stream_
```
However, `stream_` is a `protected` member so I cannot do that.
I probably don't need to specify the pragma at all, but I like to be verbose with those.
Making the stream `public` might not be the cleanest thing to do, but it would help in this case.
Thoughts?
Answers:
username_0: It's hard to test, because the code works without any `pragma` in the first place.
I amended the pragmas to directly apply, and no warning/error was produced so I assume it's fine.
Guess this was a non-issue =D
Status: Issue closed
|
iterative/dvc | 332657484 | Title: Using path relative to DVC project root
Question:
username_0: Hi guys. I was wondering whether we could make use of DVC's knowledge of the project root path.
I'm regularly referencing my /scripts folder from DVC files inside my /data folder using relative paths such as ../../../scripts/something.py, which can get quite annoying.
Would it be possible to reference paths relative to the DVC project root?
- In dependencies and outputs, paths starting with / would be handled as relative to project root (since we cannot reference paths outside the DVC project anyway)
- In commands, an environment variable could be set to be able to use $DVCROOT/scripts/something.py.
Answers:
username_1: Unfortunately I can't see a good way for dvc to setup those env vars automatically without getting into your bashrc. Maybe I'm missing something.
In the upcoming 0.9.8 you could setup a remote with the project's root directory and reference that in dependencies/outputs of your stages. I.e.:
```
$ pwd
$ /path/to/myrepo # DVC project's root directory
$ dvc remote add dvcroot /path/to/myrepo
$ dvc run -d remote://dvcroot/something.py ...
```
But it still looks a bit too long. Though, we could consider treating `dvcroot` as a scheme(i.e. dvcroot://scripts/something), which looks a bit better IMHO. That being said, this method is rather applicable to -d/-o notation and probably not to the cmd itself, as I'm not sure we should touch it. Do you specify `../../../scripts/something.py` in your cmd as well or just in -d/-o?
username_0: Yup I'm using the path to the script also in the command.
Not sure how DVC runs the command, but you can pass environment variables to commands from python like so:
```python
import subprocess, os
my_env = os.environ.copy()
my_env["DVCPATH"] = "/some/path"
subprocess.Popen("echo $DVCPATH/file.txt", env=my_env, shell=True)
# /some/path/file.txt
```
The remotes look cool, I'll look into the docs when they are ready.
username_1: Unfortunately that would not work in a general case, because you also need to be able to use that env in -d/-o notation for `dvc run` and we would not be able intercept that, because the evaluation is performed by sh and the only way to combat that would be to wrap `$DVCPATH`(or DVCROOT) in single quotes so that dvc could parse it itself which would cause a lot of confusion. E.g. here are two commands:
```
dvc run -d $DVCROOT/scripts/something.py ... python $DVCROOT/scripts/something.py
dvc run -d '$DVCROOT'/scripts/something.py ... 'python $DVCROOT/scripts/something.py'
```
First one will evaluate to:
```
dvc run -d /scripts/something.py ... python /scripts/something.py
```
because shell doesn't know DVCROOT env var. So this method would cause a lot annoyances.
On the other hand the proposed `dvc root` would have to return relative path to dvc root and would work like so:
```
$ dvc root
# ../../../
$ dvc run -d $(dvc root)/scripts/something.py ... python $(dvc root)/scripts/something.py
```
Status: Issue closed
|
coryodaniel/arbor | 187804597 | Title: Passing a root element to siblings raises casting error
Question:
username_0: value `nil` in `where` cannot be cast to type :id (if you want to check for nils, use is_nil/1 instead) in query:
from n in ArborBench.Node,
where: n.id != type(^549844, :id),
where: fragment("parent_id = ?", type(^nil, :id)),
select: n
(elixir) lib/enum.ex:1623: Enum."-reduce/3-lists^foldl/2-0-"/3
(elixir) lib/enum.ex:1247: Enum."-map_reduce/3-lists^mapfoldl/2-0-"/3
(elixir) lib/enum.ex:1247: Enum."-map_reduce/3-lists^mapfoldl/2-0-"/3 |
usegalaxy-eu/usegalaxy-eu-tools | 369890676 | Title: tool_conf.xml sample file
Question:
username_0: Hi,
I was wondering if it's possible to get the tool_conf.xml file behind usegalaxy.eu including the labels and sections with ids.
I'm asking because I'm trying to create one using all the sections in tools_iuc/europe_custom yaml.and including all the labels from usegalaxy.eu but I found some sections are duplicated (e.g FASTA/FASTQ manipulation) and with different sections in them. If I try to install tools using for example the tools_iuc yaml file it won't have the expected results in the installation, are these the files used or do you have other yaml files based on section ids?
Thanks in advance!
Ignacio
Answers:
username_1: You can find it here: https://gist.github.com/username_1/1137798bef20144262659ac96c78bc29
There are multiple tools hidden and deprecated that you probably don't want.
username_2: We only base our tool installation off of section name. We need to clean it up but have not found the time yet.
Status: Issue closed
username_2: Live version of tool conf is now available here: https://github.com/usegalaxy-eu/infrastructure-playbook/tree/master/files/galaxy/config
We have done a major re-organisation together with the other usegalaxy.* servers to clean up our toolboxes |
derailed/k9s | 566226120 | Title: Option to enter arbitrary port on PortForward
Question:
username_0: <img src="https://raw.githubusercontent.com/username_1/k9s/master/assets/k9s_small.png" align="right" width="100" height="auto"/>
<br/>
<br/>
<br/>
**Is your feature request related to a problem? Please describe.**
Before 0.15, if was possible to enter arbitrary port when port forwarding. Now, only exposed ports are available.
**Describe the solution you'd like**
It would be awesome if I was able to enter arbitrary ports as before...
**Describe alternatives you've considered**
I can of course expose those ports to the container.
**Additional context**
Answers:
username_1: @username_2 Thank you for this issue! Reworked the dialog in v0.16.0. Hopefully happier?
Status: Issue closed
username_2: @username_1 Much happier! 👯♂ Thank you very much for your work! |
jneen/parsimmon | 185720899 | Title: -2+3 = -5 and 0 - -0 = Parse failure (math.js)
Question:
username_0: `$ echo "-2+3" > test && node math.js test
[ 'Negate', [ 'Add', [ 'Number', 2 ], [ 'Number', 3 ] ] ]`
or in other words: -(2 + 3)
I would have expected (-2) + 3 instead:
`[ 'Add', [ 'Negate', [ 'Number', 2 ] ], [ 'Number', 3 ] ]`
Another one:
`$ echo "0 - -0" > test && node math.js test
Parse failure
=============
{ status: false,
index: { offset: 4, line: 1, column: 5 },
expected: [ '\'(\'', 'number' ] }`
I thought about writing something to generate random math expressions to compare the result to some other parser, but it became too complicated because for example 0--0 is not allowed in JavaScript since '--' is parsed as decrement, so I gave up.
Answers:
username_1: Good catch, thanks.
username_2: It's the problem of precedence, just move the negative to the top of table:
var table = [
{type: PREFIX, ops: operators({Negate: '-'})},
{type: BINARY_RIGHT, ops: operators({Exponentiate: '^'})},
...
when -2^-3 => ^ [ - 2 ] [ - 3 ]
and 4-2^-3 => [- 4 [ ^ 2 [ - 3] ] ]
as expected.
username_1: Arguably that parse of `-2^-3` is not right, math usually defines that as `-(2^-3)`, but I'm gonna go with it anyway because it's better than the current behavior, and personally makes more sense to me than the normal math version.
Status: Issue closed
|
numo-labs/lambda-taggable-s3-event-listener | 153420778 | Title: What Problem are we Solving?
Question:
username_0: At present we have:

However we *cannot* trigger more than one lambda function when S3 create event is fired:
<issue_closed>
Status: Issue closed |
Project-OSRM/osrm-backend | 62340429 | Title: Cannot find route between points
Question:
username_0: OSRM cannot find any route to sime point. For example, 55.726231,37.756256.
Viaroute requets/respose:
http://server:5000/viaroute?loc=55.726231,37.756256&loc=55.726231,37.759209
{"status":207,"status_message":"Cannot find route between points"}
Nearest request/response:
http://server:5000/nearest?loc=55.726231,37.756256
{"status":0,"mapped_coordinate":[55.726231,37.756256],"name":""}
This point (55.726231,37.756256) locates near service route, with is, in turn, locates near primary road.
Debug build of osrm-routed while processing viaroute request craches with message:
"Expression: vector iterator not dereferencable"
Environment:
Real machine (not VM), Windows server 2012 R2
osrm build date: March 10 2015 15:03:26
Answers:
username_1: What osrm version are you using? The develop branch?
There seems to be indeed some sort of edge case for the small-component detection here.
http://map.project-osrm.org/?hl=en&loc=55.726282,37.756324&loc=55.726231,37.759209&z=18¢er=55.726169,37.757156&alt=0&df=0&re=0&ly=763558683
The start is behind a ```lift_gate```, the coordinate lookup should find a node before the gate. Needs investigation.
username_0: username_1, thank you for your answer.
Yes, i am using last developer version of OSRM (at the moment when first message was written).
Your assumption about this OSRM behaviour reason was correct: I comments some code in node_function() in car.lua, and OSRM was able to compute route.
```lua
function node_function (node, result):
--MY COMMENT BEGIN
--[[
-- parse access and barrier tags
local access = find_access_tag(node, access_tags_hierachy)
if access ~= "" then
if access_tag_blacklist[access] then
result.barrier = true
end
else
local barrier = node:get_value_by_key("barrier")
if barrier and "" ~= barrier then
if barrier_whitelist[barrier] then
return
else
result.barrier = true
end
end
end
--]]
--MY COMMENT END
...
end
```
Also, there are some situations with similar behavior.
For example, OSRM can not find route from A to B:
http://localhost:5000/viaroute?loc=55.659624755041,37.551527188987&loc=55.612045640724,37.732799427242
... but can find reverse route, from B to A:
http://localhost:5000/viaroute?loc=55.612045640724,37.732799427242&loc=55.659624755041,37.551527188987
This is, perhaps, because one endpoint snaps to oneway road.
After commenting some oneway checks in car.lua OSRM was able to route:
```lua
function way_function (way, result)
...
--MY COMMENT BEGIN
--[[
-- check if oneway tag is unsupported
local oneway = way:get_value_by_key("oneway")
if oneway and "reversible" == oneway then
return
end
--]]
--MY COMMENT END
...
end
```
So, although there is an issue, you points me acceptable workaround. Thank you again :)
Status: Issue closed
|
GlobWetlandAfrica/GWA_models_and_workflows | 353298436 | Title: Test models, workflows and scripts in which GRASS 6 was replaced by GRASS 7
Question:
username_0: @floinn @kegro @ctottrup @mriffler In two recent commits ([1](https://github.com/GlobWetlandAfrica/GWA_scripts/commit/278671cab5284c3467edcafe37430390729df4ec), [2](https://github.com/GlobWetlandAfrica/GWA_models_and_workflows/commit/fa8afc744d2394c9c09270d1fff9afea277ae27b#diff-c57965e1ec9b418ee4bb6ddab20d67cd)) I updated models, workflows or scripts created by you to replace GRASS 6 algorithms by GRASS 7 algorithms. This should not change the model/workflow/script behaviour but just in case could you please test it and report here both in case everything is fine or if there is a problem.
You might need to install GRASS 7 using the OSGeo4W setup utility if you do not already have it on your machine.
Answers:
username_0: Another two commits that require testing: [1](https://github.com/GlobWetlandAfrica/GWA_models_and_workflows/commit/97e5b4e8db27b43ac304e19bf8cf081e0478b8cb) and [2](https://github.com/GlobWetlandAfrica/GWA_scripts/commit/526c9cd0fdccf1b2ab66105823133a151583f1cb). |
frostbytten/pythia | 546851959 | Title: Dump json from sensitivity
Question:
username_0: In order to check the corrected planting window, would be better to see created json file.
Status: Issue closed
Answers:
username_1: There is no generated JSON to dump. We could either check the resulting files or add more logging. Both of these solutions are file intensive, the later being moreso. |
romainl/vim-cool | 618112523 | Title: Allow blink cursor on current match?
Question:
username_0: `vim-slash` has this feature and it's useful, currently `vim-slash` has an issue with `*`.
Would you consider implement this feature for `vim-cool`?
Answers:
username_1: Could you describe a use case?
username_0: I use cursor shape like `_` and when surfing multiple candidates with `n` or `N` in same screen, I can't distinguish where's the cursor on.
username_0: Even with current implementation of `vim-slash`'s blink, it's still hard to find the candidate that cursor on, it only blink previous cursor block.
username_1: I think this extends beyond the purpose of this plugin, which is to toggle search highlight contextually.
Status: Issue closed
|
reisig/proyecto_aplicaciones | 130662796 | Title: Clase Imagen
Question:
username_0: Implementación de la clase Imagen.
Answers:
username_0: @ariel26 Referente a 222a94c9154420509d0d8d4bf6b11d4df61e6c32 faltan:
1. Las dimensiones de la imagen.
2. La id de la imagen, para que el administrador o el profesor pueda borrarla de la base de datos.
3. La descripción breve no debería ir, ya que el cliente no la especifico.
Hasta el momento eso es lo que veo que falta. Lo otro, crea una función que se traiga una cantidad determinada de imágenes de la base de datos y que las devuelva como un array de objetos Imagen. SI quieres puedes crear un issue de esa tarea, aunque es bien pequeña.
username_0: @ariel26 Se me olvido algo, llena la tabla repositorio con muchas imágenes, unas 500 (si son más mejor) y después la exportas y me la pasas para ir probando con ella en vez del array de imágenes que esta ahora.
Usa las mismas imágenes que tenemos ahora varias veces (en diferente orden para poder distinguir cuando se cambie de pagina en la galería) para no tener que agregar más y hacer el repositorio más pesado. |
singpath/classmentors | 183392721 | Title: Failure to join events
Question:
username_0: See error below

Answers:
username_1: The error message that I get when trying to join events is more like this one.

username_2: I am checking it now. Are both errors occurring with the current master?
username_1: Yes
username_2: @username_0 I cannot reproduce this error. Which browser?
Status: Issue closed
username_2: @username_0 if still have the issue, please reopen this issue |
75lb/command-line-args | 464746140 | Title: Allow typescript definition at @types/command-line-args of commandLineArgs to accept a readonly array
Question:
username_0: Could you please update the typescript definitions so that the `commandLineArgs` function can accept a Readonly array of options? The function definition could then look something like this:
```typescript
declare function commandLineArgs(
optionDefinitions: ReadonlyArray<commandLineArgs.OptionDefinition>,
options?: commandLineArgs.ParseOptions
): commandLineArgs.CommandLineOptions;
```
Answers:
username_1: could you post a PR on [DefinitelyTyped](https://github.com/DefinitelyTyped/DefinitelyTyped/tree/master/types/command-line-args) please? I don't use Typescript, I learned the basics a year or two ago but have forgotten everything already. If you post a PR, I will check and approve it if all is good.
Status: Issue closed
|
dotnet/roslyn | 109535669 | Title: Rationalize checking LanguageService results for consistency/invariants.
Question:
username_0: Inspired by https://github.com/dotnet/roslyn/pull/5618
We should really be pushing the detection of errors as close as possible to when they are first introduced. For example, for outlining, we should ideally be producing an error when the client code creates outlining spans.
I think we can accomplish this across all our language service API in a consistent manner by moving to the 'context' pattern. For example, for outlining, we would change the API to be:
```C#
internal interface IOutliningService : ILanguageService
{
Task AddOutliningSpansAsync(OutliningContext context);
}
struct OutliningContext
{
... properties ...
void AddOutliningSpan(TextSpan textSpan, string bannerText, bool autoCollapse)
{
// Check spans here.
}
}
```
This approach seems nicer to me than having people have to get and pass along a SourceText object to our current OutliningSpan object. The latter would be necessary if we wanted to immediately check things at OutliningSpan creation time. Now, the client doesn't need to do that. They just need to use the context object, and they'll find out immediately if they're doing something wrong.
Answers:
username_1: I completely agree!
username_0: Closing this out as this is hte general pattern we're taking as @mattwar, @username_1 and I work on making APIs public.
Status: Issue closed
|
litzvi/avc-beta | 811690403 | Title: duplicate queries - repositories
Question:
username_0: com.avc.mis.beta.repositories.RelocationRepository.findStorageMovesWithGroup(int)
com.avc.mis.beta.repositories.ProcessRepository.findUsedItemsWithGroup(int)
see how to combine
Status: Issue closed
Answers:
username_0: Storage moves has some extra fields. so hard to do for now |
DisposaBoy/GoSublime | 68206501 | Title: Add the gocode autobuild option support
Question:
username_0: gocode has an experimental `autobuild` config that can be enabled to ensure that altered packages get built for completion. That would be a possible solution for #526
Any other way to silently run `go install` would also be a good solution.
Answers:
username_1: +1
Status: Issue closed
|
cmderdev/cmder | 45469586 | Title: How to enable 256 color?
Question:
username_0: Hi, very nice project. I love it. The problem I have right now is enabling 256 color support. I have the following startup:
```
-new_console:d:C:\Users\fbm "C:\Program Files (x86)\Git\bin\sh.exe" --login -i
```
Here's what happens when I ran the 256color perl script:

Answers:
username_1: The link given above is dead. How do you enable 256 colors?
username_2: https://conemu.github.io/en/Xterm256Colors.html
username_3: I was looking into enabling 256 color from windows 10 when I SSH to linux hosts. One option I found that seems a lot simpler and cleaner is to add putty as a new task. You can find steps [here](https://www.thecrumb.com/2013/03/04/configuring-conemu-and-putty/) |
outline/rich-markdown-editor | 606315625 | Title: Block menu background
Question:
username_0: I'm using the Prosemirror fork, as it actually appears a lot more stable than 9.x!
I set the background property on the theme to be transparent, to blend in with the page. I think [this line here](https://github.com/outline/rich-markdown-editor/blob/prosemirror/src/components/BlockMenu.tsx#L336) should read:
```javascript
background-color: ${props => props.theme. blockToolbarBackground};
```
Otherwise, you wouldn't be able to set separate background colours for the editor vs the block menu. More than happy to submit a PR if needed.

Answers:
username_1: Thanks, addressed
Status: Issue closed
|
saketkc/pysradb | 1153222030 | Title: [BUG]
Question:
username_0: **Describe the bug**
gsm-to-srp is not working
**To Reproduce**
pysradb gsm-to-srp GSM2177186
**Desktop (please complete the following information):**
NAME="Ubuntu"
VERSION="18.04.5 LTS (Bionic Beaver)"
ID=ubuntu
ID_LIKE=debian
PRETTY_NAME="Ubuntu 18.04.5 LTS"
VERSION_ID="18.04"
pysradb 1.3.0
Answers:
username_1: Can you try this with the main branch? I do not see any error (could have be an intermittent issue at NCBI end):
```bash
$ pysradb gsm-to-srp GSM2177186
experiment_alias study_accession
GSE81903 SRP075720
```
Status: Issue closed
|
spring-projects/spring-boot | 397251758 | Title: Upgrade to Spring Batch 3.0.10.RELEASE
Question:
username_0: <!--
Thanks for raising a Spring Boot issue. Please take the time to review the following
categories as some of them do not apply here.
🙅 "Please DO NOT Raise an Issue" Cases
- Question
STOP!! Please ask questions about how to use something, or to understand why something isn't
working as you expect it to, on Stack Overflow using the spring-boot tag.
- Security Vulnerability
STOP!! Please don't raise security vulnerabilities here. Head over to https://pivotal.io/security to learn how to disclose them responsibly.
- Managed Dependency Upgrade
You DO NOT need to raise an issue for a managed dependency version upgrade as there's a semi-automatic process for checking managed dependencies for new versions before a release. BUT pull requests for upgrades that are more involved than just a version property change are still most welcome.
- With an Immediate Pull Request
An issue will be closed as a duplicate of the immediate pull request, so you don't have to raise an issue if you plan to create a pull request immediately.
🐞 Bug report (do not copy/paste)
Please provide details of the problem, including the version of Spring Boot that you
are using. If possible, please provide a test case or sample application that reproduces
the problem. This makes it much easier for us to diagnose the problem and to verify that
we have fixed it.
🎁 Enhancement (do not copy/paste)
Please start by describing the problem that you are trying to solve. There may already
be a solution, or there may be a way to solve it that you hadn't considered.
--> |
dsccommunity/AzureDevOpsDsc | 792787193 | Title: AzureDevOpsDsc: Common module should be built with ModuleBuilder
Question:
username_0: In the `source/Modules/AzureDevopsDsc.Common` should be built using the pattern public/private folders.
We ought to be able to build the PSM1 by merging the Public/Private/Classes/etc. folders the same way as the top module.
We should configure the `NestedModules` section in the `build.yml`
```
NestedModule:
DscResource.Common:
CopyOnly: true
Path: ./output/RequiredModules/DscResource.Common
AddToManifest: false
Exclude: PSGetModuleInfo.xml
AzDevOpsProject.Common:
CopyOnly: false
Path: ./source/Modules/AzureDevOpsDsc.Common
AddToManifest: false
```
And remove the entry `Modules` from here:
https://github.com/dsccommunity/AzureDevOpsDsc/blob/8d43d12989359b8facdf46d86011a34dfe787898/build.yaml#L5-L9
That will replace the dot-source code in the file below, and will speed up the usage of the distributed resources.
https://github.com/dsccommunity/AzureDevOpsDsc/blob/8d43d12989359b8facdf46d86011a34dfe787898/source/Modules/AzureDevOpsDsc.Common/AzureDevOpsDsc.Common.psm1#L10-L45 |
nasa/PSP | 779549022 | Title: Re-add static code analysis for pull requests
Question:
username_0: **Describe the bug**
The Travis CI runner ran cppcheck which we don't have anymore
**To Reproduce**
Submit pull request, notice that Travis CI doesn't run anymore.
**Expected behavior**
A success or failure report of a cppcheck run.
**Code snips**
Check from travis.yml
```
# Check versions
- cppcheck --version
#cppcheck flight software psp/fsw
- cppcheck --force --inline-suppr --std=c99 --language=c --error-exitcode=1 --enable=warning,performance,portability,style --suppress=variableScope --inconclusive fsw 2>cppcheck_flight_psp.txt
- |
if [[ -s cppcheck_flight_psp.txt ]]; then
echo "You must fix cppcheck errors before submitting a pull request"
echo ""
cat cppcheck_flight_psp.txt
exit -1
fi
```
**System observed on:**
n/a
**Additional context**
Part of ongoing TravisCI to GitHub Actions migration<issue_closed>
Status: Issue closed |
QuanhanSun/Deep-and-Confident-Prediction-for-Time-Series-at-Uber | 770560694 | Title: 在我的机器上运行您的程序时,我把涉及 google.colab的代码注释掉,结果出现如下错误,请问是这么回事?
Question:
username_0: 注释掉的代码
# from google.colab import drive
# drive.mount('/content/drive')
# import os
# path = "/content/drive/My Drive/SaicAILab"
# os.chdir(path)
运行错误:发生在 ’预训练模型图解‘处:
from IPython.display import SVG
from keras.utils.vis_utils import model_to_dot
SVG(model_to_dot(ae_model,show_shapes= True, show_layer_names=True, dpi=65).create(prog='dot', format='svg'))
错误提示:
('Failed to import pydot. You must `pip install pydot` and install graphviz (https://graphviz.gitlab.io/download/), ', 'for `pydotprint` to work.')
---------------------------------------------------------------------------
AttributeError Traceback (most recent call last)
<ipython-input-15-2f244cd4f33f> in <module>
2 from keras.utils.vis_utils import model_to_dot
3
----> 4 SVG(model_to_dot(ae_model,show_shapes= True, show_layer_names=True, dpi=65).create(prog='dot', format='svg'))
AttributeError: 'NoneType' object has no attribute 'create'
根据提示我安装好了pydot和graphviz 两个包后,重新运行,结果还是出同样的错:
控制台输出如下:
(Keras) F:\AnalyticsZoo\UberPaperimplement>pip install pydot
Collecting pydot
Downloading pydot-1.4.1-py2.py3-none-any.whl (19 kB)
Requirement already satisfied: pyparsing>=2.1.4 in c:\users\wxy\anaconda3\envs\keras\lib\site-packages (from pydot) (2.4.7)
Installing collected packages: pydot
Successfully installed pydot-1.4.1
(Keras) F:\AnalyticsZoo\UberPaperimplement>pip install graphviz
Collecting graphviz
Downloading graphviz-0.15-py2.py3-none-any.whl (18 kB)
Installing collected packages: graphviz
Successfully installed graphviz-0.15
(Keras) F:\AnalyticsZoo\UberPaperimplement>jupyter notebook
[I 12:28:27.235 NotebookApp] Serving notebooks from local directory: F:\AnalyticsZoo\UberPaperimplement
[I 12:28:27.236 NotebookApp] Jupyter Notebook 6.1.5 is running at:
[I 12:28:27.238 NotebookApp] http://localhost:8888/?token=<KEY>
[I 12:28:27.242 NotebookApp] or http://127.0.0.1:8888/?token=<KEY>
[I 12:28:27.243 NotebookApp] Use Control-C to stop this server and shut down all kernels (twice to skip confirmation).
[C 12:28:27.293 NotebookApp]
To access the notebook, open this file in a browser:
file:///C:/Users/wxy/AppData/Roaming/jupyter/runtime/nbserver-16972-open.html
Or copy and paste one of these URLs:
http://localhost:8888/?token=<KEY>
or http://127.0.0.1:8888/?token=d754dc8d300c895d6009fe14a08f530ac339d74dc5b2522b
[I 12:28:33.832 NotebookApp] Kernel started: 9de98f90-7991-4cec-86ce-9814f3a1fb6e, name: python3
2020-12-18 12:28:54.044539: I tensorflow/core/platform/cpu_feature_guard.cc:142] This TensorFlow binary is optimized with oneAPI Deep Neural Network Library (oneDNN)to use the following CPU instructions in performance-critical operations: AVX AVX2
To enable them in other operations, rebuild TensorFlow with the appropriate compiler flags.
[I 12:30:33.767 NotebookApp] Saving file at /Deep_and_Confident_Prediction_for_Time_Series_at_Uber.ipynb
[Truncated]
预训练模型图解
In [15]:
from IPython.display import SVG
from keras.utils.vis_utils import model_to_dot
SVG(model_to_dot(ae_model,show_shapes= True, show_layer_names=True, dpi=65).create(prog='dot', format='svg'))
('Failed to import pydot. You must `pip install pydot` and install graphviz (https://graphviz.gitlab.io/download/), ', 'for `pydotprint` to work.')
---------------------------------------------------------------------------
AttributeError Traceback (most recent call last)
<ipython-input-15-2f244cd4f33f> in <module>
2 from keras.utils.vis_utils import model_to_dot
3
----> 4 SVG(model_to_dot(ae_model,show_shapes= True, show_layer_names=True, dpi=65).create(prog='dot', format='svg'))
AttributeError: 'NoneType' object has no attribute 'create'
请教如何解决此问题?多谢!
Answers:
username_1: https://github.com/AppliedDataSciencePartners/DeepReinforcementLearning/issues/3 我查了下可能是这个问题,我本地运行的时候没有问题,也有可能是用的keras和tensorflow.keras的backend混淆了有关系,colab应该是默认tf的backend
------------------ 原始邮件 ------------------
username_2: ### 我也遇到了大概相似的问题(已解决)
***
### 我遇到问题和解决问题的过程如下:
- 在jupyter上运行如下代码之前没有问题
`from google.colab import drive`
`drive.mount('/content/drive')`
- 在导入google.colab之后jupyter无法再使用了
显示:`ValueError: signal only works in main thread`
- 在这篇博客找到了问题所在
《jupyter notebook中出现ValueError: signal only works in main thread 报错 即 长时间in[*] 解决办法》[https://blog.csdn.net/jackhh1/article/details/103435550](url)
- 上文未给出解决方案,参考另一篇博客找到了解决方案
《Jupyter Notebook "signal only works in main thread"》
[https://blog.csdn.net/loovelj/article/details/82184223](url)
- 采用了其中一部分教程
**大概如下:**
`pip uninstall google.colab`
`pip install "pyzmq==17.0.0" "ipykernel==4.8.2"`
`pip install pydot`
***
### 总得来说
`from google.colab import drive`
`drive.mount('/content/drive')`
`import os`
`path = "/content/drive/My Drive/SaicAILab"`
`os.chdir(path)`
**这些代码似乎都没有用,我把它们注释之后整体运行是没有错误的!!!**
@username_0
Status: Issue closed
|
stevekirks/openlayers-wkt-viewer | 1154819061 | Title: How to pass geometry collection?
Question:
username_0: I am working on Leaflet-WKT.js as provided on Github
https://github.com/arthur-e/Wicket/blob/master/wicket-leaflet.js
I am trying to pass geometrycollection WKT MultiPolygon collection.
Could anyone please help me to resolve this issue?
Attachment
https://drive.google.com/file/d/1wke9XbsTYYb7OQEW8Y4blJvIGbW5LgkG/view?usp=drivesdk
Console log
The geometrycollection WKT type is not yet supported. |
sabbelasichon/typo3-rector | 775066126 | Title: Breaking: #93003 - PageRenderer renders only full page
Question:
username_0: Breaking: #93003 - PageRenderer renders only full page
https://docs.typo3.org/c/typo3/cms-core/master/en-us/Changelog/11.0/Breaking-93003-LimitationOfPageRendererToOnlyRenderFullPage.html
.. include:: ../../Includes.txt
======================================================
Breaking: #93003 - PageRenderer renders only full page
======================================================
See :issue:`93003`
Description
===========
TYPO3s main API class to build a full HTML page for Frontend
and Backend rendering - :php:`PageRenderer` - previously allowed
to only render the header or footer separately, which was built
due to historical reasons when rendering content.
This is however obsolete and TYPO3 Core only renders full pages in
Frontend and Backend internally.
For this reason, PageRenderer's :php:`render` method does not accept
any method arguments anymore and always renders the complete HTML page.
In addition, the constants
* :php:`PageRenderer::PART_COMPLETE`
* :php:`PageRenderer::PART_HEADER`
* :php:`PageRenderer::PART_FOOTER`
[Truncated]
API to render the page.
Migration
=========
It is recommended for third-party extensions to use custom hooks to
process or manipulate header or footer parts.
.. index:: Backend, Frontend, FullyScanned, ext:core<issue_closed>
Status: Issue closed |
sdi-sweden/geodataportalen | 1046964176 | Title: (246, 'Finns gratis DTM (höjddata)')
Question:
username_0: **2011-03-23T06:35:21.000+00:00**
****:
Förnamn : Christian
Efternamn : Donaldson
E-postadress : <EMAIL>
Fråga: Vet Du om det finns något höjdmodell (DTM) över hela Sverige som är gratis eller kostar en rimlig slant?
Mvh
Christian
Answers:
username_0: **2011-03-23T06:35:38.000+00:00**
****:
skickat till kundcenter
username_0: **2011-04-07T14:05:14.000+00:00**
****:
Updating tickets (#201, #202, #203, #204, #205, #206, #207, #208, #209, #210, #211, #212, #213, #214, #215, #216, #217, #218, #219, #220, #221, #222, #223, #224, #225, #226, #227, #228, #229, #230, #231, #232, #233, #234, #235, #236, #237, #238, #239, #240, #241, #242, #243, #244, #245, #246, #247, #248, #249, #250, #251, #252, #253, #254, #255, #256, #257, #258, #259, #260, #261, #262, #263, #265, #266, #267, #268, #269, #270, #271, #274, #275, #276, #277, #278, #279, #280, #281, #282, #285, #286, #287, #288, #289, #290, #291, #292, #293, #294, #295, #296, #297, #298, #299, #300, #301, #302, #303, #304, #305) |
prometheus/client_java | 343166767 | Title: @EnablePrometheusEndpoint throws exception with Spring Boot 2.x
Question:
username_0: Using the EnablePrometheusEndpoint annotation in a Spring Boot 2.0.3 application that ends with the following:
`Caused by: java.lang.IllegalStateException: Failed to introspect annotated methods on class io.prometheus.client.spring.boot.PrometheusEndpointConfiguration
Caused by: java.lang.NoClassDefFoundError: org/springframework/boot/actuate/endpoint/AbstractEndpoint
Caused by: java.lang.ClassNotFoundException: org.springframework.boot.actuate.endpoint.AbstractEndpoint`
Looking at Spring Boot source, I see clearly that org.springframework.boot.actuate.endpoint.AbstractEndpoint no longer exists. The closest thing in that package is now AbstractExposedEndpoint.
If you can, please let me know whether there is any plan to support Spring Boot 2.x in the future.
Answers:
username_0: Nevermind. Further down my page of google results I found [https://github.com/prometheus/client_java/issues/345], which suggests there will be no Spring Boot 2.x support from this project because micrometer will handle the endpoint.
Status: Issue closed
username_1: Sorry to ask. Does it mean that Prometheus could not be used with Spring Boot 2.0 and above ?
username_2: Prometheus is orthogonal to the clients. This particular integration only supports 1.0, however client_java in general will work with anything on the JVM. |
jupyterlab/jupyterlab | 389578451 | Title: At autocompletion show the suggestion is method or attribute?
Question:
username_0: Now Jupyter lab after press tab only shows some suggestions. it is much better to show which one needs () and is a method or doesn't need.
It can be by adding () at the end or show initial at the first of dropdown
Answers:
username_1: Thanks, great suggestion!
username_0: @username_1 I found this tool at our office which is works at jupyter lab. I couldn't find how it is activated.
 |
PorktoberRevolution/ReStocked | 407935083 | Title: Jumbo-64 tank default color
Question:
username_0: I'd like to suggest that you make the large Rockomax fuel tank default to the orange variant, as the description still refers to the orange color.
*The largest tank available from Rockomax, the Jumbo-64 holds a vast amount of fuel in a friendly orange insulated container. Contrary to popular belief, the Jumbo-64 is NOT orange flavored and should NOT be tasted.*
Status: Issue closed
Answers:
username_1: Nah. We preserve the defaults as they currently stand. |
espnet/espnet | 718052049 | Title: Bad results when training Transformer TTS from EGS2 LJSpeech recipe
Question:
username_0: I'm experimenting with training LJSpeech Transformer TTS from the EGS2 recipe in the repo.
I'm using a custom dataset based on audios from Youtube (<NAME> voice, as it's the most readily available) for a total dataset duration of 8 hours. The audios are clean of noise and split into 6-12 second segments, much like LJSpeech.
The text has been obtained through Google Speech-to-Text and numbers/abbreviations have been expanded into words. I'm attaching a few samples of the dataset for reference.
I've experimented with both fine tuning from a pre-trained Transformer model (from model zoo, as described [here](https://github.com/espnet/espnet/tree/master/egs2/TEMPLATE/tts1)) and training from scratch.
I'm unable to achieve any decent speech when decoding, with the voice being very robotic and making no sense at all when speaking.
I've looked at the train charts for both the pre-trained LJSpeech Transformer TTS and my own results, and noticed that the bce_loss and decoder_alpha charts look very different, with the "valid" curve increasing in my dataset instead of decreasing like in LJSpeech.
I've trained both the LJSpeech dataset and my dataset from scratch (using the same config) for 12hrs on a 4x Tesla V100 GPU instance, and compared the results; I'm attaching both charts folders and also a direct comparison between those two charts.

As far as I can see, there's little to no difference between my dataset's audios/text and LJSpeech's.
Clearly it's something with my dataset, however I can't put my finger on what :confused:
[trump_sample.zip](https://github.com/espnet/espnet/files/5354404/trump_sample.zip)
[ljspeech_charts.zip](https://github.com/espnet/espnet/files/5354401/ljspeech_charts.zip)
[custom_charts.zip](https://github.com/espnet/espnet/files/5354400/custom_charts.zip)
Answers:
username_1: First, please check FAQ:
https://github.com/espnet/espnet/tree/master/egs2/TEMPLATE/tts1#why-the-model-cannot-be-trained-well-with-my-own-dataset
In the AR E2E-TTS model, the attention is the most important, the loss value is not so meaningful.
Please check the attention plot is diagonal (saved in `exp/*/att_ws/`).
Maybe the problem is the silence which does not correspond to the silence.
This makes the diagonal attention learning difficult.
username_0: Do you mean there should be a period/comma in the transcription whenever there's a silence break?
I've tried to model my dataset on the basis of LJSpeech; I've read the FAQ but I can't really spot any weird silences/long pauses in my data, the longest pause between two sentences in my dataset is 300ms - is this too long?
Spectrogram comparison:

username_1: Yes. It is better to add such a symbol.
For example, during 300 ms / (256 shfit pt / 22050 hz * 1000) = 26 frames, the model cannot decide where should be focused. This makes the attention learning very difficult.
In my case, I always remove silence at the beginning and the end of the audio using force alignment results and use a comma-like symbol for silence in the middle of the audio.
You may be able to train the model with a large reduction factor (say `reduction_factor=5`) and a large guided attention lambda.
username_0: Fantastic, thank you for the detailed reply! I'll fix the dataset as you said and report back with the results.
Status: Issue closed
username_0: Can confirm that the pauses and lack of commas were indeed the problem, I'm seeing better results across the board now!
Thank you for your help! I'll close the issue as it's now solved. |
CocoaPods/CocoaPods | 152863116 | Title: Using a CocoaPod with a dependency and then using that dependency directly causes "file not found" on import (1.0.0.rc.1)
Question:
username_0: * [x] I've read and understood the [*CONTRIBUTING guidelines and have done my best effort to follow](https://github.com/CocoaPods/CocoaPods/blob/master/CONTRIBUTING.md).
# Report
## What did you do?
Pod 1: RMNCommon
-Has subspec 'ImagePicker' which contains a dependency on PEPhotoCropEditor
-Has subspec 'Network' which contains a dependency on Reachability
Pod 2: RMNCitadelClient which contains a dependency on RMNCommon/Network
Project: test project
-Uses pod 'RMNCitadelClient'
-Uses pod 'RMNCommon/ImagePicker'
## What did you expected to happen?
Pods should install and compile.
## What happened instead?


When building test project, RMNCitadelClient is showing error 'PECropViewController.h' file not found only INSIDE the test project.

RMNCitadelClient builds successfully.

RMNCommon project compiles successfully.
Please note that RMNCitadelClient does NOT use ImagePicker subspec or use PEPhotoCropEditor at all!
## Podfile
RMNCommon Podfile
```ruby
source 'redacted specs url'
source 'https://github.com/CocoaPods/Specs.git'
platform :ios, '8.0'
use_frameworks!
xcodeproj 'RMNCommon.xcodeproj'
target 'RMNCommon' do
pod 'SVProgressHUD','0.9', :inhibit_warnings => true
pod 'PEPhotoCropEditor', '1.3', :inhibit_warnings => true
pod 'Reachability', '3.2', :inhibit_warnings => true
pod 'Base32', '~> 1.1'
pod 'UICKeyChainStore', '2.0.6'
# internally-managed specs
pod 'Haversine', '0.0.1'
end
```
RMNCommon Podspec
```ruby
[Truncated]
pod 'RMNCommon/ImagePicker', :path => "../rmn-common-ios"
pod 'RMNCitadelClient', :path => "../citadel/ios-client-sdk"
end
target 'test WatchKit 1 Extension' do
end
target 'test WatchKit 1 App' do
end
target 'testTests' do
end
target 'testUITests' do
end
```
Answers:
username_0: If I remove the ImagePicker subspec pod (comment it out):
```ruby
pod 'RMNCommon/ImagePicker', :path => "../rmn-common-ios"
```
to
```ruby
#pod 'RMNCommon/ImagePicker', :path => "../rmn-common-ios"
```
test project builds successfully.
username_1: I think I have a similar problem. I switched my project to use `use_frameworks!` and I started getting file not found from all the subspecs.
username_2: Can you please try and reduce this to a runnable project where the podspecs don't set custom frameworks search paths? Thanks!
username_0: @username_2 I removed these portions:
RMNCommon
```ruby
s.xcconfig = { 'FRAMEWORK_SEARCH_PATHS' => '"$(PODS_ROOT)/"' }
```
RMNCitadelClient
```ruby
sdk.xcconfig = {
'FRAMEWORK_SEARCH_PATHS' => '"$(PODS_ROOT)/" "$(PLATFORM_DIR)/Developer/Library/Frameworks"'
}
```
The same error occurred.

username_2: Ok, but we still need a project we can use to debug.
username_0: @username_2 Oh I see. I'll try to get it set up for you soon.
username_2: https://github.com/CocoaPods/CocoaPods/issues/5266 might fix this
Status: Issue closed
username_3: Closing, because we never got a reproducible project. |
HumanCellAtlas/dcp | 417511654 | Title: Review HCA repositories and verify that the list of readers/writers is following principle of least access
Question:
username_0: The HumanCellAtlas GitHub teams have been restructured.
To ease the transition to this structured world, everyone is currently in a team called "HCA" that has WRITE access to all repos. We want to move away from this.
Each component team lead needs to audit the GitHub teams that have access to their repos.
Ensure only appropriate teams (e.g. CZI, UCSC, etc) have write access to your component's repos.
Ensure that you approve of any collaborators.
Then remove access from team HCA.
Answers:
username_1: @username_0 - There seems to be three duplicate(?) _dcp_issues:
* https://app.zenhub.com/workspaces/dcp-backlogs-5ac7bcf9465cb172b77760d9/issues/humancellatlas/dcp/295
* https://app.zenhub.com/workspaces/dcp-backlogs-5ac7bcf9465cb172b77760d9/issues/humancellatlas/dcp/305
* https://app.zenhub.com/workspaces/dcp-backlogs-5ac7bcf9465cb172b77760d9/issues/humancellatlas/dcp/306
Not sure whether 306 is intended to track this issue for the _dcp_ repo itself since it's linked to the 295 epic.
username_0: Yup this a dup. Closing.
306 is indeed to track the dcp GitHub repo itself. I'll update the comment to reflect that.
Status: Issue closed
|
clio-lang/clio | 540526762 | Title: Prettify crash messages
Question:
username_0: Some CLI crash with an ugly error message, when an incorrect filepath is provided. These crashes should be catched, handled and displayed as a human-friendly message:
- [ ] `clio run`
- [ ] `clio ast`
- [ ] `clio highlight`
- [ ] `clio compile`
Answers:
username_0: What about some emojis?✨
username_1: Yes, I think we can use some emojis. :)
username_0: I also noticed, that this "cliopkg.toml is missing, or are you running tests" message is performed way too often. I think that has something to do with the way I have implemented it:
```js
(x => do(x))()
```
Could you also take care of that?
username_1: Yup, no problem. I'll try to reduce its frequency.
Status: Issue closed
|
zulip/zulip-terminal | 855495125 | Title: TRACKING: Event handling improvements
Question:
username_0: - [ ] #588 : Alert words handling
- [ ] #988 : Update user display settings
- [ ] Update global notifications (currently stream-based priority? #666)
- [ ] #816 Streams (add, delete, update) (#851)
- [ ] #529 Upload/Attachment events
- [ ] Submessages (e.g., polls) (#287)
- [ ] #988: User status updates.
- [ ] Custom profile fields (extension of user info popup #848)
- [ ] Handle Default stream updates (changed by org admin).
- [ ] Delete message event (first event handling, then feature support; #690)
- [ ] User group events (add/modify/delete)
- [ ] #809 : Realm emoji events (#827)
Answers:
username_0: @zulipbot add "area:event handling"
username_0: @zulipbot add "area: event handling"
username_1: Tagged with GSoC, but this is across multiple areas so as per the tracking nature of this issue, this isn't something one person will likely work on.
username_2: I get that this is not a high priority, but whenever in the future ZT adds support for handling events sent after the linkifiers in the realm have changed, we should handle the new `realm_linkifiers` events, since the `realm_filters` event type has been deprecated in feature level 54. #995 has more context. Can you update the issue description?
username_0: Thanks! Updated the issue description :) |
gnugat/redaktilo | 55152080 | Title: Flat PHP vs Redaktilo
Question:
username_0: Documentation showing the pains of flat PHP vs Redaktilo.
Answers:
username_1: Maybe something like that:
(flat php was mainly inspired by Redaktilo's internal, not sure how this reflects real flat php)
# Opening and saving a file
## Flat PHP
```php
$filename = '/path/to/file.txt';
$content = file_get_contents($filename);
$lines = preg_split('/\R/', $content);
$lineBreak = strpos($content, "\r\n") !== false ? "\r\n" : (strpos($content, "\n") !== false ? "\n" : PHP_EOL);
... // some manipulation on lines
file_put_contents($filename, implode($lineBreak, $lines));
```
## Redaktilo
```php
$editor = EditorFactory::createEditor();
$file = $editor->open('/path/to/file.txt');
... // some manipulation on lines
$editor->save($file);
```
# Navigating through the content
## Flat PHP
```php
// Jumps to the first line matching the pattern, starting from the line 10
$belowLines = array_slice($lines, 10, null, true);
$found = preg_grep('/pattern/', $belowLines);
$lineNumber = key($found);
$line = $lines[$lineNumber];
```
## Redaktilo
```php
// Jumps to the first line matching the pattern, starting from the line 10
$editor->jumpBelow($file, '/pattern/', 10);
$lineNumber = $file->getCurrentLineNumber();
$line = $file->getLine();
```
# Manipulating a Line
## Flat PHP
```php
// Add a line after a given line number
array_splice($lines, $lineNumber, 0, $newLine);
// Replacing a line with a callback
$line = $lines[$lineNumber];
$replacement = $callback($line);
$lines[$lineNumber] = $replacement
```
## Redaktilo
```php
// Add a line after a given line number
$editor->insertBelow($file, $newLine, $lineNumber);
// Replacing a line with a callback
$editor->replace($file, $callback, $lineNumber);
```
username_0: How very nice! I was thinking about re using my blog article, but your solution is way more concise. |
dblock/slack-gamebot | 128531079 | Title: Horizontal scalability?
Question:
username_0: As far as I see web sockets run in a single thread within web dyno. What if i need to increase dyno size? Will app open dup connections?
Answers:
username_1: It would, and it would be very bad.
I've been thinking about how to deal with this. For Heroku I think the easiest would be to declare the number of working dynos in ENV and distribute load that way. The next step is probably to have a more transactional way of "starting" a service and having dynos compete for locks on start, then try to rebalance at runtime automagically. That problem alone could be a great Ruby gem.
username_0: Basically we have https://github.com/schneems/the_lone_dyno which can help us to isolate thread in a single dyno while puma's workers can run in a multiple.
Did you ever count how much websocket connections server can handle in a standard Heroku dyno?
username_1: I think isolating workers to 1 dyno won't do much for slack-gamebot unless someone starts heavily using the API.
I think we should be able to do many hundreds at the very least. Right now I have an instance on http://playplay.io that isn't breaking a sweat with stable RAM at about 300MB (bot classes are singletons) with 100+ teams.
username_1: The constructs we need from each dynos perspective is:
1. Tell me what the optimal number of connections should I be starting based on the number of workers alive.
2. If I am above that number, I will stop some.
3. If I am below that number, give me the next un-started connection to start.
The system should be stable, ie. the optimal number should be a little bigger than total connections / number of workers so that connections don't constantly get shutdown and restarted elsewhere.
username_0: Yeah, I was thinking the same. Anyway we should store connection ids somewhere then. It maybe Redis and we can just store bot token there or something similar.
The only thing I aware of is that we can't see if dyno crashed and did not update that info in Redis.
username_0: Probably it can be any another NoSQL DB providing persistence.
username_1: Slack-gamebot already has a MongoDB store, you can use atomic updates and https://github.com/afeld/mongoid-locker. Would love some PRs!
username_1: I've introduced https://github.com/username_1/slack-gamebot/blob/master/Procfile.heroku which can be used to split the web and the worker process, so now web ones can be scaled horizontally. You still don't want multiple workers. |
spring-cloud/spring-cloud-gateway | 670082260 | Title: RequestHeaderFilter doesn't support shortcut form
Question:
username_0: As the title says, there's no way to shortcut RequestHeaderFilter in YAML.
The long form works:
```
filters:
- name: RequestHeaderSize
args:
maxSize: 8KB
```
But neither of the following work:
```
filters:
- RequestHeaderSize=8KB
```
...or...
```
filters:
- RequestHeaderSize=maxSize, 8KB
```
Also, this filter isn't documented at all in the reference documentation.<issue_closed>
Status: Issue closed |
BrikerMan/BMPlayer | 640077735 | Title: xcode 11.5, swift 5, carthage 0.34.0 不能编译成功
Question:
username_0: ## Check List
Thanks for considering to open an issue. Before you submit your issue, please confirm these boxes are checked.
- [ ] I have searched in [existing issues](https://github.com/username_1/BMPlayer/issues?utf8=%E2%9C%93&q=) but did not find the same one.
## BMPlayer version
## Installed with
- [X] Carthage
- [ ] Cocoapods
## Issue Description
### What
[Tell us about the issue]
log文件中显示 :
```
2020-06-17 09:38:31.628 xcodebuild[42570:2527876] [MT] iPhoneConnect: 📱<DVTiOSDevice (0x7ff8b56080c0), ., iPhone, 13.5.1 (17F80), 162bf0974572929ea8eed8b0f1b97f97e6b6f70b> == END: Underlying device preparation errors ==
error: /Users/mrzhou/app/video/Carthage/Checkouts/BMPlayer/Example/Pods/Target Support Files/BMPlayer/BMPlayer.release.xcconfig: unable to open file (in target "BMPlayer" in project "Pods") (in target 'BMPlayer' from project 'Pods')
error: /Users/mrzhou/app/video/Carthage/Checkouts/BMPlayer/Example/Pods/Target Support Files/BMPlayer/BMPlayer.release.xcconfig: unable to open file (in target "BMPlayer" in project "Pods") (in target 'BMPlayer' from project 'Pods')
error: /Users/mrzhou/app/video/Carthage/Checkouts/BMPlayer/Example/Pods/Target Support Files/BMPlayer/BMPlayer.release.xcconfig: unable to open file (in target "BMPlayer" in project "Pods") (in target 'BMPlayer' from project 'Pods')
error: /Users/mrzhou/app/video/Carthage/Checkouts/BMPlayer/Example/Pods/Target Support Files/VIMediaCache/VIMediaCache.release.xcconfig: unable to open file (in target "VIMediaCache" in project "Pods") (in target 'VIMediaCache' from project 'Pods')
error: /Users/mrzhou/app/video/Carthage/Checkouts/BMPlayer/Example/Pods/Target Support Files/VIMediaCache/VIMediaCache.release.xcconfig: unable to open file (in target "VIMediaCache" in project "Pods") (in target 'VIMediaCache' from project 'Pods')
error: /Users/mrzhou/app/video/Carthage/Checkouts/BMPlayer/Example/Pods/Target Support Files/VIMediaCache/VIMediaCache.release.xcconfig: unable to open file (in target "VIMediaCache" in project "Pods") (in target 'VIMediaCache' from project 'Pods')
warning: MobileCoreServices has been renamed. Use CoreServices instead. (in target 'VIMediaCache' from project 'Pods')
error: /Users/mrzhou/app/video/Carthage/Checkouts/BMPlayer/Example/Pods/Target Support Files/SnapKit/SnapKit.release.xcconfig: unable to open file (in target "SnapKit" in project "Pods") (in target 'SnapKit' from project 'Pods')
error: /Users/mrzhou/app/video/Carthage/Checkouts/BMPlayer/Example/Pods/Target Support Files/SnapKit/SnapKit.release.xcconfig: unable to open file (in target "SnapKit" in project "Pods") (in target 'SnapKit' from project 'Pods')
error: /Users/mrzhou/app/video/Carthage/Checkouts/BMPlayer/Example/Pods/Target Support Files/SnapKit/SnapKit.release.xcconfig: unable to open file (in target "SnapKit" in project "Pods") (in target 'SnapKit' from project 'Pods')
error: /Users/mrzhou/app/video/Carthage/Checkouts/BMPlayer/Example/Pods/Target Support Files/NVActivityIndicatorView/NVActivityIndicatorView.release.xcconfig: unable to open file (in target "NVActivityIndicatorView" in project "Pods") (in target 'NVActivityIndicatorView' from project 'Pods')
error: /Users/mrzhou/app/video/Carthage/Checkouts/BMPlayer/Example/Pods/Target Support Files/NVActivityIndicatorView/NVActivityIndicatorView.release.xcconfig: unable to open file (in target "NVActivityIndicatorView" in project "Pods") (in target 'NVActivityIndicatorView' from project 'Pods')
error: /Users/mrzhou/app/video/Carthage/Checkouts/BMPlayer/Example/Pods/Target Support Files/NVActivityIndicatorView/NVActivityIndicatorView.release.xcconfig: unable to open file (in target "NVActivityIndicatorView" in project "Pods") (in target 'NVActivityIndicatorView' from project 'Pods')
```
### Reproduce
[The steps to reproduce this issue. What is the URL you were trying to play, where did you put your code, etc.]
### Other Comment
[Add anything else here]
Answers:
username_1: 请问你的 cartfile 内容和 build 使用的命令是怎么样的? |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.