repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
pangeo-data/pangeo | 289708238 | Title: Add support for slurm to pangeo object
Question:
username_0: I like the pangeo/pbs.py module and think that it should be extended. To this end I have moved some of the code from there to a parent module I call JobQueue and created a slurm module. I've confirmed that the test_pbs still works and added a test_slurm, but I haven't managed to get the test_slurm working yet and I'm looking for help in debugging.
My fork is <EMAIL>:username_0/pangeo.git branch is addslurm
Answers:
username_1: I'm very glad to hear this. Can I recommend that you issue a PR marked
with WIP (work in progress)? The PR interface is really nice for having
conversations around code, even if a PR isn't ready for inclusion.
Status: Issue closed
username_0: opened PR #83
username_2: I like the pangeo/pbs.py module and think that it should be extended. To this end I have moved some of the code from there to a parent module I call JobQueue and created a slurm module. I've confirmed that the test_pbs still works and added a test_slurm, but I haven't managed to get the test_slurm working yet and I'm looking for help in debugging.
My fork is <EMAIL>:username_0/pangeo.git branch is addslurm
username_2: closed via #83
Status: Issue closed
|
testorgiz/tutorials | 279518287 | Title: Tutorial Page onedx-team-meeting.md Issue. TEST GREEN
Question:
username_0: Tutorial issue found: [https://github.com/testorgiz/tutorials/blob/master/tutorials/onedx-team-meeting/onedx-team-meeting.md](https://github.com/testorgiz/tutorials/blob/master/tutorials/onedx-team-meeting/onedx-team-meeting.md) contains invalid tags. Even though your tutorial was created, the invalid tags listed below were disregarded. Please double-check the following tags:
- tutorial>title
Affected server: TEST GREEN<issue_closed>
Status: Issue closed |
bazelbuild/rules_kotlin | 1162588000 | Title: Issue with "jdeps automatically generated from kt_jvm_library rules" and genrule
Question:
username_0: There is an issue with generating jdeps as a side-effect of ``kt_jvm_library`` as follows...
Given a ``kt_jvm_library`` target:
```
kt_jvm_library(
name = "com-dancer-provider-collection",
...
)
```
And a ``genrule`` target that depends on the jar created above:
```
genrule(
name = "com-dancer-provider-collection-hollow-gen",
cmd_bash = "kotlin -cp $(location :com-dancer-provider-collection) dancer.provider.collection.GenerateClassesKt $(RULEDIR)"
...
)
```
An error is generated indicating that the target ``:com-dancer-provider-collection`` has *two outputs* and therefore ``$(locations)`` should be used instead of ``$(location)``. Those two outputs are the .jar file and the jdeps. However, using ``$(locations)`` is not possible because it generates an incorrect classpath for the ``kotlin`` invocation.
This seems undesirable.
I was able to craft an ugly workaround as follows:
```
load("//:functions.bzl", "get_compile_jar")
get_compile_jar(
name = "collection-jar",
lib = ":com-dancer-provider-collection",
)
genrule(
name = "com-dancer-provider-collection-hollow-gen",
cmd_bash = "kotlin -cp $(location :collection-jar) dancer.provider.collection.GenerateClassesKt $(RULEDIR)"
...
)
```
Where ``get_compile_jar()`` is defined in ``functions.bzl`` as:
```
def _compileJarImpl(ctx):
path = ctx.attr.lib[JavaInfo].java_outputs[0].compile_jar
return [
DefaultInfo(files = depset([path])),
]
get_compile_jar = rule(
implementation = _compileJarImpl,
attrs = {
"lib": attr.label(),
},
)
```
By taking ``java_outputs[0].compile_jar`` I am able to get just the .jar file (where ``java_outputs[1]`` references the jdeps output) and set that as the singlar output of the ``get_compile_jar`` rule. This works but is ugly. I do not think that the standard ``java_library`` rule output behaves in this way, though I could be wrong.
Answers:
username_1: Notes:
jdeps is returned as part of the default file outputs.
https://github.com/bazelbuild/rules_kotlin/blob/master/kotlin/internal/jvm/impl.bzl#L42-L43
Decision needs to be made wether to keep them as part of the implicit outputs.
username_1: Will remove the jdeps from DefaultInfo.
username_1: Open Question: Where does unused dependencies get jdeps from?
username_0: @username_1 It looks like the ``JavaInfo`` structure contains a field explicitly for *jdeps*. So, the rule can generate a single output (``JavaInfo``), with this field populated:
https://github.com/bazelbuild/bazel/blob/01a46f05dc79db2313c6c8e174a5d6eab474aefc/src/test/java/com/google/devtools/build/lib/rules/java/JavaInfoRoundtripTest.java#L88 |
ArtPoon/BelleJS | 845613661 | Title: Debugging XML (JC, strict clock, constant coalescent, dated tips)
Question:
username_0: ```console
Mar 30, 2021 9:33:54 PM dr.app.beast.BeastMain <init>
SEVERE: Parsing error - poorly formed BEAST file, belle-jc-strict-dated.xml:
Error parsing '<scaleOperator>' element with id, 'null':
Scale operator can only be used on parameters constrained to be strictly positive or negative (clock.rate)
Error thrown at: dr.inferencexml.operators.ScaleOperatorParser.parseXMLObject(Unknown Source)
```
Answers:
username_0: Problem is this block:
```xml
<!-- The strict clock (Uniform rates across branches) -->
<strictClockBranchRates id="branchRates">
<rate>
<parameter id="clock.rate" value="1.0"/>
</rate>
</strictClockBranchRates>
```
`<parameter id="clock.rate" value="1.0"/>` needs to have `lower="0.0"`
Status: Issue closed
|
ddavison/rest-client | 184267503 | Title: How to search?
Question:
username_0: How do I search the returned payload for certain text strings?
Answers:
username_1: you are able to click the Open in External Editor, and search from that new editor! if you are on windows, then this is currently [not a possibility](https://github.com/username_1/rest-client/issues/42).. that bug needs to be fixed first.
Status: Issue closed
|
lowRISC/opentitan | 996533452 | Title: [Private CI] clkmgr hangs
Question:
username_0: clkmgr tests failed with Segmentation fault error, which I think it may be due to an infinite loop. Then I enabled `loopdetect` to rerun and now it has been stuck for over 35 mins.
Temporarily remove clkmgr from smoke and nightly.
Answers:
username_0: According to Synopsys, this is a tool issue fixed in VCS 2020.12-SP2-2.
It appears with VCS 2020.12-SP2-1, but I tried with VCS 2020.12-SP2 (the default version used in OpenTitan) and it's also fine.
Let me add clkmgr test back and close this issue.
username_1: i think this is closed now, @username_0 let me know if you disagree.
Status: Issue closed
username_0: I forgot to close this. Thanks @username_1. |
flutter/flutter | 254984717 | Title: Android studio does not find flutter java source code
Question:
username_0: ## Steps to Reproduce
* Create a flutter plugin project
```
$ flutter create -t plugin plugin_test
```
* Open the project (`plugin_test/example`) in android studio
* In the project view, goto `plugin_test/java/com.yourcompany.plugintest/PluginTestPlugin`
* Ctrl+click on e.g. `result.success` in line 26 and notice you get to a decompiled view, instead of the original sources.
## Flutter Doctor
[✓] Flutter (on Linux, locale en_US.UTF-8, channel master)
• Flutter at /usr/local/google/home/username_0/projects/flutter
• Framework revision baf3b45e0d (28 hours ago), 2017-09-02 22:27:17 -0700
• Engine revision 29e0c1c281
• Tools Dart version 1.25.0-dev.11.0
[✓] Android toolchain - develop for Android devices (Android SDK 26.0.1)
• Android SDK at /usr/local/google/home/username_0/Android/Sdk
• Platform android-26, build-tools 26.0.1
• Java binary at: /usr/local/google/home/username_0/installed/android-studio/jre/bin/java
• Java version OpenJDK Runtime Environment (build 1.8.0_112-release-b06)
[✓] Android Studio (version 2.3)
• Android Studio at /usr/local/google/home/username_0/installed/android-studio
• Java version OpenJDK Runtime Environment (build 1.8.0_112-release-b06)
[✓] IntelliJ IDEA Community Edition (version 2017.2)
• Flutter plugin version 17.0
• Dart plugin version 172.3968.27
[✓] Connected devices
• Nexus 5X • 00dca44ab04d9baf • android-arm • Android 7.1.2 (API 25)
• Android SDK built for x86 • emulator-5554 • android-x86 • Android 7.0 (API 24) (emulator)
Answers:
username_1: @username_0 does this sound like https://github.com/flutter/flutter-intellij/issues/1482 ?
username_2: @devoncarew Is this an IDE bug we should move over?
username_3: The work-around, copied from a similar report:
Android Studio works best when you open the Android module as a top-level project. You can easily do so by selecting the Android module then using the menu Tools > Flutter > Open Android module in Android Studio. |
adriens/chocolatey-schemacrawler | 697281333 | Title: 16.9.5
Question:
username_0: # New schemacrawler release
Please release the choco package for latest schemacrawler `16.9.5`
# Want to make release faster ?
1 - Modify [schemacrawler.properties](schemacrawler.properties) with the target version
2 - Make a *PR*
Answers:
username_0: I will work on it this evening 👍
username_0: 
username_0: 
username_0: 
Status: Issue closed
|
dominikh/go-tools | 639248036 | Title: SA4006: build tags and control flow
Question:
username_0: The latest entry in our "build tags hate us" saga: SA4006.
```
func gen() int { return 42 }
func die() { panic("good bye") }
func Fn() {
x := gen()
println(x)
x = gen()
die()
println(x)
}
```
we flag the second assignment to `x` here, because we recognize that control flow doesn't continue after `die` gets called. Things get problematic when `die` turns into `feature that is only implemented on some OSs and panics unconditionally on others thanks to build tags`. Now we flag an assignment that is only conditionally unused, depending on the build tags in use. |
kkeker/unicom24-test-task | 260971399 | Title: Разработать документацию на API
Question:
username_0: Разработать документацию в проекте в GIT в виде Markdown-документа с требованиями, входными/выходными параметрами и примерами.
Перенести это описание в публичную корпоративную WiKi и прикрепить к задаче PDF-версию для рассылки заказчикам.
Answers:
username_0: Выполнено https://github.com/username_0/unicom24-test-task/blob/master/README.md |
airavata-courses/autobots | 379494244 | Title: Create private docker container registry on jetstream
Question:
username_0: Create a private docker registry on jetstream which keeps track of the latest build images from github and can be used for kubernetes pods
Status: Issue closed
Answers:
username_0: Docker registry is hosted at js-169-230.jetstream-cloud.org:5000
Refer to documentation on [Wiki](https://github.com/airavata-courses/autobots/wiki/Project-3) for more info |
aokashi/aokashi_home | 664337056 | Title: Bulma 0.9.0 のリリースに伴うアップデートの検討
Question:
username_0: Bulma 0.9.0 がリリースされました。
いくつか変更点があるようですが、現在使用している Aokashi Home に差し替えて影響が無いか確認しておきたいです。
## 特に気になる箇所
- helpers CSS の非推奨化
- 余白を指定する CSS の追加
- これまで余白を指定するのであれば、セクション(だったかな?) クラスに頼る必要があったが、その余白が大きすぎたので、細かく指定出来るのは有り難い<issue_closed>
Status: Issue closed |
joshwatson/binaryninja-msp430 | 900794188 | Title: JLO lifting condition backwards
Question:
username_0: Explicit example can be seen in Jakarta level in Microcorruption. The JLO lifting target is backwards, the ASM is:

Meanwhile the lifting is :

Simple fix of adding LLFC.UGT condition to flags and swapping around the condition in the lifter.
Answers:
username_1: github never tells me I have issues open! I refactored some stuff for binja 3.0 and added the fixes from your PR into it (there were merge conflicts with the PR so I just manually did them). Thanks for the report!
Status: Issue closed
|
M-Yankov/SWapi-CSharp | 619569837 | Title: Update to Swapi.dev
Question:
username_0: They moved to new URI address: Need to update documentation and every hardcoded string in the project solution.
Status: Issue closed
Answers:
username_0: lol forgot to update documentation (ReadMe.md)
username_0: They moved to new URI address: Need to update documentation and every hardcoded string in the project solution.
username_0: closed with https://github.com/username_0/SWapi-CSharp/pull/20
Status: Issue closed
|
JuliaPackaging/BinaryBuilder.jl | 339184554 | Title: Add an option to not unpack sources.
Question:
username_0: On occasion, it can be useful to not unpack a src tarball (e.g. because the tarball has a bunch of data files that need to extracted to a specific directory). Add an option to (not) do that.
Answers:
username_1: Is it not better to just move the extracted files after they've already been extracted? Or are you saying you'd only want to extract a small portion of the files within the tarball?
username_0: My problem is that I don't know what the extracted files are if they get extracted to the top level directory all at once.
username_1: Your tarball doesn't contain a containing directory?

username_0: Yeah, I know :/
username_2: Fixed by #643 and #671: now users can either
* use
```julia
ArchiveSource(...; unpack_target = "subdir")
```
to unpack the tarbomb to `/workspace/srcdir/subdir`
* or use
```julia
FileSource(...)
```
and the archive will simply not be unpacked
Status: Issue closed
|
williamritchie/IRFinder | 435750899 | Title: library advice for Intron retention
Question:
username_0: could you let me know if a non-polyA library could be used for intron retention detection?
Answers:
username_1: Not recommended. I assume you mean Ribo-depletion libraries. The rationale of using polyA library is to ensure the RNA molecules detected are mature, so that we have high confident that the remaining introns in the RNA are due to intron retention (IR). In ribo-depletion libraries, it's hard to distinguish true IR from introns in nascent transcription/processing.
username_0: Thank you very much for your reply!
Could you please advice me on the threshold/cut-off for determining significance from differential IR analysis (using GLM from IRFinder)? Is the p-value < 0.05 or p-value <0.01 cut-off enough? Using test runs on few poly-A enriched datasets (with about 50 million reads), I observed there are very, very few introns (less than 20) that are significant with p-adjusted value. Do you have any idea on why is this very low?
username_1: There is no gold standard in p value/FDR cutoff. It's all about how much Type I error you want to tolerant. Actually I would say all cutoffs in bioinformatics should be case specific. Practically, DESeq2 uses a FDR cutoff at 0.1 for differential gene expression (DGE). And keep in mind another two things:
1) GLM method requires a solid estimation of variance/dispersion. With small number of samples, it is hard or the estimation can be biased.
2) 50M reads are sufficient to carry out DGE analysis. However, we're talking about introns here, which are way less covered in a library. You really should filter out introns with low sequencing depth. This also influences the total number of test you have to do and consequently influences the FDR.
username_0: Thanks for the explanation. I understand from the paper and your explanation that the intron depth should also be considered along with the differential p-values from DeSeq2. I actually tried implementing the following thresholds: p<0.05, splice depth about 4 (extracted from the quantification file column 19) and ir ratio 0.1 in at least one sample. However, for many datasets (with good coverage) explored so far I end up with very few introns significant using above thresholds. And end up with almost nothing using FDR 0.1. Does this mean in general the introns retention patterns in these datasets are not biologically important, since there are very less high-confidence IRs unlike DEGs?
username_1: Please note we hadn't implemented GLM-based method for differential IR in the paper, which is relatively stringent on small sample size. Instead, we applied Audic test. And we were working on samples with distinct cell morphologies. There we observed ~80 differential IRs. Our conclusion is, IR signature, like gene expression, is a cellular feature under strict regulation.
Your observation of few differential IR can be either as expected or intriguing, depending on what the exact comparison you've set up. Is it between different cell types? Or is it between WT and mutations (and what kind of mutation)?
Before you jump into statistical test, ask yourself do you really expect significant IR changes in terms of biology. E.g. IR is a splicing related event, which is most likely controlled by splicing factors and epigenetics. Are you in that scenario? If you believe so, then you can do some quick checks by extracting IR values (from IRFinder report) across samples. You can average IR values in both your conditions, calculate the difference and rank them. You can check what's the range of IR changes, which gene contains the most dramatic change and etc. Do these result fit your hypothesis? If not, we might need to change your hypothesis. BTW, splice depth about 4 is way to low IMO.
Please remember the purpose of this site is to report bugs, instead of discussing biology. Feel free to send me direct emails if your question is NOT about IRFinder crash and I'm more than happy to help (my email is at the end of the frontpage of IRFinder Wiki) . If you don't mind, I'll close this report for now.
Status: Issue closed
|
psrc/urbansim2 | 192442014 | Title: Lag variables
Question:
username_0: Figure out how to do lag variables in urbansim2.
Answers:
username_0: From Scott's email 12/13/2016:
We didn’t use any lag variables with urbansim 2. I believe if you were going to do so, you would probably handle this use orca tables/columns to basically archive agent tables for previous years.
Status: Issue closed
username_0: Implemented in the function add_lag_datasets in (models.py)[https://github.com/psrc/urbansim2/blob/master/psrc_urbansim/models.py]. A solution for estimation needs to be worked out, see #58. |
Tencent/vConsole | 828615307 | Title: window.fetch 第二个参数是可选的。3.4.0版本覆盖系统window.fetch后,第二个参数变成必选,导致报错
Question:
username_0: let prevFetch = (input, **init**) => {
let id = that.getUniqueID();
that.reqList[id] = {};
let item = that.reqList[id] || {};
let query = [],
url = '',
method = 'GET',
requestHeader = null;
// handle `input` content
if (tool.isString(input)) { // when `input` is a string
method = init.method || 'GET';
url = input;
requestHeader = init.headers;
} else { // when `input` is a `Request` object
method = input.method || 'GET';
url = input.url;
requestHeader = input.headers;
}
Answers:
username_1: +1.
username_2: +1
username_3: 谢谢反馈,3.4.1已经修复。
Status: Issue closed
|
graphql-boilerplates/react-fullstack-graphql | 314443712 | Title: Refactor ApolloClient instantiation to new apollo-boost syntax
Question:
username_0: The advanced boilerplate has a lot of the older non-apollo-boost way of setting up the ApolloClient. For the purposes of helping others who might be new to Apollo and GraphQL, we should update this to use the new syntax.
Lines that should be refactored: [L14-L56](https://github.com/graphql-boilerplates/react-fullstack-graphql/blob/master/advanced/src/index.js#L14-L56)
Answers:
username_1: @username_0 In here since we are using apollo-link-ws, is it possible to migrate to apollo-boost, since they still do not support websocket link in apollo-boost according to this:-
https://github.com/apollographql/apollo-client/issues/3117 |
transcode-de/cookiecutter-django-project | 60495044 | Title: Use pip-tools to manage requirements
Question:
username_0: We should use [pip-tools](https://github.com/nvie/pip-tools) to manange the requirements. This way we could remove the dependencies from the requirements files and would have an easier way to check for outdated packages. It can also be used to update the various requirements files.
Currently pip-tools has an issue if pip's configuration contains an ``--find-links`` option to install Wheels (see [issue #79](https://github.com/nvie/pip-tools/issues/79)).
Also I'm not sure if pip-tools will find the requirements files if they are located in a requirements directory.
----------------------------------------
- Bitbucket: https://bitbucket.org/transcode/transcode-django-project/issue/1
- Originally reported by: <NAME>
- Originally created at: 2014-06-04T18:05:20.009
Status: Issue closed
Answers:
username_0: We should use [pip-tools](https://github.com/nvie/pip-tools) to manange the requirements. This way we could remove the dependencies from the requirements files and would have an easier way to check for outdated packages. It can also be used to update the various requirements files.
Currently pip-tools has an issue if pip's configuration contains an ``--find-links`` option to install Wheels (see [issue #79](https://github.com/nvie/pip-tools/issues/79)).
Also I'm not sure if pip-tools will find the requirements files if they are located in a requirements directory.
----------------------------------------
- Bitbucket: https://bitbucket.org/transcode/transcode-django-project/issue/1
- Originally reported by: <NAME>
- Originally created at: 2014-06-04T18:05:20.009
username_0: _From <NAME> on 2015-03-03 16:23:53+00:00_
In future we won't use the requirements directory. We have all requirements in `setup.py` instead.
Furthermore in the linked issue meant, we could use `pip list --outdated` without having an interactive mode. IMO we can't used the `pip-dump` feature to write new package version to (requirements) file. We need pip-dump for `setup.py` requires-lists. |
matvp91/shaka-player-react | 995627767 | Title: How to play Azure Video Analyzer url?
Question:
username_0: I want to play cvr videos which is stored in Azure Video Analyzer. In the Azure Video Analyzer it has base clientEndpoint API and token . Using both i have to play video in shaka player.
Answers:
username_1: I'm sorry but I don't think I'll be of much help considering I have no experience with "Azure Video Analyzer". If you manage to get a DASH (.mpd) / HLS (.m3u8) url out of there, you may serve this in the `src` prop.
Status: Issue closed
|
theQRL/QRL | 294932226 | Title: RPC block template
Question:
username_0: There is no rpc documentation, after quick code search I see that there is no even a trace of this functionality.
Answers:
username_1: Mining pool integration is something that we are working on at the moment.
What you are refering to is the complete feature set. We will have the JSON-RPC interface for the mining pools soon.
I noticed that you've been also actively developing mining pools. Maybe we can cooperate and help you with the integration. Feel free to contact us or join our Discord channel.
username_0: I am banned on qrl reddit, discord, slack... Just thought beta is beta, where feature set is already sealed. I have no plans in visible future to develop a pool for this algo, if you claim it will work with existing pools it will also work with my solo stratum which is available for free and will probably require only small changes. My policy is to charge exorbitant amount of money from ICO projects, never reached an agreement so far lol.
Status: Issue closed
username_0: How soon is now? There is literally nothing for devs not familiar with this code to play with rpc. Is it too hard to provide a markdown docs with 3 API calls described?
username_2: Here is the node-cryptonote-pool which has been modified to work with QRL codebase.
https://github.com/username_2/node-cryptonote-pool/
Following is the customized QRL config.json file for cryptonote-pool (make sure to update pool address)
https://github.com/username_1/qrl_dpool/blob/master/config.json
username_0: Offtopic, how to join testnet from master branch or maybe launch a private one?
username_2: You need to wait for next hard fork. We have not yet hard forked network with latest codebase that supports pool mining. Although if you are interested, you can run your private network to establish the pool and setup all other stuffs related to pool mining.
username_0: That's what I am asking for, any documentation?
username_2: If you would like to start your private one, you need to override the configuration file.
https://github.com/theQRL/QRL/blob/master/README.md
This above link includes how to override config.json
Just remove the peer_ip in config.json
Use following command to start your testnet with much less pain.
qrl wallet_gen
cp wallet.qrl ~/.qrl/
username_0: ```
curl -X POST -H "Content-Type: application/json" --data '{"id":0,"jsonrpc":"2.0","method":"getblocktemplate","params":{"reserve_size":9,"wallet_address":"Q0106009aeec11a402bb2abfdc0f02dcefcfba8620395a686555e61cd31b37e7a07a4730fab798e"}' http://127.0.0.1:9009/json_rpc
```
Getting response: `@@`.
I see no http endpoint changes in your node pool fork, definitely missing something. or I am getting malformed reply due to some weird qrl node state?
username_2: You need to run this grpcProxy.py
https://github.com/theQRL/QRL/blob/master/qrl/grpcProxy.py
This will support HTTP endpoint at port 18081
username_0: Consider different port for RPC, it's a conflict with Monero, there is no reason to copy rpc port if you decided to use same algo. It's confusing and annoying, some pools will definitely run both on same server.
username_2: We will make these ports configurable, nothing will be hardcoded, and sure, will look for other ports to avoid conflict.
username_0: So this rpc service is not a part of default application? You can really make many lives easier if mining rpc will listen on 127.0.0.1:PORT by default without tons of configs. There are tons of coins and every need to add their own "special things".
username_2: Yes for now the rpc service is not a part of default application.
username_0: `python3 qrl/grpcProxy.py`
```
Traceback (most recent call last):
File "qrl/grpcProxy.py", line 227, in <module>
payment_slaves = read_slaves(config.user.mining_pool_payment_wallet_path)
File "qrl/grpcProxy.py", line 21, in read_slaves
with open(slaves_filename, 'r') as f:
FileNotFoundError: [Errno 2] No such file or directory: '/home/.qrl/payment_slaves.json'
```
Why not /home/CURRENT_USER/.qrl ?
username_0: Missing `payment_slaves.json` is also confusing, where to get it?
username_0: There is no config.json, there is config.yml in ~/.qrl. And there is no peer_ip param in it.
username_2: You need atleast two nodes, to set it up correctly.
One of the node will be a mining node, but will not be running grpcProxy.py.
You can directly start that node using command
qrl --randomizeSlaveXMSS
The another node should be a node with pool service enabled.
You may use following command into it.
qrl wallet_gen
qrl -r --host 192.168.3.11 slave_tx_generate
(make sure you enter access_type 1 )
move the newly generated slaves.json into ~/.qrl/
qrl -r --host 192.168.3.11 slave_tx_generate --otsidx 1
(make sure you enter access_type 0 )
rename this file from slaves.json to payment_slaves.json and move it to ~/.qrl/
Now run your node, wait for the 2nd node to start mining. As soon as it successfully starts mining, you need to stop the node. Override config mining_enabled to False
This will stop your node from mining alone.
Restart the node.
In order to override config. you can copy the config.yml from
https://github.com/theQRL/QRL
The default config.yaml is already mentioned there.
You can create the file ~/.qrl/config.yml
and override any of the above mentioned settings.
username_0: But it still need `'/home/.qrl/payment_slaves.json'` and ignores current user.
username_2: I have mentioned above, how to generated payment_slaves.json
Read again, my last comment.
username_0: Haven't you see that it want /home/.qrl VS cirrent /home/ME/.qrl?
username_2: You can move that to /home/.qrl/
instead of what I mentioned above
username_0: Maybe it requires a fix instead?
username_2: Before the hardfork yes, processes will be simplified from what it is now.
username_0: Ok, I am done, 8 months after scamming people with ICo you ended up with defunct script.
username_2: Good luck
username_3: perhaps try send it valid json?
you're missing a brace at the end of what you're sending.
username_2: You need to run grpcProxy.py which is inside qrl directory.
username_0: @username_3 I am not missing a brace, only in my post it's not valid. I am using my existing stratum with this crap and it simply does not support `reserve_size` which is required for job distribution otherwise same job goes to miners and they will hash same job.
```
curl -X POST -H "Content-Type: application/json" --data '{"id":0,"jsonrpc":"2.0","method":"getblocktemplate","params":{"reserve_size":9,"wallet_address":"Q0106009aeec11a402bb2abfdc0f02dcefcfba8620395a686555e61cd31b37e7a07a4730fab798e"}}' http://127.0.0.1:18081/json_rpc
{"jsonrpc": "2.0", "error": {"code": -32602, "message": "Invalid params", "data": {"message": "getblocktemplate() got an unexpected keyword argument 'reserve_size'", "args": ["getblocktemplate() got an unexpected keyword argument 'reserve_size'"], "type": "TypeError"}}, "id": 0}
```
```
curl -X POST -H "Content-Type: application/json" --data '{"id":0,"jsonrpc":"2.0","method":"getblocktemplate","params":{"wallet_address":"Q0106009aeec11a402bb2abfdc0f02dcefcfba8620395a686555e61cd31b37e7a07a4730fab798e"}}' http://1192.168.3.11:18081/json_rpc
{"jsonrpc": "2.0", "result": {"height": 5534, "difficulty": 146746, "status": "OK", "blocktemplate_blob": "cbfbd9a26d83388fa54892917f61e56985422fe0aa9bb75058db8187c4b8db9086e892f8b9c28c0000000022e6c67d71a4bdb15b92ac57baef2c506e0b071ad5a872c365fa9545edee5d7d19f984afddb4cbf136"}, "id": 0}
```
You can see `00000000` in a response, here you will put a nonce from miner, but another reserved zero bytes needed to set a pool nonce and then send this job to miner. So basically, current state of QRL is that you need to run a pool per miner which is just retarded.
username_0: This commit of him demonstrates that they simply have no clue https://github.com/username_2/node-cryptonote-pool/commit/2edca30538754c1e69b8e2b88146fb532e975519
With extra nonce option nothing will work, without extra nonce pool is not a pool.
username_2: @username_0 Aware of that, work in progress with it.
username_0: Okay cool. I have no idea why you can't make grpcproxy a part of qrl already without any extra action, we need to test software and prepare servers.
username_0: I see some commits related to extra nonce over last 3 days, is it supposed to be fixed? If yes, here is my output using 6a35a6d3150507f8e99f88263cee44ec3a9f6ccb:
```
{"result": {"status": "OK", "blocktemplate_blob": "1514dd73241d8a7dd3b49deea2d524701c6319b5c4b5234e2b7d36207005a1c95b9eb0c457a1b70000000000000000b73884b533c62aa25459cc27bea7ade3398c2a3346a12f8a68ce192c69e31bafbc001c2a1f", "reserved_offset": 0, "height": 1169, "difficulty": 30399}, "id": 0, "jsonrpc": "2.0"}
```
`reserved_offset: 0` is kinda strange. And I am getting `unexpected keyword argument 'reserve_size'` as well.
username_2: You need to wait for next hard fork & release.
username_3: @username_0 you're aware several pools (including myself) were using that modified cryptonote repo with no issues? as far back as a week (or further) and successfully solving blocks..
the job of any stratum is to divide the nonce ranges to scan equally per miner; thats not purely up to extranonce itself..
username_0: OMG stop teaching me ok? I am talking about interface. I see it clearly that there is extra 4 bytes available, hence 8 total.
username_0: Currently there is uint32 for extra nonce. I would like to specify reserve size of 8 bytes vs default 4 hence it will be possible to shard stratums with the same block template. That's how it works in Monero. I was just confused that there is no way currently to request more. According to @username_2 I assume it will be implemented soon.
username_2: It has been implemented, we were testing the pool. Commit will be available soon. |
matplotlib/matplotlib | 449787393 | Title: required interactive framework error on Backend attribute on matplotlib V3.0.3
Question:
username_0: Hi guys,
As I am running below codes, I am getting " AttributeError: type object 'Backend' has no attribute 'required_interactive_framework' ".
import statsmodels.api as sm
test = pd.DataFrame({"prediction": pred, "observed": y_validation.flatten()})
lowess = sm.nonparametric.lowess
z = lowess(pred.flatten(), y_validation.flatten())
test.plot(figsize = [14,8],
x ="prediction", y = "observed", kind = "scatter", color = 'darkred')
plt.title("KNN: Prediction Vs Test Data", fontsize = 18, color = "darkgreen")
plt.xlabel("Predicted Power Output", fontsize = 18)
plt.ylabel("Observed Power Output", fontsize = 18)
plt.plot(z[:,0], z[:,1], color = "blue", lw= 3)
plt.show()
Do you have any suggestion that can solve my issue?
Cheers,
Answers:
username_1: Can you please fill out
* Operating system:
* Matplotlib version:
* Matplotlib backend (`print(matplotlib.get_backend())`):
* Python version:
* Jupyter version (if applicable):
* Other libraries:
My knee-jerk guess is that your enviroment is corrupted and you have over-lapping installations of different versions of mpl.
username_0: Thanks a lot, @username_1. Your knee-jerk guess solved the issue I have on MLP. I uninstalled and installed. After restarting the conda, the trouble was gone.
Status: Issue closed
|
jlippold/tweakCompatible | 354080909 | Title: `FlipConvert` working on iOS 11.3.1
Question:
username_0: ```
{
"packageId": "com.julioverne.flipconvert",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "com.julioverne.flipconvert",
"deviceId": "iPhone10,2",
"url": "http://cydia.saurik.com/package/com.julioverne.flipconvert/",
"iOSVersion": "11.3.1",
"packageVersionIndexed": true,
"packageName": "FlipConvert",
"category": "Tweaks",
"repository": "julioverne's Repo",
"name": "FlipConvert",
"installed": "0.0~beta5",
"packageIndexed": true,
"packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 100% with 4 working reports.",
"id": "com.julioverne.flipconvert",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.1.0",
"shortDescription": "Port Flipswitch Toggles With CCSupport",
"latest": "0.0~beta5",
"author": "julioverne",
"packageStatus": "Working"
},
"base64": "<KEY>
"chosenStatus": "working",
"notes": ""
}
``` |
pulumi/get.pulumi.com | 892170991 | Title: Update operational practices around the get.pulumi.com bucket
Question:
username_0: If this bucket is inadvertently deleted, then we lose all production artifacts
Answers:
username_1: We currently protect it from deletion in the program. Do you have a specific suggestion? Maybe we should replicate?
username_2: Yes, I think we should add bucket replication for this. |
liamdamato1997/acunetix360 | 699264937 | Title: Vulnerability - Cross-site Scripting
Question:
username_0: **URL:** http://testhtml5.vulnweb.com/#/popular
**Name:** Cross-site Scripting
**Severity:** High
**Confirmed:** True
**ParameterName:** username
**ParameterType:** POST
**Payload:** '"--></style></scRipt><scRipt>netsparker(0x008CEB)</scRipt>
**Injection URL :**
http://testhtml5.vulnweb.com/login
You can see vulnerability details from the link below:
https://online.acunetix360.com/issues/detail/ed6dd92e252842c5d773ac3302b3916f |
microsoftgraph/msgraph-sdk-dotnet | 964661955 | Title: App install context in not exposed in Graph SDK and also not returned in the Graph Explorer.
Question:
username_0: App install Execution context property is not exposed to be set from the Microsoft Graph SDK client(Version : 3.19.0.0) for msi packages WindowsMobileMSI class. Is there any specific reason for not exposing it from Graph SDK ?
Thanks,
Sangamesh
Answers:
username_1: Hey @username_0,
The SDK is generated from the metadata that is available [here](https://graph.microsoft.com/v1.0/$metadata) and the property mentioned is not present at the moment.

Any chance you could provide more information on whether/how the property can indeed be set/retrieved?
username_0: Hi @username_1 Thanks for the quick reply. Property App install context is not available as part of metadata as shown in UI below. I wanted to know the reason why this property is not exposed as part of metadata of winsdowsMobileMSI.

Status: Issue closed
username_1: Thanks for the extra information @username_0.
As this repository specifically deals with client library related issues, we may not be in a position to know if this is a bug or something unsupported.
Please raise the same question at the link below with the appropriate tag and the relevant API owner should be able to give us more information on the status of this. (It would be great if you would also add the link to the question in this issue so that future users may find the information as well)
https://docs.microsoft.com/en-us/answers/products/graph
Closing this for now so that it may be followed up there. |
pmodels/mpich | 1081029011 | Title: use -fallow-argument-mismatch automatically when necessary
Question:
username_0: I think MPI Fortran bindings require this behavior, so if MPICH configure detects that it is necessary, why not just automatically set the flags, rather than fail and force the user to do it. Given the user has no option in this case, why bother asking?
```
checking whether flang allows mismatched arguments... yes, with -fallow-argument-mismatch
configure: error: The Fortran compiler flang does not accept programs that call the same routine with arguments of different types without the option -fallow-argument-mismatch. Rerun configure with FCFLAGS=-fallow-argument-mismatch
```
Answers:
username_1: Hopefully https://github.com/pmodels/mpich/pull/5685 will obviate the need of this flag.
Status: Issue closed
|
VATSIM-UK/UK-Sector-File | 1046699821 | Title: St Athan (EGSY) Link C Added Hold A5 Moved
Question:
username_0: # Changes
Link C added as manouevring area
Hold A5 moved
# Link to AMDT (page number on PDF)
[AMDT 2112 P972](https://nats-uk.ead-it.com/cms-nats/export/sites/default/en/Publications/AIP/EG_Amdt_A_2021_12_en_v1.pdf)
# Files to be changed
/Airports/EGSY/SMR
Answers:
username_1: @username_2 you may want to look at this one given your most recent contribution? :)
username_2: I shall certainly take a look. Currently away for a week so will have to wait until I get back.
username_1: @username_3 if you're still going to work on this one, there's also now #3927 to consider - probably best to group those
username_3: will do!
Status: Issue closed
username_4: Closed - all details now in #3927 |
FriendsOfCake/crud | 309725736 | Title: How to use one Model in two controller
Question:
username_0: Dear Friends,
I have two controllers `UserController.php` in `/Admin` and `VendorsController.php` in `/Vendors`. but these Both Controllers should use single Model `CustomersTable.php`.
Please let me to know how should i achieve this?
Status: Issue closed
Answers:
username_1: This is an issues tracker not a help forum. Please one of the official CakePHP help forum. |
pgjdbc/pgjdbc | 367889917 | Title: driver 42.2.5 does not recognize cacerts of JRE
Question:
username_0: **I'm submitting a ...**
- [x ] bug report
- [ ] feature request
**Describe the issue**
Using the driver 42.2.5 with ssl via a connection url like "jdbc:postgresql://host.domain.de:5432/dbname?ssl=true" produces error message "Could not open SSL root certificate file C:\Users\username\AppData\Roaming\postgresql\root.crt" instead of connecting. This is although the full cert and ca verification can be via cacerts of the JRE. That means, the (commercial) ca of my db certificate is contained in this file. Providing the file "C:\Users\username\AppData\Roaming\postgresql\root.crt" is a hotfix and connecting works. However, validation via cert-store of JRE should operate without that file, at least it does using the driver 42.2.2 or previous. Please note, with both driver versions I can additionally add "sslmode=verify-full" to the connection string without any change, i.e., the older connects, whereas the newer does not.
**Java Version**
OpenJDK 11
**OS Version**
Win10 and Debian Testing
**PostgreSQL Version**
10.5
**Expected behaviour**
The root.crt file should not be necessary, if the verification can be done via JRE cert store. If it cannot be done, the connection error message of 42.2.2 was better, i.e., "SSL error: PKIX path building failed: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target". It should not be a must to have root.crt as the error message from above suggests.
Thanks, Chris
Answers:
username_1: As of 42.2.5 ssl=true implies verify-full as per the release notes. If you wish to get the old behaviour use sslmode=require
username_2: I think this issue highlights another side of the equation: currently pgjdbc defaults to LibpqFactory which basically emulates libpq location for certificates.
One can switch to JDK's trust store for certificates by using `sslfactory= org.postgresql.ssl.DefaultJavaSSLFactory`
However it might make sense to teach `LibpqFactory` to check Java's trust store for certificates when files like `root.crt` are not found.
@username_1 , what do you think?
username_1: @username_2 it's not difficult to do. The challenge I see is that the documentation gets difficult as the logic is becoming more "magical"
username_2: You are right.
Let's keep current factories as is, and let's just add `org.postgresql.ssl.MagicFactory` that uses all the certs it can locate :)
username_0: I disagree. sslmode=require does not simulate the old behavior: With "ssl=true&sslmode=require" no error is displayed and the connection is successfully opened after removing the issuing ca of my cert from cacerts of JRE then! In contrast and as I wrote above, the old behavior was then the error "SSL error: PKIX path building failed: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target" and the connection was not opened. This is clearly the better solution, as it seems (and is documented) that sslmode=require does not do any ca checks. These were done before 42.2.5.
Clearly, the best solution for a Java jdbc-driver is to check both, the JRE keystore and (if exists) root.crt. The former is the behavior one expects from a Java program, right? There is no magic in that and easy to document.
B.
username_0: Setting sslfactory=org.postgresql.ssl.DefaultJavaSSLFactory works. I hope, but have not checked, that it verifies the whole ca chain and the host identity as before. Is setting this factory as default the best solution?
Chris
username_2: That depends.
Technically speaking, pgjdbc had always been using "ssl factories" to specify the locations of the certificates.
Then if you want to use Java's trust store, you need sslfactory=org.postgresql.ssl.DefaultJavaSSLFactory
As I said, pgjdbc defaults to libpq-like mode, and it might be worth improving.
For instance, consider both `root.crt` and Java's keystore for cerver validation, however for client validation only a single certificate can be used, so it is somewhat hard to predict which one to send.
In the mean time, pull requests are welcome.
username_3: @username_2 I need to deploy an AWS Lambda Java container with pgjdbc connecting to an PostgreSQL RDS instance, or in other words, I need to include AWS's root certificate into pgjdbc. If I use the `sslfactory=org.postgresql.ssl.DefaultJavaSSLFactory` and truststore as you suggest, are there any advantages/disadvantages over the default SSL Factory from pgjdbc?
username_1: @username_3 the default SSL Factory requires you to have access to the machine. I'm guessing in this case you do not. Using the default JAVA SSL Factory just uses java's SSL mechanism. We added the pgjdbc factory to be more like libpq.
username_3: Yes, I don't have any access to the machine.
What I really wanted to do is pass the certificate PEM file as a regular java resource. This way I can run my code locally or in the cloud without having to modify anything or create a truststore and a SSL Factory myself. In other words, I want it to be simple and platform independent.
I tried to load the resource URI and pass it to pgjdbc, but it says it can't find the file (maybe because the URI starts with "file:"?).
username_1: Yes, you would have to load it from the class loader, not a file
username_3: And how can I pass it then to pgjdbc?
username_1: At first thought you would have to write your own SSL handler which loads from the classpath. This should be pretty much the same as the LibPQ handler except how it loads the file
username_4: @username_3 Check out SingleCertValidatingFactory. It's for exactly this use case, namely using a pinned SSL certificate that you have in advance but is not and cannot be loaded into the JVM wide keystore.
[1]: https://github.com/pgjdbc/pgjdbc/blob/master/pgjdbc/src/main/java/org/postgresql/ssl/SingleCertValidatingFactory.java
username_3: @username_1 I will see @username_4 indication first, if it doesn't work I will try your way. Thanks guys, I appreciate it a lot.
username_4: Here's the class: https://github.com/pgjdbc/pgjdbc/blob/master/pgjdbc/src/main/java/org/postgresql/ssl/SingleCertValidatingFactory.java
You'd specify it via the `sslsocketfactory` / `sslsocketfactoryarg` connection properties.
username_3: On the `SingleCertValidatingFactoryTest` I found this piece of code, and I think my case is similar, I just have to substitute `file:` to `classpath` as the `SingleCertValidatingFactory` docs point out.
`info.setProperty("ssl", "true");
info.setProperty("sslfactory", "org.postgresql.ssl.SingleCertValidatingFactory");
info.setProperty("sslfactoryarg", "file:" + goodServerCertPath);
testConnect(info, true);`
username_3: @username_4 this solution works perfectly, thanks!
username_3: I believe this should be included in the docs.
username_1: Please consider helping us out with a Pull Request
<NAME>
username_5: @username_4 @username_3 - I have a similar requirement wherein my kubernetes pod with pgjdbc need s to be connected to an PostgreSQL IBM Cloud instance, or in other words, I need to include IBM Cloud's root certificate into pgjdbc.
I followed the same steps as mentioned in SingleCertValidatingFactoryTest -
```
String url = "jdbc:postgresql://1ea3b3dc-9606-4ee9-8df1-06532a353f84.68ea2cbd8c8d4c30b5b8450be6b8593a.databases.appdomain.cloud:31290/clouddb";
Properties info = new Properties();
info.setProperty("ssl", "true");
info.setProperty("sslfactory", "org.postgresql.ssl.SingleCertValidatingFactory");
String fileContents = loadFile ("/opt/hb/postgres_cert/PGSSLROOTCERT.crt");
LOG.info(String.format(" PostgresDBInvoker - getConnection cert details %s .", fileContents));
info.setProperty("sslfactoryarg", fileContents);
c = DriverManager.getConnection(url, info);
```
On execution, i could see in the logs that the fileContents string contains the certificate details.
```
-----BEGIN CERTIFICATE-----
MI<KEY>
-----END CERTIFICATE-----
```
But when i try to make a connection, i receive the following error -
```
[8/1/19 11:10:24:956 UTC] 00000086 SystemErr R org.postgresql.util.PSQLException: SSL error: Received fatal alert: handshake_failure
[8/1/19 11:10:24:957 UTC] 00000086 SystemErr R at org.postgresql.ssl.MakeSSL.convert(MakeSSL.java:42)
[8/1/19 11:10:24:957 UTC] 00000086 SystemErr R at org.postgresql.core.v3.ConnectionFactoryImpl.enableSSL(ConnectionFactoryImpl.java:441)
[8/1/19 11:10:24:958 UTC] 00000086 SystemErr R at org.postgresql.core.v3.ConnectionFactoryImpl.tryConnect(ConnectionFactoryImpl.java:94)
[8/1/19 11:10:24:958 UTC] 00000086 SystemErr R at org.postgresql.core.v3.ConnectionFactoryImpl.openConnectionImpl(ConnectionFactoryImpl.java:192)
[8/1/19 11:10:24:958 UTC] 00000086 SystemErr R at org.postgresql.core.ConnectionFactory.openConnection(ConnectionFactory.java:49)
[8/1/19 11:10:24:959 UTC] 00000086 SystemErr R at org.postgresql.jdbc.PgConnection.<init>(PgConnection.java:195)
[8/1/19 11:10:24:959 UTC] 00000086 SystemErr R at org.postgresql.Driver.makeConnection(Driver.java:458)
[8/1/19 11:10:24:959 UTC] 00000086 SystemErr R at org.postgresql.Driver.connect(Driver.java:260)
[8/1/19 11:10:24:960 UTC] 00000086 SystemErr R at java.sql.DriverManager.getConnection(DriverManager.java:675)
[8/1/19 11:10:24:960 UTC] 00000086 SystemErr R at java.sql.DriverManager.getConnection(DriverManager.java:219)
[8/1/19 11:10:24:960 UTC] 00000086 SystemErr R at com.ibm.iae.db.PostgresDBInvoker.getConnection(PostgresDBInvoker.java:88)
-----
-----
[8/1/19 11:10:24:963 UTC] 00000086 SystemErr R Caused by:
[8/1/19 11:10:24:963 UTC] 00000086 SystemErr R javax.net.ssl.SSLHandshakeException: Received fatal alert: handshake_failure
[8/1/19 11:10:24:963 UTC] 00000086 SystemErr R at com.ibm.jsse2.k.a(k.java:42)
[8/1/19 11:10:24:964 UTC] 00000086 SystemErr R at com.ibm.jsse2.k.a(k.java:37)
[8/1/19 11:10:24:964 UTC] 00000086 SystemErr R at com.ibm.jsse2.av.b(av.java:549)
[8/1/19 11:10:24:964 UTC] 00000086 SystemErr R at com.ibm.jsse2.av.a(av.java:715)
[8/1/19 11:10:24:964 UTC] 00000086 SystemErr R at com.ibm.jsse2.av.i(av.java:574)
[8/1/19 11:10:24:964 UTC] 00000086 SystemErr R at com.ibm.jsse2.av.a(av.java:280)
[8/1/19 11:10:24:965 UTC] 00000086 SystemErr R at com.ibm.jsse2.av.startHandshake(av.java:431)
[8/1/19 11:10:24:965 UTC] 00000086 SystemErr R at org.postgresql.ssl.MakeSSL.convert(MakeSSL.java:40)
[8/1/19 11:10:24:965 UTC] 00000086 SystemErr R ... 20 more
```
username_4: @username_5 Create a separate issue for this and try to include a reproducible example that someone else could run. Nothing stands out from the code you've shown so not sure what's printing the "SystemErr" lines. That's likely something else on the app server as this driver does not do that.
username_1: Most problems with ssl have to do with using the correct format see the
note https://jdbc.postgresql.org/documentation/head/ssl-client.html
<NAME>
username_6: But in libqp, we can override the location of the root certificate by setting the environment variable `PGSSLROOTCERT` and point the file to the system truststore for example. This seems not possible with pgjdbc.
username_1: @username_6 https://jdbc.postgresql.org/documentation/head/connect.html#connection-parameters see sslcert. This allows you to specify the path to the cert
Status: Issue closed
|
qiu8310/dot-template | 415076628 | Title: 建议扩展一下Inject
Question:
username_0: inject 感觉有两个功能比较重要
1. [ ] 支持 prepend
2. [ ] 支持自定义 匹配 开头/结尾,如字符串/正则
Answers:
username_1: 第二个功能不太明白,能举个例子么?
username_0: 现在是通过
`# INJECT_START {"key": "xxx"} #`,
`# INJECT_END #`
这两个特殊串来匹配注入的,
但是有些场景并不适合加入这种特殊串,
比如向这个json的list注入的新的内容
```json
{
"list": [
"a",
"b"
],
"obj": {
}
}
```
可以通过 `"list": [`,`],`去匹配注入。
设置这个`INJECT_START`这种可能是为了避免重复,这点在一定程度上用户可以保证这个唯一
username_1: json 文件可以通过 JSON.parse 去解析并处理,或者使用 json5(带注释的 json)
你说的方式太宽松了,很容易注入出问题,导致代码莫名奇妙被更新。
username_0: 这就限制了,标准的json中不能使用。
可能不够安全,但是提供这种能力我觉得还是有用的,只是默认不开启
username_2: 请问不使用csjon而是使用json可以使用到inject功能吗,我在文档里没有看到有关inject的指南介绍呀
username_3: 突然反应过来我最近搞的 https://github.com/makeflow/inplate 和 inject 这个功能跟有点交集,不过之前虽然经常误插入 inject 注释,倒没有实际用过。我自己搞的是命令行工具,平时方便放到 CI 里。 |
traveller59/spconv | 418220511 | Title: Not able to run stup.py
Question:
username_0: Hi @traveller59 I am trying to install this module and getting this error. Please could you have a look.
Release
-- Caffe2: CUDA detected: 9.0.176
-- Caffe2: CUDA nvcc is: /usr/local/cuda/bin/nvcc
-- Caffe2: CUDA toolkit directory: /usr/local/cuda
-- Caffe2: Header version is: 9.0
-- Found cuDNN: v7.3.1 (include: /usr/include, library: /usr/lib/x86_64-linux-gnu/libcudnn.so)
-- Autodetected CUDA architecture(s): 6.1;6.1;6.1;6.1;6.1;6.1;6.1;6.1
-- Added CUDA NVCC flags for: -gencode;arch=compute_61,code=sm_61
CMake Error at CMakeLists.txt:28 (add_subdirectory):
The source directory
/root/spconv/third_party/pybind11
does not contain a CMakeLists.txt file.
-- Configuring incomplete, errors occurred!
See also "/root/spconv/build/temp.linux-x86_64-3.6/CMakeFiles/CMakeOutput.log".
See also "/root/spconv/build/temp.linux-x86_64-3.6/CMakeFiles/CMakeError.log".
Traceback (most recent call last):
File "setup.py", line 86, in <module>
zip_safe=False,
File "/usr/local/lib/python3.6/dist-packages/setuptools/__init__.py", line 145, in setup
return distutils.core.setup(**attrs)
File "/usr/lib/python3.6/distutils/core.py", line 148, in setup
dist.run_commands()
File "/usr/lib/python3.6/distutils/dist.py", line 955, in run_commands
self.run_command(cmd)
File "/usr/lib/python3.6/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "/usr/local/lib/python3.6/dist-packages/wheel/bdist_wheel.py", line 192, in run
self.run_command('build')
File "/usr/lib/python3.6/distutils/cmd.py", line 313, in run_command
self.distribution.run_command(command)
File "/usr/lib/python3.6/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "/usr/lib/python3.6/distutils/command/build.py", line 135, in run
self.run_command(cmd_name)
File "/usr/lib/python3.6/distutils/cmd.py", line 313, in run_command
self.distribution.run_command(command)
File "/usr/lib/python3.6/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "setup.py", line 39, in run
self.build_extension(ext)
File "setup.py", line 69, in build_extension
subprocess.check_call(['cmake', ext.sourcedir] + cmake_args, cwd=self.build_temp, env=env)
File "/usr/lib/python3.6/subprocess.py", line 311, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['cmake', '/root/spconv', '-DCMAKE_LIBRARY_OUTPUT_DIRECTORY=/root/spconv/build/lib.linux-x86_64-3.6/spconv', '-DCMAKE_PREFIX_PATH=/usr/local/lib/python3.6/dist-packages/torch', '-DPYBIND11_PYTHON_VERSION=3.6', '-DSPCONV_BuildTests=OFF', '-DCMAKE_CUDA_FLAGS="--expt-relaxed-constexpr"', '-DCMAKE_BUILD_TYPE=Release']' returned non-zero exit status 1.
Answers:
username_0: solved by adding
Status: Issue closed
|
google/filament | 1079521243 | Title: How can I turn off the PBR ?
Question:
username_0: **Describe the bug**
A clear and concise description of what the bug is.
**To Reproduce**
Steps to reproduce the behavior:
1. ...
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Logs**
If applicable, copy logs from your console here. Please *do not*
use screenshots of logs, copy them as text.
**Desktop (please complete the following information):**
- OS: [e.g. iOS]
- GPU: [e.g. NVIDIA GTX 1080]
- Backend: [OpenGL/Vulkan]
**Smartphone (please complete the following information):**
- Device: [e.g. Pixel 2]
- OS: [e.g. Android Pie 9.0]
**Additional context**
Add any other context about the problem here.
Answers:
username_0: https://github.com/google/filament/files/7704288/fps15.html.zip
this is the systrace file
username_1: You can't turn off PBR. Please use a discussion instead of an issue.
Status: Issue closed
username_0: @username_1 Thank you very much !
username_0: @username_1 Could you please help see the systrace file? whitch step cost much time? |
poooi/poi | 408422775 | Title: 卡住在loading畫面
Question:
username_0: **poi 版本 / poi version: 10.2.0
**操作系统 / OS: Windows 10
**插件名和版本 / Plugin name & version:**
**你遇到了什么样的问题 / The problem you've met: 如題, 卡住在loading畫面
開發者截圖: https://imgur.com/a/LJLqsax
Answers:
username_1: @username_0 需要 console 部分的内容
username_0: https://imgur.com/a/JeUiBjR
username_1: 请看看这里的版本是否可以使用 [ci.appveyor.com/project/KochiyaOcean/poi/builds/22244818/artifacts](https://ci.appveyor.com/project/KochiyaOcean/poi/builds/22244818/artifacts)
username_0: 可以了 謝謝
Status: Issue closed
|
wix/react-native-notifications | 685768433 | Title: Android Body Text Truncated
Question:
username_0: I've been testing out Android Local Notifications which are working.
The issue I'm having is my body text is getting truncated and isn't displaying fully and there's no way to expand the notification - if you tap it it disappears - any ideas why that's the case?
Also I don't really understand what what the extra key does in the example below - isn't doing anything as far as I can see:
Notifications.postLocalNotification({
title: "Local notification",
body: "This notification was generated by the app!",
extra: "data"
});
The docs are a little light on explanation.
Answers:
username_0: Looking at the Android docs you can implement a style for BigText - is this possible at all? https://developer.android.com/reference/android/app/Notification.BigTextStyle
username_0: Think I'm going to have to look at a different option. It's nearly there for me in terms of functionality but there's not really any documentation or support for Android.
username_0: I've found a solution to the issue. You need to add this to the PushNotificaytions.java:
**.setStyle(new Notification.BigTextStyle()**
```
final Notification.Builder notification = new Notification.Builder(mContext)
.setContentTitle(mNotificationProps.getTitle())
.setContentText(mNotificationProps.getBody())
.setContentIntent(intent)
**.setStyle(new Notification.BigTextStyle()**
.bigText(mNotificationProps.getBody()))
.setDefaults(Notification.DEFAULT_ALL)
.setAutoCancel(true);
```
Status: Issue closed
username_1: It is important , is It possible to add in master branch. |
anayinfowind/leeloolxp_tracking_sso | 708733823 | Title: consider length of plugin name
Question:
username_0: Your plugin name is quite long. Moodle has some limits on the lengths of table names due to certain databases that we support and all tables implemented in your plugin must use the full frankenstyle plugin name as a prefix to the tables so that the uninstall process can clean up and remove associated tables, but also to prevent naming conflicts.
your prefix "auth_leeloolxp_tracking_sso" is 27 characters long, and the limit for table names in moodle is 28 characters, at the moment you don't have any tables in your plugin, but it will be very difficult for you to add tables to your plugin in future so you might want to look at renaming to make your future easier.
Please note - you cannot rename a moodle plugins db entry so if you do rename you will need to make a new submission to the plugins db with your new code.
this doesnt block approval in the plugins db, but I would highly reccomend you address this anyway. |
4577/loult-ng | 186918932 | Title: Refactor du serveur
Question:
username_0: wesh.
Comme j'ten avais parlé, avant de tenter d'implémenter des effets sur les voix, je voulais refactor le code pour séparer les différents composants. Il faut que je teste si mon refactor marche, mais en attendant, check juste un peu si ça te va niveau de la structure (si c'est pas le cas je vais pas continuer du coup, jvais pas faire du taff pour rien).
https://github.com/username_0/loult-ng voilà le dépot de mon fork<issue_closed>
Status: Issue closed |
frontendbr/forum | 205375083 | Title: Traduções
Question:
username_0: **N**egads,
O que vocês estão usando para tradução de contexto em sistemas grandes no frontend?
Answers:
username_1: @username_0 você diz tradução de idiomas? Se sim, geralmente uso o I18n, seja bom Rails ou JS.
Geralmente soluções no server side são melhores em termos de SEO.
username_0: E quando tu tem um SPA full js servido pelo S3 da Amazon e se comunicando por API?
Estou desenvolvendo uma plataforma fechada então SEO não é um foco.
username_1: Hum saquei! Se é SPA, presumo que o render de view é no client né? Nesse caso tem a lib do i18n pra JS. Veja se o framework que vc tá usando tem algum adapter pra i18n e manda ver! :)
username_2: Já utilizei o [i18n](http://i18next.com/) com render no client e sempre uso quando preciso, bem fácil e tranquilo. +1
username_3: Deu certo @username_0 ?
username_4: Pessoal que trabalha com i18n no front.. Usam alguma ferramenta para abrir os JSONs e traduzir?
Os arquivos serão enviados para outras pessoas traduzirem. Em outro projeto onde as traduções ficam em arquivos .PO, eles usam o Poedit para facilitar a tradução.
Estou procurando uma ferramenta nesse sentido. Vi que tem a BabelEdit que faz isso e é paga.
username_5: @username_4 Estamos estudando o [Crowdin](https://crowdin.com/), ouvi em um podcast do PodProgramar e na RD eles utilizam ele.
username_6: Queria usar esse [cara](http://i18n-tag.kolmer.net/) mas nunca tive oportunidade.
Status: Issue closed
|
camelcasetechsd/attendance-system | 114575862 | Title: Persist the year filter in holidays
Question:
username_0: - Steps to reproduce
-- go to http://attendance.local/settings/holiday/index
-- select year ( e.g. 2015 )
-- press filter
-- the listing is updating with the holidays in 2015 , but the select dropdown is reverted to 2016<issue_closed>
Status: Issue closed |
spinnaker/spinnaker | 141080681 | Title: Does it support CentOs in Bake Process ?
Question:
username_0: when I bake something I only have two options ubuntu 12 and 14, what if I want CentOs ?
Answers:
username_1: You'll need to configure whatever base image make sense for your deployment in the bakery configuration: https://github.com/spinnaker/rosco/blob/master/rosco-web/config/rosco.yml
Not sure what platform you're on, but you should be able to find `baseImages` in that file for whichever you are on.
Then you'll need to cycle rosco to pick up the changes: http://spinnaker.io/documentation/troubleshooting.html#i-changed-my-configuration-how-do-i-get-spinnaker-to-pick-up-the-modified-configuration
The ui (deck) will dynamically render the Bake Stage with the set of choices you've configured within rosco.
username_0: Thanks where is the location of the config file on my spinnaker instance ? I couldn't find it under home/ubuntu or /var/ or /etc/...
username_1: If you are on a prebuilt spinnaker image, it is likely within /opt/rosco
somewhere.
username_1: It is `/opt/rosco/config/rosco.yml`.
username_0: Thanks, yes I use a prebuild image, do I need to re-build or re-compile after I change the configurations?
username_1: You should just need to cycle rosco to pick up the config change: http://spinnaker.io/documentation/troubleshooting.html#i-changed-my-configuration-how-do-i-get-spinnaker-to-pick-up-the-modified-configuration
username_1: Seems like this is resolved. Ok to close this issue?
Status: Issue closed
|
SymbiFlow/vtr-verilog-to-routing | 457511843 | Title: Include Yosys in VtR test infrastructure
Question:
username_0: As SymbiFlow heavily depends on Yosys it would be useful to integrate it in the VtR testing infrastructure to let it possible having SymbiFlow tests being run directly in VtR.
### Todo
* [ ] Check how the testing infrastructure is built to easily inject Yosys in the flow
* [ ] Try to use Yosys in the flow with a simple test
* [ ] Make a mergable patch that includes Yosys in VtR |
wso2/product-ei | 455148786 | Title: Unable to run processcleanuptool.sh with Secure Vault enabled in BPS
Question:
username_0: **Description:**
When configuring the package-cleanup, it needs to configure process-cleanup.properties file which is in BPS_HOME/repository/conf directory. When configuring this file need to give clientTrustStorePassword & <PASSWORD>.
But if enable the Secure Vault for these, these are not the plain text passwords.
After configuring process-cleanup.properties file with Secure Vault processcleanuptool is not working as expected.
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
**OS, DB, other environment details and versions:**
**Steps to reproduce:**
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> |
Maximus5/ConEmu | 432027447 | Title: conemu/conemuc/guimacro support for obtaining lists of titles of tabs (so can script switching to them)
Question:
username_0: The goal here was to write a utility that can list a 'tree' of ConEmu windows and tabs in them, and allow activating the user-selected ConEmu window and tab. (Can think of it as a component of an intelligent task switcher app).
To do so, was trying to see if GuiMacro support using ConEmuC had options for that, but only see ways to change title but not to get it. There dont seem to be other ways of obtaining a list of windows/tabs and/or their titles from other command line options for conemu / conemuc either.
Is there an existing support for allowing something like that? What would be the possibility/effort for supporting this kind of workflow? (Basically to allow cmd-line querying for ConEmu windows/tab ids/titles and allowing cmd line support for activating those windows/tabs). Thanks!
Answers:
username_1: I'm going to create GuiMacro `Tabs 12` to return list of tabs. Example:
```
# ConEmuC -GuiMacro tabs 12
1: cmd
2: cmd - ConEmuC -GuiMacro tabs 12 (Admin)
3: mxmmsk@LENOVO: /mnt/e/Maks/username_1/Tools/vim81.base
4: powershell
```
Than you may parse the result and execute for example `ConEmuC -GuiMacro tabs 7 4` to activate tab with `powershell`. |
socketio/engine.io | 165705117 | Title: Ping timeout test fails
Question:
username_0: This ping test fails sometimes, it depends on way too small timeout of 10ms:
Uncaught Error: expected 'transport error' to equal 'ping timeout'
This is with the ws module.
Answers:
username_0: [05:59:24] Using gulpfile ~/engine.io/gulpfile.js
[05:59:24] Starting 'test'...
․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․․
46 passing (2s)
1 failing
1) server close should trigger on both ends upon ping timeout:
Uncaught Error: expected 'transport error' to equal 'ping timeout'
at Assertion.assert (node_modules/expect.js/expect.js:99:13)
at Assertion.be.Assertion.equal (node_modules/expect.js/expect.js:200:10)
at Assertion.(anonymous function) [as be] (node_modules/expect.js/expect.js:73:24)
at Socket.onClose (test/server.js:424:29)
at Socket.Emitter.emit (node_modules/component-emitter/index.js:134:20)
at Socket.onClose (node_modules/engine.io-client/lib/socket.js:705:10)
at Socket.onError (node_modules/engine.io-client/lib/socket.js:671:8)
at XHR.<anonymous> (node_modules/engine.io-client/lib/socket.js:261:10)
at XHR.Emitter.emit (node_modules/component-emitter/index.js:134:20)
at XHR.Transport.onError (node_modules/engine.io-client/lib/transport.js:65:8)
```
Status: Issue closed
username_1: Closed by https://github.com/socketio/engine.io/pull/461. |
paulloz/godot-ink | 777345142 | Title: Should we be using master or .5 tag?
Question:
username_0: Hi! Trying with master, when I create my addons folder and try to load the plugin I get:
Unable to load addon script from path: 'res://addons/username_1.ink/PaullozDotInk.cs'.
The file exists.
Any insight on next steps for troubleshooting?
Answers:
username_1: Hi,
You probably haven't built the C# project before trying to load the plugin? I'll make sure to write that more explicitly in the README.
username_2: @username_0 I'm using the latest master code and have it working. My .csproj had some extra items (i cant remember why), i cleaned it up a little and its all working for me now.
Might be worth posting yours just in-case.
@username_1 I think the readme needs some work. For example, you mention the .dll but not where to get it. I initially compiled it from source until i realized it was with the compiler.
Another question for you regarding inklecate.exe, the setting, does that require the path or does it require the path and the filename? i.e. `D:\inklecate` or `D:\inklecate\inklecate.exe`
(also as an ease of life issue, is it not possible to have a window dialog to select the path or exe instead of having to copy and paste it in?)
Awesome job on the addon btw, going to try it out properly soon :)
username_1: It already is.

username_2: @username_1 i don't see that;

And regarding my project file, no it was this;
```
<Project Sdk="Godot.NET.Sdk/3.2.3">
<PropertyGroup>
<Configuration Condition=" '$(Configuration)' == '' ">Tools</Configuration>
<LangVersion>8.0</LangVersion>
<ProjectGuid>{DB00ED65-226D-4856-AE9B-6DB9D971D0B8}</ProjectGuid>
<OutputType>Library</OutputType>
<RootNamespace>Project</RootNamespace>
<AssemblyName>Project</AssemblyName>
<CodeAnalysisRuleSet>roslynator.ruleset</CodeAnalysisRuleSet>
<GenerateAssemblyInfo>false</GenerateAssemblyInfo>
<GenerateTargetFrameworkAttribute>false</GenerateTargetFrameworkAttribute>
<TargetFramework>net472</TargetFramework>
<!--The following properties were overriden during migration to prevent errors.
Enabling them may require other manual changes to the project and its files.-->
<EnableDefaultCompileItems>false</EnableDefaultCompileItems>
<GenerateAssemblyInfo>false</GenerateAssemblyInfo>
<Deterministic>false</Deterministic>
</PropertyGroup>
<ItemGroup>
<Reference Include="System" />
<Reference Include="Ink">
<HintPath>$(ProjectDir)/ink-engine-runtime.dll</HintPath>
<Private>False</Private>
<SpecificVersion>False</SpecificVersion>
</Reference>
</ItemGroup>
<ItemGroup>
<PackageReference Include="Roslynator.Analyzers">
<Version>3.0.0</Version>
<PrivateAssets>all</PrivateAssets>
<IncludeAssets>runtime; build; native; contentfiles; analyzers; buildtransitive</IncludeAssets>
</PackageReference>
<PackageReference Include="System.Reactive.Linq" Version="4.4.1" />
<PackageReference Include="Unity.Container" Version="5.11.8" />
<PackageReference Include="Newtonsoft.Json" Version="11.0.2"/>
</ItemGroup>
<ItemGroup>
<Compile Include="scripts\**\*.cs" />
</ItemGroup>
</Project>
```
username_2: Must be a little godot bug @username_1 , i disabled the plugin and re-enabled and the folder icon appeared.
username_1: @username_0 do you have an update regarding this issue? Based on the answers here and the updated README?
Status: Issue closed
username_1: Without any news on this and as it seems to be user error, I'm closing. |
decentraland/decentraland-dapps | 1166497561 | Title: Reduce the amount of RPC requests in order to check the txs status
Question:
username_0: - Try to get the txs status by using multicall as we do for approvals
- Try to reduce the number of retries when the tx is not found. Maybe by using a 2x of wait time every time it fails.
https://images.zenhubusercontent.com/141639552/084f9926-840f-4284-aac5-e6703ec8113a/screen_recording_2022_03_11_at_11_30_41.mov |
EMFTeam/EMF | 446565255 | Title: [SWMH] Silk road aesthetic path adjustment
Question:
username_0: ( Suggested path simply excludes the extraneous seazone 1942 )
Answers:
username_1: Looks like the recent brouhaha with the `positions.txt` fixes somehow changed the naval position of the Coast of Somalia. We can just revert the position to the old and be done with it. Or just bypass Socotra altogether, which I'm not averse to :p
username_0: The port was likely moved to point into the Coast of Yemen seazone, probably yes. However, I didn't move ports for no reason -- it was to put them in reasonable / workable positions such that they don't conflict with a wonder position.
Why don't we just go with the suggested change, to exclude the Coast of Yemen seazone? It looks perfect, the change has no gameplay difference whatsoever, etc.
username_0: Also, to note: I'd be totally opposed to bypassing Socotra on the Spice Route.
username_0: I meant:
Why don't we just go with the suggested change, to exclude the **Coast of Somalia** seazone?
username_1: So long as skipping a seazone is fine, I'm fine with that. I think it is.
Any reason for opposing bypassing Socotra? From a historical standpoint IMO Silk Road extension there makes the island worthier than it was. Contemporary records make little to no mention of marketable resources. There were trades between Socotra and the outside, of course, but there are countless examples of that.
username_0: Purely from a gameplay perspective, it's just fun having Socotra be on the Spice Route / Southern Silk Road, and since the island is literally right in the way of the route, it just seems weird not to include it, frankly.
Unless the island was something like the islands of the Canarias were -- inhabited by aboriginals that most definitely did not partake in trade with the outside world -- then I don't see any reason to exclude it. It can make a fun start -- or at least make the prospect of playing on Socotra a lot more interesting, and I've seen players do just that with this exact reasoning.
username_0: As far as skipping seazones, yeah, I'm almost positive that is no issue whatsoever. Ultimately I think all the seazones on any of the trade route paths (i.e., generically, the game mechanic itself) are for decoration. The trade route system only cares about sieged or occupied provinces on the path for the purposes of determining wealth bonus along the route (downstream of the sieges/occupations), and you cannot siege or occupy a seazone.
username_2: It was mainly for gameplay reasons that i added it to the route back in the day. Socotra's place on the route were my doing.
username_2: Personally i think it's a nice little bit of flavour to an otherwise flavour thin corner of the map
Status: Issue closed
username_0: Made the change as suggested. Closing issue. |
MorkaZ/MorkazSk | 405029555 | Title: Glowing / Glowing Color
Question:
username_0: Can you make a feature where you can make an entity glow (not just players, all entities)? Also make a way to turn off the glow. And most importantly, a way to specify the color of the glow. Example syntax's:
Make %entity% glow [%string%] # Make the player glow. The %string% would be a color. If color is not specified, make the color white.
Make %entity% stop glowing # Make the entity stop glowing
Answers:
username_1: I may be incorrect but entity glowing is potion effect and should be able to be applied using "apply potion to entity" effect
username_1: I may be incorrect but entity glowing is potion effect and should be able to be applied using "apply potion to entity" effect. If it is true then there is no point in creating effect like you suggested.
username_0: I believe you can only apply potion effects to living entities. So I believe you cannot apply them to dropped items, item frames, etc.
username_1: Thanks. Then I will add it to MorkazSk in near future.
Syntax should look like this:
`make %entities% glow with color %color% for %players%`
username_1: I have checked docs and asked community if Entity (not living) is possible to have colored glowing effect. It seems that it is impossible, it can be done only with living entities. I may be incorrect but with my knowledge it is not possible to do. If you (or someone) will find a way or will show how to do it, then I will try to create effect like you suggested.
Status: Issue closed
username_1: Can you make a feature where you can make an entity glow (not just players, all entities)? Also make a way to turn off the glow. And most importantly, a way to specify the color of the glow. Example syntax's:
Make %entity% glow [%string%] # Make the player glow. The %string% would be a color. If color is not specified, make the color white.
Make %entity% stop glowing # Make the entity stop glowing
SharpSK's features for glow really suck and don't work well so a new add on with this support would be awesome.
username_1: You are right. I have added it into MorkazSk. It will be updated soon :)!
username_1: Requested feature has been added in new update.
Status: Issue closed
|
nanjizal/hxRectPack2D | 381829675 | Title: Demo of phaser texturepacker data used in phasor
Question:
username_0: Demo Source: https://github.com/CodeAndWeb/TexturePacker-Phaser/blob/master/texturepacker-demo/cityscene.json
```
{"frames": {
"background":
{
"frame": {"x":2,"y":2,"w":800,"h":400},
"rotated": false,
"trimmed": false,
"spriteSourceSize": {"x":0,"y":0,"w":800,"h":400},
"sourceSize": {"w":800,"h":400}
},
"capguy/walk/0001":
{
"frame": {"x":967,"y":2,"w":158,"h":316},
"rotated": false,
"trimmed": true,
"spriteSourceSize": {"x":15,"y":3,"w":158,"h":316},
"sourceSize": {"w":187,"h":324}
},
"capguy/walk/0002":
{
"frame": {"x":1850,"y":2,"w":168,"h":303},
"rotated": false,
"trimmed": true,
"spriteSourceSize": {"x":0,"y":8,"w":168,"h":303},
"sourceSize": {"w":187,"h":324}
},
"capguy/walk/0003":
{
"frame": {"x":1553,"y":2,"w":152,"h":307},
"rotated": false,
"trimmed": true,
"spriteSourceSize": {"x":26,"y":2,"w":152,"h":307},
"sourceSize": {"w":187,"h":324}
},
"capguy/walk/0004":
{
"frame": {"x":1707,"y":2,"w":141,"h":306},
"rotated": false,
"trimmed": true,
"spriteSourceSize": {"x":42,"y":1,"w":141,"h":306},
"sourceSize": {"w":187,"h":324}
},
"capguy/walk/0005":
{
"frame": {"x":1275,"y":2,"w":139,"h":311},
"rotated": false,
"trimmed": true,
"spriteSourceSize": {"x":33,"y":3,"w":139,"h":311},
"sourceSize": {"w":187,"h":324}
},
"capguy/walk/0006":
{
"frame": {"x":1416,"y":2,"w":135,"h":311},
"rotated": false,
"trimmed": true,
"spriteSourceSize": {"x":29,"y":8,"w":135,"h":311},
"sourceSize": {"w":187,"h":324}
},
[Truncated]
},
"capguy/walk/0008":
{
"frame": {"x":804,"y":2,"w":161,"h":319},
"rotated": false,
"trimmed": true,
"spriteSourceSize": {"x":22,"y":1,"w":161,"h":319},
"sourceSize": {"w":187,"h":324}
}},
"meta": {
"app": "http://www.codeandweb.com/texturepacker",
"version": "1.0",
"image": "cityscene.png",
"format": "INDEXED",
"size": {"w":2020,"h":404},
"scale": "1",
"smartupdate": "$TexturePacker:SmartUpdate:3f95b1feed2fbfc4385b49348c54e958:fed0f696c0a1349f4c5839d5f65ab643:292542e6853f316339413b9243f95b10$"
}
}
```
Answers:
username_0: Implemented this format as above so that the demo traces this.
Status: Issue closed
|
azine/AzineMailgunWebhooksBundle | 261614121 | Title: [RFC] Deprecate testing against old versions (PHP & framework) for v2
Question:
username_0: * Require and test against 7.1, 7.2/nightly (allowing failures against nightly)
* Test against Symfony 2.8, ^3.3, & dev-latest, for LTS & latest versions
With a move to a PHP 7 code-base, we can also move to PHPunit ^6.0.
Are people still actively using PHP 5?
Status: Issue closed
Answers:
username_1: I still use php 5.6. ;-)
A lot of applications are developed and deployed and keep running for years. So I also guess I am not alone still usng 5.6
I think all the other suggestions are fixed with your PR #17. Right? |
xamarin/Xamarin.Auth | 216005238 | Title: OAuth 1.0a (One Legge)
Question:
username_0: hi,
On OAuth1 i don't have token before beginning registered. And when i used OAuth1.getsignature without token secret the method return me a empty signature.
Can you help me?
thx
Answers:
username_1: Hi @username_0
Can I get back to you in few days (as soon as NativeUI support is published - SFSafariViewController and Custom Tabs).
username_1: @username_0
Which OAuth service provider?
I need for testing.
Status: Issue closed
username_2: No response from over a year |
swcarpentry/git-novice | 282857569 | Title: Correction to data path for Python lesson, episode 4, solution to the last exercise
Question:
username_0: Hi,
I propose a change to the Python lesson, episode 4, solution to the last exercise called "Generate Composite Statistics".
During the whole lesson the attendees are inside ~/Desktop/python_novice_inflamation/data/. Yet the solution to the exercise assumes that they are in ~/Desktop/python_novice_inflamation/.
I propose to change the solution to the exercise from:
import glob
import numpy
import matplotlib.pyplot
filenames = glob.glob('data/inflammation*.csv')
to:
import glob
import numpy
import matplotlib.pyplot
filenames = glob.glob('inflammation*.csv')
Cheers,
Katerina
Answers:
username_1: Hi, @username_0 ! Thanks for your contribution. I think you meant to open the issue at [the python lesson repo](https://github.com/swcarpentry/python-novice-inflammation). I'm closing this here, but you definitely should open it there.
Status: Issue closed
|
McJtyMods/XNet | 447345004 | Title: [Suggestion] Set Filter based on inventory
Question:
username_0: Issue description:
I would like to propose an option that I would have loved to see in XNet for Sky Factory 4. In order to improve performance I import bonsai products via XNet into Filing cabinets.
To ensure that the system knows where to put each item, I have set up a whitelist for each connector. So the system knows where to input an item instead of checking each of my 30+filing cabinets. That is a boatload of boring work though.
In AE2 Storage busses you have a button to format the connector - which means, it checks the attached inventory and creates a whitelist of all items in it. It would be amazing to have that feature in the Xnet controller.
Thank you. |
ppoffice/hexo-theme-icarus | 623656454 | Title: 最近文章图片和分类如何删掉
Question:
username_0: 请问一下 最近文章图片和分类如何删掉,没有找到入口
Answers:
username_1: @username_0 https://blog.zhangruipeng.me/hexo-theme-icarus/Widgets/icarus%E7%94%A8%E6%88%B7%E6%8C%87%E5%8D%97-%E6%8C%82%E4%BB%B6/
username_0: @username_1 作者你好,这篇文章我看过,但是并没有写 如何去掉 最新文章 显示的图片
username_1: @username_0 如果你只想隐藏最新文章的图片的话,就要修改代码了
username_0: 感谢
Status: Issue closed
|
rancher/rancher | 545622220 | Title: 自动拉起aws EC2 失败后,不能删除该节点,手动删除EC2实例后,集群监控,项目监控工作不正常
Question:
username_0: 自动拉起aws EC2 失败后,不能删除该节点,手动删除EC2实例后,集群监控,项目监控工作不正常
---

Versions Rancher `v2.3.3` UI: `v2.3.3` |
Status: Issue closed
Answers:
username_1: Please add `rancher2` to join Rancher WeChat group if you would like to ask questions in Chinese. |
BehaviorTree/BehaviorTree.CPP | 1063538314 | Title: `Any::cast` fails for references
Question:
username_0: ### Situation
I have roughly the following situation:
* I store user-structs in the blackboard, not always small ones
* I want to read them regularly (e.g. for displaying its contents in a GUI)
* => I don't want to create copies each time
So what I want to do is roughly:
```cpp
struct Foo {...}; // some larger struct; stored an instance in blackboard with key "someKey"
const BT::Any* value = myBlackboard->getAny("someKey");
const auto& entry = value->cast<const Foo&>();
```
`any_cast` supports casting to references, so intuitively this should be working. But I can only invoke
```cpp
value->cast<Foo>();
value->cast<const Foo>();
```
### Error
This does not compile. Error (filtered)
```bash
/BehaviorTree.CPP/include/behaviortree_cpp_v3/utils/expected.hpp: In instantiation of ‘class nonstd::expected_lite::expected<const Foo&, std::__cxx11::basic_string<char> >’:
/BehaviorTree.CPP/include/behaviortree_cpp_v3/utils/safe_any.hpp:137:34: required from ‘T BT::Any::cast() const [with T = const Foo&]’
/testCode.cpp:160:83: required from here
/BehaviorTree.CPP/include/behaviortree_cpp_v3/utils/expected.hpp:1224:34: error: forming pointer to reference type ‘nonstd::expected_lite::expected<const Foo&, std::__cxx11::basic_string<char> >::value_type’ {aka ‘const Foo&’}
1224 | constexpr value_type const * operator ->() const
| ^~~~~~~~
/BehaviorTree.CPP/include/behaviortree_cpp_v3/utils/expected.hpp:1229:18: error: forming pointer to reference type ‘nonstd::expected_lite::expected<const Foo&, std::__cxx11::basic_string<char> >::value_type’ {aka ‘const Foo&’}
1229 | value_type * operator ->()
| ^~~~~~~~
In file included from /BehaviorTree.CPP/include/behaviortree_cpp_v3/utils/safe_any.hpp:14,
from /BehaviorTree.CPP/include/behaviortree_cpp_v3/basic_types.h:14,
from /BehaviorTree.CPP/include/behaviortree_cpp_v3/blackboard.h:12,
from /project/libs/p90902/components/task_management/Blackboard.h:16,
from /testCode.cpp:1:
/BehaviorTree.CPP/include/behaviortree_cpp_v3/utils/expected.hpp: In instantiation of ‘union nonstd::expected_lite::detail::storage_t<const Foo&, std::__cxx11::basic_string<char> >’:
/BehaviorTree.CPP/include/behaviortree_cpp_v3/utils/expected.hpp:1383:28: required from ‘class nonstd::expected_lite::expected<const Foo&, std::__cxx11::basic_string<char> >’
/BehaviorTree.CPP/include/behaviortree_cpp_v3/utils/safe_any.hpp:137:34: required from ‘T BT::Any::cast() const [with T = const Foo&]’
/testCode.cpp:160:83: required from here
/BehaviorTree.CPP/include/behaviortree_cpp_v3/utils/expected.hpp:400:24: error: forming pointer to reference type ‘nonstd::expected_lite::detail::storage_t<const Foo&, std::__cxx11::basic_string<char> >::value_type’ {aka ‘const Foo&’}
400 | value_type const * value_ptr() const
| ^~~~~~~~~
/BehaviorTree.CPP/include/behaviortree_cpp_v3/utils/expected.hpp:405:18: error: forming pointer to reference type ‘nonstd::expected_lite::detail::storage_t<const Foo&, std::__cxx11::basic_string<char> >::value_type’ {aka ‘const Foo&’}
405 | value_type * value_ptr()
| ^~~~~~~~~
/BehaviorTree.CPP/include/behaviortree_cpp_v3/utils/expected.hpp:421:16: error: non-static data member ‘nonstd::expected_lite::detail::storage_t<const Foo&, std::__cxx11::basic_string<char> >::m_value’ in a union may not have reference type ‘nonstd::expected_lite::detail::storage_t<const Foo&, std::__cxx11::basic_string<char> >::value_type’ {aka ‘const Foo&’}
421 | value_type m_value;
```
### Reason
I looked into the code, also the error messages are quite obvious. `Any::cast()` invokes `Any::convert()`, which has a signature `template <typename DST> nonstd::expected<DST,std::string> convert(...)`.
`DST` cannot be a reference though, does not work with `nonstd::expected`.
### Suggested Fix
[Truncated]
template <typename DST>
- nonstd::expected<DST,std::string> convert(EnableUnknownType<DST> = 0) const
+ nonstd::expected<std::decay_t<DST>,std::string> convert(EnableUnknownType<DST> = 0) const
{
return nonstd::make_unexpected( errorMsg<DST>() );
}
```
### Testing
I did a quick debugging in my IDE and can confirm that
* code compiles with the proposed fixes
* `value->cast<Foo>();` returns a copy of the original
* `value->cast<const Foo&>();` returns a const ref to the same data (so no copying here)
### How to proceed
What's your opinion here? What would you suggest?
I think the last mentioned option is a bit hacky and doesn't cover all possible cases.
I was also wondering if it might've actually been a design-choice that `Any::cast` can only return copies. |
ioBroker/ioBroker.devices | 775682164 | Title: Update stable version in repo from 0.3.13 to 0.3.15
Question:
username_0: # Think about update stable version to 0.3.15
**Version**: stable=**0.3.13** (133 days old) => latest=**0.3.15** (15 days old)
**Installs**: stable=**2854** (57.05%), latest=**1277** (25.52%), total=**5003**
Click to [edit](https://github.com/ioBroker/ioBroker.repositories/edit/master/sources-dist-stable.json#L200)<issue_closed>
Status: Issue closed |
kubernetes/kubernetes | 348820380 | Title: Inject environment variables into Persistent Volume definition
Question:
username_0: Here is my first attempt of a json file in which environment variables would be injected into a definition of a Persistent volume:
{
"apiVersion": "v1",
"kind": "PersistentVolume",
"metadata": {
"name": "statefulset-pgdata-0",
"labels": {
"name": "statefulset-pgdata"
}
},
"spec": {
"capacity": {
"storage": "$CCP_STORAGE_CAPACITY"
},
"accessModes": ["$CCP_STORAGE_MODE"],
"nfs": {
"path": "$CCP_STORAGE_PATH",
"server": "$CCP_NFS_IP"
},
"persistentVolumeReclaimPolicy": "Retain",
"envFrom": {
"configMapRef": {
"name": "aws-auth"
}
}
}
}
{
"apiVersion": "v1",
"kind": "PersistentVolume",
"metadata": {
"name": "statefulset-pgdata-1",
"labels": {
"name": "statefulset-pgdata"
}
},
"spec": {
"capacity": {
"storage": "$CCP_STORAGE_CAPACITY"
},
"accessModes": ["$CCP_STORAGE_MODE"],
"nfs": {
"path": "$CCP_STORAGE_PATH",
"server": "$CCP_NFS_IP"
},
"persistentVolumeReclaimPolicy": "Retain"
}
}
{
"apiVersion": "v1",
"kind": "PersistentVolume",
"metadata": {
"name": "statefulset-pgdata-2",
"labels": {
"name": "statefulset-pgdata"
}
},
[Truncated]
io.k8s.api.core.v1.PersistentVolumeSpec;
Also, for completeness, I am including the `ConfigMap` from which I am trying to import the above environment variables as follows:
apiVersion: v1
kind: ConfigMap
metadata:
name: aws-auth
namespace: kube-system
data:
mapRoles: |
- rolearn: arn:aws:iam::999999999/hjhjjhjhjhjhhj
username: system:node:{{something}}
groups:
- system:bootstrappers
- system:nodes
CCP_STORAGE_CAPACITY: 400M
CCP_STORAGE_MODE: ReadWriteMany
CCP_STORAGE_PATH: /mnt/nfsfileshare
CCP_NFS_IP: 10.0.0.190
Answers:
username_0: @k8s-ci-robot
Do you mean like that? I dont know what sig to choose. I am guessing since it has to do with variables then maybe Architecture sig. |
electron/electron | 722820771 | Title: Apple reject build: The experience that your app provides is not sufficiently different from a web browsing experience.
Question:
username_0: Apple rejected build with below reason.
I have used many native features like;
1- Tray
2- In-App purchase
3- Shortcut keys
4- Dock
5- Welcome screen
6- Notifications
and more
Guideline 4.2.2 - Design
The experience that your app provides is not sufficiently different from a web browsing experience.
We advise incorporating additional substantive native macOS features into your application to better differentiate it from the Safari experience.
any suggestion to resolve this issue?
Status: Issue closed
Answers:
username_1: This is for you to deal with Apple. I'd suggest filing an appeal and pointing out the list of feature you referenced here. We have no say / power over the mac app store 🤷 |
sequelize/sequelize | 1083066647 | Title: include.attributes.map is not a function
Question:
username_0: <!--
If you don't follow the issue template, your issue may be closed.
Please note this is an issue tracker, not a support forum.
For general questions, please use StackOverflow:
https://stackoverflow.com/questions/tagged/sequelize.js
-->
## Issue Creation Checklist
[x] I have read the [contribution guidelines](https://github.com/sequelize/sequelize/blob/main/CONTRIBUTING.md)
## Bug Description
await db.sequelize.query(sql, {
model: Model,
attributes: { exclude: ['attribute1'] }
});
### What do you expect to happen?
Expect Model to accept the exclude attributes
### What is actually happening?
Throwing error options.attributes.map is not a function
```
options.attributes.map is not a function
```
### Environment
- Sequelize version: 6.6.5
- Node.js version: 12.22.6
## Bug Report Checklist
<!-- Please answer the questions below. If you don't, your issue may be closed. -->
### How does this problem relate to dialects?
<!-- Choose one. -->
- [x ] I think this problem happens regardless of the dialect.
- [ ] I think this problem happens only for the following dialect(s): <!-- Put dialect(s) here -->
- [ ] I don't know, I was using PUT-YOUR-DIALECT-HERE, with connector library version XXX and database version XXX
### Would you be willing to resolve this issue by submitting a Pull Request?
<!-- Remember that first contributors are welcome! -->
- [ ] Yes, I have the time and I know how to start.
- [ x] Yes, I have the time but I don't know how to start, I would need guidance.
- [ ] No, I don't have the time, although I believe I could do it if I had the time...
- [ ] No, I don't have the time and I wouldn't even know how to start.
Answers:
username_1: I don't think `query` supports an `attributes` option, and if it did, it's likely going to be an array of attributes to include and not going to support `exclude`
username_1: Confirmed that `Sequelize#query` does not support an `attributes` options. Options are passed as-is to Dialect#Query which is the reason why using `attributes` causes a crash.
Solution is to simply not pass `attributes` to `Sequelize#query`.
This option doesn't really make sense in `query` as you're already writing the SQL yourself. What was your reason for using it? |
ProyectoIntegrador2018/e_sports_ai | 443597521 | Title: Jungle Objectives
Question:
username_0: Header|2nds
---|---
name| Jungle Objectives
about| The jungler bot would kill the jungle objective to help the team
title| Jungle Objectives
labels| enhancement
assignees|
### Description
As a jungler I will effectively create an advantage for my team by killing camps and jungle (forest) objectives
---
### Acceptance criteria
The jungle will efficiently and effectively navigate the jungle (forest) and make the appropriate objectives according to the situation
- [ ] The bot correctly navigates the jungle
- [ ] The bot makes the appropriate camp or jungle objective
- [ ] The bot shows in lane presence when required
---
### Important details
This will only apply to the jungler as the jungler needs to navigate through the jungle (forest) instead of a lane and needs to kill camps instead of creeps
---
### Priority
High
---
### ToDo
- [ ] Create an appropriate navigation method for the jungles (mode_laning_”jungle chanp”.lua)
- [ ] Create an appropriate Bayesian network to define the nest bot objective in the jungle |
neutralinojs/neutralinojs | 615321181 | Title: Publish Neutralino client library to NPM
Question:
username_0: <!-- ISSUE TEMPLATE -->
<!-- (Update "[ ]" to "[x]" to check a box) -->
**Expected Behavior**
NPM has package `neutralino-client-library` to install
**Actual Behavior**
You need to bundle package by yourself
Answers:
username_1: Hello.. Thanks for suggesting a nice idea.. We thought about this idea earlier too and decided to download the client from the Github repo instead. It is chosen due to the following reasons.
- `neutralino.js` is dynamic. Neutralino server injects some JavaScript into it dynamically.
- All users may not use npm and frontend frameworks to develop Neutralinojs apps.
Right now, the Neutralinojs client will be downloaded/updated with `neu create` and `neu update` commands. Besides, it is git ignored.
Thanks again for the idea.
Status: Issue closed
|
cityofaustin/atd-data-tech | 824947964 | Title: TIA Mod Project Team Feedback 3/8/21
Question:
username_0: - [ ] Update edit case form to allow to connect to TDS cases from TIA Case Details page
- [ ] Remove Case Notes field from the portal application.
- [ ] A new feature to allow users to change the assigned reviewer
- [ ] Case Management Submissions Details Pages: Change green 'Yes' box to a check box
- [ ] On Case Management details, add Review Start Date, Review End Date.
- [ ] Move Case Management button to top left of button section
- We spoke at length about whether we should capture all segments related to a single mitigation record, or if the current design of assigning the record to one segment will work. It was decided to leave as-is. John asked that we get Jaime's team to do a GIS review at some point.
Answers:
username_0: - [ ] Update edit case form to allow to connect to TDS cases from TIA Case Details page
- [ ] Remove Case Notes field from the portal application.
- [ ] A new feature to allow users to change the assigned reviewer
- [ ] Case Management Submissions Details Pages: Change green 'Yes' box to a check box
- [ ] On Case Management details, add Review Start Date, Review End Date.
- [ ] Move Case Management button to top left of button section
- We spoke at length about whether we should capture all segments related to a single mitigation record, or if the current design of assigning the record to one segment will work. It was decided to leave as-is. John asked that we get Jaime's team to do a GIS review at some point.
username_1: I have completed all listed items. Reviewers are now able to be added, updated, and changed for the entire case, or individually for each Submission. This will allow us to create Task Pages for individual Reviewers so they may see all Cases and by extension, all Submissions they have been assigned or worked on. Form buttons on each of the Details pages have been built for this. Did a sweep over pages and forms where user information could be helpful and provided help text where needed.
Additionally multiple Development Cases can be connected to a TIA case either on the TIA Requests table or on its Details page with the Connect Cases button.
Both Reviewers and Connected Cases and also displayed on the Case Details page
complete, ready for review
username_0: All of this looks great. Closing.
Status: Issue closed
|
lovell/sharp | 345182591 | Title: Error 404 download link
Question:
username_0: Hi, If I type in my server console npm install sharp I getting error:
`info sharp Downloading https://github.com/username_2/sharp-libvips/releases/download/v8.6.1/libvips-8.6.1-freebsd-x64.tar.gz`
If i type in browser link `info sharp Downloading https://github.com/username_2/sharp-libvips/releases/download/v8.6.1/libvips-8.6.1-freebsd-x64.tar.gz` I get too error 404, but if I change freebsd to linux in link I can download sharp from browser
Linux Ubuntu server
node -v v6.12.0
npm -v 5.8.0
Answers:
username_1: You can update your node version to latest LTS version? (v8.11.3 at this time)
username_0: I have now v8.9.1 but my server work on freeBSD and I dont know how istall sharp. On localserver i work on windowsx64 and eveything work good
username_1: I dont used at freebsd but did you try to install libvips manually?
username_0: I made npm install -g libvips and this work, then I try install sharp but the same get the same error
username_2: Please see http://sharp.pixelplumbing.com/en/stable/install/#freebsd
The installation process could be improved to detect FreeBSD and provide a link to the relevant documentation so I'll leave this issue open to track that work (note to self/implementer: add a conditional statement in [here](https://github.com/username_2/sharp/blob/master/install/libvips.js#L30) somewhere).
username_2: Error messaging for FreeBSD improved in commit bd9f238
Status: Issue closed
username_2: v0.20.6 now available with improved error messaging for this scenario. |
jaredpalmer/formik | 448773334 | Title: Namespace React has no exported member ProviderExoticComponent
Question:
username_0: ## 🐛 Bug report
Formik version used : 1.5.7
### Current Behavior
<!-- If applicable, add screenshots to help explain your problem. -->
When I launch `./gradlew test` on my project with Formik 1.4.7, I have these two errors :
`ERROR in [...]/node_modules/formik/dist/connect.d.ts(3,44):
TS2694: Namespace 'React' has no exported member 'ProviderExoticComponent'.`
`ERROR in [...]/node_modules/formik/dist/connect.d.ts(3,132):
TS2694: Namespace 'React' has no exported member 'ExoticComponent'.`
### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
When I downgraded my version to Formik 1.5.4, everything works like expected.
### Suggested solution(s)
I haven't any idea about a solution.
<!-- How could we solve this bug? What changes would need to made to Formik? -->
### Additional context
<!-- Add any other context about the problem here. -->
### Your environment
| Software | Version(s) |
| ---------------- | ---------- |
| Formik | 1.5.7
| React | 16.4.2
| TypeScript | 3.0.1
| Browser | Mozilla Firefox Quantum 67.0
| npm/Yarn | npm 3.5.2
| Operating System | Ubuntu 18.04.2
Thanks for your attention
Answers:
username_1: ExoticComponent is used in a number of newer React scenarios.
`React.Fragment`: `<></>`.
- If newer Formik 1.x versions use a Fragment, and you have a version of React which hasn't defined Fragments yet, this could be why. Try updating React.
- It's possible that your version of TypeScript doesn't yet understand this format of JSX. Try updating TypeScript.
`React.Provider / React.Consumer`:
React defines Context Provider + Consumer with ExoticComponent, so if any code uses React Context, that could be the issue. I would check if that version of Formik is using Context.
And of course, make sure you have a good package.lock or yarn.lock in version control so you can roll back any of these changes for testing purposes.
username_2: @username_0 closing the issue if no information/response will be provided
username_3: I think we should start asking for the @types/react version in the issue template
username_2: Agree
username_0: Thank you for your answers, I will try it
username_4: Closing due to long inactivity
Status: Issue closed
|
mathertel/OneButton | 385292761 | Title: function to retrieve number of buttonclicks and short or long press?
Question:
username_0: hello,
how about a function to retrieve the number of buttonclicks and/or short or long press?
e.g.
```
void loop() {
button1.tick();
button2.tick();
if(button1.click()) { // returns 0 if no click - or number of clicks or long press
Serial.println(button1.click() ); // 1 for single click , 2 for doubleClick, 3 for long press
}
// or
if(button2.click() == 1) {do something ;}
if(button2.click() == 2) {do other special;}
if(button2.click() == 3) {do something else;}
delay(10);
} // loop
```
Answers:
username_1: see #50
Status: Issue closed
|
flutter/flutter | 384183619 | Title: Can't get realse apk
Question:
username_0: Output: D:\projects\flutter_work_schedule\workschedule\build\autostart\intermediates\res\merged\release\values-v28\values-v28.xml:7: error: resource android:attr/dialogCornerRadius not found.
D:\projects\flutter_work_schedule\workschedule\build\autostart\intermediates\res\merged\release\values-v28\values-v28.xml:11: error: resource android:attr/dialogCornerRadius not found.
D:\projects\flutter_work_schedule\workschedule\build\autostart\intermediates\res\merged\release\values\values.xml:956: error: resource android:attr/fontVariationSettings not found.
D:\projects\flutter_work_schedule\workschedule\build\autostart\intermediates\res\merged\release\values\values.xml:957: error: resource android:attr/ttcIndex not found.
error: failed linking references.
Total output was in the flutter_01.log
## Logs
<!--
Include the full logs of the commands you are running between the lines
with the backticks below. If you are running any "flutter" commands,
please include the output of running them with "--verbose"; for example,
the output of running "flutter --verbose create foo".
-->
```
```
<!-- If possible, paste the output of running `flutter doctor -v` here. -->
## flutter doctor -v
[√] Flutter (Channel beta, v0.11.9, on Microsoft Windows [Version 10.0.17134.407], locale en-US)
• Flutter version 0.11.9 at D:\flutter\flutter_beta\flutter
• Framework revision d48e6e433c (5 days ago), 2018-11-20 22:05:23 -0500
• Engine revision 5c8147450d
• Dart version 2.1.0 (build 2.1.0-dev.9.4 f9ebf21297)
[√] Android toolchain - develop for Android devices (Android SDK 28.0.3)
• Android SDK at C:\Users\myaxj\AppData\Local\Android\sdk
• Android NDK location not configured (optional; useful for native profiling support)
• Platform android-28, build-tools 28.0.3
• Java binary at: C:\Program Files\Android\Android Studio\jre\bin\java
• Java version OpenJDK Runtime Environment (build 1.8.0_152-release-1136-b06)
• All Android licenses accepted.
[√] Android Studio (version 3.2)
• Android Studio at C:\Program Files\Android\Android Studio
• Flutter plugin version 30.0.1
• Dart plugin version 181.5656
• Java version OpenJDK Runtime Environment (build 1.8.0_152-release-1136-b06)
[√] VS Code (version 1.29.1)
• VS Code at C:\Users\myaxj\AppData\Local\Programs\Microsoft VS Code
• Flutter extension version 2.20.0
[√] Connected device (1 available)
• ONEPLUS A3010 • 3ff8b9c0 • android-arm64 • Android 8.0.0 (API 26)
• No issues found!
```
```
Status: Issue closed
Answers:
username_1: Merging with #15646 |
dart-lang/sdk | 275101925 | Title: Make JS pseudofunction take a type argument
Question:
username_0: ```dart
external T JS<T>(...);
```
Type inference in Dart 2.0 can insert the type parameter, making some of the type annotations unnecessary.
```dart
String name = JS('', 'JSON.stringify(#)', blob);
-->
String name = JS<String>('', 'JSON.stringify(#)', blob);
```
This is equivalent to the following on Dart 1.x:
```dart
String name = JS('returns:String|Null', 'JSON.stringify(#)', blob);
```
However, the **returns:** behaviour may still be used to specify non-null types:
```dart
String name = JS('returns:String', 'JSON.stringify(#)', blob);
```
- [ ] Add definition.
- [ ] Start using types in addition to **returns:** to avoid implicit downcasts.
- [ ] Refine SSA node type to intersection of ad-hoc **returns**: specification and type argument.
- [ ] Make dart2js emit an error or warning if type argument is inconsistent with **returns:** specification.
Answers:
username_0: Closing since DDC has its own metalanguage in the first position
Status: Issue closed
|
briis/unifiprotect | 824315679 | Title: After the 2021.3.x the 'live' setting makes the camera go blank
Question:
username_0: I have both unifi and nest camera integration activated.
I only have 1 nest, the video doorbell, so I wouldn't know if this applies to other nest cameras.
I have the camera showing like this:
```
type: picture-entity
entity: camera.indgang
camera_image: camera.indgang
show_name: false
show_state: false
```
If I add
`camera_view: live`
It goes black, and doesn't show anything.
Answers:
username_0: I'm not sure if this relates to the problem, looking in the log the only one I find that could be close to anything with this is this one:
Logger: homeassistant.components.stream.worker
Source: components/stream/worker.py:130
Integration: Stream (documentation, issues)
First occurred: 7. marts 2021 17.30.45 (212 occurrences)
Last logged: 09.29.31
Error opening stream: rtsps://stream-eu1-charlie.dropcam.com:443/sdm_live_stream/CiQA2vuxr0SUhyOWYql5O57CGtwWIckN52XZRuos3[...]
username_1: There have been some issues reported with the stream component and live view, but I currently have no issues with the Unifi Protect Integration.
@username_0 Is the camera you refer to above a Unifi Camera or the Nest camera? When I look at *charlie.dropcam.com:443/sdm_live_stream/CiQA2vuxr0SUhyOWYql5O57CGtwWIckN52XZRuos3[...]* that you posted above, I am positive this has nothing to do with Unifi Protect.
username_0: Darned, sorry Bjarne, I was looking in your github to see if you had solved it, and I was meaning to post it in the nest integration github, sorry, just ignore me :-)
Status: Issue closed
|
flutter/flutter | 151913777 | Title: Color constants in Atom should be color-coded
Question:
username_0: I saw @apwilson has some plugin in Atom that colorises any hex string that it sees. We should totally do that for any Color literal in Flutter code.
cc @devoncarew
Answers:
username_1: This issue was moved to flutter/flutter-intellij#460
username_1: Moved to IntelliJ tracker.
Status: Issue closed
|
evert-arias/EasyButton | 906573270 | Title: Change the parameters for .onSequence(presses, timeout, onSequenceMatched);
Question:
username_0: Hi is it possible to redefine the parameters for the presses and timeout in this function, based on either user input. Or can they only he set up when first called. A function the change the parameters without redefining the call back would be great. |
eXpandFramework/eXpand | 445080474 | Title: HideToolBar as standalone package
Question:
username_0: The functionality now lives in system modules in HideToolBarController spawn across the agnostic, win, web assemblies.
release a one assembly standalone package and replace the controllers.
Answers:
username_1: eXpand.lab release [19.1.301.1](https://github.com/eXpandFramework/eXpand.lab/releases/19.1.301.1) includes commit [Install HideToolBar agnostic package #402](https://github.com/eXpandFramework/eXpand.lab/commit/eb1b81f2a42fe748ca824e4e3a7c0f6a4f7596a1) that relate to this task. Please test if it addresses the problem. If you use nuget add our LAB [NugetServer](https://xpandnugetserver.azurewebsites.net/nuget) as a nuget package source in VS.
Thanks a lot for your contribution.
Status: Issue closed
username_1: Closing issue for age. Feel free to reopen it at any time.
.Thank you for your contribution. |
dabeaz/ply | 104582385 | Title: Python 3.5 compatibility
Question:
username_0: Several testlex.py failures under Python 3.5:
```
% python3.5 testlex.py
....F/usr/lib/python3.5/unittest/case.py:625: ResourceWarning: unclosed file <_io.BufferedReader name=4>
outcome.errors.clear()
FFF.............F.F..................
======================================================================
FAIL: test_lex_opt_alias (__main__.LexBuildOptionTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "testlex.py", line 541, in test_lex_opt_alias
self.assert_(pymodule_out_exists("aliastab.pyo"))
AssertionError: False is not true
======================================================================
FAIL: test_lex_optimize (__main__.LexBuildOptionTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "testlex.py", line 390, in test_lex_optimize
self.assert_(pymodule_out_exists("lextab.pyo"))
AssertionError: False is not true
======================================================================
FAIL: test_lex_optimize2 (__main__.LexBuildOptionTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "testlex.py", line 443, in test_lex_optimize2
self.assert_(pymodule_out_exists("opt2tab.pyo"))
AssertionError: False is not true
======================================================================
FAIL: test_lex_optimize3 (__main__.LexBuildOptionTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "testlex.py", line 494, in test_lex_optimize3
self.assert_(pymodule_out_exists("lexdir/sub/calctab.pyo"))
AssertionError: False is not true
======================================================================
FAIL: test_lex_re1 (__main__.LexErrorWarningTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "testlex.py", line 161, in test_lex_re1
contains=True))
AssertionError: False is not true
======================================================================
FAIL: test_lex_re3 (__main__.LexErrorWarningTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "testlex.py", line 179, in test_lex_re3
contains=True))
AssertionError: False is not true
----------------------------------------------------------------------
Ran 42 tests in 1.595s
FAILED (failures=6)
```
Answers:
username_0: https://github.com/username_1/ply/pull/78
Adjusts for PEP 488 and different re error messages in Python 3.5
Status: Issue closed
|
OSGeo/gdal | 968566100 | Title: [Security Issue] Stack-based Buffer Overflow
Question:
username_0: @username_1
The `OGRFMEDriver` suffers from a stack based overflow.
Would you consider reviewing the private report and PoC here: https://huntr.dev/bounties/20a59e73-c419-4732-bcd4-7e0d27e95726/
Only your GitHub account is authorised to view it. Let me know if you encounter any issues accessing the report. Thanks!
Answers:
username_1: The FME driver is queued for removal in GDAL 3.5 (https://github.com/OSGeo/gdal/blob/3348482695c26698def5bb3d2692900b300a23f6/gdal/NEWS#L15) and is not built in any standard binary build, so I don't think we'll take action on this
Status: Issue closed
|
deps-rs/deps.rs | 1118303916 | Title: allow site domain name as synonym/specification of hosting service
Question:
username_0: Right now, the badge url requires `../github/..`. Would like it to take `../github.com/..` as well. This is especially important for `gitlab`, where for example `gitlab.freedesktop.org` is not currently supported as far as I know.
Answers:
username_1: semi related to https://github.com/deps-rs/deps.rs/issues/84 |
girder/girder_worker | 291774668 | Title: Start tracking release notes
Question:
username_0: Create a CHANGELOG, Girder has an [example](https://github.com/girder/girder/blob/master/CHANGELOG.rst) to follow.
Start enforcing PRs to update the CHANGELOG.
Update the CONTRIBUTING to mention this as part of the process. |
OpenEmu/OpenEmu | 361996683 | Title: Resize screen after 2.0.7 update
Question:
username_0: btw thank you guys so much for fixing this app for macos 10.14 love the update after waiting all summer I really appreciate the hard work put into fixing the app. Anyways I can't resize the window when playing games since 2.0.7. And if I try minimizing the window the app stops working and I have to force close the app to restart it.
Answers:
username_1: Going to need better details than that. I can't replicate the problem, so please provide exact steps.
username_0: I can resize the window after I put it into full screen and take it out of full screen by chance do you use the app better touch tool?
username_1: I do not use BetterTouchTool. I'm not sure the issue you are experiencing is related to the latest update either.
username_0: launch any game, notice the black boarder is pure black and I can't move the window at all
<img width="962" alt="screen shot 2018-09-19 at 9 50 57 pm" src="https://user-images.githubusercontent.com/4202657/45793110-2b08b380-bc56-11e8-90bd-44e69947add6.png">
after putting the screen into full screen and taking it out of full screen take a look at the board its now gray and I can move the window around
<img width="956" alt="screen shot 2018-09-19 at 9 52 31 pm" src="https://user-images.githubusercontent.com/4202657/45793153-568b9e00-bc56-11e8-9c45-f756c544c339.png">
Status: Issue closed
username_1: You need to learn to be more specific and not generalize your problems. I cannot help you if the N64 core plugin is crashing - it's due to compatibility issues with the core emulator and GPU plugin.
First:
* You need to have the Mupen64Plus 2.5.5 core plugin
* If you for some reason edited the version number locally on an older plugin, you need to delete it and redownload the latest from us in-app (OpenEmu > Preferences > Cores).
To work around your loading/crashing issue with GoldenEye:
* In the app, go to OpenEmu > Preferences > Library, click "Reset warnings"
* Open your game and click "No" when prompted to "Continue where you left off"
username_0: This has nothing to do with it crashing, and all my cores are up to date. All I am talking about is resizing the window, it happens with all the cores I launch I have to put it into full screen mode then take it out of full screen to move and resize the screen.
username_1: Provide more details. Find precise conditions this happens under with exact steps to replicate, otherwise we cannot help you. I don't know how to be more clear than that. |
DestinyItemManager/DIM | 650907811 | Title: LO: No builds found when locking or excluding (mobile)
Question:
username_0: From Discord:

Transcription:
Loadout optimizer keeps showing 0 results on mobile anytime I lock or exclude any item on mobile. Is this a known issue? Didnt see anything pinned for it in this or known issues channels

Happens with any item, also in the exclude item slot. I used the ghost in above example to limit results with different ghosts. But it displays 0 results
Hmm I just got it to work for the first time ever! Solution: have to tap in the dark grey window area next to locked item -> then tap Filters to hide the filter menu. Only then does it display and allow scrolling
Answers:
username_0: seems the webworker is waiting for you to finalize your selection on the filter before it processes anything, which it should do. So maybe make that more obvious somehow.
username_1: Ghosts aren't included anymore so I need to remove those from all the pickers.
As a side note I can't reproduce this on Safari for iPhone. It did almost catch me out as I selected a locked item with a stat lower than the minimum total stat setting. We should ignore that setting if an item is locked?
username_1: Figured it out with the help of @sundevour. The issue is when an item is locked the filter section becomes bigger than the screen and subsequently the user gets stuck in a scrollable css element. I will fix the CSS properly this week but I have added a little note on mobile to tell users to collapse filters for now.
Status: Issue closed
|
Kamiran79/bear-watcher | 626883332 | Title: Setup Files #1
Question:
username_0: User Story
As a dev, there should be all the needed files in the repo
AC
WHEN the repo is pulled down
THEN there should be all the needed files
AND dependancies should be included
Dev Notes
Add index.html, main.js, main.css
Pull in bootstrap
Background color should change, console should log, h1 should be there with page title |
shirtsgroup/physical_validation | 608651836 | Title: Enable LGTM and Codecov
Question:
username_0: ## Description
We should enable the [LGTM](https://lgtm.com/) and [Codecov](https://codecov.io/) services on this repository to gain extra lint checking, and better reporting of the coverage of our unit tests.
Likely a good first issue for @ocmadin, @username_1 or @username_2.
Answers:
username_1: Sounds good. I've only used Codecov before but I will look into LGTM.
username_2: Looking into lgtm today, will see if I can add any badges generated by lgtm to the README
username_2: Got a codecov page and a lgtm page going for physical validation:
https://lgtm.com/projects/g/shirtsgroup/physical_validation/ci/
https://codecov.io/gh/shirtsgroup/physical_validation
Adding badges and relevant additions to our continuous_integration.yaml files in this PR #72
Status: Issue closed
|
DiscordDungeons/Bugs | 287340316 | Title: 'buyfield' command not working
Question:
username_0: 
**Be sure to read [here](https://github.com/DiscordDungeons/Bugs/blob/master/FAQ.md) first.**
**Please describe your issue below.**
I'm not sure what to do here... I saw on the official server that there was a new command called 'buyfield' for crops and such. I tried buying a field in Copperfall but it gave me the error message I attached.
**If an error is provided, please copy the block into the space below.**
```js
< TypeError: message.channel.hasPerm is not a function
at Buyfield.Execute (/home/mackan/RPGBot/src/Commands/Commands/Locations/Buyfield.js:127:23)
at <anonymous> >
```
**If you have 1 screenshot or more for context of the issue, please insert into the brackets below.**
*You may use github's standard way of presenting images, or use an image uploader (like imgur).*

Status: Issue closed
Answers:
username_1: Fixed
username_0: Thank you so much! |
StefanScherer/dockerfiles-windows | 400905453 | Title: Traefik doesn't detect swarm services on overlay network windows
Question:
username_0: Hi Stefan,
First of all thank you for all your work in the windows docker community.
Background:
I've used your traefik-windows dockerfile successfully with docker compose on Windows Server 2016 and am able to view the dashboard of online services as well as use the whoami service.
However I plan on creating many more swarm services in the future and would like to be able to spin up new swarm services via CLI (docker service create ...) instead of docker-compose.
Problem:
I was able to edit your docker-compose file to work with an external overlay network and it works exactly the same. However when I attach new swarm services to the overlay network they aren't available through traefik, nor can I run a traefik service (the url of the dashboard just hangs indefinitely).
Working docker-compose.yaml with overlay network
```
NETWORK ID NAME DRIVER SCOPE
743575wiqb4p traefik_net overlay swarm
```
**traefik.toml**
```
################################################################
# api configuration backend
################################################################
[api]
entryPoint = "traefik"
dashboard = true
debug = true
################################################################
# Docker configuration backend
################################################################
[docker]
domain = "docker.localhost"
endpoint = "tcp://172.30.112.1:2375"
network = "net"
watch = true
# Use the IP address from the binded port instead of the inner network one
usebindportip = true
# increase docker api version to 1.24 with swarm-mode :-)
swarmmode = true
```
**docker-compose.yaml**
```
version: '3.0'
services:
proxy:
image: stefanscherer/traefik-windows
command: --api --docker --logLevel=DEBUG --docker.endpoint=tcp://172.30.112.1:2375
ports:
- "80:80"
- "8080:8080"
volumes:
- E:/docker/traefik:C:/etc/traefik
networks:
- default
whoami:
image: stefanscherer/whoami
labels:
- "traefik.enable=true"
- "traefik.backend=whoami"
[Truncated]
```--publish mode=host,target=8080```
```--label traefik.docker.network=traefik```
```--label traefik.port=80 (and 8080)```
``` --endpoint-mode=dnsrr```
with no luck.
**Expected result:**
The whoami2 service should show up on the traefik dashboard and I should be able to access it from /whoami2
**Actual result:**


Any help would be appreciated.
Thanks,
Brayden
Answers:
username_1: Hi @username_0 Thanks for trying Traefik on Windows.
Currently I'm trying Traefik on Windows Server 2019 / Windows 10 1809, but I had not yet time to run it in swarm mode. I first updated the docker-compose example in this repo to get rid of the TLS certs by using the named pipe.
My next steps are to try swarm mode as described in https://dockerswarm.rocks, but on a Windows Server 2019.
username_0: Hey @username_1 ,
No worries. Thanks for the link it looks promising!
username_2: This dont working in Stack file
```
type: npipe
source: \\.\pipe\docker_engine
target: \\.\pipe\docker_engine
```
username_1: @username_2 This will work in the next release of Docker EE as the PR for it got merged. I used the nightly builds to demo it in February at a conference.
username_3: @username_1 What version of Docker EE supports this?
username_1: @username_3 it‘s the 19.03 release, there are release candidates available and the final release will be in the near future.
username_3: Thanks for the info @username_1
Is running the 19.03-ee the only way to get Traefik to run on a Windows Swarm cluster? (I am currently running 18.09.7 on Windows Server 2019)
I have software that is currently working using linux containers using Traefik as the loadbalancer. I am trying to figure out the best way (most production ready) to get this working on a windows based swarm cluster. It looks like the LCOW stuff is still pretty experimental so I thought I would try the route of building windows containers for my software.
username_4: Hi folks,
Again awesome with with the docker on windows.
I still can't get the traefik service to run in a docker swarm. I'm running the EE 19.03.1 version and using the below i still get 'services.traefik.volumes.0 must be a string'
- type: npipe
source: \\.\pipe\docker_engine
target: \\.\pipe\docker_engine
Status: Issue closed
username_1: Closing this, nothing I can do in this repo. |
SeedCompany/cord-api-v3 | 701449858 | Title: Remove status from UpdateBudget
Question:
username_0: We don't want users to be able to change status directly. Only via status changes to project should this be updated.
Answers:
username_1: Okay, so it seems like we can't remove the `status` field from `UpdateBudget` because the `update()` method of the Budget Service needs to be able to use this when called internally. So instead we need to modify the `budget` field of `UpdateBudgetInput` so that it excludes the `status` field from `UpdateBudget`. Is this accurate?
username_0: Sure we can just drop the `@Field` decorator to make it private.
Status: Issue closed
|
SenorPez/project-cars-replay-enhancer | 193281232 | Title: Request for features/enhancements display driver pitting
Question:
username_0: Is it possible to show when a driver is doing a pit?

Answers:
username_1: Coming soon! The logic from the old infrastructure shouldn't be too hard to port into the current version.
Shows a red "PIT" indicator when the car enters the pits.
Shows a running clock while the car is stopped in the pits.
Shows a stopped clock while the car is leaving the pits.
See an example here:
https://youtu.be/AjCzBwsEtLs?t=457
username_0: That's looking really nice...we want it :)
Status: Issue closed
|
psychopy/psychojs | 619204876 | Title: Sync XMLHTTPRequests deprecated by chrome
Question:
username_0: Reference:
[https://www.chromestatus.com/feature/4664843055398912](https://www.chromestatus.com/feature/4664843055398912)
In Chrome, if users close the experiment before it finishes, the user will get an error of type "sync XHR in page dismissal".
Basically the server doesn't get the DELETE request from the client, although I don't know what effect this has - not closing an experiment properly.
https://github.com/psychopy/psychojs/blob/092829f325a70791685bf648c538ea0740f874e1/js/core/ServerManager.js#L193-L201
Answers:
username_1: This is most vexing, but alas, that was always going to happen.
Thank you for the detailed comment! I shall look into it this week.
Best wishes,
Alain
username_1: I am testing a solution, using the Beacon API. So far so good.
I would have preferred fetch/keepalive but there is a CORS bug in Chrome. Nothing is ever easy :-)
username_0: I think Beacon API is a good solution too as it's chrome's migration recommendation.
I just wonder whether it's 100% reliable because I saw this [one article](https://volument.com/blog/sendbeacon-is-broken) saying it's not:
According to the article, "about 30% of browsers that claim to support the beacon API failed to deliver the data to our servers when the page was closed, which is the whole purpose of the sendBeacon call. Linux fails 94.1% of the time".
But again that's just one article and there's no alternative (that i know of).
And despite the API saying the network request is still sent after the page is navigated away or the tab is closed, I cannot find documentation about how it behaves if the browser is closed altogether; in that case I don't know what it could work since the process is terminated.
I think the foolproof way is to use Beacon API + making the server tolerate not closing the experiment properly (maybe it already does? I haven't seen the code for pavlovia server) like closing it automatically after a timeout. Moreover I don't know the purpose of this network request so I don't know the effect of not receiving it. But i doubt it would affect data collection? :-)
Status: Issue closed
|
geneontology/go-ontology | 295219946 | Title: Equivalent Class Check
Question:
username_0: We need a check for equivalent classes in the ontology. This should fail the Travis checks when they are detected. We should double check to see if we have any currently and fix them.
Answers:
username_1: I thought we had something like this already? https://github.com/geneontology/go-ontology/blob/master/src/sparql/equivalent-classes-violation.sparql? Is this not working? Can you give an example of which classes are equivalent?
username_2: probably equivalent to https://github.com/geneontology/go-ontology/issues/13308
username_1: @username_2 Thanks for detecting equivalent ticket violations!
Does that sparql query I linked above not work? Do you have an example of an existing violation that we're not catching now?
username_2: Hi @username_1
In the current master, GO:0008381 and GO:0022839 have the same equivalence axiom.
I did a pull request to fix it but I havent merged it yet (https://github.com/geneontology/go-ontology/pull/15056)
So you can test your script with master.
Thanks, Pascale
username_1: https://github.com/geneontology/go-ontology/pull/15060
I think this should do it?
username_1: failing of this travis build proves this works: https://travis-ci.org/geneontology/go-ontology/builds/338712020?utm_source=email&utm_medium=notification
username_2: I cannot tell if it's fixed: it does pick up the error but still the build has worked; is this what we expect ?

Thanks, Pascale
username_1: Yeah this is fixed. I just updated my branch with your fixed master, and my check passes. Rerunning travis now.
Status: Issue closed
|
goldfire/howler.js | 299759231 | Title: Safari slow on loading audio files
Question:
username_0: Hello, we've just moved from a custom built solution to howler.js because we had encountered too many problems with handling audio.
Everything seems to be working fine and we fixed our issues, but we noticed that safari and iOS browsers are a lot slower when it comes to loading and playing the audio.
Here's our integration using vue.js
```
created() {
Vue.playerBus.$on(EVENT_LOAD_TRACK, beat => {
this.audioLoading = true
// reset event sent every time we load a new track
this.eventSent = false
this.currentAudioSecondsPlayed = 0
this.currentPosition = 0
// if we were already playling audio unload howl
if (this.howl) this.howl.unload()
const self = this
// create new howl object
this.howl = new Howl({
src: [`https://clickandrap.com/s3/beat_play/${beat.id}/`],
html5: true,
onplay() {
self.audioLoading = false
self.trackTotalDuration = this._duration
// Start upating the progress of the track.
requestAnimationFrame(self.step.bind(self))
}
})
// play the sound
this.howl.play()
})
Vue.playerBus.$on(EVENT_AUDIO_TOGGLE, audioPaused => {
if (audioPaused === true) this.howl.pause()
else this.howl.play()
})
}
```
Website (dev server): https://giacomogalletti.com
What can we do to improve the experience on safari and iOS browsers?
Answers:
username_1: same here
Status: Issue closed
username_2: Closing due to inactivity. If this is still an issue, reply with more info and I'll reopen.
username_3: I tried it with Ionic, works perfect while testing on chrome but as soon as I build for android and run it on an actual device the files aren't played instantly, it takes about 10 seconds for the song to play depending on bandwidth
username_4: 求解决,谢谢了 |
Ralith/openxrs | 978085027 | Title: Dynamically loading a runtime does not work per spec
Question:
username_0: The `Entry::load_from` function seems to be loading function pointers in a non spec compliant way.
Looking at this code:
https://github.com/username_1/openxrs/blob/master/openxr/src/entry.rs#L67
Take a look at the documentation for runtime loading:
https://github.com/KhronosGroup/OpenXR-SDK-Source/blob/master/specification/loader/runtime.adoc#loaderruntime-interface-negotiation
It looks to me like an openxr implementation should only export a single function `XrNegotiateRuntimeRequest ` and then this function will provide a function pointer to load the remaining symbols.
I ran into this when trying to manually load Steam's OpenXR implementation. Running dumpbin on `vrclient_x64.dll` reveals that `xrNegotiateLoaderRuntimeInterface` is the only exported OpenXR function.
Answers:
username_1: (emphasis added)
Status: Issue closed
|
EBWiki/EBWiki | 527202119 | Title: ActiveRecord::StatementInvalid: [safely] PG::UndefinedColumn: ERROR: column visits.visit_token does not exist
LINE 1: SELECT "visits".* FROM "visits" WHERE "visits"."visit_token...
^
: SELECT "visits".* FR
Question:
username_0: View details in Rollbar: [https://rollbar.com/username_0/EBWiki/items/2234/](https://rollbar.com/username_0/EBWiki/items/2234/)
```
ActionController::InvalidAuthenticityToken: ActionController::InvalidAuthenticityToken
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_controller/metal/request_forgery_protection.rb", line 195, in handle_unverified_request
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_controller/metal/request_forgery_protection.rb", line 223, in handle_unverified_request
File "/app/vendor/bundle/ruby/2.5.0/gems/devise-4.7.1/lib/devise/controllers/helpers.rb", line 255, in handle_unverified_request
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_controller/metal/request_forgery_protection.rb", line 218, in verify_authenticity_token
File "/app/vendor/bundle/ruby/2.5.0/gems/activesupport-5.0.7.2/lib/active_support/callbacks.rb", line 382, in block in make_lambda
File "/app/vendor/bundle/ruby/2.5.0/gems/activesupport-5.0.7.2/lib/active_support/callbacks.rb", line 169, in block (2 levels) in halting
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/abstract_controller/callbacks.rb", line 12, in block (2 levels) in <module:Callbacks>
File "/app/vendor/bundle/ruby/2.5.0/gems/activesupport-5.0.7.2/lib/active_support/callbacks.rb", line 170, in block in halting
File "/app/vendor/bundle/ruby/2.5.0/gems/activesupport-5.0.7.2/lib/active_support/callbacks.rb", line 454, in block in call
File "/app/vendor/bundle/ruby/2.5.0/gems/activesupport-5.0.7.2/lib/active_support/callbacks.rb", line 454, in each
File "/app/vendor/bundle/ruby/2.5.0/gems/activesupport-5.0.7.2/lib/active_support/callbacks.rb", line 454, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/activesupport-5.0.7.2/lib/active_support/callbacks.rb", line 101, in __run_callbacks__
File "/app/vendor/bundle/ruby/2.5.0/gems/activesupport-5.0.7.2/lib/active_support/callbacks.rb", line 750, in _run_process_action_callbacks
File "/app/vendor/bundle/ruby/2.5.0/gems/activesupport-5.0.7.2/lib/active_support/callbacks.rb", line 90, in run_callbacks
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/abstract_controller/callbacks.rb", line 19, in process_action
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_controller/metal/rescue.rb", line 20, in process_action
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_controller/metal/instrumentation.rb", line 32, in block in process_action
File "/app/vendor/bundle/ruby/2.5.0/gems/activesupport-5.0.7.2/lib/active_support/notifications.rb", line 164, in block in instrument
File "/app/vendor/bundle/ruby/2.5.0/gems/activesupport-5.0.7.2/lib/active_support/notifications/instrumenter.rb", line 21, in instrument
File "/app/vendor/bundle/ruby/2.5.0/gems/activesupport-5.0.7.2/lib/active_support/notifications.rb", line 164, in instrument
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_controller/metal/instrumentation.rb", line 30, in process_action
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_controller/metal/params_wrapper.rb", line 248, in process_action
File "/app/vendor/bundle/ruby/2.5.0/gems/searchkick-3.1.3/lib/searchkick/logging.rb", line 209, in process_action
File "/app/vendor/bundle/ruby/2.5.0/gems/activerecord-5.0.7.2/lib/active_record/railties/controller_runtime.rb", line 18, in process_action
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/abstract_controller/base.rb", line 126, in process
File "/app/vendor/bundle/ruby/2.5.0/gems/actionview-5.0.7.2/lib/action_view/rendering.rb", line 30, in process
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_controller/metal.rb", line 190, in dispatch
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_controller/metal.rb", line 262, in dispatch
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_dispatch/routing/route_set.rb", line 50, in dispatch
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_dispatch/routing/route_set.rb", line 32, in serve
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_dispatch/routing/mapper.rb", line 16, in block in <class:Constraints>
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_dispatch/routing/mapper.rb", line 46, in serve
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_dispatch/journey/router.rb", line 39, in block in serve
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_dispatch/journey/router.rb", line 26, in each
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_dispatch/journey/router.rb", line 26, in serve
File "/app/vendor/bundle/ruby/2.5.0/gems/actionpack-5.0.7.2/lib/action_dispatch/routing/route_set.rb", line 727, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/newrelic_rpm-5.2.0.345/lib/new_relic/agent/instrumentation/middleware_tracing.rb", line 92, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/rack-pjax-1.1.0/lib/rack/pjax.rb", line 12, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/newrelic_rpm-5.2.0.345/lib/new_relic/agent/instrumentation/middleware_tracing.rb", line 92, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/newrelic_rpm-5.2.0.345/lib/new_relic/rack/agent_hooks.rb", line 30, in traced_call
File "/app/vendor/bundle/ruby/2.5.0/gems/newrelic_rpm-5.2.0.345/lib/new_relic/agent/instrumentation/middleware_tracing.rb", line 92, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/newrelic_rpm-5.2.0.345/lib/new_relic/rack/browser_monitoring.rb", line 32, in traced_call
File "/app/vendor/bundle/ruby/2.5.0/gems/newrelic_rpm-5.2.0.345/lib/new_relic/agent/instrumentation/middleware_tracing.rb", line 92, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/bullet-5.9.0/lib/bullet/rack.rb", line 12, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/newrelic_rpm-5.2.0.345/lib/new_relic/agent/instrumentation/middleware_tracing.rb", line 92, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/rack-host-redirect-1.3.0/lib/rack/host_redirect.rb", line 18, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/newrelic_rpm-5.2.0.345/lib/new_relic/agent/instrumentation/middleware_tracing.rb", line 92, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/remotipart-1.4.2/lib/remotipart/middleware.rb", line 32, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/newrelic_rpm-5.2.0.345/lib/new_relic/agent/instrumentation/middleware_tracing.rb", line 92, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/warden-1.2.8/lib/warden/manager.rb", line 36, in block in call
File "/app/vendor/bundle/ruby/2.5.0/gems/warden-1.2.8/lib/warden/manager.rb", line 34, in catch
File "/app/vendor/bundle/ruby/2.5.0/gems/warden-1.2.8/lib/warden/manager.rb", line 34, in call
File "/app/vendor/bundle/ruby/2.5.0/gems/newrelic_rpm-5.2.0.345/lib/new_relic/<issue_closed>
Status: Issue closed |
pmndrs/jotai | 998650006 | Title: Destructuring lost reference
Question:
username_0: All tests are the same thing, the difference is destructuring of result of renderHook
https://codesandbox.io/s/focused-glitter-9ixgx?file=/src/ok.test.ts
Is this issue related to Jotai? Or @testing-library/react-hooks?
Answers:
username_1: I'm not super familiar with testing-library/react-hooks, but destructuring `current` doesn't seem correct. Does it work with `React.useState`?
username_0: Good question... Really `React.useState` not work too. But destructuring is correct, I add a new test on codesandbox with destructuring step by step. Thanks for help, I go open issue in testing-library/react-hooks
Status: Issue closed
|
facebook/react-native | 274378344 | Title: Android - Dropdown Picker - changing Picker.Item "label" is falsely triggering onValueChange
Question:
username_0: <!--
PLEASE DON'T DELETE THIS TEMPLATE UNTIL YOU HAVE READ THE FIRST SECTION.
-->
### Is this a bug report?
Yes
<!--
If you answered "Yes":
We expect that it will take you about 30 minutes to produce a high-quality bug report.
While this may seem like a lot, putting care into issues helps us fix them faster.
For bug reports, it is REQUIRED to fill the rest of this template, or the issue will be closed.
If you answered "No":
We use GitHub Issues exclusively for tracking bugs in React Native. If you're looking for help,
the Community page at http://facebook.github.io/react-native/support.html list various resources
that should help you get started.
Now scroll below!
-->
### Have you read the [Contributing Guidelines](https://facebook.github.io/react-native/docs/contributing.html)?
Yes
### Environment
<!--
In your terminal run `react-native info` and paste its contents here. Next, specify your target platform, like this:
Environment:
OS: macOS Sierra 10.12.6
Node: 8.4.0
Yarn: 0.27.5
npm: 5.4.0
Watchman: 4.7.0
Xcode: Xcode 8.3.3 Build version 8E3004b
Android Studio: 2.3 AI-162.4069837
Packages: (wanted => installed)
react-native: 0.48.1 => 0.48.1
react: 16.0.0-alpha.12 => 16.0.0-alpha.12
Target Platform: iOS (10.3)
***Make sure you are on v0.48.0 or greater of react-native, otherwise you may get this error:
Unrecognized command 'info'
Run react-native --help to see list of all available commands
-->
Environment:
OS: Windows 10
Node: 8.4.0
Yarn: 0.19.1
npm: 5.4.2
Watchman: Not Found
[Truncated]
Once you’re done, copy and paste the link to the Snack or a public GitHub repository below:
-->
https://snack.expo.io/Hyr2Aucyz
<!--
What happens if you skip this step?
Someone will read your bug report, and maybe will be able to help you,
but it’s unlikely that it will get much attention from the team. Eventually,
the issue will likely get closed in favor of issues that have reproducible demos.
Please remember that:
* Issues without reproducible demos have a very low priority.
* The person fixing the bug would have to do that anyway. Please be respectful of their time.
* You might figure out the issues yourself as you work on extracting it.
Thanks for helping us help you!
--> |
Landmaster/PlusTiC | 238284191 | Title: [MC 1.11.2] PlusTiC 5.0.1.5 java.lang.reflect.InvocationTargetException on client load
Question:
username_0: I get the following after adding PlusTiC 5.0.1.5 to my Minecraft 1.11.2 client. Loading on a server starts with no issue.
crash-report:
[crash-2017-06-23_18.09.55-client.txt](https://github.com/Landmaster/PlusTiC/files/1099328/crash-2017-06-23_18.09.55-client.txt)
Disabling PlusTiC and my client loads fine.
Status: Issue closed
Answers:
username_0: Sorry, mis-filed, looks like it's PurpleIRC. Opened https://github.com/cnaude/PurpleIRC-forge/issues/11 . |
MehediH/Tweeten | 284610931 | Title: Allow Users to keep their own Settings
Question:
username_0: On first/every startup Tweeten is setting the users theme to be dark and font size to be large.
Instead of doing this, leave the users setting as is and maybe open a modal saying that you have recommended settings (dark theme + bigger font) and allow users setting it that way.
Having my settings changed by Tweeten is not what I really want.
Answers:
username_1: Each application has its own default settings for certain things. While I understand you may not want your TweetDeck settings being overwritten by us, Tweeten is still going to apply the recommended options that work best for the design of the app.
But I do agree with showing a warning about the recommended settings -- will look into adding this for new users during the onboarding experience.
Status: Issue closed
username_1: Fixed https://github.com/username_1/Tweeten/releases/tag/4.3.1 |
JSQLParser/JSqlParser | 174112326 | Title: This sql should work but failed to parse
Question:
username_0: SELECT choice.response_choice_id AS uuid,
choice.digit AS digit,
choice.text_response AS textResponse,
choice.voice_prompt AS voicePrompt,
choice.action AS action,
choice.contribution AS contribution,
choice.order_num AS order,
choice.description AS description,
choice.is_join_conference AS joinConference,
choice.voice_prompt_language_code AS voicePromptLanguageCode,
choice.text_response_language_code AS textResponseLanguageCode,
choice.description_language_code AS descriptionLanguageCode,
choice.rec_phrase AS recordingPhrase
FROM response_choices choice
WHERE choice.presentation_id = ?
ORDER BY choice.order_num
Answers:
username_1: Any exception? My first guess is keyword "order" as column name. Cheers Tobias
username_0: I lost the exception. Yes, it complains ORDER is incorrect... You can simply try the above sql to reproduce the problem.
username_1: Order is a hard column name or alias to parse. It would introduce lookaheads in various positions of the parser:
select * from tab1 order by test
Is order an alias or start of order by??
So I decided at the moment to not fix this.
Status: Issue closed
|
ros2/rosidl_typesupport | 486451390 | Title: Decouple idl generation pipeline from rmw implementations
Question:
username_0: ## Feature request
#### Feature description
<!-- Description in a few sentences what the feature consists of and what problem it will solve -->
Currently, [rosidl_generate_interfaces from rosidl_cmake](https://github.com/ros2/rosidl/blob/36bc907e149b1e55d76f6203d62f993158759998/rosidl_cmake/cmake/rosidl_generate_interfaces.cmake) can't be called before the rmw implementations are built.
Decoupling idl generation from the `rmw` implementations will allow to use the same message abstractions at rmw level for implementing https://github.com/ros2/design/pull/250.
#### Implementation considerations
<!-- Relevant information on how the feature could be implemented and pros and cons of the different solutions -->
Currently, [rosidl_typesupport_c](https://github.com/ros2/rosidl_typesupport/blob/38eb801f1f856a503676bb79875786b3a3b6d92d/rosidl_typesupport_c/package.xml#L25) and [rosidl_typesupport_cpp](https://github.com/ros2/rosidl_typesupport/blob/38eb801f1f856a503676bb79875786b3a3b6d92d/rosidl_typesupport_cpp/package.xml#L25) have a build dependency on `rmw_implementation`, which depends on the available implementations (see [1](https://github.com/ros2/rmw_implementation/blob/cf709e83e5d34291c52145cebf9381c7cb0f2dac/rmw_implementation/package.xml#L18-L20) [2](https://github.com/ros2/rmw_implementation/blob/cf709e83e5d34291c52145cebf9381c7cb0f2dac/rmw_implementation/package.xml#L30)).
I first believed that they depend on `rmw_implementation`, because the typesupports are being registered there (e.g.: [rmw_fastrtps_cpp](https://github.com/ros2/rmw_fastrtps/blob/099f9eed9a0f581447405fbd877c6d3b15f1f26e/rmw_fastrtps_cpp/CMakeLists.txt#L116-L118)), but I later realize that that resource isn't needed in `rmw_typesupport_*` packages nor in [rosidl_cmake](https://github.com/ros2/rosidl/blob/36bc907e149b1e55d76f6203d62f993158759998/rosidl_cmake/package.xml).
If that the case, we could simply delete that dependency.
For using the idl pipeline generation (calling `rosidl_generate_interfaces` macro), you may want to use the default generators by adding a build tool dependency on `rosidl_default_generators`, which depends on the `rmw_typesupport_*` and also in [rosidl_generator_py](https://github.com/ros2/rosidl_defaults/blob/926ce4d8f460fd8d7bb8c25544c2ebe294e691c1/rosidl_default_generators/package.xml#L15-L22). The latest, also depends on the [rmw_implementation package](https://github.com/ros2/rosidl_python/blob/1b88adf2b7add927a091b16ff08bb164f977e3cd/rosidl_generator_py/package.xml#L25), because [it is using](https://github.com/ros2/rosidl_python/blob/1b88adf2b7add927a091b16ff08bb164f977e3cd/rosidl_generator_py/cmake/rosidl_generator_py_get_typesupports.cmake#L18) the previously commented resource registered by the rmw implementations.
I think that we could avoid using that resource, and directly query [this other](https://github.com/ros2/rosidl_typesupport_fastrtps/blob/master/rosidl_typesupport_fastrtps_c/CMakeLists.txt#L61).
As I don't need the python generated file, I will only add a dependency on the packages I need instead of using `rosidl_default_generators`.
Answers:
username_0: Closing, as I've already found a way of solving the problem. See:
- https://github.com/username_0/rmw_dds_common/commit/6c45d57d5629a39ce6dd4a8179e8caf3c46956d6
- https://github.com/ros2/rmw_fastrtps/pull/312
Status: Issue closed
username_1: ## Feature request
#### Feature description
<!-- Description in a few sentences what the feature consists of and what problem it will solve -->
Currently, [rosidl_generate_interfaces from rosidl_cmake](https://github.com/ros2/rosidl/blob/36bc907e149b1e55d76f6203d62f993158759998/rosidl_cmake/cmake/rosidl_generate_interfaces.cmake) can't be called before the rmw implementations are built.
Decoupling idl generation from the `rmw` implementations will allow to use the same message abstractions at rmw level for implementing https://github.com/ros2/design/pull/250.
#### Implementation considerations
<!-- Relevant information on how the feature could be implemented and pros and cons of the different solutions -->
Currently, [rosidl_typesupport_c](https://github.com/ros2/rosidl_typesupport/blob/38eb801f1f856a503676bb79875786b3a3b6d92d/rosidl_typesupport_c/package.xml#L25) and [rosidl_typesupport_cpp](https://github.com/ros2/rosidl_typesupport/blob/38eb801f1f856a503676bb79875786b3a3b6d92d/rosidl_typesupport_cpp/package.xml#L25) have a build dependency on `rmw_implementation`, which depends on the available implementations (see [1](https://github.com/ros2/rmw_implementation/blob/cf709e83e5d34291c52145cebf9381c7cb0f2dac/rmw_implementation/package.xml#L18-L20) [2](https://github.com/ros2/rmw_implementation/blob/cf709e83e5d34291c52145cebf9381c7cb0f2dac/rmw_implementation/package.xml#L30)).
I first believed that they depend on `rmw_implementation`, because the typesupports are being registered there (e.g.: [rmw_fastrtps_cpp](https://github.com/ros2/rmw_fastrtps/blob/099f9eed9a0f581447405fbd877c6d3b15f1f26e/rmw_fastrtps_cpp/CMakeLists.txt#L116-L118)), but I later realize that that resource isn't needed in `rmw_typesupport_*` packages nor in [rosidl_cmake](https://github.com/ros2/rosidl/blob/36bc907e149b1e55d76f6203d62f993158759998/rosidl_cmake/package.xml).
If that the case, we could simply delete that dependency.
For using the idl pipeline generation (calling `rosidl_generate_interfaces` macro), you may want to use the default generators by adding a build tool dependency on `rosidl_default_generators`, which depends on the `rmw_typesupport_*` and also in [rosidl_generator_py](https://github.com/ros2/rosidl_defaults/blob/926ce4d8f460fd8d7bb8c25544c2ebe294e691c1/rosidl_default_generators/package.xml#L15-L22). The latest, also depends on the [rmw_implementation package](https://github.com/ros2/rosidl_python/blob/1b88adf2b7add927a091b16ff08bb164f977e3cd/rosidl_generator_py/package.xml#L25), because [it is using](https://github.com/ros2/rosidl_python/blob/1b88adf2b7add927a091b16ff08bb164f977e3cd/rosidl_generator_py/cmake/rosidl_generator_py_get_typesupports.cmake#L18) the previously commented resource registered by the rmw implementations.
I think that we could avoid using that resource, and directly query [this other](https://github.com/ros2/rosidl_typesupport_fastrtps/blob/master/rosidl_typesupport_fastrtps_c/CMakeLists.txt#L61).
As I don't need the python generated file, I will only add a dependency on the packages I need instead of using `rosidl_default_generators`.
username_1: I will reopen this since the previous work around has reached its limits with recent changes. The goal described by this ticket is what is imo the correct approach to solve ros2/rmw_dds_common#8.
username_1: * [x] ros2/rmw_dds_common#12
* [x] ros2/rosidl_typesupport#62
* [ ] `rosidl_generator_py` must not depend on `rmw_implementation`: https://github.com/ros2/rosidl_python/blob/a98f9378ce345dbb4fe628723af8273210ee004a/rosidl_generator_py/package.xml#L25
username_0: I've cancelled both linux jobs, as there are a lot of failures in the macos and Windows ones.
It seems to be related with the change in `rosidl_generator_py`.
username_1: The `cpplint` / `uncrustify` warnings were resolved by https://github.com/ros2/rmw_fastrtps/pull/364/commits/ff<PASSWORD>dae4d2d1<PASSWORD>afcff<PASSWORD>.
The `flake8` warnings are unrelated and targeted by ros2/rclpy#539.
Status: Issue closed
|
akkadotnet/akka.net | 104704738 | Title: InvalidActorNameException upon actor restart
Question:
username_0: A while back we noticed some processes weren't being executed anymore in our application. Upon some investigation we noticed that several actors had crashed, which happens from time to time, but this time the actors couldn't be restarted due to a `InvalidActorNameException`.
This has happened once within a few months. I'm not sure if this is an akka.net issue or a bad implementation on our side. But anyway, just to be sure I thought I'd post it here. The names of several systems have been altered and this is one of several actors that crashes. They all produced similar/same log outputs as this one.
All the actors crashes due to a bad HTTP request to some service. Actors crashes. Couldn't be restarted so those particular processes stopped. All the other processes kept running just fine.
```
2015-08-28 07:12:54.826 +00:00 [Error] "Unknown Http Error: BadGateway"
System.Exception: Unknown Http Error: BadGateway (`SomeApplication/user/Mother/UpdateAddresses` crashes)
2015-08-28 07:12:55.154 +00:00 [Error] "Error while creating actor instance of type Akka.FSharp.Actors+FunActor`2[System.Object,System.Object] with 1 args: (SomeApplication.Actors.Addresses+updateActor@136)"
[akka://SomeApplication/user/Mother/UpdateAddresses]: Akka.Actor.PostRestartException: Exception post restart (System.Exception) ---> System.Exception: Error while creating actor instance of type Akka.FSharp.Actors+FunActor`2[System.Object,System.Object] with 1 args: (SomeApplication.Actors.Addresses+updateActor@136) ---> System.Reflection.TargetInvocationException: Exception has been thrown by the target of an invocation. ---> Akka.Actor.InvalidActorNameException: Actor name "UpdateAddresses" is not unique!
at Akka.Actor.Internal.NormalChildrenContainer.Reserve(String name)
at Akka.Actor.ActorCell.<>c__DisplayClass53.<ReserveChild>b__52(IChildrenContainer c)
at Akka.Util.Internal.InterlockedSpin.Swap[T](T& reference, Func`2 updater)
at Akka.Actor.ActorCell.ReserveChild(String name)
at Akka.Actor.ActorCell.MakeChild(Props props, String name, Boolean async, Boolean systemService)
at Akka.Actor.ActorCell.ActorOf(Props props, String name, Boolean isAsync, Boolean isSystemService)
at Akka.Actor.ActorCell.ActorOf(Props props, String name)
at SomeApplication.Actors.Addresses.Actor.updateApiActor[a,b](Actor`1 mailbox) in C:\SomeFolder\AddressesActor.fs:line 33
at [email protected](Actor`1 mailbox) in C:\SomeFolder\AddressesActor.fs:line 136
at Akka.FSharp.Actors.FunActor`2..ctor(FSharpFunc`2 actor)
--- End of inner exception stack trace ---
at System.RuntimeMethodHandle.InvokeMethod(Object target, Object[] arguments, Signature sig, Boolean constructor)
at System.Reflection.RuntimeConstructorInfo.Invoke(BindingFlags invokeAttr, Binder binder, Object[] parameters, CultureInfo culture)
at System.RuntimeType.CreateInstanceImpl(BindingFlags bindingAttr, Binder binder, Object[] args, CultureInfo culture, Object[] activationAttributes, StackCrawlMark& stackMark)
at System.Activator.CreateInstance(Type type, BindingFlags bindingAttr, Binder binder, Object[] args, CultureInfo culture, Object[] activationAttributes)
at System.Activator.CreateInstance(Type type, Object[] args)
at Akka.Actor.Props.ActivatorProducer.Produce()
at Akka.Actor.Props.NewActor()
--- End of inner exception stack trace ---
at Akka.Actor.Props.NewActor()
at Akka.Actor.ActorCell.CreateNewActorInstance()
at Akka.Actor.ActorCell.<>c__DisplayClass48.<NewActor>b__47()
at Akka.Actor.ActorCell.UseThreadContext(Action action)
at Akka.Actor.ActorCell.NewActor()
at Akka.Actor.ActorCell.FinishRecreate(Exception cause, ActorBase failedActor)
--- End of inner exception stack trace ---
2015-08-28 07:12:55.185 +00:00 [Error] "Object reference not set to an instance of an object."
System.NullReferenceException: Object reference not set to an instance of an object.
at Akka.Event.DeadLetterListener.Receive(Object message)
at Akka.Actor.ActorBase.AroundReceive(Receive receive, Object message)
at Akka.Actor.ActorCell.ReceiveMessage(Object message)
at Akka.Actor.ActorCell.Invoke(Envelope envelope)
2015-08-28 07:12:55.201 +00:00 [Error] "Object reference not set to an instance of an object."
System.NullReferenceException: Object reference not set to an instance of an object.
at Akka.Event.DeadLetterListener.Receive(Object message)
at Akka.Actor.ActorBase.AroundReceive(Receive receive, Object message)
at Akka.Actor.ActorCell.ReceiveMessage(Object message)
at Akka.Actor.ActorCell.Invoke(Envelope envelope)
```
Answers:
username_1: We need some way to reproduce this.
This specific line in the stacktrace looks suspicious to me:
```
2015-08-28 07:12:55.185 +00:00 [Error] "Object reference not set to an instance of an object."
System.NullReferenceException: Object reference not set to an instance of an object.
```
Was `null` passed as a message somewhere?
username_0: That's probably the F# `unit` we send to some actors. Like the implementation in #1285. These actors are called using the scheduler.
username_2: Whats the status on this issue? |
chasonr/nethack-3.6.0-patches | 124600120 | Title: Mac OS X fails to build with SDL or QT installed from homebrew
Question:
username_0: Thanks for an awesome job on this porting!
Building on Mac OS X (El Capitan, but I don't believe it matters), I noticed that the macosx10.10 hint file doesn't support SDL2 or QT that was installed via homebrew. Nor, for QT, does it support setting a custom QTDIR (this came from commit 6062fb1c).
For SDL, I just went ahead and installed the official SDL frameworks in /Library/Frameworks, so I'm not sure what exactly needs to be done; for QT, I removed the hardcoded setting QTDIR from the hints file, and changed:
QT_FRAMEWORKS_DIR = $(QTDIR)/Frameworks # /Library/Frameworks
I'm not sure what the best way to support either official installations or homebrew is (I'm pretty good with make on Linux systems, but I'm still figuring out how exactly Mac OS puts everything together), so I'm not submitting a patch. Just pointing out something that might be helpful!
Again, from someone whose imagination doesn't quite stretch far enough to visualize a @ or a ( or ", thanks for letting me enjoy Nethack 3.6!
Answers:
username_1: Thank you for noting what commit you used; I've done lots more since that one, though I don't think any of them address the issues you've noted.
My Mac is still running Yosemite and has SDL2, SDL2_image and various Qt 4 packages under /Library/Frameworks. By "homebrew", perhaps you mean that you built one or more of these packages from source?
Could you attach the hints file that finally worked for you?
username_0: Just to be clear; the commit I was using was e6c14f4. Commit 6062fb1 was where you added a hardcoded QTDIR.
Homebrew is http://brew.sh ; it's a pretty popular system for installing packages on Mac OS (similar to MacPorts, if you've ever used that). The biggest change for this context is that everything is installed under /usr/local; for example, I have
lrwxr-xr-x 1 mikemi admin 48 Dec 30 13:41 /usr/local//Frameworks/QtCore.framework -> ../Cellar/qt/4.8.7_2/Frameworks/QtCore.framework
for QtCore.framework.
The hints file I used for using QT from homebrew (like I said; for SDL2, I just downloaded the two .framework files (SDL2 and SDL2_image) and copied them to /Library/Frameworks is attached (renamed because github won't let me upload a file with that "extension")
[macosx10.10.txt](https://github.com/username_1/nethack-3.6.0-patches/files/76480/macosx10.10.txt)
Status: Issue closed
|
appeltel/cpython_assembly | 306536591 | Title: Handle EXTENDED_ARG
Question:
username_0: The assembler should be able to handle large arguments and insert EXTENDED_ARG instructions as needed, updating jump locations as it goes. This requires a rewrite of the (currently very simple/sloppy) system which believes that each instruction takes 2 bytes.
I think it also makes sense not to have the assembly language itself accept EXTENDED_ARG as an instruction in the source, but to assemble them as needed based on the arguments provided and locations.
Answers:
username_0: First attempt in 7125cf9, needs more testing but it could be ok...
username_0: On second thought, it would be a good idea to allow EXTENDED_ARG so that people can experiment and break things, as well as potentially allow for dumping in the output of the dis module and making their own changes.
For the dis module output, EXTENDED_ARG is present, but argument values are given as > 255. There could be a `.noextend` directive that simply truncates arguments over 255. This would allow for using dis output assuming junk before the instruction is ignored.
username_0: A better option than `.noextend` might be to wrap `dis.dis` and munge the output into the assembler format, adding consts, names, etc, and leaving arguments as numerical values.
Status: Issue closed
|
akxcv/vuera | 352946607 | Title: [Feature] Total conversion of a project React -> Vue or the other way around
Question:
username_0: Is it possible to use the babel plugin `vuera `in the future, convert a React source code to Vue or the other way around? Definitely not just during use.
Answers:
username_1: Hi! This sounds quite complicated as I don't think you can reliably and completely map React and Vue to one another. Maybe it's possible to some extent, but definitely not with vuera or its babel plugin. You could try to implement a library of your own that would do that!
Closing this for now, but feel free to continue the discussion if you have any thoughts or questions.
Status: Issue closed
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.