repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
scribbletune/johann | 705230673 | Title: show octave on note (guitar & piano)
Question:
username_0: **Description**
better to show octave on note useful when play in choir
e.g. in standard choir usually use SATB (Sopran Alto Tenor Bass) which every vocal characteristic (male/female) has it's own range (base on octave)
**Implementation**
perhaps use button to show octave info for show or hidden the info (recommended because not every musicians care about octave can hide it, but for the detail musicians that must play within choir can show the octave)
or
simply put octave info without option to show or hide octave info
**Examples**
which fretboard must be press to play melody with tenor vocal on major third tuning

**ref on piano**
https://en.wikipedia.org/wiki/Tenor
thanks |
Rajawali/Rajawali | 247378793 | Title: .OBJ parser always prints debug messages even if debugging is turned off
Question:
username_0: #### Rajawali Version or Branch
Bombshell 1.1.777
#### Device and Android Version
Nexus. Android 5.0+
Samsung Note 4
#### Summary
When turning off logging in Rajawali, the object parsing doesn't respect it and keeps printing output to the logcat
#### Steps to Reproduce
[Describe reproducable steps and sample code]
#### Trace or Log Output
[include stacktrace(s) if relevant to the issue. Engine crashes can not be debugged without a stacktrace or at minimum reproduction code.]
Answers:
username_1: What approach are you using when "turning off logging in Rajawali"? |
LeonHeidelbach/ttv_adEraser | 775034963 | Title: chat bug
Question:
username_0: Hey,
when i enable this addon i can't copy and paste from twitch chat anymore.
Status: Issue closed
Answers:
username_1: Thank you for letting me know. This is actually something I have completely missed while testing but was true and a quick fix. I have just pushed an update that fixes this problem. |
ibm-messaging/mq-mqi-nodejs | 960237005 | Title: Cannot find
Question:
username_0: I simply installed 'ibmmq' package using "npm i ibmmq" and tried to run amqsconn.js from samples. But I'm getting this below error:
Cannot find MQ C library.
Has the C client been installed?
Have you run setmqenv?
Answers:
username_1: You don't say which platform you're running on.
Status: Issue closed
username_0: Hi ,
I was working on windows.
My issue has been resolved. |
030/n3dr | 468714865 | Title: docker not lauching
Question:
username_0: Hello guy,
I've try to use your soft, because i want to download some repository from our older nexus3 (test) to our new production nexus3
but when i launch your docker through docker-compose and the docker creating but i got an error and no docker stay alive. I don't know if i'm doing something wrong or not ? i will link you my config / steps
1/
create a /opt/n3dr/.n3dr.yaml (with chmod 400)
create a /opt/n3dr/download/ (folder)
2/ create my docker-compose.yml
```
version: '3.7'
services:
n3dr:
image: utrecht/n3dr:3.0.0
volumes:
- /opt/n3dr/.n3dr.yaml:/home/n3dr/.n3dr.yaml
- /opt/n3dr/download:/download
command: repositories -n http://myservernexusonlan:8081/nexus/ -u admin -v ${NEXUS_API_VERSION} -b
```
and when i do my docker-compose up i got this ?
```
docker-compose up
WARNING: The NEXUS_API_VERSION variable is not set. Defaulting to a blank string.
Recreating n3dr_n3dr_1 ... done
Attaching to n3dr_n3dr_1
n3dr_1 | Error: unknown shorthand flag: 'v' in -v
n3dr_1 | Usage:
n3dr_1 | n3dr repositories [flags]
n3dr_1 |
n3dr_1 | Flags:
n3dr_1 | -b, --backup Backup artifacts from all repositories
n3dr_1 | -c, --count Count the number of repositories
n3dr_1 | -h, --help help for repositories
n3dr_1 | -a, --names Print all repository names
n3dr_1 |
n3dr_1 | Global Flags:
n3dr_1 | -d, --debug Enable debug logging
n3dr_1 | -p, --n3drPass string The Nexus3 password
n3dr_1 | -n, --n3drURL string The Nexus3 URL
n3dr_1 | -u, --n3drUser string The Nexus3 user
n3dr_1 |
n3dr_1 | unknown shorthand flag: 'v' in -v
n3dr_n3dr_1 exited with code 1
```
and i got no more logs nothing.
Answers:
username_0: So i advance, i create the docker using the docker file ;). But when i try to connect to my nexus nothing appen even if i put right credential
i add context path "/nexus" (that we keep from old nexus) because we are migrating from 2 to 3
```
docker run -v /opt/n3dr/.n3dr.yaml:/home/n3dr/.n3dr.yaml -v /opt/n3dr/download:/download n3dr repositories -n http://mynexus:8081/nexus/ -c -u admin -v v1 -d
```
Answer :
```time="2019-07-16T15:50:36Z" level=debug msg="Consult repositories URL: http://myservernexus:8081/nexus/" func=n3dr/cli.Nexus3.repositories file="/go/n3dr/cli/repositories.go:15"
0```
username_1: Could you use the latest n3dr? I.e. 3.1.0.
username_1: @username_0 Did 3.1.0 solve the issue?
username_1: @username_0 Could you use the command line tool and check whether the issue persists?
username_1: Suggestion is to remove docker support as it contains too many issues.
username_1: Docker support removed
Status: Issue closed
username_1: Hello guy,
I've try to use your soft, because i want to download some repository from our older nexus3 (test) to our new production nexus3
but when i launch your docker through docker-compose and the docker creating but i got an error and no docker stay alive. I don't know if i'm doing something wrong or not ? i will link you my config / steps
1/
create a /opt/n3dr/.n3dr.yaml (with chmod 400)
create a /opt/n3dr/download/ (folder)
2/ create my docker-compose.yml
```
version: '3.7'
services:
n3dr:
image: utrecht/n3dr:3.0.0
volumes:
- /opt/n3dr/.n3dr.yaml:/home/n3dr/.n3dr.yaml
- /opt/n3dr/download:/download
command: repositories -n http://myservernexusonlan:8081/nexus/ -u admin -v ${NEXUS_API_VERSION} -b
```
and when i do my docker-compose up i got this ?
```
docker-compose up
WARNING: The NEXUS_API_VERSION variable is not set. Defaulting to a blank string.
Recreating n3dr_n3dr_1 ... done
Attaching to n3dr_n3dr_1
n3dr_1 | Error: unknown shorthand flag: 'v' in -v
n3dr_1 | Usage:
n3dr_1 | n3dr repositories [flags]
n3dr_1 |
n3dr_1 | Flags:
n3dr_1 | -b, --backup Backup artifacts from all repositories
n3dr_1 | -c, --count Count the number of repositories
n3dr_1 | -h, --help help for repositories
n3dr_1 | -a, --names Print all repository names
n3dr_1 |
n3dr_1 | Global Flags:
n3dr_1 | -d, --debug Enable debug logging
n3dr_1 | -p, --n3drPass string The Nexus3 password
n3dr_1 | -n, --n3drURL string The Nexus3 URL
n3dr_1 | -u, --n3drUser string The Nexus3 user
n3dr_1 |
n3dr_1 | unknown shorthand flag: 'v' in -v
n3dr_n3dr_1 exited with code 1
```
and i got no more logs nothing.
username_1: https://github.com/wagoodman/dive
Status: Issue closed
username_1: Resolved in 6.0.3. |
serenaf/ember-workshop | 268101142 | Title: Basic tests
Question:
username_0: - Acceptance tests for route
- [ ] nav bar
- Component integration tests
- [ ] Item list
- [ ] item component
- [ ] Comment component
- [ ] Nav bar
- Unit tests
- [ ] Item model
- [ ] Comment model |
indico/indico | 170919386 | Title: ImportError: No module named sanitizer
Question:
username_0: After installing indico using `easy_install` I got
```
mod_wsgi (pid=31701): Target WSGI script '/opt/indico/htdocs/indico.wsgi' cannot be loaded as Python module.
mod_wsgi (pid=31701): Exception occurred processing WSGI script '/opt/indico/htdocs/indico.wsgi'.
Traceback (most recent call last):
File "/opt/indico/htdocs/indico.wsgi", line 20, in <module>
from indico.web.flask.app import make_app
File "/usr/lib/python2.7/site-packages/indico-1.2.1rc7-py2.7.egg/indico/web/flask/app.py", line 35, in <module>
from MaKaC.webinterface.pages.error import WErrorWSGI
File "/usr/lib/python2.7/site-packages/indico-1.2.1rc7-py2.7.egg/MaKaC/webinterface/pages/error.py", line 20, in <module>
from MaKaC.webinterface.pages.base import WPDecorated
File "/usr/lib/python2.7/site-packages/indico-1.2.1rc7-py2.7.egg/MaKaC/webinterface/pages/base.py", line 26, in <module>
import MaKaC.webinterface.wcomponents as wcomponents
File "/usr/lib/python2.7/site-packages/indico-1.2.1rc7-py2.7.egg/MaKaC/webinterface/wcomponents.py", line 43, in <module>
from MaKaC import conference
File "/usr/lib/python2.7/site-packages/indico-1.2.1rc7-py2.7.egg/MaKaC/conference.py", line 25, in <module>
from MaKaC.common.utils import formatDateTime
File "/usr/lib/python2.7/site-packages/indico-1.2.1rc7-py2.7.egg/MaKaC/common/utils.py", line 54, in <module>
from indico.util.string import truncate
File "/usr/lib/python2.7/site-packages/indico-1.2.1rc7-py2.7.egg/indico/util/string.py", line 29, in <module>
import bleach
File "/usr/lib/python2.7/site-packages/bleach/__init__.py", line 8, in <module>
from html5lib.sanitizer import HTMLSanitizer
ImportError: No module named sanitizer
```
Seems like html5lib has been refactored quite a bit, and a workaround was introduced in bleach 1.4.3 (https://github.com/mozilla/bleach/commit/532463ef194be11f2b73af78a65b0016b68000c1), but `easy_install indico` installed bleach 1.4 for me, not 1.4.3.
A requirement for bleach 1.4.3 was actually already introduced in https://github.com/indico/indico/commit/f252ee11b54010f0ac51bb65337e8976ddcc7532, so then I guess installing indico from git master should work. Seems like a good idea to push a new rc soon then.
Answers:
username_1: That's actually a dependency problem. Installing the latest html5lib (?)
should fix it, IIRC. We'll release a fixed package next week.
username_1: v1.2.1rc8 should fix this.
Status: Issue closed
|
videojs/video.js | 109157347 | Title: videojs.Flash undefined
Question:
username_0: I just tried the HLS getting started: http://videojs.github.io/videojs-contrib-hls/
I get errors that `videojs.Flash` is undefined.
I included video-js.css, video.min.js, videojs-media-sources.js and videojs-hls.js.
Any ideas what I might be missing?
Status: Issue closed
Answers:
username_0: So the solution is to use video.js 4.x
username_1: I just updated to the latest version of videojs and still experiencing this issue?
Any suggestions on what to do?
username_2: If you grade to the latest contrib-hls, it works with latest videojs 5. |
ungaralex/ase17-Group4-2-AttendanceTrackingSystem | 283251506 | Title: Final TODOs before first deadline
Question:
username_0: **Webinterface**:
- Date check enable
- Remove hardcoded tests
- Convert date ID to real date based on milliseconds
- Integration test
**Pi** (Friday):
- Integration test
**App**:
- Error output (wrong date, etc.)
- (Rescale QR code)
- **GMail login**
- Integration test
**Documentation** @Fraethy :
- Completed vs. due features
- Architecture
- Deployment diagram
- Use case diagram
- Lessons learnt and tradeoffs
- Future work<issue_closed>
Status: Issue closed |
statamic/ideas | 773970372 | Title: static caching to exclude pages using regex
Question:
username_0: Right now you can only use a wildcard to exclude URLs matching the last part of the URL.
https://statamic.dev/static-caching#excluding-pages
It would be nice to use regex for exclude rules for better flexibility.
_I'd be happy to work on a PR for this if this is something others are interested in._ |
Penetrum-Security/Maltree-Issue-Repo | 775499666 | Title: Maltree Issue (1560394e8eca1d1)
Question:
username_0: Python version `2.717`
Traceback:
```
Traceback (most recent call):
File "/home/malcore/bin/devel/maltree/entry/main.py", line 132, in main
is_elf(malware_sample_name),
File "/home/malcore/bin/devel/maltree/lib/settings.py", line 1114, in is_elf
res = open(r"{}".format(filename), "rb").read(4) == "\x7fELF"
[Errno 2] No such file or directory: '/home/malcore/bin/devel/maltree/temp_files/_MALWARE__MoeZzCTRaUnyxOrsrQDN.tmp'
```
Running platform: `Linux-5.4.0-53-generic-x86_64-with-Ubuntu-18.04-bionic`<issue_closed>
Status: Issue closed |
netty/netty | 235723409 | Title: java.lang.ClassNotFoundException: jdk.internal.misc.Unsafe
Question:
username_0: ### Expected behavior: No stack trace regarding this class
### Actual behavior:
```
D 06-06 11:01:33.811 pool-1-thread-2 i.n.u.i.l.InternalLoggerFactory:71 |::] Using SLF4J as the default logging framework
D 06-06 11:01:33.819 pool-1-thread-2 i.n.u.i.PlatformDependent0:76 |::] -Dio.netty.noUnsafe: false
D 06-06 11:01:33.823 pool-1-thread-2 i.n.u.i.PlatformDependent0:71 |::] sun.misc.Unsafe.theUnsafe: available
D 06-06 11:01:33.824 pool-1-thread-2 i.n.u.i.PlatformDependent0:71 |::] sun.misc.Unsafe.copyMemory: available
D 06-06 11:01:33.825 pool-1-thread-2 i.n.u.i.PlatformDependent0:71 |::] java.nio.Buffer.address: available
D 06-06 11:01:33.826 pool-1-thread-2 i.n.u.i.PlatformDependent0:71 |::] direct buffer constructor: available
D 06-06 11:01:33.828 pool-1-thread-2 i.n.u.i.PlatformDependent0:76 |::] java.nio.Bits.unaligned: available, true
D 06-06 11:01:33.834 pool-1-thread-2 i.n.u.i.PlatformDependent0:91 |::] jdk.internal.misc.Unsafe.allocateUninitializedArray(int): unavailable
java.lang.ClassNotFoundException: jdk.internal.misc.Unsafe
at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
at io.netty.util.internal.PlatformDependent0$6.run(PlatformDependent0.java:295)
at java.security.AccessController.doPrivileged(Native Method)
at io.netty.util.internal.PlatformDependent0.<clinit>(PlatformDependent0.java:288)
at io.netty.util.internal.PlatformDependent.getSystemClassLoader(PlatformDependent.java:895)
at io.netty.util.internal.PlatformDependent.isAndroid0(PlatformDependent.java:919)
at io.netty.util.internal.PlatformDependent.<clinit>(PlatformDependent.java:70)
at io.netty.util.ConstantPool.<init>(ConstantPool.java:32)
at io.netty.util.AttributeKey$1.<init>(AttributeKey.java:27)
at io.netty.util.AttributeKey.<clinit>(AttributeKey.java:27)
....
at com.google.inject.internal.cglib.reflect.$FastConstructor.newInstance(FastConstructor.java:40)
at com.google.inject.internal.DefaultConstructionProxyFactory$1.newInstance(DefaultConstructionProxyFactory.java:60)
at com.google.inject.internal.ConstructorInjector.construct(ConstructorInjector.java:85)
at com.google.inject.internal.ConstructorBindingImpl$Factory.get(ConstructorBindingImpl.java:254)
at com.google.inject.internal.ProviderToInternalFactoryAdapter$1.call(ProviderToInternalFactoryAdapter.java:46)
at com.google.inject.internal.InjectorImpl.callInContext(InjectorImpl.java:1031)
at com.google.inject.internal.ProviderToInternalFactoryAdapter.get(ProviderToInternalFactoryAdapter.java:40)
at com.google.inject.Scopes$1$1.get(Scopes.java:65)
at com.google.inject.internal.InternalFactoryToProviderAdapter.get(InternalFactoryToProviderAdapter.java:40)
at com.google.inject.internal.SingleParameterInjector.inject(SingleParameterInjector.java:38)
at com.google.inject.internal.SingleParameterInjector.getAll(SingleParameterInjector.java:62)
at com.google.inject.internal.ConstructorInjector.construct(ConstructorInjector.java:84)
at com.google.inject.internal.ConstructorBindingImpl$Factory.get(ConstructorBindingImpl.java:254)
at com.google.inject.internal.ProviderToInternalFactoryAdapter$1.call(ProviderToInternalFactoryAdapter.java:46)
at com.google.inject.internal.InjectorImpl.callInContext(InjectorImpl.java:1031)
at com.google.inject.internal.ProviderToInternalFactoryAdapter.get(ProviderToInternalFactoryAdapter.java:40)
at com.google.inject.Scopes$1$1.get(Scopes.java:65)
at com.google.inject.internal.InternalFactoryToProviderAdapter.get(InternalFactoryToProviderAdapter.java:40)
at com.google.inject.internal.FactoryProxy.get(FactoryProxy.java:54)
at com.google.inject.internal.ProviderToInternalFactoryAdapter$1.call(ProviderToInternalFactoryAdapter.java:46)
at com.google.inject.internal.InjectorImpl.callInContext(InjectorImpl.java:1031)
at com.google.inject.internal.ProviderToInternalFactoryAdapter.get(ProviderToInternalFactoryAdapter.java:40)
at com.google.inject.Scopes$1$1.get(Scopes.java:65)
at com.google.inject.internal.InternalFactoryToProviderAdapter.get(InternalFactoryToProviderAdapter.java:40)
at com.google.inject.internal.SingleFieldInjector.inject(SingleFieldInjector.java:53)
at com.google.inject.internal.MembersInjectorImpl.injectMembers(MembersInjectorImpl.java:110)
at com.google.inject.internal.MembersInjectorImpl$1.call(MembersInjectorImpl.java:75)
at com.google.inject.internal.MembersInjectorImpl$1.call(MembersInjectorImpl.java:73)
at com.google.inject.internal.InjectorImpl.callInContext(InjectorImpl.java:1024)
at com.google.inject.internal.MembersInjectorImpl.injectAndNotify(MembersInjectorImpl.java:73)
at com.google.inject.internal.MembersInjectorImpl.injectMembers(MembersInjectorImpl.java:60)
[Truncated]
D 06-06 11:01:33.842 pool-1-thread-2 i.n.util.internal.CleanerJava6:71 |::] java.nio.ByteBuffer.cleaner(): available
```
### Steps to reproduce
We were using netty 4.0.x branch for long time and once we switch to 4.1.x, this stack trace appeared.
### Netty version: 4.1.11.Final
### Netty tcnative version: 2.0.1.Final
### JVM version (e.g. `java -version`)
openjdk version "1.8.0_121"
OpenJDK Runtime Environment (build 1.8.0_121-8u121-b13-0ubuntu1.16.04.2-b13)
OpenJDK 64-Bit Server VM (build 25.121-b13, mixed mode)
### OS version (e.g. `uname -a`)
Ubuntu 16.04 and Mac 10.11.4
Status: Issue closed
Answers:
username_1: @username_0 ah this was fixed recently and is harmless:
https://github.com/netty/netty/commit/0ee49e6d668bd100320003b7d6b287bc289db28f
Upgrade to 4.1.12.Final and you will not see it anymore
username_0: Great, thank you! |
sharat87/ti | 135350294 | Title: Reports
Question:
username_0: Sorry to be so slow, but how exactly can I read my json file to generate beautiful html reports?
I am on Ubuntu 14.04
Status: Issue closed
Answers:
username_1: Hey, sorry for not responding sooner. You're not slow, I am. The HTML reports suggestion in the README is just an idea. I did something like that a few years ago but nothing that I shared with the project. As all the information you need is in the JSON file, such a report is certainly possible to generate.
Thanks. |
Tencent/ncnn | 450196737 | Title: arm端int8量化后速度慢了
Question:
username_0: ARM Cortex-A53,build armv7 library
模型 | 模型大小 | ARM端前向耗时ms
-- | -- | -- | --
P 输入400*300 | 27KB | mean:18.01
R 输入24*24 | 401KB | mean:1.71
O 输入48*48 | 1.6MB | mean:5.67
P-int8 输入400*300 | 7KB | mean:23.28
R-int8 输入24*24 | 105KB | mean:2.43
O-int8 输入48*48 | 398KB | mean:7.42
这个速度是否正常?
Answers:
username_1: A53 为啥要用 armv7 library
username_0: 路径依赖😄
username_2: 你好,请问量化时image这一项需要多少图片,我用的是一张,量化后的模型好像不能检测
username_0: 我用了几万张,能检测,效果不好。
username_2: 你训练好的caffe模型可以分享一下吗,我用A53跑的mobilenetssd(chuanqi305的模型)检测一张图结果要700ms
username_0: 链接:https://pan.baidu.com/s/1x0vyzF1PN8Bx1HK1Q8hyCg 密码:<PASSWORD>
username_2: @username_0 请教一下我量化的过程中为什么一直出现couldn‘t find any detctions.
我这次用了5000张图片
Status: Issue closed
username_3: https://zhuanlan.zhihu.com/p/370689914
20210507版本大幅优化int8推理速度,建议更新 |
docker/hub-feedback | 159925503 | Title: Missing tags/builds in pulled list
Question:
username_0: Hi,
I'm using bitbucket in combination with docker to create automated builds.
This worked perfectly till yesterday.
I first thought I might have too much tags/builds in the hub. So I deleted first few versions.
But still whenever I try to pull the list of build version in Synology, I only get the previous versions (not even ':latest'), although all of them are visible on the hub.docker.com website.
Any ideas?
It's quite urgent, I need to put the website online by tomorrow.
Answers:
username_0: I just realised there is another thread/bug with the same issues:
#687
Status: Issue closed
|
zalando/zalando.github.io | 333155279 | Title: Link and design review of docs
Question:
username_0: ## Docs section review
There appears to have been issues with the design and the links in the docs section of the site. Short term, read through the content and check all links, secondly investigate why the design looks weird on some screens.
Longer team implement a link checker for deploys on jekyll.
### Tasks
These are the concrete tasks to complete before the issue is marked as done:
- [ ] check all links
- [ ] review design
- [ ] typos, correctness of content etc
### Expected delivery
Being able to republish the docs section<issue_closed>
Status: Issue closed |
google/ksp | 1119565578 | Title: Extended Java interface defines Any as a supertype
Question:
username_0: Consider the following:
Named.java
```
public interface Named {
String getName();
}
```
Test.kt
```
interface Test : Named {
fun setName(name: String)
}
```
In this case getting the class declaration for `Named` then calling superTypes will return a single super type `Any`. This is different from Kotlin classes where super types would be empty. I understand that this is likely because all Java classes extend from Object, but the behavior being different for classes compiled with Java compared to Kotlin makes working with KSP more difficult because you have to exclude Any when working with super types under this condition. You have to exclude Object when looking at super types when working with Java annotation processors, but at least it's consistent.
I think if `Any` is listed as a super type for Java classes it should also be for Kotlin classes as well, or neither of them should report `Any` as a super type<issue_closed>
Status: Issue closed |
elasticfence/elasticsearch-http-user-auth | 128832807 | Title: Can't work on Elasticsearch 2.1.1
Question:
username_0: Exception in thread "main" java.lang.IllegalArgumentException: Plugin [http-user-auth-plugin] is incompatible with Elasticsearch [2.1.1]. Was designed for version [2.0.0]
at org.elasticsearch.plugins.PluginInfo.readFromProperties(PluginInfo.java:117)
at org.elasticsearch.plugins.PluginsService.getPluginBundles(PluginsService.java:302)
at org.elasticsearch.plugins.PluginsService.<init>(PluginsService.java:108)
at org.elasticsearch.node.Node.<init>(Node.java:146)
at org.elasticsearch.node.Node.<init>(Node.java:128)
at org.elasticsearch.node.NodeBuilder.build(NodeBuilder.java:145)
at org.elasticsearch.bootstrap.Bootstrap.setup(Bootstrap.java:178)
at org.elasticsearch.bootstrap.Bootstrap.init(Bootstrap.java:285)
at org.elasticsearch.bootstrap.Elasticsearch.main(Elasticsearch.java:35)
Refer to the log for complete error details.
how to fix it ?
Answers:
username_1: Hi @username_0 please use the [2.1 branch](https://github.com/elasticfence/elasticsearch-http-user-auth/tree/2.1) of Elasticfence for Elasticsearch 2.1.1
Status: Issue closed
|
Bostwickenator/FtpPhotosUploader | 302883284 | Title: FTP_Uploader_1970_01_01
Question:
username_0: Hi,
Querying the date may have a problem because the created folder is: FTP_Uploader_1970_01_01
Please fix it if possible! Thanks :-)
Answers:
username_1: Hi, Please make sure your device has the correct time and date.
username_0: The date is set correctly. I can not even set an unrealistic date. The camera's language is set to Hungarian. I don't know if it affects...
username_0: I just noticed that ADB shell date query returns 1970. In the camera settings, the date is set correctly and photos are also dated. The two dates are not synchronized.
username_1: Odd, have you tried to set the time yet again in the menus?
Since you mention you've established a ADB connection, it may be an idea to try set datetime thu ADB Shell, using for ex. [these instructions](https://www.codeday.top/2017/07/28/31323.html) *(untested by me).
Another idea would be to connect the camera to your home wifi and see if it syncs the system date/time thu the internet (ntp).
I should probably note that I am unaffiliated with this project, I'm just trying to be helpful.
username_2: was that useful? setting a datetime in adb?
username_3: I have the sam problem. Date an time in the cam menu is current. How to set date an time in adb shell?
thanks Axel |
spdx/LicenseListPublisher | 824021951 | Title: Quotes in template keyword text is not handled correctly
Question:
username_0: Since double quotes are used to delineate the keyword values, having them inside the value itself trips up the matching rule parsing.
This is a specification issue, adding this issue here to track implementation once the issue is resolved - see https://github.com/spdx/spdx-spec/issues/497 for details.
Status: Issue closed
Answers:
username_0: This turned out to not be the problem - the original issue was a semi-colon which can be escaped with a backslash |
scylladb/seastar | 166761280 | Title: seastar pipe missing data
Question:
username_0: pipe implements by queue, it return a promise when the queue is full, if I push n(n>1) data to the fully queue, it just save the last promise, and the last promise would set value, the data we push to the queue except the last one would never notify any more, in other word, they are missing.
Answers:
username_1: So normally, you won't even have access to one pipe writer from multiple fibers.
It is interesting to consider whether we can relax the "single producer" assumption, and concede that even though we expect each producer to wait after blocking on a write, there may be several independent producers and we wish to allow all of them to block on a write() and wake up separately as space becomes available on the queue. Definitely doable. The biggest question is whether we can do this without losing performance or gaining too much complexity.
Another option is to keep a pipe as a single-producer feature (and uncopyable), and introduce a new feature (pipe writer multiplexor?) which multiplexes several pipe writers into one copyable pipe writer. This is similar to what we have done in shared_future, which multiple fibers can make a copy of and use.
There haven't been too many uses of seastar::pipe until now, so definitely we can still rethink its design and its purpose and what additional or changed abstractions can be useful for Seastar applications.
P.S. even in the current single-producer implementation, we can make it more helpful in helping the seastar::pipe user catch mistakes in the way it is used. For example, `queue<T>::not_empty()` and `not_full()` could assert (or throw an exception) if the promise is already set - instead of happily overwriting it with a new promise. |
dividab/tsconfig-paths-webpack-plugin | 765194008 | Title: can not resolve to correct path when used with webpack5
Question:
username_0: here is a minimum repo to show this bug: https://github.com/username_0/tsconfigplugin-issue-with-webpack-5
The test file in packages/fullscreen couldn't be correctly resolved. If we check the output bundle js file we will see
```
/***/ "../packages/fullscreen/src/entry.ts":
/*!*******************************************!*
!*** ../packages/fullscreen/src/entry.ts ***!
\*******************************************/
/***/ ((__unused_webpack_module, exports, __webpack_require__) => {
"use strict";
function __export(m) {
for (var p in m) if (!exports.hasOwnProperty(p)) exports[p] = m[p];
}
Object.defineProperty(exports, "__esModule", ({ value: true }));
console.log('packages/fullscreen/src/entry.ts loaded');
__export(__webpack_require__(/*! ./test */ "./src/test.ts"));
```
noted that on the last line, the file path is `./src/test.ts`, which resolves to the file `application/src/test.ts`. While the correct resolve path should be like `../packages/fullscreen/src/test.ts` (in webpack 4).
I suppose this has something to do with how webpack5 distinguishes files. But I don't have the knowledge to solve this bug. Wish someone could look into it. For now, I'm planning to use webpack resolve.alias instead of this plugin.
Answers:
username_1: same problem, fallbacked to resolve.alias :(
username_2: I too have switched to resolve aliases to work around this; I had this this previously and assumed it was my fault, but there's something going on in the plugin. It's not too much code to read paths and produce aliases, and the output is the same (but it actually works in cases where the file names conflict between paths).
username_3: Any progress? |
Toxblh/MTMR | 790892149 | Title: Is there a way to disable skip song when touching the Spotify currently playing
Question:
username_0: I always accidentally skip songs while typing. Was wondering if you can stop the track skipping when touching the song that's currently being played via Spotify?
Answers:
username_1: As far as I know, the standard action of tapping the currently playing song, is to play the next track. You could change the behaviour, so it either does nothing, or just pauses and replays the song, after another tap.
If you don't want anything to happen, upon tapping, just replace this, with yours:
```
// Spotify
{
"type": "appleScriptTitledButton",
"source": {
"inline": "if application \"Spotify\" is running then\rtell application \"Spotify\"\rreturn \" \" & (get artist of current track) & \" – \" & (get name of current track)\rend tell\rend if\rreturn \"\"\r"
},
"refreshInterval": 1,
"image": {
"base64": "<KEY>"
}
},
```
If you just want it to pause / replay upon tapping on it, replace yours with the following:
```
// Spotify
{
"type": "appleScriptTitledButton",
"source": {
"inline": "if application \"Spotify\" is running then\rtell application \"Spotify\"\rreturn \" \" & (get artist of current track) & \" – \" & (get name of current track)\rend tell\rend if\rreturn \"\"\r"
},
"action": "appleScript",
"actionAppleScript": {
"inline": "if application \"Spotify\" is running then\rtell application \"Spotify\"\rif player state is playing then\rpause\relse if player state is paused then\rplay\rend if\rend tell\rend if\r"
},
"refreshInterval": 1,
"image": {
"base64": "<KEY>"
}
},
```
username_2: So here's what I have, feel free to frankenstein it and have it so some things do other things. The inline part is what it will do, and the thing above it is the action that causes it. The top one is just showing song/author, top-middle is tap to play/pause, bottom-middle is hold to skip. And the one below idk to be honest but include it.
I'm using a beta of spotify that's optimized for m1, so if you are using the normal version then just remove the '.app' part from all the instances of spotify.
// Spotify
//If playing shows name/artist, when not says tap to play
{
"type": "appleScriptTitledButton",
"source": {
"inline":
"if application \"Spotify.app\" is running then\rtell application \"Spotify.app\"\rif player state is playing then\rreturn (get artist of current track) & \" – \" & (get name of current track)\relse\rreturn \" Tap to Play || \"\rend if\rend tell\rend if\rreturn \"\"\r"
},
//tap to pause/play
"action": "appleScript",
"actionAppleScript": {
"inline":
"if application \"Spotify.app\" is running then\rtell application \"Spotify.app\"\rif player state is playing then\rpause track\relse\rplay\rend if\rend tell\rend if\r"
},
//hold to skip
"longAction": "appleScript",
"longActionAppleScript": {
"inline":
"if application \"Spotify.app\" is running then\rtell application \"Spotify.app\"\rif player state is playing then\rnext track\rend if\rend tell\rend if\r"
},
"refreshInterval": 1,
"image": {
"base64":
"<KEY>"
} |
microsoft-search/pnp-modern-search | 942046721 | Title: Adding Field to Search Filters
Question:
username_0: **Version used**
Ex: 4.2
**Question**
I'm trying to add a field that can be used as a filter in the Search Filters Web part. I created two, just need one to work. The first one that I made called Topic was a lookup field from another SP list. It allows multiple selections. It does not show up as an option to add as a filtered field in the Search Filters Web part. I think this is either because that it is a lookup field and/or because it allows for multiple selections.
To test this out, I created a second field called Topic1. This is a choice field (radio buttons) that only allows one selection. I also told SP to index this field. However, it also does not show up as an option in the Search Filters Web part.
My question would be, how do I get custom Document Library fields to be usable in the Search Filters Web part?
**Screenshots**

**Additional context**
Thanks!
Answers:
username_1: Hi @username_0,
The filters are based on the search results from the `PnP - Search Results` Web Part.
First of all, the `PnP - Search Filters` and `PnP - Search Results` Web Parts must be linked together. Then, if you want to refine a custom field, ensure that your search schema is configured well. Here some links about the SharePoint search schema:
- https://docs.microsoft.com/en-us/sharepoint/manage-search-schema
- https://docs.microsoft.com/en-us/sharepoint/search/search-schema-overview
Basically, once your search schema is configured, you can use the **Debug** view of the `PnP - Search Results` to ensure that your new Refinable/Alias is available in the properties of your items.
Don't forget to specify your _Refinable_ into the **Selected properties** of your `PnP - Search Results` Web Part.
Let us know if this information helps you :)
username_0: I'm probably not understanding all of the steps that may be needed. I added both Topic and Topic1 as managed properties in the Site Collection that are mapped to crawled properties of the same name. I waited a day and they still do show up for the PnP - Search Results properties section nor in the PnP- Search Filters section. When I added these to the site collection search managed properties in SP neither could be selected as "refinable" and maybe that is the problem. That checkbox was grayed out for both Topic and Topic1 managed properties.
username_2: @username_0
Might be this is going to help you.
- Filter will work on data sources like if you're getting result from Libraray1 for example that choice or lookup column should be present inside that library only.
- Don't create a fresh Managed property .Use existing one like in your case **RefinableString00** or some other properties can help that has **refinable enabled by default** and map to crawled property you want to.

- Use that managed property in your filter webpart (pass statically).
- I think that's enough to work.
Let us know if this information helps you :)
username_3: Chasquad,
I had a similar issue, created custom list managed meta data property in my case.
The mapping of ows_fieldname was mapped to managed property in SPO Search Admin at tent level, waited days for update in pnp debug, no result for RefinableString03. I decided to then replicate the mapping in search schema both at tenant level, and site collection administration level, within 15 minutes, pnp debug mode, voila RefinableString03 was showing values....
username_0: I've tried so many different avenues now that I think that I need to take a moment and go back and undo everything that I've tried and then start over. I think that I'll start with a Site Column called Topic and then add that Site Column to a document library and populate that data. Then in both tenant admin and site settings add the same (if possible) RefineableText00 and map that to the Site Column? Will the Site Column show up with ows_<SiteColumnName>? And then let it sit and churn for a few days before even trying to add this to the PNP Search Filter and Results web parts?
username_3: Yes, example of what I did!
Created site columns, and content type in content type hub, but you can create your site column and content type in your local sure sure collection, I prefer central management of everything.
So, managed meta data in my case, created a column called Entity, managed meta data column, pulling data from term store. I then added to content type and library, populated library with data.
Waited a day, then checked crawled property entitled ows-entity, then I had mapped to RefinableString03 example, at tenant level, then also at site collection level.
Next day, I checked search box and searched for entity:then value, also RefinableString03:value.
Try this, and yes I would start again.
username_2: It is as simple as that I have mentioned.
Create on Tenant level and also reindex your search on site collection level under search offline option.
Same problem I was facing and it work on first go. 😊
username_3: The only issue I have, not sure if you come across. You're aware title is not sortable, therefore I tried to map ows_title to a RefinableStringXX managed property, reindex, crawled, waited, nothing.....no joy.
It's all good having pnp search to build solutions, but business need to order by title in ascending...
username_0: I've removed everything, that I can remember, that I tried out. I then created a new Site Column (different name than before), add it to the Document Library, populated that data, added the RefineableString01 to both the site collection and tenant search, and initiated the reindexing of the site collection search. Now I wait until Monday to check it out!
Hopefully this work and I thank everyone for your help!
username_0: This worked. Thanks so much!
Status: Issue closed
|
andywer/threads.js | 146798055 | Title: Returning an array from thread
Question:
username_0: Hi @username_1!
An issue actually related with https://github.com/audreyt/node-webworker-threads/issues/99.
I tried both [webworker-threads](https://npmjs.org/package/webworker-threads) and [threads](https://npmjs.org/package/threads) libs to implement multithreading in [nogl-shader-output](https://npmjs.org/package/nogl-shader-output), and both packages get very slow at returning an array from threads. It takes 800ms on my laptop to return 300×300 array from the thread.
Whether something I don’t know or is there a way to speed things up?
Thank you.
Answers:
username_1: Hey Deema!
Did you use https://github.com/username_1/threads.js#transferable-objects? Passing large arrays is by default very slow, since the whole array is copied when sending messages between the threads. But you can explicitly transfer the ownership of an array to the other thread when sending data to it (see link). This is much faster, but the sending thread cannot access the array anymore until the other thread gives the array ownership back.
username_1: I forgot to ask: Are you working in a browser or a node.js setting?
username_0: @username_1 in node.
username_1: Yeah, my suggested approach was also browser-related... I don't have a Node.js solution by hand right now. The best I could come up with when googling it was this: https://github.com/audreyt/node-webworker-threads/issues/58
But that issue seems quite dead...
Status: Issue closed
username_1: Closing for now.
This is a rather tough one. Might be easy to fix once node supports web workers (nodejs/node#13143). |
jupyter-widgets/ipywidgets | 345563621 | Title: Developer install leads to installing non-dev jupyter notebook
Question:
username_0: I was following `dev_install.md` to create the ipywidgets environment. This creates a jupyter notebook version like 5.6.0, as opposed to 5.7.0.dev0 (like seen in contributing doc for the notebook project https://github.com/jupyter/notebook/blob/master/CONTRIBUTING.rst).
If this is fine, can the developer install doc indicate that? The current second bullet point makes it sound like we should be using our own build of jupyter notebook for developing ipywidgets. Maybe I'm wrong, but I was hoping using a local build will make it for new contributors like me to trace where issues lie through browser debuggers. |
lbryio/lbry-desktop | 770869277 | Title: "Mature" content is being censored
Question:
username_0: I noticed that content which is somehow labeled as 'mature' does not show up on my channel unless people activate the option to display mature content in their profile settings.
Effectively those videos are being censored because they don't even show up in the channel list of videos. It would be better to still show a thumbnail for those videos with a placeholder image and the video title, and a text that says that it contains mature content. That way, people would still see that the video exists and could click on it to find the link to enable mature content in their profile.
By hiding the videos completely people never know it exists.
I also don't see why my videos are labeled as 'mature'. Here is an example https://lbry.tv/@
Answers:
username_1: The sex tag is treated as mature content, you can remove it if the content isn't mature. Any of the pink tags will do that.
We won't be changing how this works as users need to enable mature content before they are able to see it. There's a separate issue filed that should alert them when there is mature content but it's not displayed.
Status: Issue closed
username_0: Thanks for the details. But why is just the tag 'sex' being considered as mature content? Don't people below 18 years also get sex education in school? |
UnquietCode/LogMachine | 69737493 | Title: LOG-3: Add option to synchronize LogMachine output to music.
Question:
username_0: Modify the LogMachine output (the ajax version) to accept an incoming audio stream, or integrate with a media player and echo nest api, then synchronize the log output to the bpm of the song currently playing. In fact, all of that can be handled with the Python api. There does exist a Java api now too.
I think doing this in javascript might be good, because then it's just a matter of making the AJAX queries in a time-synchronized fashion. That should be trivial.
--------------------------------------------------
Imported from JIRA
Originally reported by: [username_0](https://github.com/username_0) |
bblanchon/ArduinoJson | 172642289 | Title: json string variable type error
Question:
username_0: I use ARDUINO IDE 1.6.11 from arduino.cc
I met errors in my program.
I can show these errors by JsonParserExample (included in library)
char json[] = "{\"sensor\":\"gps\",\"time\":1351824120,\"data\":[48.756080,2.302038]}"; -->>original.
signed char json[] = "{\"sensor\":\"gps\",\"time\":1351824120,\"data\":[48.756080,2.302038]}"; -->>error.
unsigned char json[] = "{\"sensor\":\"gps\",\"time\":1351824120,\"data\":[48.756080,2.302038]}"; -->>error.
uint8_t json[] = "{\"sensor\":\"gps\",\"time\":1351824120,\"data\":[48.756080,2.302038]}"; -->>error.
int8_t json[] = "{\"sensor\":\"gps\",\"time\":1351824120,\"data\":[48.756080,2.302038]}"; -->>error.
change time key value.
char json[] = "{\"sensor\":\"gps\",\"time\":135,\"data\":[48.756080,2.302038]}";
and change time variable. long to char.
unsigned char time = root["time"]; -> OK.
signed char time = root["time"]; -> OK.
uint8_t time = root["time"]; -> OK.
int8_t time = root["time"]; -> OK.
but..
char time = root["time"]; -> error.
I'm poor English. I beginning programming.
I don't know why I met these errors.
Answers:
username_1: `char`, `signed char` and `unsigned char` are three different types in C++.
A string literal is an array of `char`, not `signed char` or `unsigned char`.
That's why you have an error when changing the type of the variable `json`.
The second issue is a feature of ArduinoJson.
The `char` type has been intentionally disabled to avoid an ambiguity.
Imagine you have this input `{'time':'8'}`, was should be the value of `root["time"]` as a `char`?
I see two possible answers.
Either you see `char` as an integer; in that case, the result should be `8`.
Or you see `char` as a character; in that case, the result should be `56` (the ASCII code of character `8`)
I highly encourage you to use `char` to store characters, and `int8_t` or `uint8_t` to store integers.
Status: Issue closed
username_1: Note: [this was added to the FAQ](https://github.com/username_1/ArduinoJson/wiki/FAQ#why-jsonvariant-cannot-be-converted-to-char) |
allista/hangar | 577272002 | Title: Compatibility problem with Aircraft Carrier Accessories mod
Question:
username_0: This incompatibility is particularly problematic because the Aircraft Carrier Accessories mod recommends and expects users to use this mod to store aircraft on carriers.
The "Aircraft Tailhook2" part, and presumably other tail hook parts from that mod, do not work on vessels spawned using this mod. Using the "deploy" function of such a part changes the label of the menu item as though the hook is deployed, but otherwise has no effect. Quicksaving and quickloading causes the affected parts to again work as expected.
In the installation in which this issue occurs, tail hook parts have been verified to work as expected on vessels not spawned by the Hangar mod.
#### Log:
[KSP.log](https://github.com/allista/hangar/files/4301338/KSP.log)
This log file is somewhat cluttered with other events. If it would greatly help, I can perform a test on an installation only containing the two mods involved in this issue. |
Akhtar-Joomun/awesome-books | 945522820 | Title: Morning session peer to peer code review
Question:
username_0: Hey Akhtar. Great job on this project and nice implementation of the feature as were stated by the project description. Everything seemed perfect so we didn't feel the need to suggest anything! Great job again. |
antleaf/jct-user-stories | 640858858 | Title: As a RESEARCHER, I want to be able to get search results for journal names with special characters whether I’m using a special keyboard, inserting special characters, or not using special characters.
Answers:
username_1: @paulwalk @richard-jones I've discussed this on other projects before, there are some things to consider as it is not a straightforward decision. Supporting search results for special characters is not hard in itself, and neither is supporting search without special characters, but supporting either OR both requires a decision about what experience is intended for the end user, and what the service wants to deliver. See below:
* in the default case, we store the journal name that we find in the world, e.g. whatever it is on crossref, or other sources we pull from. Whatever that journal name is, is the one a user could find by typing it in. If it contains an umlaut for example, and the user types an umlaut, they would get back that journal name in an autocomplete. For users who do use such languages, this would be the best match for what the user actually wants. But for users who do not use that language and who type the name without an umlaut, they would not get the result back (because, really, they are mis-spelling the title).
* a user who does use special characters, for example in a language that uses umlauts in journal title names, and using a keyboard that supports those characters (or by manually inserting them) would not necessarily expect to see results that do not include those special characters - e.g. if the journal name contains a special character and the user types that character, do they really want to be shown journal names that are NOT what they asked for?
* if we do want to support users who might expect search results to return journals with an umlaut in the title even if they did not type an umlaut in the title, is supporting these users more important than providing a lower quality search experience for users who are typing the journal name correctly, as in the case described above?
* if we do want to support matching on characters with umlauts (for example) and also on the same characters without the umlaut, then yes, we can do that, once a decision is made in relation to the caveats above. However, we do need to know in advance, as it requires extra pre-processing of the data. It is not a lot of extra work to do, but would need to be considered in scheduling.
* keep in mind that even if we did do asciifold or similar to allow these sorts of matches, it is still possible that ANY user, whether typing languages with special characters or not, may mistype or misspell a journal title. So the result we really want for the user experience may better be supported by fuzzy matching regardless of character types, which is a query-time optimisation, rather than a data pre-processing task.
username_2: @username_1 when you've discussed this on other projects has it included people whose native languages use special characters and what has been the outcome of the discussions on other projects?
Would fuzzy matching operate within the autocomplete suggestions and incorporate mistypes and special character usage (or not) within the results returned?
username_1: @username_2 I've provided similar technical points but haven't been involved in the decisions of those projects - one of them was DOAJ though, I think, so @richard-jones may know what they chose for their user base, which does have some users with native languages that contain special characters.
About fuzzy matching, there's an issue that already relates to it, #36 which is in future requirements. I'd need to check if fuzzy matches would incorporate special characters, but I believe it would.
Note, there will always be some limit to what "special characters" we can match. If we are talking about things like umlauts in human languages, that sort of thing should be fine. But if we are talking about things like special mathematical characters that certain users may use in the titles of things, we probably can't match those, because they usually get mangled during the ingest process before reaching us, e.g. crossref etc don't get them accurately from the publishers that submit them. However, this is something I've seen at article title level rather than at journal title level, but I'd guess there could be some obscure mathematical journal in the world somewhere that uses special characters that we will never be able to match.
I think the best solution for now is that when we know a journal title contains special characters (because we receive it that way from crossref for example) then we will store it that way. If a user then searches with the special characters, we will match it. Also, if the UI supports autocompletion then while the user is typing they will see possible titles appear anyway, and this may aid in them being able to select the correct one (and this also aids in minimising typing/spelling mistakes). Later, in future requirements, we can also add fuzzy matching if feedback from the first release indicates that users are too often having trouble finding the correct journal. Does that sound suitable to you?
username_2: @username_1 that seems like a logical suggestion. Would be good to know @richard-jones what DOAJ went with. |
sbt/sbt | 361462985 | Title: NullPointerException from okhttp3 during resolution
Question:
username_0: I still get this error on v1.2.1.
1) Clean out my ivy cache
2) Clone https://github.com/apache/incubator-toree
3) Run make release
I get
```[error] sbt.librarymanagement.ResolveException: unresolved dependency: com.fasterxml.jackson.core#jackson-databind;2.6.5: Resolution failed several times for dependency: com.fasterxml.jackson.core#jackson-databind;2.6.5 {test=[default(compile)]}::
[error] Resolution failed several times for dependency: com.fasterxml.jackson#jackson-parent;2.6.2 {}::
[error] Resolution failed several times for dependency: com.fasterxml#oss-parent;24 {}::
[error] java.lang.NullPointerException at okhttp3.internal.http.RetryAndFollowUpInterceptor.followUpRequest(RetryAndFollowUpInterceptor.java:278)
[error] java.lang.NullPointerException at okhttp3.internal.http.RetryAndFollowUpInterceptor.followUpRequest(RetryAndFollowUpInterceptor.java:278)
[error] java.lang.NullPointerException at okhttp3.internal.http.RetryAndFollowUpInterceptor.followUpRequest(RetryAndFollowUpInterceptor.java:278)
[error] java.lang.NullPointerException at okhttp3.internal.http.RetryAndFollowUpInterceptor.followUpRequest(RetryAndFollowUpInterceptor.java:278)
[error]
[error] java.lang.NullPointerException at okhttp3.internal.http.RetryAndFollowUpInterceptor.followUpRequest(RetryAndFollowUpInterceptor.java:278)
[error] java.lang.NullPointerException at okhttp3.internal.http.RetryAndFollowUpInterceptor.followUpRequest(RetryAndFollowUpInterceptor.java:278)
[error] java.lang.NullPointerException at okhttp3.internal.http.RetryAndFollowUpInterceptor.followUpRequest(RetryAndFollowUpInterceptor.java:278)
[error]
[error] java.lang.NullPointerException at okhttp3.internal.http.RetryAndFollowUpInterceptor.followUpRequest(RetryAndFollowUpInterceptor.java:278)
[error] java.lang.NullPointerException at okhttp3.internal.http.RetryAndFollowUpInterceptor.followUpRequest(RetryAndFollowUpInterceptor.java:278)
[error] java.lang.NullPointerException at okhttp3.internal.http.RetryAndFollowUpInterceptor.followUpRequest(RetryAndFollowUpInterceptor.java:278)
[error]
```
when I turn on debug, I see this
```
^[[0m[^[[0m^[[0mdebug^[[0m] ^[[0m^[[0m tried /Users/gxc772/.ivy2/local/com.fasterxml/oss-parent/24/jars/oss-parent.jar^[[0m
^[[0m[^[[0m^[[0mdebug^[[0m] ^[[0m^[[0m tried https://repo1.maven.org/maven2/com/fasterxml/oss-parent/24/oss-parent-24.jar^[[0m
^[[0m[^[[0m^[[0mdebug^[[0m] ^[[0m^[[0mCLIENT ERROR: Not Found url=https://repo1.maven.org/maven2/com/fasterxml/oss-parent/24/oss-parent-24.jar^[[0m
^[[0m[^[[0m^[[0mdebug^[[0m] ^[[0m^[[0m tried http://repository.apache.org/snapshots/com/fasterxml/oss-parent/24/oss-parent-24.jar^[[0m
^[[0m[^[[0m^[[0mdebug^[[0m] ^[[0m^[[0mproblem occurred while resolving dependency: com.fasterxml#oss-parent;24 {} with public: java.lang.NullPointerException^[[0m
^[[0m[^[[0m^[[0mdebug^[[0m] ^[[0m^[[0m at okhttp3.internal.http.RetryAndFollowUpInterceptor.followUpRequest(RetryAndFollowUpInterceptor.java:278)^[[0m
```
SBT is tried to download the jar for a POM type module. Obviously, POM modules don't have jars, so it tries every repository until the last one and gives this exception
_Originally posted by @username_1 in https://github.com/sbt/sbt/issues/3519#issuecomment-422498579_
Answers:
username_1: Is anyone else able to reproduce this?
username_0: This worked for me on my machine.
username_2: Reproduced on a CentOS 7, sbt-1.2.8, behind a proxy.
Workaround: added `-Dsbt.gigahorse=false` to by sbt options.
username_0: I'm going to preemptively close this since sbt 1.3.0 uses Coursier internally.
Status: Issue closed
|
gatsbyjs/gatsby | 569739797 | Title: Migrate to import * as React from 'react' and remove commonjs interop import
Question:
username_0: # Description
Reactjs has merged https://github.com/facebook/react/pull/18102 and are moving into es module exports so the correct way of importing react will be `import * as React from 'react'`, they will keep the default exports to keep everything working as expecting.
We should update our documentation and internal code to move to the new import statement.
### Expected result
All react imports should be `import * as React from 'react'`
### Actual result
All react imports should be `import React from 'react'`
More info:
- https://github.com/facebook/react/pull/18102
- https://github.com/facebook/react/pull/18106
- https://github.com/facebook/react/pull/18094
Answers:
username_1: # Description
Reactjs has merged https://github.com/facebook/react/pull/18102 and are moving into es module exports so the correct way of importing react will be `import * as React from 'react'`, they will keep the default exports to keep everything working as expecting.
We should update our documentation and internal code to move to the new import statement.
### Expected result
All react imports should be `import * as React from 'react'`
### Actual result
All react imports should be `import React from 'react'`
More info:
- https://github.com/facebook/react/pull/18102
- https://github.com/facebook/react/pull/18106
- https://github.com/facebook/react/pull/18094
username_2: Hi I'm new to the Gatsby repo. This seems like a good first issue to work on and get started here.
I see the first attempt at this issue had multiple test failures. Can you guide me what directories require to be migrated and why the early tests failed?
Status: Issue closed
|
test-kitchen/kitchen-ec2 | 165976946 | Title: "Waiting for SSH Service" loop Error(Ubuntu Image)!!
Question:
username_0: Hi Im was trying to create a ubuntu instance using kitchen-ec2.
It could successfully create EC2 instance. But after that it is getting into a loop like this,
`
-----> Starting Kitchen (v1.8.0)
-----> Creating <default-ubuntu-1604>...
Detected platform: ubuntu version 16.04 on x86_64. Instance Type: t2.micro. Default username: ubuntu (default).
If you are not using an account that qualifies under the AWS
free-tier, you may be charged to run these suites. The charge
should be minimal, but neither Test Kitchen nor its maintainers
are responsible for your incurred costs.
Instance <i-00fdc6c27c54f82cd> requested.
Polling AWS for existence, attempt 0...
Attempting to tag the instance, 0 retries
EC2 instance <i-00fdc6c27c54f82cd> created.
Waited 0/300s for instance <i-00fdc6c27c54f82cd> to become ready.
Waited 5/300s for instance <i-00fdc6c27c54f82cd> to become ready.
Waited 10/300s for instance <i-00fdc6c27c54f82cd> to become ready.
Waited 15/300s for instance <i-00fdc6c27c54f82cd> to become ready.
Waited 20/300s for instance <i-00fdc6c27c54f82cd> to become ready.
Waited 25/300s for instance <i-00fdc6c27c54f82cd> to become ready.
EC2 instance <i-00fdc6c27c54f82cd> ready.
Waiting for SSH service on 10.11.19.67:22, retrying in 3 seconds
Waiting for SSH service on 10.11.19.67:22, retrying in 3 seconds
Waiting for SSH service on 10.11.19.67:22, retrying in 3 seconds
Waiting for SSH service on 10.11.19.67:22, retrying in 3 seconds
Waiting for SSH service on 10.11.19.67:22, retrying in 3 seconds
Waiting for SSH service on 10.11.19.67:22, retrying in 3 seconds
Waiting for SSH service on 10.11.19.67:22, retrying in 3 seconds
Waiting for SSH service on 10.11.19.67:22, retrying in 3 seconds
Waiting for SSH service on 10.11.19.67:22, retrying in 3 seconds`
Providing the my kitchen.yml Here,
[kitchen.txt](https://github.com/test-kitchen/kitchen-ec2/files/367854/kitchen.txt)
Answers:
username_1: Based on those IPs it's likely because you aren't associating the public IP which means you'd need to be in the same subnet/VPC on EC2 for that to work. That or the security group doesn't have the right allowances.
username_0: @username_1 thanks for the response.
The subnet i have provided is a private subnet. And its reachable through our hosted VPN. And also we have a NAT configured for private subnet to talk to internet.
i could manually SSH to the instance through ssh -i "sysadmin.pem" [email protected].
username_1: I haven't been able to repro this as an issue - are you still having trouble with this?
username_2: I'm having the same problem, and I think I've tracked it down.
I have this in my .kitchen.yml:
```yaml
transport:
ssh_key: /secrets/kitchen.pem
```
But but the debug logs show this:
`D [SSH] opening connection to [email protected]<{:user_known_hosts_file=>"/dev/null", :paranoid=>false, :port=>22, :compression=>false, :compression_level=>0, :keepalive=>true, :keepalive_interval=>60, :timeout=>15, :keys_only=>true, :keys=>["/builds/cookbooks/nulogy-rabbitmq/.kitchen/docker_id_rsa"], :auth_methods=>["publickey"], :forward_agent=>true, :user=>"ubuntu"}>
`
Specifically:
`:keys=>["/builds/cookbooks/nulogy-rabbitmq/.kitchen/docker_id_rsa"]`
So it doesn't seem to be using the key that I specified, and fails to authenticate.
Any idea why it's ignoring my transport setting?
username_3: @tescalada, you can do `kitchen create -l debug` and it would output more stuff to debug the SSH connection.
username_4: detailed log:
D [SSH] connection failed (#<Net::SSH::AuthenticationFailed: Authentication failed for user [email protected]>)
Waiting for SSH service on 127.0.0.1:2222, retrying in 3 seconds
D [SSH] shutting previous connection [email protected]<{:user_known_hosts_file=>"/dev/null", :paranoid=>false, :port=>"2222", :com
pression=>false, :compression_level=>0, :keepalive=>true, :keepalive_interval=>60, :timeout=>15, :keys_only=>true, :keys=>["/c/ForTesting/motd/.kitchen/kitchen-vagrant/default-ubuntu-1604/.vagrant/machines/default/virtualbox/private_key"]
, :auth_methods=>["publickey"], :logger=>#<Logger:0xcb9ffd0 @level=4, @progname=nil, @default_formatter=#<Logger::Formatter:0xcb9ff88 @
datetime_format=nil>, @formatter=nil, @logdev=#<Logger::LogDevice:0xcb9ff58 @shift_period_suffix=nil, @shift_size=nil, @shift_age=nil,
@filename=nil, @dev=#<IO:<STDERR>>, @mon_owner=nil, @mon_count=0, @mon_mutex=#<Thread::Mutex:0xcb9ff10>>>, :password_prompt=>#<Net::SSH
::Prompt:0xcb9fef8>, :user=>"vagrant"}>
this is going in a loop
username_5: @username_0 like @username_1 mentioned the behavior you're seeing sometimes occurs when:
- the security group is misconfigured (maybe you selected an SG not part of the subnet/vpc you're launching in?)
- the ssh key you're attempting to use isn't available locally or the path is incorrect (try to eliminate shortcuts like ~ in your path statement in .kitchen.yml)
- the username is not set properly for the distribution you're using (I believe by default kitchen-ec2 uses ec2-user but for distributions like ubuntu you'll need to adjust transport.username to match ubuntu)
You might want to try to adjust your transport section to look like:
```
transport:
ssh_key: /home/someuser/someuser.pem
username: ubuntu
```
username_1: Closing this one out as it's a few "Hop Ons" which may or may not be related (notable one of the logs points to kitchen-vagrant which is another driver entirely). The advice given here covers how to troubleshoot this problem but this isn't a bug or even an active question any longer.
Status: Issue closed
|
rhymesg/Particle_Filter | 523276824 | Title: Fig. 6. Time history of covariance for Case 1.
Question:
username_0: I want to reproduce Fig. 6. Time history of covariance for Case 1.
I run the original code to get results for case1 setting and I used 1000 particles, only for PF for simplicity. But my result is different than the result provided in Fig6 in the paper. In Fig6 covariance is small but in my result covariance is not small and there are fluctuations, i.e., covariance is not decreasing by the simulation time. How can i reproduce the original result presented in Fig6?
Answers:
username_1: Thank you for showing interest in the paper. I'm afraid I would not be able to run the code today. But I guess if the covariance is big and the RMSE is not, it is possibly because of the filter setting:
`% filter setting`
`sig_init = 40; % std for initial P`
`sig_proc = 8; % std for Q`
`sig_meas = 25; % std for R`
Tuning the filter by manipulating these values would help. I guess decreasing sig_proc to 4 or something would give better results. Judging from `sig_v = 2; % std of actual process noise` I think the uploaded code is to run Case 2.
username_1: Also, I found that I omitted the DEM resolution setting at the uploaded code. I hope setting like the following works for you.
`%% setting`
`load('../DEM/DB_part.mat')`
`DEM.resolution = 30;`
username_0: Thank you very much for your quick response. Setting sig_proc to 4 worked for me, now time history of covariance for Case 1 is decreasing.
DEM.resolution = 30; is also loaded by 'DB_part.mat', it is not ommitted, it works fine.
Status: Issue closed
|
pythonindia/inpycon-blog | 880889189 | Title: Typo in Writing a proposal for PyCon India? Our Mentorship team is here to help! blog post
Question:
username_0: * **Do you want to request a *feature* or report a *bug*?**
Bug
Blog post link: https://in.pycon.org/blog/2021/announce-mentorship.html
The Reviewing Proposals section of the blog post
```
If you don't get any response in 48 hours, please send a ollow up message to #wg/mentorship on Zulip.
```
it should be
```
If you don't get any response in 48 hours, please send a follow up message to #wg/mentorship on Zulip.
```
Answers:
username_1: Thanks @username_0 for pointing this out. Please do create a PR for patching this up if you can.
username_0: @username_1 sure will do that
Status: Issue closed
username_1: Closing in favor of #369 |
isee4xai/iSeeOnto | 1139311500 | Title: Explainer Ontology Data Type
Question:
username_0: The Explainer ontology has a Data Type class, with subclasses for Categorical, Image, Numerical, Ordinal, and instances with the same names. This could be confusing, so we should decide which to use and remove those that are not needed. |
nestjs/graphql | 665534255 | Title: Nested Types Not Converted in Type Mapping
Question:
username_0: <!--
PLEASE HELP US PROCESS GITHUB ISSUES FASTER BY PROVIDING THE FOLLOWING INFORMATION.
ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION.
-->
## I'm submitting a...
<!--
Please search GitHub for a similar issue or PR before submitting.
Check one of the following options with "x" -->
<pre><code>
[ ] Regression <!--(a behavior that used to work and stopped working in a new release)-->
[*] Bug report
[ ] Feature request
[ ] Documentation issue or request
</code></pre>
## Code
**Entity:**
```
@ObjectType()
@Entity()
export class Order extends BaseEntity {
@Field(type => Point)
@ValidateNested()
@Column('point')
startPoint: Point;
@Field(type => Point)
@ValidateNested()
@Column('point')
endPoint: Point;
@Field(type => Int)
@IsInt()
@IsPositive()
@Column({ default: null })
deliveryCharges: number;
}
```
**Point Interface:**
```
@ObjectType()
export class Point {
@IsNumber()
x: number;
@IsNumber()
y: number;
}
```
[Truncated]
<!-- Describe how the issue manifests. -->
The scalar properties are converted correctly into InputType but the nested PointType is not converted, Giving:
Error: Cannot determine a GraphQL input type for the "startPoint".
## Expected behavior
<!-- Describe what the desired behavior would be. -->
The Point Type should also be Converted
## Environment
<pre><code>
Nest version: 7.4.1
<!-- Check whether this is still an issue in the most recent Nest version -->
For Tooling issues:
- Node version: 13.8 <!-- run `node --version` -->
- Platform: Linux <!-- Mac, Linux, Windows -->
Answers:
username_1: Please provide a minimum reproduction repository.
username_0: Here you go:
https://github.com/username_0/nestjsGraphQLIssue
username_1: I've pulled your repo and I don't think this is an issue with the schema generator. `Toy` class is decorated as an `ObjectType` and - since it's being used by `Cat` which is an `ObjectType` as well - we cannot simply convert `Toy` to an input type as this would break the `Cat` type. Likewise, we cannot generate another `Toy` input type based on the object type as this would lead to the following error: `Schema must contain uniquely named types but contains multiple types named "Toy"`. To fix this, you should redefine this property and the `Toy` type using a different name to make sure these types won't collide with each other.
Status: Issue closed
|
kubesphere/ks-installer | 482299021 | Title: 阿里云ubuntu16.04安装kubesphere遇到问题
Question:
username_0: 阿里云环境 ubuntu 16.04.6,多节点离线部署问题:
1、由于阿里云环境的安全组中有些端口没有开放,导致push镜像超时
```
The push refers to repository [172.16.31.10:5000/istio/mixer]
Get http://172.16.31.10:5000/v2/: net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)
解决方法:在安全组中把需要的端口打开即可,比如镜像仓库5000。
```
2、由于脚本中考虑的ubuntu的版本为16.04.4和16.04.5,没有16.04.6版本iso包。
```
E: Failed to fetch file:/kubeinstaller/apt_repo/16.04.6/iso/Packages File not found - /kubeinstaller/apt_repo/16.04.6/iso/Packages (2: No such file or directory)
E: Some index files failed to download. They have been ignored, or old ones used instead
Some packages could not be installed. This may mean that you have\nrequested an impossible situation or if you are using the unstable\ndistribution that some required packages have not yet been created\nor been moved out of Incoming.\nThe following information may help to resolve the situation:\n\nThe following packages have unmet dependencies:\n software-properties-common : Depends: python3-software-properties (= 0.96.20.8) but it is not going to be installed\n",
解决方法:重新制作了16.04.6版本iso包已解决,后面依赖包冲突在新的包中也解决了。
包的下载地址为:https://kubesphere-installer.pek3b.qingstor.com/offline/ubuntu-16.04.6-server-amd64.iso
```
Answers:
username_0: 3、pyopenssl库的问题
```
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: AttributeError: 'module' object has no attribute 'SSL_ST_INIT'
fatal: [master]: FAILED! => {"attempts": 5, "changed": false, "module_stderr": "Traceback (most recent call last):\n File \"<stdin>\", line 113, in <module>\n File \"<stdin>\", line 105, in _ansiballz_main\n File \"<stdin>\", line 48, in invoke_module\n File \"/tmp/ansible_apt_payload_eT0UnG/__main__.py\", line 269, in <module>\n File \"/tmp/ansible_apt_payload_eT0UnG/ansible_apt_payload.zip/ansible/module_utils/urls.py\", line 95, in <module>\n File \"/usr/local/lib/python2.7/dist-packages/urllib3/contrib/pyopenssl.py\", line 46, in <module>\n import OpenSSL.SSL\n File \"/usr/lib/python2.7/dist-packages/OpenSSL/__init__.py\", line 8, in <module>\n from OpenSSL import rand, crypto, SSL\n File \"/usr/lib/python2.7/dist-packages/OpenSSL/SSL.py\", line 118, in <module>\n SSL_ST_INIT = _lib.SSL_ST_INIT\nAttributeError: 'module' object has no attribute 'SSL_ST_INIT'\n", "module_stdout": "", "msg": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 1}
解决方法:
rm -rf /usr/lib/python2.7/dist-packages/OpenSSL
rm -rf /usr/lib/python2.7/dist-packages/pyOpenSSL-0.15.1.egg-info
sudo pip install pyopenssl
``` |
portainer/portainer | 337393246 | Title: Inconsistent action buttons/links in task/container lists and task detail page
Question:
username_0: <!--
Thanks for reporting a bug for Portainer !
Do you need help or have a question? Come chat with us on Slack http://portainer.io/slack/ or gitter https://gitter.im/portainer/Lobby.
Before opening a new issue, make sure that we do not have any duplicates
already open. You can ensure this by searching the issue list for this
repository. If there is a duplicate, please close your issue and add a comment
to the existing issue instead.
Also, be sure to check our FAQ and documentation first: https://portainer.readthedocs.io
-->
**Bug description**
The action links (logs, console, info, stats, inspect) are not consistently presented in the various task/container lists. For example:
- The containers page has all four action icons, even for stopped containers where "stats" and "console" throw an error that complains "container... is not running" and "inspect" throws "unable to inspect the container"
- The nested service task lists show a "logs" action icon for "ready" and "starting" tasks, but it is not clickable and the cursor does not change to a pointer.
- The nested service task lists never show "stats" or "inspect" action icons, even though those actions are available if I click through to the task detail page.
- The nested service task lists do not show a "console" action icon for "starting" tasks. These are generating logs and they are running. They are just waiting for the first health check to pass, which could be several minutes after a task is started. I can click through to the task detail page then click the "console" action.
**Expected behavior**
Everywhere I see a task/container in a list, I should see all currently active action buttons/links and they should work consistently. Any actions buttons/links that are not available should either not be visible at all, or should be shown but greyed out and clearly disabled.
Answers:
username_0: Also, the task list on the service detail page still shows the text style action links, instead of just the icon style that is used in most/all other task lists.
username_1: I would like to work on this.
username_2: @username_1 feel free to open a PR :)
Status: Issue closed
|
pac4j/play-pac4j | 86108109 | Title: Upgrade to pac4j 1.8
Question:
username_0: Yes, I'll start that next week.
Answers:
username_1: Hi @username_0 , is there an ETA on this upgrade?
username_0: Yes, I'll start that next week.
username_1: Cool!
username_1: Any news on this @username_0 ?
username_0: Yes, I'm working on it. I've made a huge refactoring of the library. It's almost done, I need to review everything again, test a little more and write the full documentation. Hope it will be ready for the end of this week or the beginning of the next one.
username_1: That's great news!
username_0: PR: https://github.com/pac4j/play-pac4j/pull/73
username_0: @username_1 : any feedback will be appreciated on the new v2 (#73), demos have been updated as well: https://github.com/pac4j/play-pac4j-java-demo/pull/19 & https://github.com/pac4j/play-pac4j-scala-demo/pull/18
Status: Issue closed
username_0: PR: https://github.com/pac4j/play-pac4j/pull/77
Status: Issue closed
|
tylernguyen/x1c6-hackintosh | 869916012 | Title: Screen backlight don't work
Question:
username_0: On apple logo appears screen backlight don't work, I hardly see the logo and sistem have been booted, but brightness is very very low. Does anyone know what's the problem?
Answers:
username_1: @username_0
Sorry for the late reply,
Can you provide a picture? perhaps a video?
I would also need a copy of your EFI. and settings. Your particular problem seems unique.
username_2: I have same issue since last commit
username_1: It's like related to https://github.com/username_1/x1c6-hackintosh/issues/129
Try the newest commit and let me know.
username_3: @username_1 I can confirm that backlight control is still not working.
username_1: @username_3
Please reset your EFI to the repo and clear CMOS. My backlight is functioning normally.
If not,
Please provide a copy of your EFI and appropriate boot and OpenCore logs
username_4: I also have the problem in The latest update.
username_1: I can't diagnose the problem myself because backlight control is working on my personal machine.
Please let me know the last working commit.
username_1: @username_4 @username_3 @username_0 @username_2
Please try this patch and let me know if that fixes your problem.
[SSDT-PNLF.aml.zip](https://github.com/username_1/x1c6-hackintosh/files/6543500/SSDT-PNLF.aml.zip)
username_4: <img width="300" alt="image" src="https://user-images.githubusercontent.com/68903281/119630833-e8a1c680-be41-11eb-98e8-e2494a5aca5d.png">
It still doesn't work.
username_1: Do you have hidpi patch enabled?
username_4: I just use this tool, the link is https://github.com/xzhih/one-key-hidpi, no others patches are applied.
username_5: Not sure if it is connected, but my ability to control has also disappeared. Hotkeys don't do anything despite Karabiner recording backlight-increment/decrement. And on the display settings no brightness slider is present
I also have used the same script as username_4 in the past.

username_4: <img width="868" alt="image" src="https://user-images.githubusercontent.com/68903281/120878073-a5d3b180-c5ec-11eb-9c56-bc85c6b8ab92.png">
After adding this ACPI section, Everything gets fine for me. @username_1
username_4: @username_1 The latest SSDT-PNLF.aml still has problem for adjusting brightness, But the last SSDT-PNLF.aml works well. The OpenCore Resource directory path changed in OpenCore 0.7, but latest update doesn't update this directory, I change this for myself.
username_1: @username_4
Did you remove AppleBacklightSmoother?
username_6: Yes,I remove it.
username_4: @username_1 Yes, I do removed AppleBacklightSmoother
username_1: Hm..
Can I get a copy of your EFI?
username_4: @username_1 Of course, The link is: https://github.com/username_4/x1-carbon-gen6.git. I modified some files and configurations base on your repository, if you get this copy and need me to delete this repository , just tell me please.
thank you.
By the way, I solved the HDMI output(two displays) and black screen problems for myself by change some parameters. I use this laptop to do my daily work. It's very good at present. thank you for your perfect project.


username_1: @username_4
Take a look at the latest commit and let me know.
username_4: Okay, I will verify this update if I have free time.
username_4: I update to the latest commit, It has no problem now.
Status: Issue closed
|
tel/saltine | 552623428 | Title: Adding support for generichash (Blake2b)
Question:
username_0: I would like to add `generichash` support to this library via binding to libsodium's `crypto_generichash` - would you be willing to review/give feedback on a pull request adding this binding? I am developing a haskell library for [Paseto tokens](https://paseto.io) and need a [binding to Blake2b](https://github.com/paragonie/paseto/blob/master/docs/01-Protocol-Versions/Version2.md#encrypt).
Answers:
username_1: Thank you very much for the PR, I will take a look at this next week, once my workload has dropped to reasonable levels.
username_0: @username_1 Friendly reminder to please take a look at the pull request.
username_1: Sorry for the delay, I promise I'll get to it today.
username_1: Merged.
Status: Issue closed
|
forumthemer/eles-phpbb-forum-theme | 308261086 | Title: got warning
Question:
username_0: [phpBB Debug] PHP Warning: in file [ROOT]/phpbb/di/container_builder.php on line 146: require(./cache/production/autoload_4335734bbdd20f586549a504dff5f80c.php): failed to open stream: No such file or directory
Every time i login with moderator access this thing happen,
This never happen before using else theme.
I checked the path and i found out that the file autoload_4335734bbdd20f586549a504dff5f80c.php is missing on the directory, for the mean time solution i re-uploaded the file, but everytime i loging again with mod access it happen again...
any solution for this? what went wrong?
Answers:
username_1: You have to turn off DEV MODE from the coreframe panel :)
That's why it is advised in the documentation to disable it in production
Status: Issue closed
|
elastic/elasticsearch-net | 109132068 | Title: No overload for method 'Map' takes 0 arguments
Question:
username_0: According to the NEST documentation, you can persist attribute based mapping like this:
* var response = this.ConnectedClient.Map<ElasticSearchProject>();
(see: http://nest.azurewebsites.net/nest/indices/put-mapping.html)
However, I'm getting an error saying: No overload for method 'Map' takes 0 arguments
Answers:
username_1: I just think the documentation needs to be updated.
Status: Issue closed
username_2: ++ Several documentation pages are horribly lacking behind. We are working on a solution for 2.0 where all our code examples are actual code that has to compile or the build will fail. |
calacademy-research/antcat | 128618261 | Title: Canceling name field popup adds padding to the buttons
Question:
username_0: [reproduction steps should not modify the database, but test in dev anyways]
To reproduce behavior: edit a taxon and click on the name button, click cancel, click name button again, and then cancel again. The third time the name button is clicked, the buttons look slightly different. Every time this step is repeated extra passing is added to the buttons.
This could very well only be an issue with extra padding, but considering #117 (which may actually modify the database), this is worth investigating.
Answers:
username_0: Code has been rewritten.
Status: Issue closed
|
jhlau/doc2vec | 238167166 | Title: save model in non-binary format
Question:
username_0: How can I save the model in non-binary format?
Thank you.
Status: Issue closed
Answers:
username_1: I don't think the doc2vec code provides a native function for saving doc2vec model in non-binary format. You can of course manually pull out the weights and save them yourself.
username_0: I do not really understand what you mean by "pulling out the weights".
What I did now is using the training documents as test documents (since we are here doing unsupervised classification, there should not be a problem with that) and then run the infer_test.py script.
Is that what you had in mind?
username_1: Ah I see. You can do what you are doing now, but the vectors themselves might actually bit a little different when you're re-inferring them (the inference procedure is basically a pseudo-training step with randomly initialised document vector).
If all you're looking are the train document vectors, it's saved in the model and you can get them by doing something as follows:
model = g.Doc2Vec(docs, size=vector_size, window=window_size, min_count=min_count, sample=sampling_threshold, workers=worker_count, hs=0, dm=dm, negative=negative_size, dbow_words=1, dm_concat=1, pretrained_emb=pretrained_emb, iter=train_epoch)
vector = m.docvecs[0] #vector is the document vector for the first document
username_1: For more information, you can refer to the code: https://github.com/username_1/gensim/blob/develop/gensim/models/doc2vec.py#L261
username_0: Thank you for these helpful information! |
sfztools/sfizz | 578462292 | Title: C++11 compatible?
Question:
username_0: This is a general question. Is there interest to keep code C++11 compatible?
Otherwise it makes it difficult to get the plugin into the MOD units.
In MOD we are still using gcc 4.9, as it can build up to c++11 and that has been enough so far for all the plugins.
Answers:
username_1: Hi! Yes! we have compatibility libraries in C++11 for filesystem, string views and char conversion, which are the key elements we use in C++17. We spent some time working on compatibility some months ago, I'm not sure we went as low as gcc 4.9 though.
username_0: Oh, good to know :)
I have to update cmake to 3.5 still, but that is much easier than updating the compiler.
Currently with master branch I have to following errors:
```
src/sfizz/Defaults.h:71:47: error: 'constexpr sfz::Range<Type>::Range(Type, Type) [with Type = int]' called in a constant expression
constexpr Range<int> bendRange { -8192, 8192 };
^
src/sfizz/Range.h:40:15: note: 'constexpr sfz::Range<Type>::Range(Type, Type) [with Type = int]' is not usable as a constexpr function because:
constexpr Range(Type start, Type end) noexcept
^
src/sfizz/SfzHelpers.h:202:16: error: invalid return type 'void' of constexpr function 'constexpr void sfz::multiplyByCents(float&, int)'
constexpr void multiplyByCents(float& base, int modifier)
^
src/sfizz/Opcode.h: In function 'absl::optional<bool> sfz::readBooleanFromOpcode(const sfz::Opcode&)':
src/sfizz/Opcode.h:98:20: error: 'constexpr uint64_t hash(absl::string_view, uint64_t)' called in a constant expression
case hash("off"):
^
```
username_1: The void one is gcc 5 apparently, a bit sad but well. The last one may be because of the `absl::string_view` object, the first one would be constexpr constructor being disallowed?
I seem to remember a MOD Docker environment, should I use this to try and correct these build errors maybe?
username_0: You can do it with the mod-plugin-builder, I added a sfizz plugin package in the `sfizz` branch.
The docker stuff is only needed in case you are not running Linux.
I am checking which changes we need to do, to get a confirmation that they are okay with you.
I am fine with sending you patches, no worries on that.
username_2: You can reproduce this just by setting CMAKE_CXX_STANDARD to 11.
I remember some restrictions on C++11, where constexpr functions can only consist of a single return statement I believe.
username_0: Yeah, that is the biggest issue at the moment.
I have to replace a few constexpr functions with inline ones. The `hash` one is specially nasty.
username_1: @username_0 having constexpr hashing for string literals in switch functions is the key thing needed from what you posted. The above `string_view` overload of the hash function is something I think I added, which probably does not need to be `constexpr`. If you call `hash("string literal")` I would guess the `constexpr uint64_t hash(const char*, ...)` overload would be called? This one should be C++11?
username_1: Sorry, checking the source, I realize there is no such overload because the `string_view` one handles it. Hopefully adding the overload will be enough. If you open a PR with the rest I can work on this part later today if you have no time :slightly_smiling_face:
username_0: I am not sure I follow with the method that needs to be overloaded..
username_2: None, because constexpr string_view works in C++11. It's not the root of the problem.
It's just needed to make the function have a single return statement, by rewriting hash to use ternary conditionals.
username_2: See: https://godbolt.org/z/TtNPHK
username_0: that seems to make it go further, but the switch case fails.
so I have this:
```
constexpr uint64_t hash(absl::string_view s, uint64_t h = Fnv1aBasis)
{
return s.length() > 0 ? hash( { s.data() + 1, s.length() - 1 }, (h ^ s.front()) * Fnv1aPrime ) : 0;
}
```
but that fails with:
```
src/sfizz/Opcode.h: In function 'absl::optional<bool> sfz::readBooleanFromOpcode(const sfz::Opcode&)':
src/sfizz/Opcode.h:100:5: error: duplicate case value
case hash222("on"):
^
src/sfizz/Opcode.h:98:5: error: previously used here
case hash222("off"):
^
```
username_1: You have to put `h` instead of `0` at the end of the ternary.
We may all be doing the same thing in parallel so I suggest to open a PR and work it out there. Note that it seems the `atomic_queue` used in the background loader may need work too.
username_0: ah yes!
so far I am only seeing how far I can make it build
username_0: btw, gcc4 does not have any constexpr on the std math functions, so a lot of stuff has to be use `inline` instead of `constexpr`
username_1: I made a `constexpr` min/max using a ternary too, it works out and that should be the only constexpr thing needed for maths (even then it's just a sanity check on the constexpr `Range` which is not really critical...).
username_1: This rule about only a single return statement does not encourage readable code lol, I'm checking out what will be needed in the `atomic_queue` and it will be ugly :D
username_0: The Range constexpr I was able to get around by using this:
```
constexpr Range(Type start, Type end) noexcept
: _start(start)
, _end(start > end ? start : end)
{
}
```
username_1: Ho no..
```
/home/paul/source/sfizz/src/sfizz/Synth.cpp:88:35: error: use of ‘auto’ in lambda parameter declaration only available with ‘-std=c++14’ or ‘-std=gnu++14’
```
This will sting xD
username_0: These are the changes I made so far.
```
diff --git a/src/sfizz/Buffer.h b/src/sfizz/Buffer.h
index 757fe9c..e37b634 100644
--- a/src/sfizz/Buffer.h
+++ b/src/sfizz/Buffer.h
@@ -274,9 +274,9 @@ public:
return counter;
}
private:
- static constexpr auto AlignmentMask { Alignment - 1 };
- static constexpr auto TypeAlignment { Alignment / sizeof(value_type) };
- static constexpr auto TypeAlignmentMask { TypeAlignment - 1 };
+ static constexpr auto AlignmentMask = Alignment - 1U;
+ static constexpr auto TypeAlignment = Alignment / sizeof(value_type);
+ static constexpr auto TypeAlignmentMask = TypeAlignment - 1U;
static_assert(std::is_arithmetic<value_type>::value, "Type should be arithmetic");
static_assert(Alignment == 0 || Alignment == 4 || Alignment == 8 || Alignment == 16, "Bad alignment value");
static_assert(TypeAlignment * sizeof(value_type) == Alignment, "The alignment does not appear to be divided by the size of the Type");
diff --git a/src/sfizz/EGDescription.h b/src/sfizz/EGDescription.h
index cc65551..ceba155 100644
--- a/src/sfizz/EGDescription.h
+++ b/src/sfizz/EGDescription.h
@@ -134,9 +134,12 @@ struct EGDescription
* @param velocity
* @return float
*/
- float getStart(const SfzCCArray &ccValues, uint8_t velocity [[maybe_unused]]) const noexcept
+ float getStart(const SfzCCArray &ccValues, uint8_t velocity) const noexcept
{
return Default::egPercentRange.clamp(ccSwitchedValue(ccValues, ccStart, start));
+
+ // unused
+ (void)velocity;
}
/**
* @brief Get the sustain level with possibly a CC modifier and a velocity modifier
diff --git a/src/sfizz/MathHelpers.h b/src/sfizz/MathHelpers.h
index aa99bec..60aec45 100644
--- a/src/sfizz/MathHelpers.h
+++ b/src/sfizz/MathHelpers.h
@@ -127,9 +127,7 @@ inline float midiNoteFrequency(const int noteNumber)
template<class T>
constexpr T clamp( T v, T lo, T hi )
{
- v = std::min(v, hi);
- v = std::max(v, lo);
- return v;
+ return std::max(std::min(v, hi), lo);
}
template<int Increment = 1, class T>
@@ -152,17 +150,17 @@ constexpr ValueType linearInterpolation(ValueType left, ValueType right, ValueTy
}
template <class Type>
-constexpr Type pi { static_cast<Type>(3.141592653589793238462643383279502884) };
+constexpr Type pi() { return static_cast<Type>(3.141592653589793238462643383279502884); };
template <class Type>
-constexpr Type twoPi { static_cast<Type>(2) * pi<Type> };
[Truncated]
ASSERT(noteNumber < 128);
ASSERT(noteNumber >= 0);
@@ -485,7 +485,10 @@ void sfz::Synth::noteOff(int delay, int noteNumber, uint8_t velocity [[maybe_unu
for (auto& voice : voices)
voice->registerNoteOff(delay, noteNumber, replacedVelocity);
- noteOffDispatch(delay, noteNumber, replacedVelocity);
+ return noteOffDispatch(delay, noteNumber, replacedVelocity);
+
+ // unused
+ (void)velocity;
}
void sfz::Synth::noteOffDispatch(int delay, int noteNumber, uint8_t velocity) noexcept
```
I will stop here since you seem to be doing the same :)
The warnings can be ignored, like unknown [[maybe_unused]] stuff, I just did it locally here so I could better see results.
the `atomic_queue::round_up_to_power_of_2` really seems like going to be the biggest pain
username_1: Yes we'll need to patch this atomic_queue... Ho well. I'll work on updating all of this. Thanks for the report and the patches!
We might also need a gcc 4.9 CI run to check that we're not introducing problems.
username_2: I'll suggest to rather a MOD plugin builder CI, based on the Docker image.
username_1: Changes have been merged in #111 and we have a CI check in #112 , so I'm closing this for now. I'll let you open other issues as needed! Thanks.
Status: Issue closed
|
jlippold/tweakCompatible | 339155703 | Title: `SpringChanger` working on iOS 11.3.1
Question:
username_0: ```
{
"packageId": "com.wizages.springchanger",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "com.wizages.springchanger",
"deviceId": "iPhone8,1",
"url": "http://cydia.saurik.com/package/com.wizages.springchanger/",
"iOSVersion": "11.3.1",
"packageVersionIndexed": true,
"packageName": "SpringChanger",
"category": "Tweaks",
"repository": "BigBoss",
"name": "SpringChanger",
"packageIndexed": true,
"packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 100% with 1 working reports.",
"id": "com.wizages.springchanger",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.0.7",
"shortDescription": "change colors of respring screen",
"latest": "0.3",
"author": "wizages",
"packageStatus": "Working"
},
"base64": "<KEY>
"chosenStatus": "working",
"notes": ""
}
``` |
scantist3/WebGoat-Legacy | 725120002 | Title: CVE-2018-19362
Question:
username_0: <details><summary>CVE description</summary>FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the jboss-common-core class from polymorphic deserialization.</details> |
at-internet/atinternet-apple-sdk | 225925898 | Title: Sementic Versionning PLEASE
Question:
username_0: Please please please consider support proper [sementic versioning](http://semver.org).
Because now your minor update break the API and my implementation.
Why does a minor update 1.0.8 break my code ?
```
no visible @interface for 'Tracker' declares the selector 'setSiteId:completionHandler:'
[_tracker setSiteId:xitiSiteID.intValue completionHandler:nil];
~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
no visible @interface for 'Tracker' declares the selector 'setSecureModeEnabled:completionHandler:'
[_tracker setSecureModeEnabled:YES completionHandler:nil];
~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
no visible @interface for 'Tracker' declares the selector 'setLog:completionHandler:'
[_tracker setLog:at_log completionHandler:nil];
~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
no visible @interface for 'Tracker' declares the selector 'setSecuredLog:completionHandler:'
[_tracker setSecuredLog:at_logSSL completionHandler:nil];
~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
no visible @interface for 'Tracker' declares the selector 'setDomain:completionHandler:'
[_tracker setDomain:at_logDomain completionHandler:nil];
~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
no visible @interface for 'Tracker' declares the selector 'setStringParam::'
[_tracker setStringParam:key :[value addingPercentEncoding]];
~~~~~~~~ ^~~~~~~~~~~~~~
6 errors generated.
```
Answers:
username_0: I already ask the same thing a couple month ago... https://github.com/at-internet/atinternet-apple-sdk/issues/5
username_1: sorry it was not supposed to break objective-c it's a mistake from my part because swift optional parameters are sadly not supported by ObjC. I will update the repo to fix the compatibility
username_0: Great thanks,
But do it in a new version of the SDK please.
Thanks,
username_1: However the "setStringParam" method is unchanged, did the @discardableResult break it ?
username_1: In fact i'm going to remove 1.0.8 from Cocoapods and put this version to 2.5.0 to solve #2
Status: Issue closed
username_1: With the 'new' version we will be more careful with Versioning |
syndesisio/syndesis | 529845493 | Title: Operation paths in warnings look wonky
Question:
username_0: ## This is a...
<pre><code>
[ ] Feature request
[x] Regression (a behavior that used to work and stopped working in a new release)
[ ] Bug report <!-- Please search GitHub for a similar issue or PR before submitting -->
[ ] Documentation issue or request
</code></pre>
## Description

Could be related to the recent OpenAPI change, I don't think they were like this before that; do you think that's so @christophd? |
apache/apisix | 962550241 | Title: request help: java如何集成apisix
Question:
username_0: ### Issue description
有个不明白。springboot项目怎么集成apisix呢,有没有demo?
### Environment
Request help without environment information will be ignored or closed.
* apisix version (cmd: `apisix version`):2.8
* OS (cmd: `uname -a`):Linux slave1 3.10.0-514.el7.x86_64 #1 SMP Tue Nov 22 16:42:41 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
* OpenResty / Nginx version (cmd: `nginx -V` or `openresty -V`):nginx version: openresty/1.19.3.2
* etcd version, if have (cmd: run `curl http://127.0.0.1:9090/v1/server_info` to get the info from server-info API):2.7.3
* apisix-dashboard version, if have:2.7
* luarocks version, if the issue is about installation (cmd: `luarocks --version`):3.4.0
Answers:
username_1: Try to learn https://github.com/apache/apisix-java-plugin-runner.
Status: Issue closed
username_0: 你的这是自定义插件吧,比如鉴权抽成一个服务 |
electron/electron-api-demos | 190859273 | Title: Trojan detected
Question:
username_0: * What operating system are you using? Win10
* What version of Node.js is on your system? N/A
When I try to extract the file, it alerted me that Trojan has been detected in the zip file....
Answers:
username_1: What file and what anti-virus software. It is 100% a false positive FYI
username_2: I downloaded electron-api-demos-windows.zip and it just alerted me that trojan detected.
Status: Issue closed
|
lewellen/gaussianMixtureModel | 271704064 | Title: Undefined Reference problem using Nsight and Cmake
Question:
username_0: Hi,
I am working on GMM and finding your code very helpful! I want to write a program that can test the performance of different approaches of EM. I am not familiar with the makefile, so basically I import the source files in `src/` `lib/` `include/` into a Nsight project and let it generate makefile automatically. However when I tried to compile it, the compiler presented #Undefined reference# issues as blow:
```
......
Building file: ../lib/util.c
Invoking: NVCC Compiler
/usr/local/cuda-8.0/bin/nvcc -I/home/ubuntu/CMU/MyCode/emTimeTest/include -O2 -std=c++11 -gencode arch=compute_50,code=sm_50 -odir "lib" -M -o "lib/util.d" "../lib/util.c"
/usr/local/cuda-8.0/bin/nvcc -I/home/ubuntu/CMU/MyCode/emTimeTest/include -O2 -std=c++11 --compile -x c -o "lib/util.o" "../lib/util.c"
nvcc warning : The 'compute_20', 'sm_20', and 'sm_21' architectures are deprecated, and may be removed in a future release (Use -Wno-deprecated-gpu-targets to suppress warning).
Finished building: ../lib/util.c
Building target: emTimeTest
Invoking: NVCC Linker
/usr/local/cuda-8.0/bin/nvcc --cudart static --relocatable-device-code=false -gencode arch=compute_50,code=compute_50 -gencode arch=compute_50,code=sm_50 -link -o "emTimeTest" ./src/timeNumPoints.o ./lib/barrier.o ./lib/component.o ./lib/cudaGmm.o ./lib/datFile.o ./lib/gmm.o ./lib/kmeans.o ./lib/linearAlgebra.o ./lib/parallelGmm.o ./lib/seqGmm.o ./lib/util.o
./src/timeNumPoints.o: In function `main':
timeNumPoints.c:(.text.startup+0x127): undefined reference to `cudaFit'
./lib/cudaGmm.o: In function `cudaFit(double const*, unsigned long, unsigned long, unsigned long, unsigned long)':
cudaGmm.cc:(.text+0x81): undefined reference to `initGMM(double const*, unsigned long, unsigned long, unsigned long)'
cudaGmm.cc:(.text+0x1f6): undefined reference to `gpuGmmFit(double const*, unsigned long, unsigned long, unsigned long, double*, double*, double*, double*, double*, unsigned long)'
collect2: error: ld returned 1 exit status
make: *** [emTimeTest] Error 1
makefile:60: recipe for target 'emTimeTest' failed
23:00:31 Build Finished (took 1s.870ms)
```
Also, I tried to write a CMakelist to do the compilation, it seems has same problem.
```
[100%] Linking CXX executable gaussianMixtureModel
CMakeFiles/gaussianMixtureModel.dir/src/timeNumPoints.cc.o: In function `main':
/home/ubuntu/gaussianMixtureModelCMake/src/timeNumPoints.cc:45: undefined reference to `cudaFit'
libmy_c_lib.a(kmeans.c.o): In function `kmeans':
/home/ubuntu/gaussianMixtureModelCMake/lib/kmeans.c:44: undefined reference to `vecAddInPlace'
collect2: error: ld returned 1 exit status
CMakeFiles/gaussianMixtureModel.dir/build.make:100: recipe for target 'gaussianMixtureModel' failed
make[2]: *** [gaussianMixtureModel] Error 1
CMakeFiles/Makefile2:142: recipe for target 'CMakeFiles/gaussianMixtureModel.dir/all' failed
make[1]: *** [CMakeFiles/gaussianMixtureModel.dir/all] Error 2
Makefile:83: recipe for target 'all' failed
make: *** [all] Error 2
```
I have stuck for days. Would you mind to help me to find where the problem is ? Thank you! |
naser44/1 | 134793769 | Title: كيف أنهى رجل الأعمال البريطاني سايمون بينر حياته في عيادة انتحار وفضيحة BBC
Question:
username_0: <a href="http://ift.tt/1SDANZh">كيف أنهى رجل الأعمال البريطاني “سايمون بينر” حياته في عيادة انتحار وفضيحة BBC</a> |
jlippold/tweakCompatible | 431848652 | Title: `BreadCrumbArrow` working on iOS 12.1.1
Question:
username_0: ```
{
"packageId": "com.joemerlino.breadcrumbarrow",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "com.joemerlino.breadcrumbarrow",
"deviceId": "iPhone10,6",
"url": "http://cydia.saurik.com/package/com.joemerlino.breadcrumbarrow/",
"iOSVersion": "12.1.1",
"packageVersionIndexed": true,
"packageName": "BreadCrumbArrow",
"category": "Tweaks",
"repository": "BigBoss",
"name": "BreadCrumbArrow",
"installed": "0.0.1-5",
"packageIndexed": true,
"packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 100% with 1 working reports.",
"id": "com.joemerlino.breadcrumbarrow",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.1.5",
"shortDescription": "Removes breadcrumbs text and keeps cellular data.",
"latest": "0.0.1-5",
"author": "<NAME>",
"packageStatus": "Working"
},
"base64": "<KEY>2<KEY>
"chosenStatus": "working",
"notes": ""
}
```<issue_closed>
Status: Issue closed |
blb-vertnet/blb | 369982595 | Title: Monthly VertNet data use report for 2017-11, resource blb_all
Question:
username_0: Your monthly VertNet data use report is ready!
You can see the HTML rendered version of the reports with this link:
http://tools-usagestats.vertnet-portal.appspot.com/reports/f11db245-3f9f-4fc6-a0cc-12b4124d081b/201711/
Raw text and JSON-formatted versions of the report are also available for
download from this link. In addition, a copy of the text version has been
uploaded to your GitHub repository, under the "Reports" folder. Also, a full
list of all reports can be accessed here:
http://tools-usagestats.vertnet-portal.appspot.com/reports/f11db245-3f9f-4fc6-a0cc-12b4124d081b/
You can find more information on the reporting system, along with an
explanation of each metric, here:
http://www.vertnet.org/resources/usagereportingguide.html
Please post any comments or questions to:
http://www.vertnet.org/feedback/contact.html
Thank you for being a part of VertNet. |
micjahn/ZXing.Net | 288027584 | Title: Support for Unity 2017.2+ ?
Question:
username_0: Does the latest release from October work with Unity 2017.2+ ?
I am trying to create an app for the HoloLens with a QR scanner, but the project does not seem to build without errors. It works in the editor with a webcamera as camera source.
Before I start debugging for days, I was wondering if 2017.2 is even supported?
If no, would it help to build the project from source?
Currently, I am using 2017.2.1p1 and Vuforia 6.5, but will be upgrading to 2017.3 and Vuforia 7 later on.
Answers:
username_1: I didn't try it. And I don't have a development environment with unity at the moment.
Which errors occur?
username_0: Copypaste from Console:

I am currently in the process of building ZXing.net from source for Unity, 3.5 and 4.6 to in VS2017 to test.
username_1: I think it comes from QrCodeEncodingOptions.cs.
Please try to build it with an additional preprocessor directive "PORTABLE" or remove the class ErrorLevelConverter from the sources.
username_0: I got it to work on the HoloLens by removing ErrorLevelConverter and commenting out the reference to it in QrCodeEncodingOptions.cs.
Thanks for your help!
I'll try your edited suggestion as well. I'll let you know if it works.
username_1: Can you look at the last checkin
https://github.com/username_1/ZXing.Net/commit/a1abc545bdd0ef16e71d49a0e574b43cff89684c
if it works for you?
username_0: Build completed with a result of 'Succeeded' !
I'm guessing I should've changed both those lines, as I only changed the first one.
username_0: BTW, follow-up question. is it possible to extract the x and y coordinates of the center of the detected QR code in the image? Attempting to raycast in the direction of the detected QR code, and need to know the relative position of a detected QR code in an image. |
adazzle/react-data-grid | 186485363 | Title: Delete a row Or Empty the table
Question:
username_0: [ ] bug report
[x ] feature request
[ ] support request => Please do not submit support request here, instead see
**Current behavior**
The toolbar of the table only has one button which is "Add Row"
**Expected/desired behavior**
1. Delete the row which isSelect
2. Empty the whole table
PS: Or there have already been the ways to solve my problems??? But I didn't find out???
Answers:
username_1: Hi @username_0.
The toolbar is exposed as an addon, you can either use it or not it's up to you.
The grid is supposed to be a powerful core to support multiple use cases, I don't think it makes much sense to implement that as the rows themselves are handled outside the grid, is something that you have full control of and the grid doesn't need to know about the way you're storing you're rows.
What you can do is create your own toolbar (or extend the one we offer in the addons) and create a new button to clear all the rows. The best way to do it will depend on the way you're storing the rows but basically what you will need to do is pass an empty row object (array, immutable whatever) and re-render the grid.
For the deletion the concept is the same, you handle the way the rows are stored outside the grid, get the new rows object after deletion and re-render
Cheers
Status: Issue closed
|
sonicmax/ChromeLL-2.0 | 48137134 | Title: Convert background script into event page
Question:
username_0: https://developer.chrome.com/extensions/event_pages
https://developer.chrome.com/extensions/event_pages#transition
"Event pages are available in the stable channel as of Chrome 22, and the performance advantages are significant, especially on low-power devices."
Status: Issue closed
Answers:
username_0: ChromeLL would lose important parts of its functionality without the ability to use the webRequest API, and it doesn't look like declarativeWebRequest will be moving to stable any time soon. Will reopen if anything changes.
username_0: looks like the WebRequest API restriction may be removed: https://code.google.com/p/chromium/issues/detail?id=447270
username_0: https://developer.chrome.com/extensions/event_pages
https://developer.chrome.com/extensions/event_pages#transition
"Event pages are available in the stable channel as of Chrome 22, and the performance advantages are significant, especially on low-power devices."
Status: Issue closed
username_0: The glacial pace of declarativeWebRequest API development makes it very unlikely that I will ever get around to doing this |
ant-design/ant-design | 549928435 | Title: docs: 各组件中 API 中 string 的一致性
Question:
username_0: - [ ] I have searched the [issues](https://github.com/ant-design/ant-design/issues) of this repository and believe that this is not a duplicate.
### Reproduction link
[https://next.ant.design/components/button-cn/](https://next.ant.design/components/button-cn/)
### What is expected?
- 描述方式一致
### What is actually happening?
- 描述方式不一致
| Environment | Info |
|---|---|
| antd | 4.0.0-rc.1 |
| React | The latest |
### 各组件中描述 概览
| 组件 | 属性 | 类型定义方式 | 备注 |
| ------------ | ----------------------------------- | ------------------------------------------------------------ | :---------------------------------------------------- |
| Typography | type | `secondary`, `warning`, `danger` | |
| Grid | align 、justify | `top` `middle` `bottom` | |
| Menu | mode、theme | string: `vertical` `horizontal` `inline` | |
| Steps | type 等 | string | 可选值在说明中列出 |
| Cascader | expandTrigger、popupPlacement、size | string | 可选值在说明中,其中 popupPlacement 的类型写成了 Enum |
| DatePicker | mode、picker | `date`, `week`, `month`, `year` | 我看描述,是不是 4.0 已经取消了 mode 呢? |
| Form | labelAlign、layout 、size | 'left' \| 'right' | size 使用 `small \| middle \| large` |
| InputNumber | size | string | 说明中未列出可选 |
| Input | size | | 同 Steps 组件 |
| Mentions | placement | 'top' \| 'bottom' | |
| Radio | size、buttonStyle | `large` \| `default` \| `small` | |
| Switch | size | | 同 Steps 组件 |
| TreeSelect | size | | 同 Steps 组件 |
| Upload | listType | | 同 Steps 组件 |
| Collapse | expandIconPosition | `left` | 可选值在说明中列出,类型中写了默认值 |
| Carousel | dotPosition | | 同 Steps 组件 |
| Card | size | | 同 Radio 组件 |
| Calendar | mode | | 同 Steps 组件 |
| Descriptions | size、layout | | 同 Radio 组件 |
| List | size、position | | size 同 Radio 组件,position 同 From 组件 |
| Tree | expandAction | | 同 Steps 组件 |
| Tooltip | placement、trigger | | 同 Steps 组件 |
| Timeline | mode、position | | 同 Radio 组件 |
| Tabs | size、tabPosition、type | | 同 Radio 组件 |
| Table | size | default \| middle \| small | |
| Alert | type | | 同 Steps 组件 |
| Drawer | placement | | 同 From 组件 |
| Progress | status | | 同 Steps 组件 |
| Result | status | `'success' \| 'error' \| 'info' \| 'warning'\| '404' \| '403' \| '500'` | |
| Spin | size | | 同 Steps 组件 |
| Skeleton | shape、size | Enum{ 'large', 'small', 'default' } | |
| Divider | orientation、type | enum: `left` `right` | |
### 各组件描述 总结
| 分类 | 类型 | 说明 |
| ---- | ---------------------------------- | ---------- |
| 1 | string | 列出可选值 |
| 2 | default \| middle \| small | |
| 3 | 'default' \| 'middle' \| 'small' | |
| 4 | `default` \| `middle` \| `small` | |
| 5 | `default` `middle` `small` | |
| 6 | `default`, `middle`, `small` | |
| 7 | string: `default` `middle` `small` | |
### 疑问
- 是否需将各组件中 描述信息保持一致?
- 各组件采用相同定义方式?
- 还是根据不同情况(如意义、可选值数量)进行分别定义?
<!-- generated by ant-design-issue-helper. DO NOT REMOVE -->
Answers:
username_1: 第 4 种感觉好点
username_2: size 这种可以抽出来放到公用的,直接跳过去
username_0: 有些可选值多的,那一项会列出很多
username_1: `size` 内部是同一个接口,文档里不建议跳走。没多少字,用户还要切页面反而不太方便。
username_0: ### Tip
- List:grid props 英语 API 中 有 size 属性,中文无
- Form:validateMessages属性 英文 API 中显示中文,未翻译
- Pagination:showLessItems属性 中文 API 中显示英文
- 有些组件文档中有 “版本”,有些没有,并且有的均为空,该定义是否在 4.0 中取消?或是正式版本该值 再开始维护?
username_3: 版本 是当初为4.0清理掉了,相当于从0开始就有的。和3.x无关了。
username_0: @username_3 👌
Status: Issue closed
username_0: @username_3 要不要 我提个 PR 把所有API 中 “版本” 都去掉?
username_0: 还有那 2 个 中英文 未翻译的 要怎么处理?
username_0: 还有一个 中 英文 API 不一致
username_3: 😂,可以给PR 打[WIP]。
username_3: 4.0 正式版以后新增的API还是需要它的呀。 |
facebookincubator/submitit | 766284645 | Title: 喀什汽车站哪有特殊服务的洗浴p
Question:
username_0: 喀什汽车站妹子真实找上门服务【╋薇⒈O⒎⒎⒈⒐O⒐】月日上午,习总书记在庆祝中华人民共和国成立周年大会上讲到,“在这里,我代表党中央、全国人大、国务院、全国政协和中央军委,向一切为民族独立和人民解放、国家富强和人民幸福建立了不朽功勋的革命先辈和烈士们,表示深切的怀念!”近日,由华策影视、安徽广电集团五星东方影视、天润大美影视出品的史诗大剧《外交风云》,正是回首波澜壮阔的不凡外交征程,深切缅怀守家卫国的外交先驱,凝聚坚不可摧的奋进力量。一直以来,主旋律题材创作都“重武轻文”,外交题材剧更是寥若晨星。《外交风云》凭借稀缺题材和热血燃情,实现主旋律剧的“年轻化落地”,领跑国庆档成功出圈,观众直呼看不够,强烈要求三集连播。北京电视台积极协调各方《外交风云》从月日开始重磅加料由原来日播两集变更为日播三集让更多观众能够在国庆假期畅快淋漓地一睹“外交天团”风采。稀缺题材,外交博弈大解密从新中国成立之初打破西方孤立封锁到改革开放后逐步建立起全方位多层次的对外关系新格局,再到如今建立对各种形式的伙伴关系、参与几乎所有政府间国际组织和多项国际公约,中国不仅打破西方对新中国的敌视围堵,亦打破了零和博弈、赢者通吃的旧思维,以中国之道参与塑造国际新秩序。《外交风云》将镜头对准中国外交破冰之旅,在影视作品中鲜见。被贴上“中国硬核外交”标签的《外交风云》,在观众尤其年轻观众中反响强烈,他们沉迷剧中的解密,用弹幕、超话、表情包等时代符号互动交流,并纷纷为其打,“这场没有硝烟的纷争,我们终将铭记。”目前已播出的剧情中,该剧硬核解密中国核武器背后的故事。中国希望苏联在尖端武器的研制上能给中国军队以帮助,但两国在意识形态、对时代和国际形势以及核武器的态度等问题上产生了重大分歧。革命、建设,沧桑岁月引人深思回味。历史的车轮滚滚向前,人民军队在战斗中成长,在继承中创新,在建设中发展,已经由过去单一军种的军队发展成为诸军兵种联合的强大军队,由过去“小米加步枪”武装起来的军队发展成为基本实现机械化、加快迈向信息化的强大军队。网友疯狂转发和阅兵式对比图,除了飞机再也无需飞两遍外,祖国还有了火箭军、东风核导弹等现代化新型武装力量。抚今追昔,《外交风云》点燃了观众的拳拳爱国心,众多网友表示,“这盛世,如你所愿!”热血燃情,实现“年轻化落地”“华北之大,已经安放不得一张平静的书桌了。”当年,清华大学救国会这样发出告全国民众书。那时的延安,被称为“中国的耶路撒冷”,“延安的城门成天开着,成天有从各个方向走来的青年,背着行李,燃烧着希望,走进这城门。”物换星移,热血,依旧是青春的精神特质。慷慨激昂、风骨铿然,《外交风云》的燃情热血在献礼剧中突出重围。朝鲜内战爆发后,美国杜鲁门政府悍然派兵进行武装干涉,发动对朝鲜的全面战争,并不顾中国政府多次警告,越过三八线,直逼中朝边境的鸭绿江和图们江,出动飞机轰炸我国东北边境城市和乡村,把战火烧到了新生的中华人民共和国国土之上。在此危急关头,应朝鲜党和政府的请求,中国共产党高瞻远瞩,审时度势,毅然决然地作出了抗美援朝、保家卫国的历史性决策,以大无畏的英雄气概果敢承担起保卫和平的历史使命。在已播出的剧情中,“最可爱的人”凯旋而归。让人不禁想起,阅兵式上的老兵方阵,从前,他们浴血奋战,守卫这万里锦绣;今天,他们热血未凉,坚守这永不褪色的信仰。投仍邓卣镣招喜晾锤毕绰捞伪弊桶吭林怖伪竟勾厦嘲至页https://github.com/facebookincubator/submitit/issues/882?17219 <br />https://github.com/facebookincubator/submitit/issues/870?34779 <br />https://github.com/facebookincubator/submitit/issues/858?51313 <br />https://github.com/facebookincubator/submitit/issues/846?28213 <br />https://github.com/facebookincubator/submitit/issues/834?81335 <br />mjfahpoldntdohbarsdtpahsbfvxnxkuobx |
MicrosoftDocs/azure-docs | 955421248 | Title: This doesn't exactly say how to use ForceDelete
Question:
username_0: [Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 67566e6a-ffb4-4adf-73b5-3d6c0c58cd05
* Version Independent ID: 9c9a6f63-ec87-0e45-9d0e-f5030041ab06
* Content: [Add and verify custom domain names - Azure Active Directory](https://docs.microsoft.com/en-us/azure/active-directory/enterprise-users/domains-manage)
* Content Source: [articles/active-directory/enterprise-users/domains-manage.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/active-directory/enterprise-users/domains-manage.md)
* Service: **active-directory**
* Sub-service: **enterprise-users**
* GitHub Login: @curtand
* Microsoft Alias: **curtand**
Answers:
username_1: Hello @username_0 , thanks for taking time to share your valuable feedback! We are taking a look into this and will get back to you soon.
Status: Issue closed
username_1: There is a[ similar question ](https://social.msdn.microsoft.com/Forums/vstudio/en-US/3cef5e2f-2123-4437-ac59-af90c63f0167/forcedelete-custom-domain?forum=WindowsAzureAD)around the same.
Go to Home > Azure Active Directory > Custom Domain Names > Click into the domain name > Delete

Ensure that all of your resources attached to the domain are deleted first. You should be able to remove the account from the directory and add it to a new one.
Also note that, you need to delete the custom domain using a Global Administrator account that is either tied to the default domain (onmicrosoft.com) or a custom domain (mydomainname.com) to delete the custom domain.
We will now proceed to close this thread. If there are further questions regarding this matter, please tag me in your reply. We will gladly continue the discussion and we will reopen the issue. |
xtermjs/xterm.js | 560659869 | Title: latest firefox has context menu on canvas
Question:
username_0: <!-- ⚠️ Please search existing issues to avoid creating duplicates. ⚠️ -->
<!-- Describe the bug here. -->
## Details
- Browser and browser version:
- OS version:
- xterm.js version:
### Steps to reproduce
1. fire up firefox, version >= 72.0.2 (tested on windows 10, though i'm pretty sure it behaves the same on mac and linux. did not test on older version of FF)
2. go to demo page https://xtermjs.org/
3. right click inside the terminal
EXPECTED: "normal" context menu (back/forward, copy, paste etc.)
OBSERVED: depending where in the terminal you click, you get either the normal context menu or the image context menu.
context menu is very important in terminal, since keyboard shortcut for cut/paste do not work (in any browser, afaik).
this seems to be specific to firefox: at least chrome behaves as expected.
i did not test any other browser.
Terminal.ts, line 421 (on commit 3d2d35f), says:
`
// Handle right click context menus
if (Browser.isFirefox) {
// Firefox doesn't appear to fire the contextmenu event on right click
this.register(addDisposableDomListener(this.element, 'mousedown', (event: MouseEvent) => {
if (event.button === 2) {
rightClickHandler(event, this.textarea, this.screenElement, this._selectionService, this.options.rightClickSelectsWord);
}
}));
} else {
this.register(addDisposableDomListener(this.element, 'contextmenu', (event: MouseEvent) => {
rightClickHandler(event, this.textarea, this.screenElement, this._selectionService, this.options.rightClickSelectsWord);
}));
}
`
i assume at some point, firefox did not have context menu for "canvas" elements. apparently, it does now, so this code has to change: maybe the "else" now works for FF, so eht condition can be changed to something like "is firefox && browser version >= 72
maybe FF still needs special handling, but the current handling is wrong.
peace.
Answers:
username_1: @username_0 Cannot repro under linux with FF 72. I get for any renderer (canvas/DOM/webgl) the right context menu.
username_0: thanks for the prompt response.
>>>> Cannot repro under linux with FF 72. I get for any renderer
(canvas/DOM/webgl) the right context menu.
(i did not know renderer is selectable - i encountered this bug as a user,
not as a developer, so the code to create the terminal is not mine.
in both sites where i encountered it (the "real" one and xtermjs demo page)
use canvas).
some suggestion for reproducing:
1) reproducible easily and repeatedly (though not consistently - see
below) on win10. if i'll manage to repro on linux (not next to my linux box
ATM), i'll inform you or update the ticket. theoretically, maybe there's a
diff between FF 72 and 72.0.2, but it does not sound likely.
2) more importantly: you should right click all over the terminal,
repeatedly - the behavior is inconsistent. my guess is that there's some
kind of race bendition between the "mousedown" event and the "contextmenu"
event, and the result of this race is not consistent. make sure to click
inside the "canvas" element - in the demo, some of the visual "terminal"
area can be outside the canvas..
i tried to set FF event breakpoint, for mouse events (contextmenu and
click, and later for all of them), but FF did not stop anywhere. i may be
doing something wrong, or FF is doing something wrong, or "terminal"
somehow suppress debugging (i doubt it). chrome happily stops for
"contextmenu" and "mousedown".
on further digging, it may not even depend on the point: in the demo, it
seems that the terminal can be wider than the canvas:
in this case, right-clicking will be "good" in the area outside the canvas,
but still, inconsistent behavior: sometimes it's the standard FF context
menu, with the icons at top (back/forward/refresh/bookmark), and sometimes
it's a "naked" context menu, w/o those icons.
it seems there's a fierce competition going on between the "contxtmenu"
event and the "mousedown" event. since the code only handles "mousedown"
for firefox, whenever the "contextmenu" event fires first, it is handled by
the default handler of "canvas", which is very similar to that of an img.
i'm guessing the handler eats up the "mousedown" event in these cases (or
not - however, something does, since only one menu is opened per each
right-click).
this actually _may be_ a new thing, and it also may be specific to win10 -
i know nothing about the message pump and how the browser creates events.
it is possible that the source of the event (and of the competition) is the
OS itself.
*tl;dr*: right-click around and a lot - as the report states, the behavior
is not consistent.
i originally thought it depends on the point you click, but on further
experiments, even the same point is inconsistent.
FF has context menu for canvas for a long time now, so the comment in the
code is definitely wrong.
peace
username_1: code is definitely wrong.
FF is not part of our development tests, so there might be outdated workarounds. Our automated tests run with puppeteer (chromium engine). Not sure if they could be extended to FF somehow. The tests heavily rely on the devtools protocol, I think FF does not support it to a workable extend. Selenium would require a serious rewrite of the tests while providing less testable information, not a viable way for us. So yeah, FF heavily relies on the user side when it comes to testing.
username_2: Not this one https://github.com/xtermjs/xterm.js/pull/2632, that was closed
username_3: Encountered the same problem here.
I also discovered a workaround for it: if you add `contenteditable="true"` to the `xterm-cursor-layer` element, it should pop the context menu for text editing tool instead. But I don't really know about the implications, so I don't recommend it until after the pros here has actually figured this problem out. |
bbc/digital-paper-edit-firebase | 883520939 | Title: Improve transcribing process in UI
Question:
username_0: **Context:**
As part of #221, @nganthi has improved the transcribing process. See design [here](https://app.zeplin.io/project/5ce554083480691e05af2f74/screen/608bd86820f3fe2fd5d4f67a) (relates to Transcripts section only).
**This ticket is to implement the following features:**
- [x] Add green tick icon next to available transcripts (check font awesome for a good match)
- [x] Show the "uploading progress bar" and offers the ability to cancel the upload. (Also remove the react-fetch-progressbar dependency as no longer in use) [cancelling moved to #248]
- [x] Show the approximate transcribing time (use media duration for now - #240 covers estimation) and offers the ability to cancel the transcribing process [cancelling moved to #248]
- [x] Show the "queued" status if the media is queued by PSTT
- [x] Show the error message if upload or transcribing fails and offers the ability to re-upload or re-transcribe [moved to #244]
- [x] Show file deleted message if media has expired (see https://github.com/bbc/digital-paper-edit-firebase/issues/195)
**Developers' notes:**
- For icons, continue using the Font Awesome library (audio, video, cancel, retry, alert icons)
**Out of scope:**
- Notify the users when their transcription is ready.
- Changing the font<issue_closed>
Status: Issue closed |
tianqing370687/tianqing370687.github.io | 284637871 | Title: 浅谈一下MySQL中的count(*),count(1),count(column)的区别
Question:
username_0: ## 浅谈一下MySQL中的count(*),count(1),count(column)的区别
### 试验



### 分析



从上面的结果我们可以看出select count(*)和select count(1)是等价的,但是count(column)不同,即使column是主键。用主键的查询用时会更长,如果这里的column是一个允许为NULL的值,查询时间会更长。select count(column)会过滤掉column=null的结果。
### 补充
每种数据库(mysql、oracle等)都会有数据来描述表的信息,包括表的名字、表的大小等信息,这些信息被称为元数据,所以当需要查询一个表中有多少列时应该先想到的是mysql是如何存储表的元数据的。
MySQL有个库叫information_shcema,这个库里面有个表叫TABLES,这个表有个字段TABLE_ROWS,所以
```mysql
SELECT TABLE_ROWS FROM information_schema.`TABLES` WHERE TABLE_NAME = 'table name'
```
能粗略查到那个表的统计量。
如果,觉得这个值不够精确,想要更精确的,那就效仿这种方法,自己建立一个表存储表名和表对应的总量(可以称为计数器),每次取总量的时候就直接查这个表的字段就可以了,当然那个百万表在做增删的时候记得加减这个计数器。
## 参考&引用
[<font color="#87CEFA">请教一下大家,关于 MySQL 百万数据量的 count(*) 查询如何优化?</font>](https://www.zhihu.com/question/34781415)
## 更新时间
发布时间 : 2017-12-27 |
cypress-io/cypress | 777804717 | Title: browser option : chrome not found
Question:
username_0: I'm trying to run my testcases through cypress though the following command :
`npx cypress open`
its not showing chrome as option.
any idea how I can solve this issue ?
firefox and electron is working fine.

It used to work fine previously. My currently running Version 4.12.1
Answers:
username_1: Please see our [troubleshooting guide on Launching Browsers](https://on.cypress.io/troubleshooting#Launching-browsers)
Closing due to inactivity.
Status: Issue closed
|
chanzer0/NBA-TopShot-Chrome-Extension | 856766135 | Title: Not showing in moments screen
Question:
username_0: Does not seem to work for me. Some things are greyed out. attached screen snip and build below.
Google Chrome is up to date
Version 89.0.4389.114 (Official Build) (64-bit)

Status: Issue closed
Answers:
username_1: Hi, @username_0. Unfortunately, I have discontinued support for this extension as I have joined a full-time team dedicated to making NBA Top Shot tools and features.
I hope that in the near future you will be able to use these tools to their fullest potential. Until then, there are currently no plans for updating the current repository.
All the best,
Sean.
username_0: Oh ok.
What version did it last support.
Get Outlook for Android<https://aka.ms/AAb9ysg> |
maximwebb/ellas-vs-maxims | 427338114 | Title: Limit zombies and plants
Question:
username_0: - ZombieSpawner shouldn't spawn infinitely and only spawn in lanes
- only limited number of plants should be available to place (show number available)
--> related to level design/end of game
Answers:
username_1: Do we want to come up with some deterministic way of placing zombies (ie. not randomly)? That way we can actually design levels properly
Status: Issue closed
|
raveitoutofficial/raveitout | 415196279 | Title: Add overrides for heart subtraction
Question:
username_0: Apparently #SONGTYPE is the standard for PIU simfiles.
Answers:
username_1: Pasting what I mentioned over at Facebook to help
"From what I can observe, they use
#SONGTYPE:ARCADE for regular (2 Hearts)
#SONGTYPE:SHORTCUT for short (1 Heart)
#SONGTYPE:REMIX for long (3 Hearts) and
#SONGTYPE:FULLSONG for extra-long (4 Hearts)
Those terms of course are based on the Andamiro PIU stuff so a new standard can be set by the RIO developers if you want. Just put it in the Song Structure Documentation txt and then the .ssc's can follow suit :)"
username_0: Added... Sort of.
Hiding overridden songs from the music wheel requires a custom wheel, but for now they will subtract the correct amount of hearts. |
ikedaosushi/tech-news | 440788290 | Title: Promiseとはなんぞやできるだけわかりやすく説明してみたついでにasync/awaitも)
Question:
username_0: Promiseとはなんぞや?!できるだけわかりやすく説明してみた(ついでにasync/awaitも)<br>
PromiseやAsync/awaitという言葉をよく耳にするようになってきましたが、どういう仕組みか理解せず苦しんでいる方も多いのではないでしょうか?<br>
http://bit.ly/2Y6FMXX |
netbox-community/netbox | 953452795 | Title: Visual bug. After edited or deleted custom field, bulk create and edit will keep the custom field in display field (not attribute)
Question:
username_0: ### NetBox version
v2.11.7 (tested on demo instance too)
### Python version
3.9
### Steps to Reproduce
1. Add custom field "vmid" in virtualization (admin)
2. Edit custom field "vmid" name in virtualization (admin)
3. Add custom field "core" name in virtualization (admin)
4. Delete custom field "core" name in virtualization (admin)
5. Bulk edit or delete in virtualization page
### Expected Behavior

### Observed Behavior

Answers:
username_1: #7310 might be similar to this, but not quite same. anyway, i hope this issue will be resolved sometime soon.
Status: Issue closed
|
Thorium-Sim/thorium | 356142732 | Title: None
Question:
username_0: The way the icon system currently works, the icon and the label are already shown on the targeting screen. The only thing missing would be a picture, which isn't necessary for identification (especially if Targeting works with Sensors to identify targets). Clicking would make targeting on a touchscreen more difficult, so I'm going to close this as 'Won't Fix'
Answers:
username_0: The way the icon system currently works, the icon and the label are already shown on the targeting screen. The only thing missing would be a picture, which isn't necessary for identification (especially if Targeting works with Sensors to identify targets). Clicking would make targeting on a touchscreen more difficult, so I'm going to close this as 'Won't Fix' |
meilisearch/documentation | 587814096 | Title: FAQ question summary
Question:
username_0: Currently, the FAQ questions are visible on the left in the sidebar. They can be confused with titles and the users might not see them.
We should find a way to generate a summary at the beginning of the FAQ so that the user can select the question he/she wants.<issue_closed>
Status: Issue closed |
RaphaelRoyerRivard/MicroMachine | 602613498 | Title: Improve the position of scans
Question:
username_0: efine maximum_list_point_proche
define average_distance
foreach point in points
list_point_proche.add point
define ordered_points //ordonne les points en ordre de distance par rapport à point, n'ajoute pas les points à une distance > 2*scan size
foreach closest_point in ordered_points
//Calcule moyenne de list_point_proche + closest_point
if average contains all points + closest_point
list_point_proche.add closest_point
else if list_point_proche.count > maximum_point_proche.count
maximum_list_point_proche = list_point_proche
break;
else
break //Stop computing this point, its not the best one
clear list_point_proche
return //average of maximum_list_point_proche |
calpoly-csai/swanton-chat-bot | 782707547 | Title: Add logo to welcome message
Question:
username_0: Jeanine wants the logo attached here added to the card with "Welcome to Swanton Chat", at the top in the center. We should also change the text "Welcome to Swanton Chat" to "Welcome To Swanton Poppy Chat".
<issue_closed>
Status: Issue closed |
GoogleChromeLabs/gulliver | 287804014 | Title: Some audit results don't make sense
Question:
username_0: Just looking at the report for https://simpl.info:

Not sure why these three audits are coming up as false:
- The home page https://simpl.info/index.html doesn't require JavaScript to load, and only uses well supported HTML and CSS, so I'm not sure why 'Site works cross-browser' is false.
- There are no 'page transitions', so not sure where the Page transitions don't feel like they block on the network' warning is coming from.
- Every page on the site does have a unique URL.
The performance score is only 41 — which doesn't tally with the site's performance measured using Lighthouse from Chrome DevTools. Is that because some audits failed?

Also, the Best Practices score isn't 100, but there's no reason given:

Answers:
username_1: @username_0 I will investigate it tomorrow, I should also test and deploy LH 2.7.0 soon.
username_1: The 3 checks that come as false are the "Additional items to manually check" in the Lighthouse report, we need to remove them or change the format we use to display them.
I will do few changes to the whole section, ideally we should find a way to export our json to the Lighthouse report viewer.
username_0: Thanks @username_1!
username_2: Hello,
I was wondering the exact same thing for those 3 audits being 'false' for https://similarworlds.com !!
When I used Lighthouse through Google Chrome, I was getting score 100.
I'm glad this is being looked at :)
Thank you very much.
Status: Issue closed
username_1: We now link to the full LH report in the lighthouse-viewer, our lighthouse-viewer is for LH 2.9.1 and tomorrow our reports will match that version. |
spesmilo/electrum | 669461817 | Title: Multisig wallet setup needs better UX
Question:
username_0: Going through a multisig creation process is confusing with Electrum. For example, if you have 3 different devices you want to use to setup a multisig wallet, you may not be able to plug in all the wallets at once, so there should be some way to asynchronously set up a multisig wallet. However, electrum seems to assume that all the information is available right away all at once. It asks you to create a seed, gives the master public key, and then immediately asks for all the other master public keys. Well what if I don't have them yet?
It seems like Electrum is assuming that you'll have single-seed wallets that you can use the master public key for. However, if a user just wants to create a new multisig wallet, they're likely to just hit "create multisig wallet" and then get stuck one it asks for the master public keys for other devices.
There needs to be a better flow like the following:
1. User begins creation of a new wallet,
2. chooses multisig wallet and m of n,
3. chooses seed location (eg connected hardware wallet, create seed on device, etc),
4. chooses segwit vs legacy
5. user writes down the seed
At this point the wallet should be created. You should be able to exit the app and return to it and select the multisig wallet you created. The interface should tell you the multisig wallet still needs master public keys to be set up and should provide a way for the user to enter them.
As it stands, if you don't have access to ALL the master public keys on creation of your multisig wallet, you can't even cancel the process or go back. Killing the app loses all your progress and you then have to go through the restoring process if you want to use the seed you created (and hopefully wrote down). This is pretty confusing for a user, and is incredibly inconvenient to the point of being unusable if there are multiple parties trying to create the multisig wallet and aren't doing it all at the same time (eg because they have a more secure process for backing up their seed than writing it on a piece of paper in full view of other people).
Another option, that might even be better than what I described above is to treat multisig wallets as extensions of single-seed wallets, where to create a multisig wallet you would start with a single-seed wallet, choose an option to create a multisig wallet which would ask for master public keys, then would open it as a separate wallet at that point. Similarly to access a multisig wallet, you would go into your single-seed wallet, and choose the multisig wallet within it. A single-seed wallet could even have multiple multisig wallets it is part of. The best way to create multisig wallets is to already have single-seed wallets you can get the master public keys of. Why not make this explicit and move multisig wallet creation as a subwallet of standard wallets?
Answers:
username_0: So continuing to actually go through this, its even worse.
1. Adding a Trezor wallet as a cosigner doesn't actually display any Zpub key when adding it as a multisig cosigner (like it does for Coldcard).
2. Mobile electrum has nowhere that displays its Zpub during the creation of the wallet.
With those two things, it makes it impossible for me to actually create the multisig wallet I want to create because I can't get the public keys I need to enter. This is pretty awful. The only workaround I can think of is to add other fake cosigners with keys generated on the desktop computer so you can create a wallet you don't want, get the Zpubs from that wallet, then restart and create the full multisig wallet once you have access to all the Zpubs.
This is not really a workable UI as it stands. Please let me know if you would like some mock ups of different UI and I can do that. I won't have time to actually look into building electrum and implementing for about a month probably.
username_0: And I noticed that if you try to "use a master key" in the first step, it asks for an xpub or an xprv, and NOT a Zpub. Why??? Why is this asymmetric?
It looks like it does actually accept a Zpub, but doesn't mention that it can. Once you enter something, it shows it to you again as your master public key and tells you to "share it with your cosigners". That isn't an appropriate next step, since you are likely creating a watching only wallet, or at least a a wallet with only external cosigners.
username_2: I believe this commit caused the regression that breaks the display of Zpub on Kivy: fbc539e2cc
username_1: Sorry. This was an unintended side effect of fixing the back button issue. In the future I will try to test more. |
BetaMasaheft/Documentation | 785099768 | Title: linked past format
Question:
username_0: we have a model, already since some time, to expose linked pasts data from our data. see email from <NAME> of 14.1.2020 (!!) .
that should be added as a module and added to the void list of available formats, including the observation on the lookup table for the zotero pointers, which should also be stored in the triplestore for easier reference
Answers:
username_0: https://whgazetteer.org/
http://blog.whgazetteer.org/2021/08/05/version-2/
https://whgazetteer.org/tutorials/create_lptsv/
https://whgazetteer.org/static/files/WHG_Asian_History_Lesson.pdf
username_0: archiving for no immediate need, may be reopened later
Status: Issue closed
|
OHDSI/Atlas | 361535259 | Title: Duplicate records in Data Source Drug Report
Question:
username_0: ### Expected behavior
Drug prevalence report in data source reports will have one row per a unique concept_id.
### Actual behavior
For some of the concepts full duplicates (e.g. the same concept_id,concept_name, prevalence, counts) exist.
Example:
40231925 Acetaminophen 325 MG / Oxycodone Hydrochloride 5 MG Oral Tablet
Atlas v 2.4.0, 2.4.3
Answers:
username_1: 
username_2: I think this is likely happening, because drugs are mapped up into the ATC
hierarchy, and this drug may be mapped to multiple locations. If you open
the other hidden columns, you'll see it.
username_1: @username_2 , anyway, the difference between rows should be visible on initial screen
username_2: I may have made an inappropriate choice when I built the hierarchy query to
allow for a one-to-many relationship. If we want to force concepts to have
a one-to-one relationship with only one path through the hierarchy, then
this wouldn't occur. I'm open to whatever approach the community feels is
most appropriate.
username_0: @username_2 where are those hidden columns? They hide too well :)
And then I'm not sure what you're talking about. One drug to many ATC-s is one problem (which will be fixed soon I believe), the multi-way hierarchy is another. I've somehow found that the latter annoys me as well, so we might need to discuss it indeed.
username_3: Related to OHDSI/WebAPI#347
username_4: This isn't an ATLAS bug, will be resolved within WebAPI.
Status: Issue closed
username_2: I am still observing this same problem, that we closed as resolved.

username_2: ### Expected behavior
Drug prevalence report in data source reports will have one row per a unique concept_id.
### Actual behavior
For some of the concepts full duplicates (e.g. the same concept_id,concept_name, prevalence, counts) exist.
Example:
40231925 Acetaminophen 325 MG / Oxycodone Hydrochloride 5 MG Oral Tablet
Atlas v 2.4.0, 2.4.3
username_3: The root of the issue here is the way that the concept_hierarchy table is constructed as detailed in the [init_concept_hierarchy.sql](https://github.com/OHDSI/WebAPI/blob/master/src/main/resources/ddl/results/init_concept_hierarchy.sql#L77) query in WebAPI. In that query, there is the rollup of each drug to the ATC hierarchy which can sometimes lead to different hierarchy paths which then appears as duplicates in ATLAS.
To take the example from the screenshot above, here is an excerpt of the return value returned from `/WebAPI/cdmresults/<source_key>/drug` for concept_id == 19080128
```
{
"conceptId": 19080128,
"conceptPath": "CARDIOVASCULAR SYSTEM||AGENTS ACTING ON THE RENIN-ANGIOTENSIN SYSTEM||ACE inhibitors, plain||Lisinopril||Lisinopril 10 MG Oral Tablet",
"numPersons": 2200303,
"percentPersons": 0.02564,
"recordsPerPerson": 8.82079
}, {
"conceptId": 19080128,
"conceptPath": "NA||NA||NA||Lisinopril||Lisinopril 10 MG Oral Tablet",
"numPersons": 2200303,
"percentPersons": 0.02564,
"recordsPerPerson": 8.82079
}
```
As noted by the `conceptPath` property, the ATC rollup for this particular drug results in 2 different paths and thereby results in a duplicated display to reflect these paths. There are a few ideas that come to mind:
- Remove the hierarchy portion of the returned value from WebAPI to eliminate duplicates in the table display (for drug the hierarchy is not part of the table display).
- Collapse these hierarchy entries into a single item inside of ATLAS and remove the NA values. If there is > 1 hierarchy entry, we can then make a computed column to show the hierarchy. For example:
1. CARDIOVASCULAR SYSTEM -> AGENTS ACTING ON THE RENIN-ANGIOTENSIN SYSTEM -> ACE inhibitors, plain
2. <Another ATC 1> -> <A different ATC 3> -> <A different ATC 5>
@username_2 let me know what you'd prefer to see given this information. Thanks!
username_3: From discussion with @username_2, let's plan to remove the hierarchy portion from the table display.
Status: Issue closed
|
ibis-project/ibis | 723832604 | Title: Rationalize duplicated functions from backend compilers
Question:
username_0: Yes, sorry, looks like clickhouse doesn't use Impala, it's bigquery, omnisci and spark for what I see. The CI should hopefully tell anyway.
Don't worry about the code already duplicated for now. Feel free to create an issue to take care of avoiding that duplication of functions later. But I think avoiding weird dependencies among backends is enough work for now, and will be extremely valuable.
_Originally posted by @datapythonista in https://github.com/ibis-project/ibis/pull/2472#issuecomment-710302802_
Answers:
username_1: PRs welcome.
Status: Issue closed
|
ktsn/vuex-toast | 407614436 | Title: support for vuex 3 ?
Question:
username_0: README states the requirement `Vuex >= 2.0`
I'm using vuex 3.0 and getting this warning
`[email protected] requires a peer of vuex@^2.0.0 but none is installed. You must install peer dependencies yourself.`
Is this lib compatible with vuex 3 ?<issue_closed>
Status: Issue closed |
wraybowling/curvemixer | 25262986 | Title: rotate and scale via 2d marker
Question:
username_0: allow users to place marker(s) like in Blender by which transformation operations can be made relative to.
Status: Issue closed
Answers:
username_0: duplicate of 29
username_0: allow users to place marker(s) like in Blender by which transformation operations can be made relative to. |
mr-bobz/talarts | 114599097 | Title: The Posts displayed on Search results do not have a continue reading link
Question:
username_0: 
Only pages show the link
Answers:
username_1: I see it working when the content is truncated. It is showing all contents in the list when more link is missing
username_0: Thanks @username_1 . That makes sense. As the link is lost, are we able to truncate the search results to certain characters so that every result links to a detailed page?

username_1: Why do you want to show detailed page link when there is no more information on the page/post?
username_0: The users expectation is to take to the actual page. As we have to keep the user experience consistent (especially for accessibility) this link is necessary.
username_0: This is the normal user experience. Title + content excerpt.

Status: Issue closed
|
mcordts/cityscapesScripts | 753177717 | Title: coloredlogs should only install the EvalObjectDetection3d logger
Question:
username_0: In the current version, the evalObjectDetection3d.py file will install the coloredlogs for the root logger "once included".
When I was trying to use the library in a larger project, it overwrote the logging level of other loggers in other libraries.
I would propose to install the logger only for the evaluator logger, with only a change in line 76 of evalObjectDetection3d.py.
```python
coloredlogs.install(level='INFO', logger=logger)
```
I could open a pull request if that is an accepted idea. |
migger/bosh-yandex-cpi-release | 407109903 | Title: Установить bosh director на другом облаке
Question:
username_0: https://github.com/cloudfoundry/bosh-deployment/blob/d0fb8eb4eb82aef502af44ac0628d6e778558065/aws/cpi.yml#L77
В этой строчке переменные из запуска create-env вставляются в таску
Здесь надо сделать то же самое
https://bosh.io/docs/init-aws/
Шаг 2
Там показано как передавать переменные из командной строки
Status: Issue closed
Answers:
username_0: https://github.com/cloudfoundry/bosh-deployment/blob/d0fb8eb4eb82aef502af44ac0628d6e778558065/aws/cpi.yml#L77
В этой строчке переменные из запуска create-env вставляются в таску
Здесь надо сделать то же самое
https://bosh.io/docs/init-aws/
Шаг 2
Там показано как передавать переменные из командной строки
Status: Issue closed
|
canonical-ols/talisker | 434671681 | Title: Remove werkzeug as a dependency
Question:
username_0: We used werkzeug for WSGI responses and for local thread/greenlet context, but it would be good to remove it. This would mean some simple WSGI boilerplate for endpoints, but also need a replacement for the local context.
The obvious choice for this is the new stdlib contextvars, (with pre 3.7 backport), so we should see if that would enable dropping werkzeug
Answers:
username_0: I had a look at this, and while we don't use werkzeug.local any more, we do still use it for a bunch of wsgi/http stuff that's non-trivial, like content negotiation, so we'd need to implement that ourselves if we want to remove it |
byuflowlab/Mex.jl | 1049215734 | Title: unexpected crashes
Question:
username_0: great that this project is still being updated! I just tried out with the newest release candidate for julia which again exports options, so the build script works.
I tried to use this package in the past to write bindings for matlab against a julia optimization library (I have attached a toy version
[testmex.zip](https://github.com/byuflowlab/Mex.jl/files/7508486/testmex.zip)
). I want to call a julia function from matlab, and supply it with function handles. Then in turn, the julia function will call those matlab function handles with data.
This works, as you can see. However, if you run it often enough (matlab 2018a, ubuntu linux), I get an unexpected crash. The fact that this is sporadic makes me suspect the julia garbage collector, is there something to be mindful about when using mex?
I've tried debugging this but I'm hopelessly out of my depth, I never succesfully attached gdb to the matlab process.
Answers:
username_1: I've also had occasionally crashing when using the `call_matlab` function. I don't know if this is occurs due to a limitation imposed on Mex files by MATLAB or due to improper data handling on the Julia side and I haven't been able to pinpoint the root cause of this issue. For not being very large, this package is quite complicated since it deals with two higher level languages and their respective lower level interfaces. I've often thought it would be nice to have someone who is an expert in Mex files and/or embedding Julia take a look at this package. I don't consider myself to be an expert in either.
As a workaround, try to replace all instances of `call_matlab` with equivalent calls to MATLAB using MATLAB.jl. The downside of this approach is that it limits the code you can call to what is available to the MATLAB engine session created by MATLAB.jl. Hopefully, this works for your application.
username_0: Well the attached project is primarily a test on the stability of mex.jl, on a fork that was made quite a while ago. Honestly if the issue is only in the call_matlab function, then everything I really need for my actual purposes already works! I will experiment a bit with some tests that do not work with call_matlab.
Though if the issue is really only there, and probably being a GC issue - seeing how sporadic the crashes are - I can even try to create a minimal working(crashing) example. I won't have time for the coming 3 weeks, but if I have anything I will try to propose a fix or at the very least post it somewhere hereabouts!
In any case thanks a lot for updates to this package!
username_0: ok the easiest crash I can find is:
```
jl.eval('invalidsyntax()')
jl.eval('GC.gc()')
```
this is perhaps not very important, but pretty much any error you make in julia will later on lead to an unexpected crash when garbage collection triggers.
A totally unrelated minimal example is this:
```
function test(args::Vector{MATLAB.MxArray})
out = jvalue(args[3])
temp = args[2];
for i in 1:out
temp = call_matlab(1,"feval",[args[1],temp])
GC.gc()
end
temp
end
```
`jl.mex('test',@(x) x*x,rand(5),1)`
will run without any problems, even when called multiple times. However this
`jl.mex('test',@(x) x*x,rand(5),2)`
immediately comes crumbling to the ground.
username_0: I feel rather silly, as the second example is actually also an error.
```
temp = call_matlab(1,"feval",[args[1],temp])
```
call_matlab will return a vector, which gets put into temp. temp then gets again given to call_matlab, which is where it goes wrong.
it should've been
```
temp = call_matlab(1,"feval",[args[1],temp])[1]
```
which works just fine?
So then, it mostly seems that julia errors in mexed calls or matlab errors in callbacks will lead to some invalid memory somewhere and then crashes later on.
username_1: I tried to dig into this, but can't figure out where the issue is. It seems to just affect the `call_matlab` function though. The simplest example I can find is
```matlab
jleval call_matlab(1, "sin", 1.0)
```
Note that this function doesn't error, but this does cause a segmentation fault when MATLAB is closed (which can be seen if MATLAB is run from the terminal). |
ThirteenAG/WidescreenFixesPack | 256737849 | Title: [NFSMW] HUD crop on 16:10 monitor
Question:
username_0: In-race HUD doesn`t fit in 16:10 ratio (I have 1680x1050 resolution). It`s cropping by left and right sides.
Please fix it.
And thanks for your work, you awesome.
Answers:
username_1: Dupe of #311, #214, #190 and more I think.
They decided to use vanilla military (1.85:1) aspect ratio HUD to fix widescreen mode. The elements are movable like NFSU and NFSU2 widescreen fixes (also in groups, which is easier), but I don't think they will implement it there too.
username_2: See https://github.com/username_3/WidescreenFixesPack/issues/423#issuecomment-370203252
Status: Issue closed
|
GlowstoneMC/Glowstone | 56932332 | Title: Scoreboard command [bug]
Question:
username_0: 01:18:14 [INFO] username_0 issued command: /scoreboard teams
01:18:14 [SEVERE] Exception while executing command: /scoreboard teams
org.bukkit.command.CommandException: Unhandled exception executing 'scoreboard t
eams' in org.bukkit.command.defaults.ScoreboardCommand(scoreboard)
at org.bukkit.command.SimpleCommandMap.dispatch(SimpleCommandMap.java:18
4)
at net.glowstone.GlowServer.dispatchCommand(GlowServer.java:1055)
at net.glowstone.entity.GlowPlayer$2.run(GlowPlayer.java:1297)
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.util.concurrent.FutureTask.run(Unknown Source)
at net.glowstone.scheduler.GlowTask.run(GlowTask.java:163)
at net.glowstone.scheduler.GlowScheduler.pulse(GlowScheduler.java:182)
at net.glowstone.scheduler.GlowScheduler.access$100(GlowScheduler.java:2
5)
at net.glowstone.scheduler.GlowScheduler$2.run(GlowScheduler.java:111)
at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source)
at java.util.concurrent.FutureTask.runAndReset(Unknown Source)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.
access$301(Unknown Source)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.
run(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.lang.Thread.run(Unknown Source)
Caused by: java.lang.NullPointerException
at org.bukkit.command.defaults.ScoreboardCommand.execute(ScoreboardComma
nd.java:80)
at org.bukkit.command.SimpleCommandMap.dispatch(SimpleCommandMap.java:18
0)
... 15 more
Answers:
username_1: Scoreboards are not implemented.
Status: Issue closed
username_2: Closing as a duplicate of #81 (currently implementation is #459) |
google/tsunami-security-scanner-plugins | 1015522443 | Title: PRP: Request alibaba nacos(1.4.0) 'NACOS-ISSUE #4463' execute arbitrary SQL without authentication
Question:
username_0: Hi,
I want to contribute to the tsunami scanner with a plugin to detect alibaba nacos(<=1.4.0) 'NACOS-ISSUE #4463'execute arbitrary SQL without authentication.
Vulnerability details:
alibaba/nacos#4463
When the nacos version is less than or equal to 1.4.0, it can be accessed without authentication and execute arbitrary SQL queries, which leads to the disclosure of sensitive information.
It can be seen from the issue that this vulnerability has been fixed, but there is no CVE number.
Type:
- CWE-89: Improper Neutralization of Special Elements used in an SQL Command
The vulnerability should have a HIGH or CRITICAL severity rating if there is already a CVE ID assigned (CVSS score >= 7.0). yes
The vulnerability should be relatively new and have already been patched. yes
The vulnerability should have a relatively large impact radius. yes
The vulnerability should be remotely exploitable without authentication and user interaction. yes
The detector should provide a reliable false-positive free detection report. yes
The detector should have good unit test coverage. Google's open source projects should be thoroughly tested and there is no exception for the Tsunami project. yes
The detection capability should be easy to verify using both vulnerable and fixed Docker images. yes
Answers:
username_0: dockerfile: https://github.com/username_0/nacos-docker
username_0: 1. git clone https://github.com/username_0/nacos-docker.git
2. build.sh
3. run.sh |
fieldmuseum/EMu-Documentation | 1159011227 | Title: ebibliography Topic 8: Shared Contributor Role lookup
Question:
username_0: DISCUSSION POINTS: Contributor Role is shared with emultimedia. Will need to reach and agreement with the emultimedia group as to how shared values should/should not be used.
Answers:
username_0: DECISION(S): For bibliography records we will use a subset of the Dublin Core terms in the [Relator Terms and Dublin Core Elements](http://memory.loc.gov/diglib/loc.terms/relators/dc-contributor.html). See Established Standards above.
DATA CHANGES:
[Data Changes Doc](https://docs.google.com/spreadsheets/d/1pEyMm7J9t5wD9Xf3CZ6_VGBmFZloEtVLWviZ54OmAkE/edit?usp=sharing)
Pre new development:
Post development: |
oozcitak/xmlbuilder-js | 321804712 | Title: Is `skipNullValues` working?
Question:
username_0: With it enabled I am still getting something like '<CityName/>' which is supposed to be discarded.
Answers:
username_1: I'm getting empty tags when keys are set to null values. Seems like this isn't working totally.
username_1: All references to this options seem to simply be comments. Is it even implemented?
username_2: The flag is actually named `skipNullNodes`. Sorry for the typo; fixed in e7c4a61896cb8007020da4af89ed0b02e4d9fd15 and added to the wiki.
Status: Issue closed
|
cockroachdb/docs | 381234281 | Title: run sysbench
Question:
username_0: `sysbench` is another workload like `kv`, which stresses per-node performance (as opposed to TPCC which stresses us moreso in terms of scale). We should include some docs on how to use `sysbench` with Cockroach.
Answers:
username_1: Bash script from @nvanbenschoten:
```
# Config
export CLUSTER=$USER-perf
# For write workload
export WORKLOAD=oltp_insert
export CONCURRENCY_PER_LOAD_GEN=160
# For read workload
export WORKLOAD=oltp_point_select
export CONCURRENCY_PER_LOAD_GEN=72
# Create cluster
roachprod create $CLUSTER -n=6 --clouds aws --aws-machine-type-ssd=c5d.9xlarge --geo --aws-zones='us-east-1a,us-east-1b,us-east-1c'
roachprod stage $CLUSTER cockroach
roachprod install $CLUSTER:4-6 sysbench
# Start CockroachDB
roachprod start $CLUSTER:1-3
sleep 20
# Prepare benchmark
for i in {1..3}; do
roachprod run $CLUSTER:1 -- "./cockroach sql --insecure -e \"CREATE DATABASE sysbench${i}\""
roachprod run $CLUSTER:4 -- "sysbench ${WORKLOAD} \
--db-driver=pgsql \
--pgsql-host={pghost:1} \
--pgsql-port=26257 \
--pgsql-user=root \
--pgsql-password= \
--pgsql-db=sysbench${i} \
--report-interval=1 \
--time=120 \
--threads=256 \
--tables=4 \
--table_size=1000000 \
--create_secondary=false \
--auto_inc=false \
prepare"
done
# Set lease_preferences to manually balance leases between nodes
roachprod run $CLUSTER:1 -- "./cockroach sql --insecure -e \"alter database sysbench1 CONFIGURE ZONE USING constraints = COPY FROM PARENT, lease_preferences = '[[+zone=us-east-1a]]';\""
roachprod run $CLUSTER:1 -- "./cockroach sql --insecure -e \"alter database sysbench2 CONFIGURE ZONE USING constraints = COPY FROM PARENT, lease_preferences = '[[+zone=us-east-1b]]';\""
roachprod run $CLUSTER:1 -- "./cockroach sql --insecure -e \"alter database sysbench3 CONFIGURE ZONE USING constraints = COPY FROM PARENT, lease_preferences = '[[+zone=us-east-1c]]';\""
# Wait for leaseholder conformance (run until all queries return true)
for i in {1..3}; do
roachprod run $CLUSTER:1 -- "./cockroach sql --insecure -e \"select count(*) = 0 as leases_balanced from crdb_internal.ranges where database_name = 'sysbench${i}' and lease_holder != ${i};\""
done
# Run benchmark from three load generators concurrently
## In terminal 1 (remember env vars)
roachprod run $CLUSTER:4 -- "sysbench ${WORKLOAD} \
--db-driver=pgsql \
--pgsql-host={pghost:1} \
--pgsql-port=26257 \
--pgsql-user=root \
--pgsql-password= \
--pgsql-db=sysbench1 \
[Truncated]
roachprod run $CLUSTER:6 -- "sysbench ${WORKLOAD} \
--db-driver=pgsql \
--pgsql-host={pghost:3} \
--pgsql-port=26257 \
--pgsql-user=root \
--pgsql-password= \
--pgsql-db=sysbench3 \
--report-interval=1 \
--time=120 \
--threads=${CONCURRENCY_PER_LOAD_GEN} \
--tables=4 \
--table_size=1000000 \
--create_secondary=false \
--auto_inc=false \
run"
# Merge performance numbers
## Throughput = SUM("X per sec.")
## Avg Latency = AVG("avg: X")
```
username_1: @johnrk, closing this for now. If this work comes out of your Product Area problems to solve or priorities in the future, we can create a new issue with more context.
Status: Issue closed
|
ccxt/ccxt | 302589798 | Title: inconsistent timeframes format
Question:
username_0: ATTENTION!!!
MUST READ THIS BEFORE SUBMITTING ISSUES:
https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-submit-an-issue
The issue is that the format of the exchange property timeframes is NOT consistent.
```
exchange_ids = ['binance', 'bitfinex', 'hitbtc', 'kucoin', 'kraken', 'poloniex', 'gdax', 'gemini', 'huobipro', 'okex']
exchanges = [getattr(ccxt, str.lower())() for str in exchange_ids]
timeframes = [ex.timeframes for ex in exchanges if hasattr(ex, 'timeframes')]
print(timeframes)
```
The above code prints out some examples. `timeframes` is a hash and the values of it do not follow a format across exchanges. For example, some are like '1m' or '5m' that are no different from the keys. Some others are the actual duration in seconds which arguably is the desired format. Some others are just more descriptive keys such as '1min', '1day' and etc.
- OS:
OSX 10.13.3
- Programming Language version:
Python 3.6.4
- CCXT version:
1.10.760
- Exchange:
binance
bitfinex
hitbtc
kucoin
kraken
poloniex
gdax
gemini
huobipro
okex
- Method:
the property timeframes
Answers:
username_1: `timeframes` is a hash, where keys are unified and values are exchange-specific, so, I don't see any inconsistencies there. You should not rely on the values under the keys and their formats, as they are exchange-specific. The keys are exchange-agnostic and unified and can be safely used across exchanges, namely, the "1m", "5m", "15m", "1h", and so on – those are standardized. So, I don't see an issue here, it's a matter of interpretation. If you need to get the duration in seconds, use the parseTimeframe/parse_timeframe method. Hope it helps. Let us know if it does not answer your question. Thx!
Status: Issue closed
username_0: makes sense. Thanks! |
openfl/lime | 27183676 | Title: Evolve the joystick API
Question:
username_0: I believe we should have a higher level gamepad API, that works above the existing joystick events.
Rather than subscribing to be notified when a device hits a certain key, this API should allow for profiles in order to translate events into buttons, and (ideally) should allow polling, if you wish to see which controllers are connected, listen to specific controllers only, at any point in time, check the state of a particular controller, etc.
Partially this will require implementation in Lime, and will also require new APIs to manage it either at a Lime level or in OpenFL.
Answers:
username_0: Not going to track TODO items as issues :smile:
Status: Issue closed
|
thomthom/skippy | 290270861 | Title: [Question] extensio.json and custom templates (and the scope of Skippy)
Question:
username_0: Hi!
In my personal view the extension loader file from Skippy's standard template is quite complex. Especially for really small plugins and example meant for new developers (who haven't reached the point of using Skippy themselves). I think the whole boilerplate for loading extension info from the json file can add unwanted complexity and noise.
I tried making a simpler template with all extension information directly in the loader, [similar to this](https://github.com/username_0/flatten-to-plane/blob/master/src/ene_flatten_to_plane.rb), but noticed the extension.json file wasn't a part of the template and was generated anyway.
Is this file planned to have a more integrated role in Skippy in the future? E.g. having commands for changing the project author, copyright, license etc and have both skippy.json *and* extension.json updated?
This also leads to the question of the scope of Skippy. So far I've used it solely to install dependencies. It's not obvious to me how the project author, description or license are relevant for Skippy. Does Skippy use this information somehow now, or is planned to do so in the future?
Answers:
username_1: The template system is just a stub right now. Eventually there will be commands to install templates.
Same for the info in `skippy.json` - it's not hooked up yet, but the idea was to allow that to be used in things like templates etc.
The README describe some of the main goals:
* Template system to quickly scaffold new extensions. This is something developers can customize to suit their own style.
* Automation - being able to quickly write custom skippy commands per project. (As well as having skippy ship with some out of the box common commands.)
* Basic library/dependency management.
username_1: As for `extension.json`, the template generater needs options to handle that. |
viniciussanchez/dataset-serialize | 855306128 | Title: Samples\Lazarus\basic\basic.lpi can not compile, TArray<string> not availble yet in FPC?
Question:
username_0: Hi,
When I try to compile the project basic.lpi I get errors.
```
unit DataSet.Serialize.Samples.Basic;
(line 47) FDataSetPrefix: TArray<string>;
```
From what I've seen on Lazarus Forums
` TArray<string> may not yet be implemented in FPC?`
Is there a work around for Lazarus yet available please?
```
unit DataSet.Serialize.Samples.Basic;
Compile Project, Target: basic.exe: Exit code 1, Errors: 3
DataSet.Serialize.Config.pas(47,21) Error: Identifier not found "TArray"
DataSet.Serialize.Config.pas(47,27) Error: Error in type definition
DataSet.Serialize.Config.pas(47,27) Fatal: Syntax error, ";" expected but "<" found
```
Lazarus 2.0.8 r62944 FPC 3.0.4 x86_64-win64-win32/win64
Kind regards,
Paul
Answers:
username_0: For the sake completeness I installed the latest installation ..
`Lazarus 2.0.12 r64642 FPC 3.2.0 x86_64-win64-win32/win64`
And everything compiles very nicely thank you.
I do not know exactly which release of Lazarus/FPC introdices TArray<string> etc.,
but the above version does work.
Might be worth pitching that in the Read Me adding a section for Lazarus as well as Delphi?
Thnaks again for very fine work.
username_1: Hi ... thanks for reporting.
I only tested DataSet-Serialize using Lazarus in version 2.0.10
Status: Issue closed
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.