repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
Monty15/Capstone | 284806249 | Title: Viewport meta tag
Question:
username_0: https://github.com/Monty15/Capstone/blob/master/index.html#L4-L9
Remember to include the `viewport` meta tag to properly control your website's viewport width.
```html
<meta name="viewport" content="width=device-width, initial-scale=1.0">
```
💡 This is particularly important for any mobile responsive website. |
vladbalmos/mitzasql | 741618854 | Title: Database names with dashes cause error 1064 (42000)
Question:
username_0: **Describe the bug**
When selecting a database with a dash "-" in the name the following error is displayed.
1064 (42000): You have an error in your SQL syntax; check the manual that corresponds to your MySQL server version for the right syntax to use near '-test' at line 1
**To Reproduce**
Steps to reproduce the behavior:
1. Create a database with name "wordpress-test"
2. Open Mitzasql and navigate to this database
3. When trying to enter the db you can see the error.
4. See error
**Expected behavior**
Tables inside the database would be displayed.
**Screenshots**
https://i.imgur.com/LsCfzAQ.png
**Desktop (please complete the following information):**
- OS: Debian Buster
- Python version 3.7.3
Answers:
username_1: Good find, thanks!
Status: Issue closed
username_0: Works like a charm. Thanks! |
utPLSQL/utPLSQL | 446432444 | Title: Add JSON object comparision
Question:
username_0: Since JSON is gaining a popularity and since Oracle 12.2 we got a proper support of JSON in database I thought it would b a good idea to implement a JSON comparison matcher.
I've did some initial work for cursor equality and will be adding a more methods.<issue_closed>
Status: Issue closed |
zalando/patroni | 778628705 | Title: Available implementations: """ + ', '.join(sorted(set(available_implementations)))) patroni.exceptions.PatroniFatalException: 'Can not find suitable configuration of distributed configuration store\nAvailable implementations: kubernetes, raft'
Question:
username_0: Hi ,i get same error like this:
2021-01-05 13:15:11,134 INFO: Failed to import patroni.dcs.consul
2021-01-05 13:15:11,139 INFO: Failed to import patroni.dcs.etcd
2021-01-05 13:15:11,142 INFO: Failed to import patroni.dcs.etcd3
2021-01-05 13:15:11,149 INFO: Failed to import patroni.dcs.exhibitor
2021-01-05 13:15:11,162 INFO: Failed to import patroni.dcs.zookeeper
Traceback (most recent call last):
File "patroni.py", line 6, in
main()
File "/home/kingbase/patroni/patroni/init.py", line 170, in main
return patroni_main()
File "/home/kingbase/patroni/patroni/init.py", line 138, in patroni_main
abstract_main(Patroni, schema)
File "/home/kingbase/patroni/patroni/daemon.py", line 98, in abstract_main
controller = cls(config)
File "/home/kingbase/patroni/patroni/init.py", line 29, in init
self.dcs = get_dcs(self.config)
File "/home/kingbase/patroni/patroni/dcs/init.py", line 107, in get_dcs
Available implementations: """ + ', '.join(sorted(set(available_implementations))))
patroni.exceptions.PatroniFatalException: 'Can not find suitable configuration of distributed configuration store\nAvailable implementations: kubernetes, raft'
My enviroment like this:
Python 3.6
pip3 20.3.3
pip3 list
Package Version
boto 2.49.0
certifi 2020.12.5
chardet 4.0.0
click 7.1.2
dnspython 2.0.0
flake8 3.8.4
idna 2.10
importlib-metadata 3.3.0
kazoo 2.8.0
mccabe 0.6.1
patroni 2.5
pip 20.3.3
prettytable 2.0.0
psutil 5.8.0
psycopg2 2.5.4
pycodestyle 2.6.0
pyflakes 2.2.0
pysyncobj 0.3.7
python-consul 1.1.0
python-dateutil 2.8.1
python-etcd 0.4.5
PyYAML 5.3.1
requests 2.25.1
setuptools 28.8.0
six 1.15.0
typing-extensions 3.7.4.3
urllib3 1.26.2
wcwidth 0.2.5
wheel 0.36.2
ydiff 1.2
[Truncated]
#timezone: 'PRC'
#lc_messages: 'C'
#lc_monetary: 'C'
#lc_numeric: 'C'
#lc_time: 'C'
#full_page_writes: on
#synchronous_commit: on
#wal_log_hints: on
#synchronous_standby_names: ''
#max_replication_slots: 10
create_replica_methods:
- basebackup
basebackup:
max-rate: '100M'
tags:
nofailover: false
noloadbalance: false
clonefrom: false
nosync: false
Answers:
username_1: **python-etcd** module in order to use Etcd as DCS
username_0: May i run it by source code?
If i user python2.7 ,this is ok!
username_1: It doesn't matter how you run it, `python-etcd` module must be installed if you want to run Patroni with etcd.
username_0: python-etcd 0.4.5 in the pip3 list . I am sure i install it .
username_1: Magic doesn't exists.
username_0: python-etcd 0.4.5 in the pip3 list . I am sure i install it.
It in the list:
Package Version
boto 2.49.0
certifi 2020.12.5
chardet 4.0.0
click 7.1.2
dnspython 2.0.0
flake8 3.8.4
idna 2.10
importlib-metadata 3.3.0
kazoo 2.8.0
mccabe 0.6.1
patroni 2.5
pip 20.3.3
prettytable 2.0.0
psutil 5.8.0
psycopg2 2.5.4
pycodestyle 2.6.0
pyflakes 2.2.0
pysyncobj 0.3.7
python-consul 1.1.0
python-dateutil 2.8.1
**### python-etcd 0.4.5**
PyYAML 5.3.1
requests 2.25.1
setuptools 28.8.0
six 1.15.0
typing-extensions 3.7.4.3
urllib3 1.26.2
wcwidth 0.2.5
wheel 0.36.2
ydiff 1.2
zipp 3.4.0
username_1: I totally don't undestand what you are doing. The only purpose of release.sh is publishing Patroni on PyPI.
username_0: I need change some code for our database ,the database base on postgresql, so i need build the package to publish some users.
username_1: Perfect, then you should be qualified enough to figure out what is wrong with _your system_ and why python can't find/load certain modules.
The recommended way of installing Patroni either from PyPI (pip install patroni[$EXTRAS]) or via PGDG packages (apt-get/yum install). We can't be responsible for something that you do on your own.
Status: Issue closed
|
chargebee/chargebee-dotnet | 422252320 | Title: Upgrade from 2.6.7 to 2.6.9 -> Timestamp errors
Question:
username_0: Hi,
I've updated from version 2.6.7 -> 2.6.9.
Everything compiled fine and everything seemed fined after a few tests.
Then I received a webhook from ChargeBee and everything went down...
When creating a new webhook event from an io stream, I get this error:
`OccurredAt = 'WebhookEvent.OccurredAt' threw an exception of type 'System.TypeInitializationException'`
This is how I create the webhookEvent object:
` Public Sub ProcessRequest(context As HttpContext) Implements IHttpHandler.ProcessRequest
Dim Response As String = Nothing
Dim ResponseCode As HttpStatusCode = HttpStatusCode.OK
Try
' Input
Using StreamReader As New IO.StreamReader(context.Request.InputStream)
InputData = StreamReader.ReadToEnd()
End Using
If String.IsNullOrWhiteSpace(InputData) Then Throw New ArgumentNullException("The data cannot be null.")
' Create the webhook event
WebhookEvent = New ChargeBee.Models.Event(InputData)
[...]
Catch ex as Exception
End Try
[...]
End Sub
`
I checked and the timestamp is fine in the IO stream (1552915864).
All the modules are up to date, everything else is working fine, I just can't make the webhook object creation works.
If I use 2.6.7, this works flawlessly, in fact it's been working flawlessly all the way back with version 1.x.x.
Please fix for the next release
Answers:
username_1: Stumbled upon same issue. Using v2.8.0 and it still fails for any date.
username_2: Still broken in 2.8.5
username_3: Still broken in 2.8.7
username_4: **Update**
I've developed a small fix for the conversion between Unix TimeStamp and UTC DateTime.
Couldn't find how provide a solution to creators, so I pushed the fix to my personal space. https://github.com/username_4/chargebee-dotnet/tree/bugfix/patch-timestamp-datetime-conversion
Hope it helps
username_4: **Update**
I've managed to fix this and now on my way to fork and provide the solution I've found to work.
username_4: ### **[SOLVED]**
**TL;DR;**
Be sure to call **ChargeBee.Api.ApiConfig.Configure** before accessing DateTime or Timestamp fields to avoid unreasonable Exceptions, even when unit testing or offline.
**Summary:**
Accessing DateTime and Timestamp fields fail because the timeout for an HttpClientAdapter can't be zero.
Adding one sentence to @username_3 [comment ](https://github.com/chargebee/chargebee-dotnet/issues/18#issuecomment-737283847) makes it work.
```
ChargeBee.Api.ApiConfig.Configure("fake-sitename", "fake-apikey");
var cbEvent = new Event(jsonProvidedAsString);
var termEnd = cbEvent.Content.Subscription.CurrentTermEnd;
```
ChargeBee Api needs to be initialised before accessing its model. Otherwise it will throw exception when creating **HttpClient** on **ApiUtil** constructor. The same class contains **ConvertFromTimestamp** and **ConverttoTimestamp** methods triggering unexpected errors
https://github.com/chargebee/chargebee-dotnet/blob/1d6934b645c31e9b3e17cc0fbbbf34d582334f0c/ChargeBee/Api/ApiUtil.cs#L21
ApiConfig.ConnectionTimeout is zero until `ChargeBee.Api.ApiConfig(<string>,<string>)` is run, being first assigned here:
https://github.com/chargebee/chargebee-dotnet/blob/1d6934b645c31e9b3e17cc0fbbbf34d582334f0c/ChargeBee/Api/ApiConfig.cs#L45
Thereby, if **ApiConfig.Configure** is never invoked it would always raise an exception because the timeout for an HttpClientAdapter can't be zero.
Thank you @username_3 on how to reproduce. and @username_0, @username_1, @username_2 for also reporting
username_4: #27 addresses this.
username_5: This has been addressed in v2.10.0
Status: Issue closed
|
bazelbuild/bazel | 166594096 | Title: Easy to mistakenly install bazel without specifying "--user"
Question:
username_0: When installing Bazel for the first time, I made the fairly obvious mistake of omitting the "--user" flag. The error I got back suggested I should use "sudo," which I did. Since I don't seem to be the only person to make this mistake (cf the title of #962), perhaps if the installer is about to error out due to insufficient permissions, it would be better to say something like:
`If you are just installing this for yourself, use --user, but if you would like it to be accessible to everyone on the machine, ensure you have access to...`
Answers:
username_1: The error I get does not mention sudo, it says:
```
The Bazel installer must have write access to /usr/local/bin!
Usage: bazel-0.3.0-jdk7-installer-linux-x86_64.sh [options]
```
and then prints the usage string.
What version of the installer are you using?
username_0: My mistake, I simply interpreted the lack of permissions as "I need to use sudo." I then avoided looking a few lines down to see the help text that referenced "--user".
Again, it's a bit of an embarrassing mistake, I only filed the issue because I found someone else who made the same mistake. So I'm just suggesting a slight change to the message.
Status: Issue closed
|
xmake-io/xmake | 796521346 | Title: Rebuild local Qt-dependent package with project Qt version?
Question:
username_0: I have Qt installed on Ubuntu via apt, as well as an older version installed via their web installer. I also have a Qt-dependent package in a local xrepo repo.
We can set the version of Qt to build the project with:
```
xmake f --qt=~/Qt/5.11.3
```
But the package is unaware of the change so continues to use the system-installed Qt. This causes version mismatch errors.
Is there or could there be a solution to this? Some way of using the currently set `qmake` in the package?
Answers:
username_1: What build system does this package use? xmake? cmake or autoconf?
username_0: It just uses `qmake` and `make`, here's what it looks like at the moment:
```lua
package("qscintilla")
set_homepage("https://www.riverbankcomputing.com/software/qscintilla/intro")
set_description("QScintilla is a port to Qt of <NAME>'s Scintilla C++ editor control.")
set_urls("https://www.riverbankcomputing.com/static/Downloads/QScintilla/$(version)/QScintilla-$(version).zip")
add_versions("2.11.6", "ddd0945d90bbf9394e0d4a41cfeb5bd7c1a6b918c827aa90d4396ea3da0be9a9")
on_install("linux", function(package)
os.execv("qmake", { }, { curdir = "./Qt4Qt5"})
os.execv("make", { "-j8" }, { curdir = "./Qt4Qt5" })
os.cp("./Qt4Qt5/libqscintilla2_qt5.so.15.0.0", path.join(package:installdir("lib"), "libqscintilla2_qt5.so"))
os.cp("./Qt4Qt5/libqscintilla2_qt5.so.15.0.0", path.join(package:installdir("lib"), "libqscintilla2_qt5.so.15.0.0"))
os.cp("./Qt4Qt5/libqscintilla2_qt5.so.15.0.0", path.join(package:installdir("lib"), "libqscintilla2_qt5.so.15"))
os.cp("./Qt4Qt5/Qsci/*.h", package:installdir("include/Qsci"))
end)
```
username_1: ```lua
on_install("linux", function(package)
local qtdir = get_config("qt")
end)
```
username_0: Thanks, that resolves the correct path, but `os.execv` doesn't like it:
```lua
local qtdir = get_config("qt")
os.execv(qtdir .. "/gcc_64/bin/qmake", { }, { curdir = "./Qt4Qt5"})
```
`xmake f -cvD --qt=~/Qt/5.11.3` error:
```bash
error: @programdir/core/sandbox/modules/os.lua:387: execv(~/Qt/5.11.3/gcc_64/bin/qmake ) failed(255)
stack traceback:
[C]: in function 'error'
[@programdir/core/base/os.lua:787]: in function 'raise'
[@programdir/core/sandbox/modules/os.lua:387]: in function 'execv'
[./xmake-pkgs/packages/q/qscintilla/xmake.lua:12]: in function 'script'
[...gramdir/actions/require/impl/actions/../utils/filter.lua:125]: in function 'call'
[@programdir/actions/require/impl/actions/install.lua:168]:
[C]: in function 'trycall'
[@programdir/core/sandbox/modules/try.lua:121]: in function 'try'
[@programdir/actions/require/impl/actions/install.lua:127]: in function 'action_install'
[@programdir/actions/require/impl/package.lua:831]: in function 'jobfunc'
[@programdir/modules/private/async/runjobs.lua:193]:
[C]: in function 'trycall'
[@programdir/core/sandbox/modules/try.lua:121]: in function 'try'
[@programdir/modules/private/async/runjobs.lua:186]: in function 'cotask'
[@programdir/core/base/scheduler.lua:317]:
=> install qscintilla 2.11.6 .. failed
error: @programdir/modules/private/async/runjobs.lua:217: @programdir/actions/require/impl/actions/install.lua:256: install failed!
stack traceback:
[C]: in function 'error'
[@programdir/core/base/os.lua:787]: in function 'raise'
[@programdir/actions/require/impl/actions/install.lua:256]: in function 'catch'
[@programdir/core/sandbox/modules/try.lua:127]: in function 'try'
[@programdir/actions/require/impl/actions/install.lua:127]: in function 'action_install'
[@programdir/actions/require/impl/package.lua:831]: in function 'jobfunc'
[@programdir/modules/private/async/runjobs.lua:193]:
[C]: in function 'trycall'
[@programdir/core/sandbox/modules/try.lua:121]: in function 'try'
[@programdir/modules/private/async/runjobs.lua:186]: in function 'cotask'
[@programdir/core/base/scheduler.lua:317]:
stack traceback:
[C]: in function 'error'
@programdir/core/base/os.lua:787: in function 'raise'
@programdir/modules/private/async/runjobs.lua:217: in function 'catch'
@programdir/core/sandbox/modules/try.lua:127: in function 'try'
@programdir/modules/private/async/runjobs.lua:186: in function 'cotask'
@programdir/core/base/scheduler.lua:317: in function <@programdir/core/base/scheduler.lua:315>
```
username_1: `~/xx`, you need use path.translate or path.absolute for qmake path.
username_0: Yay, that's it - thanks!
Status: Issue closed
|
nornir-automation/nornir | 336520004 | Title: Closing NAPALM connections automatically
Question:
username_0: When using NAPALM plugins connections stay open and can cause scripts to hang while waiting for the connections to timeout.
Forcing the connections to close using a task helps with this.
```python
task.host.connections["napalm"].close()
```
Maybe this could be done automatically? It'll require to find a way to know which connections are not useful anymore.
Answers:
username_1: How would we know that you didn't want to use the connection any more (i.e. in a subsequent task)?
The process is probably to have the code writer to explicitly close the connection in some way.
username_0: Yep that's complicated. I opened the issue because @username_2 wanted to track it.
I don't see any simple way to do this. Maybe proposing a task closing all NAPALM connections could be an answer to this.
username_2: I thought we could look into creating a “close” task so it’s easier for the user to explicitly close the connections or maybe do it automatically via the garbage collector by implementing some code in the __del__ magic method.
username_3: I think the close task, or perhaps something like a context manager for Nornir could make sense. There might be scenarios where users don't want the connections to be terminated. (I've seen Huawei switches where a local script running on them stops running if the ssh connection dies).
We could have some task which just iterates all of the hosts' connection dictionaries and calls a disconnect function. The main issue is perhaps that all of the connection plugins be consistent and support the same close method.
username_2: For reference, some work that will enable this is being done in #189
username_4: @username_2
I think this is fixed with #195 and could be closed? |
freedomofpress/securedrop-ux | 408375462 | Title: Online/Offline UI Differences
Question:
username_0: ### Problem
When a journalist views the Client in offline mode, what does that look like?
### Considerations
- How will parts of the UI requiring network connectivity be shown to communicate their non-functionality in Offline Mode?
- Will something outright go away, or be greyed-out?
- Where might there be suggestions or other affordances made to nudge a user to Sign In?
- When accessing the client for the first time in Offline mode, what might that experience look like?
### Acceptance Criteria
- Clickable wireframe demonstrating interactivity
- All state changes clearly itemized/outlined
- Spec'd wireframe, or multiple comments here?
- Get on the needs testing punchlist
---
This is a sub-task within #18 and #31 and #17
Answers:
username_0: First Rev (submitted on Gitter, 07 Feb): CLICK ON TOP BAR to toggle back and forth between offline and online. Delete functionality for a Source Account also goes away, as does the drop-down menu on the Messages pane. https://invis.io/X4QEP5Y9EAT#/345671069_0_New_-_DR_On-_OFFLINE
username_0: ### Feedback (cut-and-pasted from Gitter):
<NAME>
@eloquence
Feb 07 17:12
I'm a fan! I like avoiding the standard "grayed out UI controls" pattern which can be very frustrating for users, and this also takes up less space
<NAME> Alter
@username_0
Feb 07 17:15
Oh—fwiw, clarification with the above—in the wireframe, the paper airplane dealie is the "Send" button. I believe the existing Client has a proper button with the word "Send" on it
@eloquence Yay! Ya... the "omg, why won't this friggin' work?!" frustration quite sux.
<NAME>
@username_1
Feb 07 17:20
hey @username_0 this looks awesome
<NAME>
@username_0
Feb 07 17:24
Coolio!
<NAME>
@username_1
Feb 07 17:24
trying to find the image of what will be on the ... menu
<NAME>
@username_0
Feb 07 17:33
@username_1 You mean the content in that menu?
I think @eloquence will eventually do that menu as its own Issue/Story
"Export All Messages & Files" and "Delete Source" are the only two items therein, for now.
^ ...or probz "Export Source." We never did much testing on that menu.
username_0: (crap, sorry that @'d everyone!)
username_1: @username_0 - Hey just saw this and noticed that the conversation in gitter was left open-ended. I noticed while toggling back and forth between offline and online mode that you get a `. . .` options menu next to the date when online, so I was wondering if you had created any wireframes for this menu. I assume it'll include a way to logout? Is there anything else?
username_0: @username_1 Nope, logging-out is handled from the user avatar having a dropdown-ish type thing. The dot-dot-dot menu you cite currently only has "Export All Submissions" and "Delete Account" on it, as of now. Older wireframes should show more functionality on it; I also don't really know what users may want in it, so could easily pack it with a dozen account-level options like "Export all files, delete all local files" etc... but wanna keep it to just those two account-level things, for now.
Anything pertaining to the Journalist's authentication or connectivity happens in the top bar.
username_0: I'd forgotten this ticket still existed!
**Explorations**, here: [» Invision Mox (semi-interactive) «](https://invis.io/BSRX8TUN8K3)
**Final direction**, in Zeplin (ignore "Empty" pane content): [Public Screen](https://scene.zeplin.io/project/5c807ea562f734bd2756b243/screen/5cd35ec0df6a8967aedc0d95)
Status: Issue closed
username_0: Closing, cuz this is done? @eloquence feel free to reopen if you disagree. Wishing this had gone into a review, but there is only so much team bandwidth and time... :) |
DoctorVanGogh/ExtendedStorage | 247234384 | Title: Add toggle for user/real storage settings
Question:
username_0: implemented in [branch](/username_0/ExtendedStorage/tree/feature/settings-debug)

Status: Issue closed
Answers:
username_0: implemented in [branch](/username_0/ExtendedStorage/tree/feature/settings-debug)

Status: Issue closed
|
moby/moby | 758436900 | Title: IPv6 options refinement: --ipv6 behaving better out of the box (concerns both options and default behavior regarding IPv6 in custom networks)
Question:
username_0: I am just an outsider and not really a code contributor to docker/moby myself, so I just hope to get the discussion going. Basically I'm saying, just take my suggestions as a vague input but I shouldn't be the person to make any of the final calls, in the end I'm just an end-user as many others. Hopefully, this ticket will help the right people to come to the right conclusions!
Answers:
username_1: I have two related suggestions:
- In addition to setting `--fixed-cidr-ipv6` to a private subnet when using ipv6 NAT, I suspect that `default-address-pools` should default to having a private subnet available too (maybe even *without* using ipv6 NAT), so you can create a custom network with `--ipv6` and have it assigned an address without having to manually manage this, just like with ipv4.
- It might be useful if some option would be added (maybe even enabled by default) to get ipv6 enabled on user-created networks by default. It seems that you need to manually specify `--ipv6` to `network create` right now, it would be good if this could be made the default (distribution default or system default).
username_0: I like those ideas! I agree in particular with it being a good suggestion that the IPv6 being enabled by default with a new option to override that when it's enabled for the daemon. I think the point of having this finally pretty "auto-magical" IPv6 NAT that works like almost IPv4 should be to have everyone do less manual tweaks to get it to "just work" on existing setups, including those with custom networks, so that seems like a really good idea to me.
username_0: Any updates? I'm a bit concerned this might take another five years to get looked at, at which point people will probably be stung badly if these options are changed. Now seems like the best time if any to still integrate this.
username_2: Just to add to the quorum. Having `--ip6tables` as a non-experimental flag would certainly help me sell the feature to my company. We are currently using `docker-ipv6nat` https://github.com/robbertkl/docker-ipv6nat/issues/65.
To add something to this discussion (I don't know if this should be a separate ticket or not), docker-compose default created network should also be IPv6 enabled when `--ipv6` and `--ip6tables` are set. As it stands the difference of behaviour between docker and docker-compose is not very intuitive. A docker container using the default network is IPv6 enabled, but a container started by docker-compose isn't.
username_3: I was recently wondering about this, myself.
I'd like to do IPv6 NAT on my Docker swarms, without hacking [robertkl/docker-ipv6nat to work](https://github.com/robbertkl/docker-ipv6nat/issues/12) -- if that hack still works at all |
juju/python-libjuju | 1118809392 | Title: TypeError: unsupported format string passed to NoneType.__format__
Question:
username_0: Function "_print_status_apps" in file juju/status.py, below piece of code should check if app.workload_version or app.charm_channel is None.
https://github.com/juju/python-libjuju/blob/cbeae7063f8a19dc1e1ed3638f8025735f6d9bc0/juju/status.py#L116
If a charm is deployed from local rather than from charmstore, workload_version and app.charm_channel may be None
Traceback (most recent call last):
File "./lib/test_lib_juju.py", line 26, in <module>
main()
File "./lib/test_lib_juju.py", line 7, in main
juju_status = juju_get_status("bcache", keep_relations_info=True, jsfy=False, format_status=True)
File "~lib/lib_juju.py", line 57, in juju_get_status
juju_status = loop.run(get_juju_status(model_name, keep_relations_info, jsfy, format_status))
File "~venv/lib/python3.8/site-packages/juju/jasyncio.py", line 118, in run
raise task.exception()
File "~lib/lib_juju.py", line 108, in get_juju_status
juju_status = await formatted_status(model)
File "~venv/lib/python3.8/site-packages/juju/status.py", line 61, in formatted_status
result_str += _print_status_apps(result_status)
File "~venv/lib/python3.8/site-packages/juju/status.py", line 116, in _print_status_apps
result_str += limits.format(
TypeError: unsupported format string passed to NoneType.__format__
Task was destroyed but it is pending!
Answers:
username_1: Changes from #622 should fix this. Please reopen if needed.
Status: Issue closed
|
CosmicMind/Material | 144433209 | Title: Depth Does Not Change on Pulse
Question:
username_0: When a view pulses (and scale), the shadow remains the same and does not present a change of depth. Scale I presume is there to simulate the view getting closer to the fingertip.
Status: Issue closed
Answers:
username_1: The shadow slightly spreads, due to the scale effect. I believe you could add an animation by subclassing the pulse view and adding animations to the touch handlers that spreads the shadow, and contracts it. I am not going to make this a priority right now, but we are planning some nice additions with 3D touch and such, and that is when this issue will be considered at a high priority.
Feel free to continue the dialog on this, though for cleanup reasons, I will close the issue. |
PistonDevelopers/conrod | 35726002 | Title: Widget Ideas
Question:
username_0: Here is a space to collect Widget Type ideas.
This list is taken directly from the [ofxUI] (https://github.com/rezaali/ofxUI) readme as a basis for ideas. Any more ideas are welcome!
- Buttons (push, state, toggle, image, label)
- Button Matrices
- Dropdown Menus
- Labels
- Sliders (rotary, range, vertical, horizontal)
- Number Dials (aka Spinners)
- 2D Pads
- Text Input Areas
- Image Sliders
- Image Buttons
- Image Color Sampler
- Value Plotters
- Moving Graphs
- Waveform & Spectrum Graphs
- Radio Toggles
- Text Areas
- Sortable List
Status: Issue closed
Answers:
username_0: Closing in favour of posting unique issues for drafting/discussing/requesting each widget. |
google/yapf | 1084716435 | Title: [Question]yapfignore not working in pyproject.toml
Question:
username_0: Thank you for reading this issue.
I'm currently working on a Django project and I want to use `yapf` for formatting.
However, as I am trying to use `pyproject.toml` for excluding files, it does not works.
```toml
[tool.yapfignore]
ignore_patterns = [
'**/migrations]*.py',
'manage.py',
'venv/*',
]
```
On the other hand, what makes it interesting is that the command line behaves differently and successfully ignores the files I want.
```sh
yapf --recursive --exclude '**/migrations/*.py' --exclude 'manage.py' --exclude 'venv/*' .
```
I still cannot figure out what would be the problem with this discrepancy.
At the moment, a part my dependencies is like the following:
```txt
pylint==2.12.2
pylint-django==2.4.4
pylint-plugin-utils==0.6
yapf==0.31.0
```
and my current Python version is `3.9.9`.
If there is any problem with the way I make this issue post, please don't hesitate to point it out!
Thank you in advance!
Answers:
username_0: It still doesn't work, so I had a workaround, `.yapfignore`.
```
**/migrations/*.py
manage.py
venv/*
```
I have to admit that the project would be a bit more complex, but I think `pyproject.toml` is currently not working for excluding files and directories. Any suggestions?
username_1: Hi @username_0
Related to https://github.com/google/yapf/issues/955 , I think yapf version is the problem.
Latest version (0.32.0) of yapf released 26 Dec (ref: https://pypi.org/project/yapf/#history).
In my local PC, yapf ignore feature correctly worked. Please check.
Status: Issue closed
username_0: @username_1
Thank you for your reply! I appreciate your kindness.
Oh, I have missed the issue https://github.com/google/yapf/issues/955
I have checked and now it worked.
Now this issue is resolved so I close it. |
JuliaTime/TimeZones.jl | 733844499 | Title: Parsing a datetime string including "GMT" stopped working in v1.5.0
Question:
username_0: `TimeZones.ZonedDateTime` used to work prior to release 1.5.0 when GMT was included. An error referring to legacy timezones is now returned. However, I'm not aware that such legacy timezones can be specified when parsing a string.
```julia
julia> dt_string_1 = "Sat, 31 Oct 2020 11:31:54 GMT"
"Sat, 31 Oct 2020 11:31:54 GMT"
julia> TimeZones.ZonedDateTime(dt_string_1, "e, d u Y H:M:S Z")
ERROR: ArgumentError: Unable to parse string "Sat, 31 Oct 2020 11:31:54 GMT" using format dateformat"e, d u Y H:M:S Z". The time zone "GMT" is of class `TimeZones.Class(:LEGACY)` which is currently not allowed by the mask: `TimeZones.Class(:FIXED) | TimeZones.Class(:STANDARD)`
Stacktrace:
[1] TimeZone(::SubString{String}, ::TimeZones.Class) at C:\Users\jerem\.julia\packages\TimeZones\fr1IP\src\types\timezone.jl:65
[2] TimeZone at C:\Users\jerem\.julia\packages\TimeZones\fr1IP\src\types\timezone.jl:46 [inlined]
[3] ZonedDateTime(::Int64, ::Int64, ::Int64, ::Int64, ::Int64, ::Int64, ::Int64, ::SubString{String}) at C:\Users\jerem\.julia\packages\TimeZones\fr1IP\src\types\zoneddatetime.jl:132
[4] parse at C:\Users\jerem\AppData\Local\Programs\Julia-1.5.2\share\julia\stdlib\v1.5\Dates\src\parse.jl:285 [inlined]
[5] ZonedDateTime(::String, ::DateFormat{Symbol("e, d u Y H:M:S Z"),Tuple{Dates.DatePart{'e'},Dates.Delim{String,2},Dates.DatePart{'d'},Dates.Delim{Char,1},Dates.DatePart{'u'},Dates.Delim{Char,1},Dates.DatePart{'Y'},Dates.Delim{Char,1},Dates.DatePart{'H'},Dates.Delim{Char,1},Dates.DatePart{'M'},Dates.Delim{Char,1},Dates.DatePart{'S'},Dates.Delim{Char,1},Dates.DatePart{'Z'}}}) at C:\Users\jerem\.julia\packages\TimeZones\fr1IP\src\parse.jl:88
[6] ZonedDateTime(::String, ::String; locale::String) at C:\Users\jerem\.julia\packages\TimeZones\fr1IP\src\parse.jl:101
[7] ZonedDateTime(::String, ::String) at C:\Users\jerem\.julia\packages\TimeZones\fr1IP\src\parse.jl:101
[8] top-level scope at REPL[96]:1
```
Base function works when specifying the presence of Z, although the notion of timezone is ignored. This seems a viable approach for my use case. I'm just wondering whether it wouldn't be more convenient to keep supporting the parsing of GMT in `ZonedDateTime`?
```
julia> Dates.DateTime(dt_string_1, "e, d u Y H:M:S Z")
2020-10-31T11:31:54
```<issue_closed>
Status: Issue closed |
Tunous/Dawn | 654393250 | Title: Highlighted text permanently shown
Question:
username_0: After long pressing and dragging to select text, a magnification bubble pops up and remains there until I restart the app.

Answers:
username_1: This might be rom related, too. Can you post exact rom version you are on?
username_0: My ROM is stable MIUI 12.
MIUI Global 12.0.1 Stable 172.16.58.3(QFKEUXM) |
tensorflow/tensorflow | 387369750 | Title: [ppc64le] //tensorflow/lite/experimental/micro unit test fail.
Question:
username_0: You can assign this issue to me, as I'm about ready to submit a PR to fix this.
**System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): N/A
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux ppc64le Ubuntu 16.04
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A
- TensorFlow installed from (source or binary): source
- TensorFlow version (use command below): commit <PASSWORD> from Dec 4th, 2018
- Python version: 2.7
- Bazel version (if compiling from source): 0.15.0
- GCC/Compiler version (if compiling from source): gcc version 5.4.0 20160609 (Ubuntu/IBM 5.4.0-6ubuntu1~16.04.10)
- CUDA/cuDNN version: N/A
- GPU model and memory: N/A
**Describe the current behavior**
All tensorflow/lite/experimental/micro unit test fail
See: https://powerci.osuosl.org/job/TensorFlow_PPC64LE_CPU_Build_and_Test/31/testReport/
**Describe the expected behavior**
All unit test pass
**Code to reproduce the issue**
Unit test invoked by Jenkins:
`./tensorflow/tools/ci_build/ci_build.sh cpu --dockerfile tensorflow/tools/ci_build/Dockerfile.cpu.ppc64le ./tensorflow/tools/ci_build/linux/ppc64le/cpu/run_py2.sh`
`./tensorflow/tools/ci_build/linux/ppc64le/cpu/run_py2.sh` can be modified to just run `//tensorflow/lite/experimental/micro/...`
Also the getting started section here recreates two of the issues:
https://github.com/tensorflow/tensorflow/tree/master/tensorflow/lite/experimental/micro
**Other info / logs**
Running the getting started example: make -f tensorflow/lite/experimental/micro/tools/make/Makefile test_micro_speech
This warning below is flagged and run the test runs it seg faults. As I understand the warning is because your not allowed to pass a string constant in this case. (It works however on x86)
```
g++ -O3 -DNDEBUG --std=c++11 -g -DTF_LITE_STATIC_MEMORY -I. -Itensorflow/lite/experimental/micro/tools/make/../../../../../ -Itensorflow/lite/experimental/micro/tools/make/../../../../../../ -Itensorflow/lite/experimental/micro/tools/make/downloads/ -Itensorflow/lite/experimental/micro/tools/make/downloads/gemmlowp -Itensorflow/lite/experimental/micro/tools/make/downloads/flatbuffers/include -I -I/usr/local/include -c tensorflow/lite/experimental/micro/examples/micro_speech/micro_speech_test.cc -o tensorflow/lite/experimental/micro/tools/make/gen/linux_ppc64le/obj/tensorflow/lite/experimental/micro/examples/micro_speech/micro_speech_test.o
In file included from tensorflow/lite/experimental/micro/examples/micro_speech/micro_speech_test.cc:22:0:
tensorflow/lite/experimental/micro/examples/micro_speech/micro_speech_test.cc: In function 'int main(int, char**)':
./tensorflow/lite/experimental/micro/testing/micro_test.h:95:51: warning: ISO C++ forbids converting a string constant to 'va_list {aka char*}' [-Wwrite-strings]
micro_test::reporter->Report("Testing %s", #name); \
^
tensorflow/lite/experimental/micro/examples/micro_speech/micro_speech_test.cc:28:1: note: in expansion of macro 'TF_LITE_MICRO_TEST'
TF_LITE_MICRO_TEST(TestInvoke) {
^~~~~~~~~~~~~~~~~~
```
Fixing it and running the example: make -f tensorflow/lite/experimental/micro/tools/make/Makefile test identifies the same problem in another place:
```
g++ -O3 -DNDEBUG --std=c++11 -g -DTF_LITE_STATIC_MEMORY -I. -Itensorflow/lite/experimental/micro/tools/make/../../../../../ -Itensorflow/lite/experimental/micro/tools/make/../../../../../../ -Itensorflow/lite/experimental/micro/tools/make/downloads/ -Itensorflow/lite/experimental/micro/tools/make/downloads/gemmlowp -Itensorflow/lite/experimental/micro/tools/make/downloads/flatbuffers/include -I -I/usr/local/include -c tensorflow/lite/experimental/micro/micro_error_reporter_test.cc -o tensorflow/lite/experimental/micro/tools/make/gen/linux_ppc64le/obj/tensorflow/lite/experimental/micro/micro_error_reporter_test.o
tensorflow/lite/experimental/micro/micro_error_reporter_test.cc: In function 'int main(int, char**)':
tensorflow/lite/experimental/micro/micro_error_reporter_test.cc:24:56: warning: ISO C++ forbids converting a string constant to 'va_list {aka char*}' [-Wwrite-strings]
error_reporter->Report("~~~%s~~~", "ALL TESTS PASSED");
^
[Truncated]
LD_LIBRARY_PATH=/usr/local/cuda/extras/CUPTI/lib64:/usr/local/nvidia/lib:/usr/local/nvidia/lib64 \
OMP_NUM_THREADS=1 \
PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/snap/bin \
PWD=/proc/self/cwd \
PYTHON_BIN_PATH=/usr/bin/python2 \
PYTHON_LIB_PATH=/usr/local/lib/python2.7/dist-packages \
TF_CUDA_CLANG=0 \
TF_CUDA_COMPUTE_CAPABILITIES=3.7 \
TF_CUDA_VERSION=9.2 \
TF_CUDNN_VERSION=7 \
TF_NCCL_VERSION='' \
TF_NEED_CUDA=1 \
TF_NEED_OPENCL_SYCL=0 \
TF_NEED_ROCM=0 \
external/local_config_cuda/crosstool/clang/bin/crosstool_wrapper_driver_is_not_gcc -o bazel-out/ppc-opt/bin/tensorflow/lite/experimental/micro/examples/micro_speech/feature_provider_test_binary -Wl,-no-as-needed -pie -Wl,-z,relro,-z,now '-Wl,--build-id=md5' '-Wl,--hash-style=gnu' -no-canonical-prefixes -fno-canonical-system-headers -B/usr/bin -Wl,--gc-sections -Wl,@bazel-out/ppc-opt/bin/tensorflow/lite/experimental/micro/examples/micro_speech/feature_provider_test_binary-2.params)
/usr/bin/ld: bazel-out/ppc-opt/bin/tensorflow/lite/experimental/micro/examples/micro_speech/libpreprocessor_reference.a(preprocessor.o): undefined reference to symbol 'cos@@GLIBC_2.17'
//lib/powerpc64le-linux-gnu/libm.so.6: error adding symbols: DSO missing from command line
```
Passing the link option "-lm" resolves this error. The makefile already does this.
Answers:
username_1: Thanks for your PR on this (and sorry for the slow response).
username_0: closing old issues I opened. This issue is resolved.
Status: Issue closed
|
jenkinsci/office-365-connector-plugin | 637152225 | Title: "View Build" option is broken in Microsoft Teams via Jenkins
Question:
username_0: Hi Team,
Could you please help is resolving following issue?
In Jenkins pipeline, we have added Notification Webhook for "Office 365 Connector". When build is notified to Microsoft teams, there is "View Build" button which is having below URL:
http://<JENKINS_URL>/jenkins/job/<JOB_NAME>/job/<JOB>/50/display/redirect which is redirecting to: http://<JENKINS_URL>/jenkins/blue/organizations/jenkins/<JOB_NAME>%2F<JOB>/detail/<JOB>/50/ and throwing 400 error
but actual URL is having "pipeline" in the end.
http://<JENKINS_URL>/jenkins/blue/organizations/jenkins/<JOB_NAME>%2F<JOB>/detail/<JOB>/50/pipeline
How to resolve this? Where to modify this URL? Please help,
Thanks and Regards,
<NAME>
Answers:
username_1: I don't use Blue ocean and so the plugin may not work with this. If you provide all details so I can reproduce the problem maybe I would be able to help or fix it
username_0: Thanks for reply. This issue resolved. I disabled one plugin "Display URL for Blue Ocean" and result is coming fine. I think that plugin is causing some confusion with Teams URL.
Thanks once again. We can close this
username_2: Ideally, that link should be customizable, with the default being what it is now.
We're currently getting hit by this too.
username_1: Not sure what do you mean
username_3: We experience this issue as well: when clicking 'View Build' the url 'https://teams.microsoft.com/null' is opened instead of the Build detail page within jenkins.
username_4: Problem noticed yesterday. Yesterday system status:
```
The connector configuration was old (outlook in URL).
jdk: 8
Windows
```
I thought we should update the JDK version and connector. So the current status:
```
Jenkins: 2.277.4
Plugin: 4.15.0
jdk: OpenJDK 11.0.2
Windows
All other plugins are updated.
The connector URL looks like: https://[name].webhook.office.com/webhookb2/[some GUID]/JenkinsCI/[some GUID]
```
But the problem still exists.
username_5: We have a similar issue with the "View Pull Request" Button. It points to "_https://teams.microsoft.com/null_".
Can I set, the url manually or is this a known bug?
Thanks!
username_6: I have URL's created manually that have the same behavior. Teams on my mobile works, but not desktop. Also, not everybody experienced the issue at once. I'd assume it is a Teams issue
username_5: Thanks for your reply, that's interesting. Indeed it also works for my mobile version of Teams...
username_1: Have you tried to regenerate hook url? I have done it and some problems have been solved
username_7: We have the same problem too. I've regenerate hook url but the same problem still occurs.
username_8: Yes. Multiple times. Issue persists. The webhook seems to be fine since, as others have mentioned, if you click the button on the mobile Teams app it seems to work fine.
username_9: Just jumping in the conversation. Same problem here (MSTeams Connector)
I also have a HTTP Post Request that uses a custom Payload and had to use markdown with a static URL and a variable since the default option didn't work either (custom Connector)
username_1: Then maybe you have regenerated different hook? For regenerated url problems have gone. For old one I see problem on desktop and mobile
username_1: OK, I found one more interesting thing: when updating connection configuration (without updating url) desktop Teams is fixed, mobile on team card is fixed but activity card on mobile still displays that message.
username_9: No URL was changed, I've even tried to make a new one with the same configuration, just to see if there was any mistake made in the past, but no luck. Same result.
I actually don't have any error on the webhook configuration.
username_10: same problem too, but mobile no problem.
username_1: Guys, have you follow this instruction https://docs.microsoft.com/pl-pl/microsoftteams/office-365-custom-connectors ?
This provides guideline that solves problem for me
username_11: yeah, same problem too. but mobile is work.
-> https://teams.microsoft.com/null
username_12: I'm wondering if this problem is caused by connector plugin, or not.
I have tested some URL patterns against section "Jenkins URL" in Jenkins configuration page.
Results:
URL | Reachable
:--|:-:
`http://192.168.2.181:8484/` | No
`http://dd7.local:8484/` | Yes
`http://dd7:8484/` | No
## Pattern 1: `http://192.168.2.181:8484/`

build something.

`View Build` leads to `https://teams.microsoft.com/null`

## Pattern 2: `http://dd7.local:8484/`

build something.

`View Build` leads to `http://dd7.local:8484/job/TeamsTest/2/display/redirect`

## Pattern 3: `http://dd7:8484/`

build something.

`View Build` leads to `https://teams.microsoft.com/null`

username_11: I think I found a way to solve this problem.
If it's not a secure website (e.g. https), it seems to go to a teams.microsoft.com/null page.
Therefore, I think IP should be set up as an external network and apply SSL.
username_8: I have SSL setup, https://<IP Address>:8443
How do I setup the IP as an external network? |
getsentry/sentry-javascript | 403459096 | Title: Morgan and @sentry/node with the Express Handler not working well together
Question:
username_0: <!-- Requirements: please go through this checklist before opening a new issue -->
- [x] Review the documentation: https://docs.sentry.io/
- [x] Search for existing issues: https://github.com/getsentry/sentry-javascript/issues
- [x] Use the latest release: https://github.com/getsentry/sentry-javascript/releases
## Package + Version
- [x] `@sentry/node`
### Version:
```
4.5.3
```
## Description
When using the morgan logging middleware, alongside `app.use(Sentry.Handlers.errorHandler() as express.ErrorRequestHandler)`, whenever there is an async controller that throws an Error, no futher requests can be made, and the following stack trace is shown:
```
GET /api/grin/stats 500 54.544 ms - -
Error: connect ECONNREFUSED 172.16.17.32:8080
at Object._errnoException (util.js:1022:11)
at _exceptionWithHostPort (util.js:1044:20)
at TCPConnectWrap.afterConnect [as oncomplete] (net.js:1198:14)
TypeError: this.app.get is not a function
at IncomingMessage.ip (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/request.js:350:24)
at getip (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/morgan/index.js:466:14)
at logger (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/morgan/index.js:107:26)
at Layer.handle [as handle_request] (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/router/layer.js:95:5)
at trim_prefix (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/router/index.js:317:13)
at /Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/router/index.js:284:7
at Function.process_params (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/router/index.js:335:12)
at next (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/router/index.js:275:10)
at jsonParser (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/body-parser/lib/types/json.js:110:7)
at Layer.handle [as handle_request] (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/router/layer.js:95:5)
```
This does not happen if I comment out `app.use(morgan('dev'))`, or if I comment out `app.use(Sentry.Handlers.errorHandler() as express.ErrorRequestHandler)`... I just can't use both at the same time, and I'd like to do that.
Appreciate any guidance here.
Answers:
username_1: We are experiencing weird things after updating from 4.5.0 -> 4.5.3, the context is different but the underlying `this.app.get` error is the same.
username_0: @username_1 does downgrading to 4.5.0 fix the issue?
username_1: @username_0 yeah it looks like 4.5.0 works as expected
username_2: Seeing the same thing with bunyan (2.0.2) and express-bunyan-logger (1.3.2). Downgrading to 4.5.0 fixes it.
username_3: Also seeing this. https://github.com/getsentry/sentry-javascript/issues/1859 looks like the same issue
username_4: Same issue with Feathers and Winston
https://github.com/feathersjs/feathers/issues/1183
username_5: I'm not using anything like Winston or Morgan, but I'm also experiencing this issue.
Related bits:
```
Sentry.init({
beforeSend(event, hint) {
const {
originalException,
} = hint
const {
message,
} = originalException
if (message.match(/NotAuthenticated/i)) {
return null
}
return event
},
dsn: env.SENTRY_DSN,
})
// ...
app.use(Sentry.Handlers.requestHandler())
// ...
app.use((error, request, response, next) => {
if (!error.stack.match(/NotAuthenticated/i)) {
err(error.stack)
}
next(error)
})
app.use(Sentry.Handlers.errorHandler())
app.use(express.errorHandler())
```
Will check if downgrading fixes the issue.
username_6: Fixed in `4.5.4` – https://github.com/getsentry/sentry-javascript/releases/tag/4.5.4
Sorry for the inconvenience :)
Status: Issue closed
username_7: The issue still exists in `"@sentry/node": "^4.6.2"` using as Express middleware
```TypeError: req.emit is not a function
at Socket.socketOnTimeout (_http_server.js:427:48)
at Socket.emit (events.js:197:13)
at Socket.EventEmitter.emit (domain.js:446:20)
at Socket._onTimeout (net.js:447:8)
at listOnTimeout (timers.js:327:15)
at processTimers (timers.js:271:5)```
username_6: Really?... can you provide a repro please? Every scenario I tested worked just fine.
username_7: I think sentry tries to attach something to res object asynchronously, but the connection is already closed which crashes the api.
```
const api = express();
sentry.init({ dsn: process.env.SENTRY, environment: process.env.DOMAIN });
api.use(sentry.Handlers.requestHandler());
....
api.use(sentry.Handlers.errorHandler());
api.use((err, req, res, next) => {
let { message, statusCode } = err;
if (err.name === 'UnauthorizedError') {
statusCode = 401;
message = 'Token is invalid.';
}
res.status(statusCode || 400).json({ error: message });
});
```
username_6: Works just fine with your repro ¯\_(ツ)_/¯

username_7: not sure, that was the logic before we removed it. Might be something related to our own controllers. |
lipis/flag-icons | 1165239735 | Title: Reorder flags with the Unknown flag first on the list
Question:
username_0: My suggestion is that you should reorder the flag classes by placing the unknown flag (`xx`) before all other flags.
In terms of CSS precedence, if `.fi-fr` is placed on the same element as `.fi-xx`, the unknown flag will override the French flag due to the fact that the French flag is declared earlier than the Unknown flag. By declaring the Unknown flag first, means that every other flag will override it if they reside on the same element.
### So why would you want to declare both flags on the same element?
In my use case, I have a system that stores the 2-letter country code against items in a database. This allows me to display the relevant flag-icon on the page by adding the flag class to the span element (forgive me, I am using Vue):
```vue
<span class="fi" :class="[`fi-${countryCode}`]"></span>
```
Unfortunately, the data may contain invalid 2-letter country codes. (eg. `vb`). To combat this, I would like it to display the "Unknown" flag as a fallback flag.
```vue
<span class="fi fi-xx" :class="[`fi-${countryCode}`]"></span>
```
For a country code of `vb`, this would be translated to `class="fi fi-xx fi-vb"`. Since `.fi-vb` doesn't exist, this gets ignored and `.fi-xx` will produce the "Unknown" flag.
For a country code of `fr`, this would be translated to `class="fi fi-xx fi-fr"`. If `.fi-fr` was declared after `.fi-xx` then the French flag would override the "Unknown" flag and page would display the French flag.
Unfortunately, with `.fi-xx` declared last, the above example will show the "Unknown" flag instead.
Please accept this proposal as I think this would be a benefit for a lot of people using your library. |
Sage-Bionetworks/ChallengeWorkflowTemplates | 580913878 | Title: get_submission.cwl makes assumptions about output from challengeutils downloadsubmission
Question:
username_0: This line: https://github.com/Sage-Bionetworks/ChallengeWorkflowTemplates/blob/master/get_submission.cwl#L34
The workflow assumes that this is exactly where the file will be written and how it will be named. But `challengeutils downloadsubmission` already provides this information as part of its output.
I think this fix is as simple as replacing that line with something like:
```
$(JSON.parse(self[0].contents)['file_path'])
```
Answers:
username_1: @username_0, Thanks for reporting. Did you get an error when running your workflow? This is because `challengeutils` itself renames the submission file:
https://github.com/Sage-Bionetworks/challengeutils/blob/master/challengeutils/__main__.py#L86
username_0: Interesting, lines like that seem to work otherwise: https://github.com/Sage-Bionetworks/ChallengeWorkflowTemplates/blob/master/get_submission.cwl#L41
username_1: Ahh, i forgot that CWL requires the File output to be in a dictionary format like so:
```
"results": {
"location": "file:///Users/thomasyu/sage_projects/dream/ChallengeWorkflowTemplates/results.json",
"basename": "results.json",
"class": "File",
"checksum": "sha1$af6758eac5c5bf4eefbd1ac836982ffa8b621051",
"size": 222,
"path": "/Users/thomasyu/sage_projects/dream/ChallengeWorkflowTemplates/results.json"
}
```
Therefore the line in `outputEval` would have to mimic that structure above.
username_1: Closing as I don't think there is a way to create the correct output blob to match what is required from a `File` type output
Status: Issue closed
|
vcd94xt10z/zionphp | 412513484 | Title: Modulo test
Question:
username_0: Módulo para testar serviços web, funcionalidades:
- cadastrar um fluxo de chamadas de URL com ou sem dados, status http esperado etc;
- fazer botão simples para rodar os testes e exibir o resultado;
Exemplo
POST http://teste.com/usuario/salvar
Status 201 sem response body
PUT http://teste.com/usuario/1
Status 204 sem response body
GET http://teste.com/usuario/1
Status 200 com response body
DELETE http://teste.com/usuario/1
Status 202 sem response body<issue_closed>
Status: Issue closed |
godotengine/godot | 354752076 | Title: Tile selector: add option to resize tile entries
Question:
username_0: It is weird that even with a `list` mode, images are still horribly big.
I mean tiles like in gridmap object, we should be able to resize it, for example:

Option to disable icons to get more space would be useful too.
Answers:
username_1: maybe an EditorSetting could be added to allow chose the maximum icon size, but it will scale the texture uniformly, you example looks horrible imo :P
username_2: example looks terrible, please don't
Status: Issue closed
|
PlantGenIE/PlantGenIE | 1101178537 | Title: Display expression values
Question:
username_0: What value should be used in expression tools? is it TPM+2? or TPM+1?
Currently we have TPM+1 in all our tools to avoid special cases as follows.
```
log2(TPM+0.9)=-0.15 ∴ TPM=0
log2(TPM+1)=0 ∴ TPM=0
log2(TPM+2)=1 ∴ TPM=0
```
Answers:
username_0: yes, we display log2(TPM + 1) for all our expression tools. exHeatmap allow users to download expression values but that is also the same log2(TPM + 1) referred as absolute expression value. |
mapgears/ol3-google-maps | 154990007 | Title: Problem with getStroke
Question:
username_0: Hello,
when i load for the first time my olgm map, appear in console this error
"ol3gm.js:244 Uncaught TypeError: d.getStroke is not a function"
and the dots of polygon on layer are very bigger than normal.
What is the problem?
How can i fix it?
thank you

Answers:
username_1: the same problem
username_2: Would you please provide a small demo featuring the issue, maybe with JSFiddle ?
username_1: the problem in style function.When i use style object problem solved .Why don't i use style function?
username_2: Style functions are not supported, as documented in the limitations: https://github.com/mapgears/ol3-google-maps/blob/master/LIMITATIONS.md#style-functions
There's currently no plan to support them either.
username_0: Sorry but i don't use a style object neither manipulate style.
The problem is present in this function but i don't understand why (ol3gm.js - row 239).
function fa(a, b) {
var c = null,
d = C(a);
if (d) {
var c = {},
e = d.getStroke();
if (e) {
var f = e.getColor();
f && (c.strokeColor = x(f), f = B(f), null !== f && (c.strokeOpacity = f));
(e = e.getWidth()) && (c.strokeWeight = e)
}
if (e = d.getFill())
if (e = e.getColor()) c.fillColor = x(e), e = B(e), null !== e && (c.fillOpacity = e);
if (f = d.getImage()) {
d = {};
e = {};
if (f instanceof ol.style.Circle) {
e.path = google.maps.SymbolPath.CIRCLE;
var h = f.getStroke();
if (h) {
var k = h.getColor();
k && (e.strokeColor = x(k));
e.strokeWeight = h.getWidth()
}
if (h = f.getFill())
if (h = h.getColor()) e.fillColor =
x(h), h = B(h), e.fillOpacity = null !== h ? h : 1;
(f = f.getRadius()) && (e.scale = f)
} else f instanceof ol.style.Icon && ((h = f.getSrc()) && (e.url = h), h = f.getScale(), (k = f.getAnchor()) && (e.anchor = void 0 !== h ? new google.maps.Point(k[0] * h, k[1] * h) : new google.maps.Point(k[0], k[1])), (k = f.getOrigin()) && (e.origin = new google.maps.Point(k[0], k[1])), f = f.getSize()) && (e.size = new google.maps.Size(f[0], f[1]), void 0 !== h && (e.scaledSize = new google.maps.Size(f[0] * h, f[1] * h)));
Object.keys(d).length ? c.icon = d : Object.keys(e).length && (c.icon =
e)
}
0 === Object.keys(c).length ? c.visible = !1 : void 0 !== b && (c.zIndex = 2 * b)
}
return c
};
username_2: As I said, you can't use style functions with OLGM. It's not currently suported to do so. See the limitations.
username_0: Ok, thank tou for your answer but this is a bug because the code is inside ol3gm.
Can you suggest me a workaround, please?
thank you
username_2: This is not a bug. You must not use StyleFunctions with OLGM as it is stated in the limitations. There is no other workaround than "don't use StyleFunctions" with OLGM. I'm sorry that I can't be of more help regarding this for now.
If you absolutely need this, then you're welcome to create a Pull Request adding this support or you can contact info at mapgears dot com if you wish to fund the development of this.
To my knowledge, I think it's impossible to support style functions with OLGM.
Status: Issue closed
username_2: I'm closing this. See the limitations: https://github.com/mapgears/ol3-google-maps/blob/master/LIMITATIONS.md#style-functions |
RasaHQ/rasa | 604214539 | Title: Makefile:43: recipe for target 'install' failed
Question:
username_0: While trying to install install dependencies using the make install command, I am getting the following error.
```
kamaldeep@kamaldeepsingh:~/git/testing/rasa1$ curl -sSL https://raw.githubusercontent.com/python-poetry/poetry/master/get-poetry.py | python
Retrieving Poetry metadata
Latest version already installed.
kamaldeep@kamaldeepsingh:~/git/testing/rasa1$ make install
poetry run python -m pip install -U 'pip<20'
The virtual environment found in /home/kamaldeep/.cache/pypoetry/virtualenvs/rasa-LHgLSZoI-py3.6 seems to be broken.
Recreating virtualenv rasa-LHgLSZoI-py3.6 in /home/kamaldeep/.cache/pypoetry/virtualenvs/rasa-LHgLSZoI-py3.6
[CalledProcessError]
Command '['/home/kamaldeep/.cache/pypoetry/virtualenvs/rasa-LHgLSZoI-py3.6/bin/python', '-Im', 'ensurepip', '--upgrade', '--default-pip']' returned non-zero exit status 1.
Makefile:43: recipe for target 'install' failed
make: *** [install] Error 1
```
Seems like there is some issue in the creation of the Virtual environment. Please help with the same.
Status: Issue closed
Answers:
username_0: While trying to install install dependencies using the make install command, I am getting the following error.
```
kamaldeep@kamaldeepsingh:~/git/testing/rasa1$ curl -sSL https://raw.githubusercontent.com/python-poetry/poetry/master/get-poetry.py | python
Retrieving Poetry metadata
Latest version already installed.
kamaldeep@kamaldeepsingh:~/git/testing/rasa1$ make install
poetry run python -m pip install -U 'pip<20'
The virtual environment found in /home/kamaldeep/.cache/pypoetry/virtualenvs/rasa-LHgLSZoI-py3.6 seems to be broken.
Recreating virtualenv rasa-LHgLSZoI-py3.6 in /home/kamaldeep/.cache/pypoetry/virtualenvs/rasa-LHgLSZoI-py3.6
[CalledProcessError]
Command '['/home/kamaldeep/.cache/pypoetry/virtualenvs/rasa-LHgLSZoI-py3.6/bin/python', '-Im', 'ensurepip', '--upgrade', '--default-pip']' returned non-zero exit status 1.
Makefile:43: recipe for target 'install' failed
make: *** [install] Error 1
```
Seems like there is some issue in the creation of the Virtual environment. Please help with the same.
Status: Issue closed
|
dotse/bbk | 587993286 | Title: Don't report 99 MBit on a 100-200Mbit connection as 'BAD'.
Question:
username_0: There need to be some sort of limit based in percentage of the connection to determine the speed, not just some hard limits.
Answers:
username_1: Have you checked the Readme? https://frontend.bredbandskollen.se/download/README.txt
'If the subscription can be identified, the download result will be evaluated as
"GOOD", "ACCEPTABLE", or "BAD". The result is considered BAD if the speed is
below the lower end of the expected speed range as stated by the service
provider. It is considered ACCEPTABLE if it is above the lower end, but still
clearly below the middle, of the expected speed range. If the result is better,
it is considered GOOD.
E.g. if the expected range is 150-250 Mbit/s, a result below 150 is BAD; a
result between 150 and 190 is ACCEPTABLE, and a result above 190 is GOOD.'
The "limit based in percentage" you're looking for is the 40% target that decides what is needed to get a GOOD result: 250 - 150 = 100 Mbit. 40% of 100 Mbit is 40 Mbit. 150 + 40 = 190 Mbit.
username_0: Exactly and that's wrong in so many ways. 99,9% is NEVER a "bad" speed which implies it's unacceptable and hence makes people call customer support.
username_2: We are getting the connection from the ISP - if we have the possibility.
If you ISP says you connection should be 100-200 MB and you don't get up to the lowest limit we do consider that bad.
username_0: And that is flawed since it's usually "up to" in the agreements and there might be other circumstances that lowers the speed. To actually know the connection quality you would have to have iperf server instances at the ISPs.
This creates confusion which creates unnecessary support calls for those working in the field.
username_3: I just randomly found this issue, and I'd just like to say this makes no sense.
If the ISP considers 99 Mbit/s an acceptable speed on the connection that they sell, they should then sell the connection as a 99-200 Mbit/s connection, not a 100-200 Mbit/s connection.
If the ISP promises 100-200 Mbit/s, then anything below 100 Mbit/s is unacceptable. (Assuming, for now, that the measurement server is well enough connected to the ISP's network to be a valid measurement of the connection speed itself.)
Imagine, if I went to the supermarket, and I bought a box of milk that is specified to be 100-200 cl, and then I come home and see I only got 99 cl. In this case, the amount of milk that I received is out of spec, even if it's only 1% out of spec. |
febobo/JS-EveryDay-Question | 425260959 | Title: JS每日一题:前端优化手段有哪些
Question:
username_0: 20190118问:
前端优化手段有哪些?
- 静态资源合并压缩(js,css, images)
- 请求数量优化
- Gzip压缩
- 带宽优化
- CDN
- 就近节点,减少DNS查找
- 按需加载
- lazyload
- 减少请求
- 骨架屏
- 优化白屏
- web缓存
- 缓存ajax数据
- 减少重绘和重排
- 批量更新DOM样式
- 页面结构
- 将样式表放在顶部,将脚本放在底部,尽早刷新文档的输出 |
openenclave/openenclave | 843626477 | Title: Ability to make an ECALL from kernel mode
Question:
username_0: Currently another team in Microsoft is using OP-TEE directly, rather than building on Open Enclave SDK, because OE does not currently support the ability to make an ECALL from kernel mode. This forces them to jump through lots of extra hoops to write the enclave that OE would have solved for them, and limits their ability to work on other TEEs in the future.
Currently in OE, one would have to first call up into user mode and have user mode make the ECALL, which loses extra perf (on Arm at least) and causes extra complexity and this approach isn't obvious either.
Hence this is a feature request to support making an ECALL directly from kernel mode, which I am filing on the other team's behalf.
Answers:
username_0: This is peripherally related to issue #3804, but that is just for attestation, and that request came from a different team.
username_1: On SGX, user mode context is required to make an ecall.
username_1: Capture scenario in design doc. |
AlvaroLarumbe/espalet-android | 202401501 | Title: When clicking on the refresh button there is no feedback to the user
Question:
username_0: It will be good to show spinner icon or at least make images blink once.
Answers:
username_1: Look at this: https://github.com/username_1/espalet-android/commit/b1030608674f4b0bbdb25932a19c5c86215feae3#diff-350204edff7826e85e2fac268b3e263cR61
username_0: Enough
Status: Issue closed
|
invenia/Intervals.jl | 524508865 | Title: Range indexing fails
Question:
username_0: ```
julia> using Dates, TimeZones, Intervals
julia> st = HE(ZonedDateTime(2014,1,1, tz"America/Winnipeg"))
AnchoredInterval{-1 hour,ZonedDateTime}(ZonedDateTime(2014, 1, 1, tz"America/Winnipeg"), Inclusivity(false, true))
julia> r = st:Hour(1):(st + Day(1))
AnchoredInterval{-1 hour,ZonedDateTime}(ZonedDateTime(2014, 1, 1, tz"America/Winnipeg"), Inclusivity(false, true)):1 hour:AnchoredInterval{-1 hour,ZonedDateTime}(ZonedDateTime(2014, 1, 2, tz"America/Winnipeg"), Inclusivity(false, true))
julia> collect(r)[1:2]
2-element Array{AnchoredInterval{-1 hour,ZonedDateTime},1}:
AnchoredInterval{-1 hour,ZonedDateTime}(ZonedDateTime(2014, 1, 1, tz"America/Winnipeg"), Inclusivity(false, true))
AnchoredInterval{-1 hour,ZonedDateTime}(ZonedDateTime(2014, 1, 1, 1, tz"America/Winnipeg"), Inclusivity(false, true))
julia> r[1:2]
Error showing value of type StepRangeLen{AnchoredInterval{-1 hour,ZonedDateTime},AnchoredInterval{-1 hour,ZonedDateTime},Hour}:
ERROR: MethodError: Cannot `convert` an object of type Hour to an object of type ZonedDateTime
Closest candidates are:
convert(::Type{T}, ::Interval{T}) where T at /Users/ericdavies/.julia/packages/Intervals/52k6l/src/interval.jl:124
convert(::Type{T}, ::AnchoredInterval{P,T}) where {P, T} at /Users/ericdavies/.julia/packages/Intervals/52k6l/src/anchoredinterval.jl:174
convert(::Type{T}, ::T) where T at essentials.jl:167
...
Stacktrace:
[1] AnchoredInterval{-1 hour,ZonedDateTime}(::Hour) at /Users/ericdavies/.julia/packages/Intervals/52k6l/src/anchoredinterval.jl:79
[2] step(::StepRangeLen{AnchoredInterval{-1 hour,ZonedDateTime},AnchoredInterval{-1 hour,ZonedDateTime},Hour}) at ./range.jl:501
[3] show(::IOContext{REPL.Terminals.TTYTerminal}, ::StepRangeLen{AnchoredInterval{-1 hour,ZonedDateTime},AnchoredInterval{-1 hour,ZonedDateTime},Hour}) at ./range.jl:712
[4] show(::IOContext{REPL.Terminals.TTYTerminal}, ::MIME{Symbol("text/plain")}, ::StepRangeLen{AnchoredInterval{-1 hour,ZonedDateTime},AnchoredInterval{-1 hour,ZonedDateTime},Hour}) at ./show.jl:7
[5] display(::REPL.REPLDisplay{REPL.LineEditREPL}, ::MIME{Symbol("text/plain")}, ::StepRangeLen{AnchoredInterval{-1 hour,ZonedDateTime},AnchoredInterval{-1 hour,ZonedDateTime},Hour}) at /Users/ericdavies/.julia/packages/OhMyREPL/GFHgr/src/output_prompt_overwrite.jl:6
[6] display(::REPL.REPLDisplay, ::Any) at /Users/ericdavies/repos/julia1p2/usr/share/julia/stdlib/v1.2/REPL/src/REPL.jl:136
[7] display(::Any) at ./multimedia.jl:323
```
Range indexing appears to be taking a `StepRange` and making a `StepRangeLen`, which we don't handle effectively.
I think this might be a bug in `step(::StepRangeLen)`.
Answers:
username_0: I think that bug is there, but we can also avoid it by never making a `StepRangeLen`. We can prevent that by defining the traits `OrderStyle` and `ArithmeticStyle` to be the same as `AbstractTime`. We should also add `RangeStepStyle` while we're at it.
username_0: See https://github.com/JuliaLang/julia/blob/master/base/range.jl and https://github.com/JuliaLang/julia/blob/master/base/traits.jl
username_1: As of [Julia 1.5.0-DEV.207](https://github.com/JuliaLang/julia/pull/34563) (7d92a3aaed) this issue appears fixed:
```julia
julia> r[1:2]
AnchoredInterval{Hour(-1),ZonedDateTime}(ZonedDateTime(2014, 1, 1, tz"America/Winnipeg"), Inclusivity(false, true)):Hour(1):AnchoredInterval{Hour(-1),ZonedDateTime}(ZonedDateTime(2014, 1, 1, 1, tz"America/Winnipeg"), Inclusivity(false, true))
```
Do we want to try and fix this on earlier versions of Julia?
username_0: We should figure out the correct traits. If that fixes it for older versions of Julia, that would be a bonus.
username_1: Another example of this issue but on Intervals 1.3 you just see a deprecation warning:
```julia
julia> using Intervals
julia> x = AnchoredInterval{-1}(1):1:AnchoredInterval{-1}(5)
AnchoredInterval{-1,Int64,Open,Closed}(1):1:AnchoredInterval{-1,Int64,Open,Closed}(5)
julia> y = x[1:4];
julia> y
┌ Warning: `convert(T, interval::AnchoredInterval{P,T})` is deprecated for intervals which are not closed with coinciding endpoints. Use `anchor(interval)` instead.
│ caller = AnchoredInterval at anchoredinterval.jl:95 [inlined]
└ @ Core ~/.julia/dev/Intervals/src/anchoredinterval.jl:95
┌ Warning: `convert(T, interval::AnchoredInterval{P,T})` is deprecated for intervals which are not closed with coinciding endpoints. Use `anchor(interval)` instead.
│ caller = AnchoredInterval at anchoredinterval.jl:95 [inlined]
└ @ Core ~/.julia/dev/Intervals/src/anchoredinterval.jl:95
AnchoredInterval{-1,Int64,Open,Closed}(1):AnchoredInterval{-1,Int64,Open,Closed}(1):AnchoredInterval{-1,Int64,Open,Closed}(4)
```
username_1: I attempted to use traits to fix this problem but that didn't actually seem to be the source of the failure. I instead implemented the fix in #132 and made a [new issue for implementing traits](#133).
Status: Issue closed
|
reuixiy/hugo-theme-meme | 1025410447 | Title: 关于shortcodes的扩展支持
Question:
username_0: https://hugoloveit.com/zh-cn/theme-documentation-extended-shortcodes/
LoveIt主题的admonition很漂亮,如果不难的话希望加上这个feature XD
Answers:
username_1: 自己写了一些:https://github.com/username_1/shortcodes
演示:https://oi-io.me/tech/hugo-shortcodes/#note
username_0: @username_1
非常酷!
等下班了试一下,感谢!
要是能PR到meme就好拉。 @reuixiy |
geocollections/turvas | 577375294 | Title: Vaatluspunkti kaardivaade
Question:
username_0: Vaatluspunkti kaardivaates kuvatakse ala punkte, see on mugav ala uurimisel, aga vajalik oleks konkreetse valitud punkti eristamine kas suurema markeriga vms moel ning lisaks püsiv label kaardil.
Püsiv label võiks olla alati kui on ainult üks aktiivne punkt kaardil - st ka proovi detailvaates.<issue_closed>
Status: Issue closed |
NVIDIA/VideoProcessingFramework | 639325937 | Title: Single GPU decoding some video
Question:
username_0: Hi, @username_1 :
I looked at examples: SampleDecodeMultiThread.py, and tested our GPU(Single v100) decoding to the FPS 500.
But I want a single GPU to decode multiple videos simultaneously.
I found that they were serial-running decoded:
one video, decode 6s ; two video , decode 12s
How to make two video decoding process can call GPU decoding in parallel?
Answers:
username_1: Hi @username_0
To my best knowledge, this happens because of GIL.
You can run multiple sessions in parallel threads and / or sub processes but you don't see much (or any) of a performance gain.
Please look at the similar issue here:
https://github.com/NVIDIA/VideoProcessingFramework/issues/62#issuecomment-641767610
I'm running out of depth here (don't have too much of a Python development expertise) so if you can try mentioned approach and come back as you get results that would be really great.
username_0: thx @username_1
the #62 give me perfect solution!
Status: Issue closed
username_2: @username_0 Hi, I just adopt the multuprocessing method and the result is just the same as multi thread method.
Have you achieve the improvements? could you just share the snip code for that?
username_0: @username_2 sry,I have given up the research.
username_2: Thanks for your reply.
And I just want to know ,did you achieve idea twice improvement by using multiprocessing rather than multithread?
username_0: no,only multithread |
Azure/blobxfer | 309790896 | Title: AttributeError: 'NoneType' object has no attribute 'primary_endpoint'
Question:
username_0: Hi,
I'm trying to use the High Level API to upload files to Azure Blob and I started from [this](https://blobxfer.readthedocs.io/en/latest/80-blobxfer-python-library/) from the documentation. I think I'm missing something because I don't understand how come I'm getting this kind of error, which is thrown by `create_destination_id`
Here is the complete trace:
```
Traceback (most recent call last):
File ".../upload.py", line 61, in <module>
uploader.start()
File "...\venv\lib\site-packages\blobxfer\operations\upload.py", line 1167, in start
self._run()
File "...\venv\lib\site-packages\blobxfer\operations\upload.py", line 1079, in _run
create_destination_id(ase._client, ase.container, ase.name)
File "...\venv\lib\site-packages\blobxfer\operations\upload.py", line 185, in create_destination_id
return ';'.join((client.primary_endpoint, container, name))
AttributeError: 'NoneType' object has no attribute 'primary_endpoint'
```
These are the blobxfer parameters printed before the execution:
```
============================================
Azure blobxfer parameters
============================================
blobxfer version: 1.1.1
platform: Windows-10-10.0.16299-SP0
components: CPython=3.6.4 azstor.blob=1.0.0 azstor.file=1.0.0 crypt=2.2.2 req=2.18.4
transfer direction: local -> Azure
workers: disk=1 xfer=1 md5=1 crypto=0
log file: blobxfer.log
resume file: None
timeout: connect=30.0 read=30.0
mode: block
skip on: fs_match=False lmt_ge=False md5=True
delete extraneous: False
overwrite: True
recursive: True
rename single: False
access tier: hot
chunk size bytes: 4194304
one shot bytes: 33554432
strip components: 0
store properties: attr=True md5=True
rsa public key: None
local source paths: ..\sourcefolder
============================================
```
The code is like it follows, whereas config is a yaml file structured similarly to the sample configuration yaml [here](https://blobxfer.readthedocs.io/en/latest/sample_config.yaml).
```
upload_options = config['upload'][0]['options']
concurrency_dict = config['concurrency']
skip_on_options = upload_options.pop('skip_on', None)
store_file_properties = upload_options.pop('store_file_properties', None)
endpoint = "core.windows.net"
concurrency = blobxfer.api.ConcurrencyOptions(**concurrency_dict)
[Truncated]
endpoint)
source = blobxfer.api.LocalSourcePath()
source.add_path('../sourcefolder')
destination = blobxfer.api.AzureDestinationPath()
destination.add_path_with_storage_account('container', 'storageaccountname')
spec = blobxfer.api.UploadSpecification(upload_options, skip_on_options, source)
spec.add_azure_destination_path(destination)
uploader = blobxfer.api.Uploader(general_options, creds, spec)
uploader.start()
```
I feel like I'm missing something trivial.
Any help is appreciated.
Thank you
Answers:
username_1: This call `destination.add_path_with_storage_account('container', 'storageaccountname')` must have a `creds` entry for `storageaccountname`. Please double check that is the case.
You can also test getting the associated storage account with:
```
creds.get_storage_account(destination[0].lookup_storage_account('container'))
```
username_0: Hi @username_1,
from what I see in the implementation of add_path_with_storage_account in `bloxfer/operations/azure/__init__.py`:
```
def add_path_with_storage_account(self, remote_path, storage_account):
# type: (SourcePath, str, str) -> None
"""Add a path with an associated storage account
:param SourcePath self: this
:param str remote_path: remote path
:param str storage_account: storage account to associate with path
"""
if len(self._path_map) >= 1:
raise RuntimeError(
'cannot add multiple remote paths to SourcePath objects')
rpath = blobxfer.util.normalize_azure_path(remote_path)
self.add_path(rpath)
self._path_map[rpath] = storage_account
```
both the parameters should be strings? That's why I used the storage account string.
If I try to use `creds.get_storage_account(destination[0].lookup_storage_account('container'))` to check for the associated storage account, I get the following error:
```
Traceback (most recent call last):
File ".../upload.py", line 55, in <module>
print(creds.get_storage_account(destination.lookup_storage_account('container')))
File "...\venv\lib\site-packages\blobxfer\operations\azure\__init__.py", line 84, in get_storage_account
return self._storage_accounts[name]
KeyError: <blobxfer.operations.azure.StorageAccount object at 0x0000022CB4C85C88>
```
username_1: That's correct, it should be a string type.
Sorry, my check was off, try: `creds.get_storage_account(spec.destinations[0].lookup_storage_account('container'))`
username_0: Ok, I checked and the output is correct: the lookup of the storage account allows me to access the credentials I've set (name, endpoint, key, clients and so on)
```
sa = creds.get_storage_account(spec.destinations[0].lookup_storage_account('container'))
print(sa.name, sa.endpoint, sa.key, sa.block_blob_client.primary_endpoint)
```
outputs correctly:
`storageaccountname core.windows.net storageaccountkey storageaccountname.blob.core.windows.net`
So, looking for the reason as to why ase._client would be None, I found the ase._client yielded by `_generate_destination_for_source` in `blobxfer/operations/upload.py` was None, because `populate_from_local` didn't assign the correct client to ase._client, as the mode passed "auto" (string) wasn't recognized as `StorageModes.Auto`.
So, in the end, it was a really silly oversight on my part and used `mode=StorageModes.Auto` when creating the UploadOptions.
Thank you, @username_1, for your help.
Status: Issue closed
|
badcamp/camp_distro | 311469327 | Title: Text for Volunteer Page
Question:
username_0: ## Got Skills? Join Us + Volunteer!
**Yes, you! BADCamp** would not be possible without love + support from volunteers!
### Volunteer opportunities include:
Marketing + communications
Day-of-coordination
Registration desk help
Store assistants + managers
Room monitors + coordinators
Stuff movers + badge checkers
Room setup + maintenance
AV, Internet + Photographers
Signage distributors
Venue setup + teardown
---WebForm --- |
curiouslearning/workshop_drag_into_place_1 | 229487646 | Title: User testing reflection (due Wednesday 5/31)
Question:
username_0: After you have completed user testing, please reflect on the experience and complete the assignment as [outlined here](https://docs.google.com/document/d/1uOWlF4pZqYn4c9yVkLotlZ3LFPSHrzdLvJbpeWvgfUQ/edit#heading=h.iyktjiy2ixr2) using the [following document as a template](https://docs.google.com/document/d/1VozsQ_IJCJ9cNHszWFiNQBSkpJ8jgd7sgB-_xkXGy94/edit?usp=sharing).
When you have finished your write-up, you can [submit it as a Google Doc into this folder](https://drive.google.com/drive/folders/0B1Q2Y5s6EhXLMkNtRjBwcmVta1E?usp=sharing).
Please note, unlike previous task cards, this will be due midweek on Wednesday 5/31 at 11:59PM as we wanted to give everyone more time to focus on completing the prototype and running a user test. |
spring-cloud/spring-cloud-config | 129642893 | Title: Support placeholders for searchPaths on remote repos
Question:
username_0: To support a more flexible Git repo structure it would be good if the {placeholder} support was supported in the searchPaths option.
Ideally we should be able to support having a single repo with config store under application (and profile) structure. Current implementation requires change to the searchPaths entry on the config server whenever a new application is added.
```
spring:
cloud:
config:
server:
git:
uri: https://xxxx/config-repo.git
searchPaths: {application}
```
Status: Issue closed
Answers:
username_0: I've tried this with 1.1.0.M5 but doesn't seem to work with either https or ssh git repo's.
(application.yml)
```
spring:
cloud:
config:
server:
git:
#uri: ssh://xxxxx:7999/mic/cloudconfig.git
uri: https://xxxxxx/scm/mic/cloudconfig.git
cloneOnStart: true
searchPaths: {application}
```
```
(cloudconfig repo)
├── application.properties
├── application.yml
└── demo-app
├── demo-app-dev.yml
├── demo-app-prod.yml
├── demo-app.yml
├── devcloud.js
└── simple.conf
```
Test - use URL: http://localhost:8888/demo-app/dev
Expected result: 3 resource files returned (application.properties, application.yml, demo-app-dev.yml)
Actual result: 2 resource files returned (application.properties, application.yml)
username_1: Works for me. You need to quote the search paths value in YAML ("{" is not a valid literal). Maybe that's the problem?
username_0: quoting in the yaml file fixed my issue. Thanks.
```
spring:
cloud:
config:
server:
git:
#uri: ssh://xxxxx:7999/mic/cloudconfig.git
uri: https://xxxxxx/scm/mic/cloudconfig.git
cloneOnStart: true
searchPaths: "{application}"
```
username_2: The solution with quotes is currently not working for me
username_1: Probably a different issue then. This one is closed (you might have noticed?). It would be best if you can create a sample that reproduces the issue you see and then open another issue.
username_3: I was also making same mistake, after quoting in doublesquotes it worked for me |
flutter/flutter | 349883943 | Title: flutter run is error!
Question:
username_0: I tried to initialize the project is correct,
I use the vscode editor,
but, i Excuting an order `flutter run -v`,
prompt : Connection timed out !
**I tried running it a dozen times, still the same error, which made me unable to understand how to use flutter!**
@ianloic @paulcbetts @sgraham @sethladd @kevmoo


Answers:
username_1: Please add the output of `flutter doctor -v`.
username_1: How did you create the project?
Does this also happen if you do
```
flutter create my_project
cd my_project
flutter run
```
in a terminal window?
username_0: I have found the reason for this problem, because the network problem causes the dependency to be downloaded. thanks
Status: Issue closed
username_1: Thanks for the update!
Sounds like this can be closed then. |
explosion/spaCy | 296125234 | Title: How to convert simple NER format to spacy json
Question:
username_0: Currently I have data in the simple training data / offset format as shown in the docs:
```
TRAIN_DATA = [
('Who is <NAME>?', {
'entities': [(7, 17, 'PERSON')]
}),
('I like London and Berlin.', {
'entities': [(7, 13, 'LOC'), (18, 24, 'LOC')]
})
]
```
In order to utilize the CLI trainer, this data needs to be in spacys internal json training data format. to do this I am currently converting the simple offset format to conllner, then using the spacy cli converter.
Is there a better way to directly get offset / simple format directly to spacy json format? Using the cli trainer seems to be able to train much faster than using the method specified [here](spacy/examples/training/train_ner.py)
## Your Environment
* **spaCy version:** 2.0.5
* **Platform:** Linux-4.4.0-1049-aws-x86_64-with-debian-stretch-sid
* **Python version:** 3.6.4
* **Models:** en_core_web_lg, en
Answers:
username_1: Yes, there's a [`gold.biluo_tags_from_offsets`](https://spacy.io/api/goldparse#biluo_tags_from_offsets) function that converts the entity offsets to a list of per-token BILUO tags:
```python
from spacy.gold import biluo_tags_from_offsets
doc = nlp(u'I like London.')
entities = [(7, 13, 'LOC')]
tags = biluo_tags_from_offsets(doc, entities)
assert tags == ['O', 'O', 'U-LOC', 'O']
```
username_0: Is there a way to save the bilou tags in the json format used by cli
trainer?
username_2: I think we currently lack a writer that does exactly that. I've been finding the current data-format situation frustrating too. Needs improvement.
username_3: I have also the same problem in training NER.
@username_2 @username_0
Is there any json sample training data for NER?
username_1: You can find an the structure and an example of the training JSON format here:
https://spacy.io/api/annotation#json-input
So you should be able to write a script that produces data in this format, using the BILUO tags. I agree that spaCy should make this easier (and, as @honnial mentioned above, we're also thinking about overhauling the training format in general to make it easier to work with).
username_3: Thanks @username_1
That is a very good idea, I'm looking forward to it.
username_4: Same question that I posted on the chat channel today. I want my program to point to training data in a separate file rather than having it in the code. I currently have as follows based on the examples on the website
<img width="1102" alt="screen shot 2018-04-03 at 2 44 06 pm" src="https://user-images.githubusercontent.com/4360712/38277517-838f5b46-374d-11e8-9b27-0b8331a3ad9b.png">
username_5: @username_4, recently I was looking for similar need. I figured a way to put the content to a separate file instead of in code.
`inpt_text = open("text_file","r")`
`TRAIN_DATA = inpt_text.read().replace('\n ', '')`
`TRAIN_DATA = ast.literal_eval(' '.join(TRAIN_DATA.split()))`
**text_file content format:**
[
('Who is <NAME>?', {
'entities': [(7, 17, 'PERSON')]
}),
('I like London and Berlin.', {
'entities': [(7, 13, 'LOC'), (18, 24, 'LOC')]
})
]
Plz share if there are any other approaches to maintain TRAIN_DATA in a separate file instead of in code.
username_1: @username_4 @username_5 How you solve this depends on which file formats you like to work with. `ast.literal_eval` isn't *bad* – but it's also a very uncommon way to pass data around. And the transformations (replacing, splitting, joining) you have to do to read in the file can very easily lead to bugs.
A simpler way might be working with JSON. Here's a an example of a file ```data.json```:
```json
[
["Who is <NAME>?", {"entities": [[7, 17, "PERSON"]]}],
["I like London and Berlin", {"entities": [[7, 13, "LOC"], [18, 24, "LOC"]]}]
]
```
JSON doesn't have tuples, so you need to use lists everywhere. In Python, you can then do the following:
```python
import json
import io
with io.open('/path/to/data.json', encoding='utf8') as f:
train_data = json.load(f)
```
Also remember that you can always re-write the training script to format your data differently. If you're working with large training sets, we also recommend to use the `spacy train` command instead – this will take care of batching up your examples correctly for optimal performance, and it'll also give you much more useful command line output.
username_6: I am also about to write a script to convert from inline training format to CLI training format. However I train the model on sentences. So each paragraph entry will contain exactly one sentence instead. Is that fine? Since the NER model trains on context which are usually no more than sentences, I do not see a reason for paragraphs to be input with their raw text and then sentences. I think paragraphs are there for Dependency Parsing. Is that right?
username_6: I am pasting an answer derived from [stackoverflow](https://stackoverflow.com/questions/48916768/how-do-i-convert-simple-training-style-data-to-spacys-command-line-json-format) which is not showing up easily in searches. Ofcourse you will have to improvise a bit if you have multiple paragraphs.
```
def getcharoffsetsfromwordoffsets(doc,entities):
charoffsets = []
for entity in entities:
span = doc[entity[0]:entity[1]]
charoffsetentitytuple = (span.start_char, span.end_char, entity[2])
charoffsets.append(charoffsetentitytuple)
return charoffsets
def convertspacyapiformattocliformat(nlp, TRAIN_DATA):
docnum = 1
documents = []
for t in TRAIN_DATA:
doc = nlp(t[0])
charoffsetstuple = getcharoffsetsfromwordoffsets(doc,t[1]['entities'])
tags = biluo_tags_from_offsets(doc, charoffsetstuple)
ner_info = list(zip(doc, tags))
tokens = []
sentences = []
for n, i in enumerate(ner_info):
token = {"head" : 0,
"dep" : "",
"tag" : "",
"orth" : i[0].string,
"ner" : i[1],
"id" : n}
tokens.append(token)
sentences.append({'tokens' : tokens})
document = {}
document['id'] = docnum
docnum+=1
document['paragraphs'] = []
paragraph = {'raw': doc.text,"sentences" : sentences}
document['paragraphs'] = [paragraph]
documents.append(document)
return documents
```
username_7: @username_1 is "spacy convert" command the solution for this issue? The reason why I am asking this is I don't see a converter for BILOU. Is that something currently mot supported?
username_8: Hello everyone,
I have the same question as @username_7 about the input data format which I have to pass to convert function in order to train my NER for my custom entities.
Right now, my data looks like simple NER format with about 10 labels, when I try to pass it to the converter function it give me this error: Unknown format Can't find converter for json.
Could you please help me: what am I doing wrong?
username_2: In v2.1.0a1, this is now finally be easier. Thanks for your patience with the annoying format, which has a few extra levels of nesting so that it can accomodate whole document annotations, along with paragraph structure.
In v2.1.0a1, there's now a new function `spacy.gold.docs_to_json`. This function takes a list of `Doc` objects and outputs properly formed json data, for use in `spacy train`. You just have to set up the `Doc` objects to have the annotations you want. For named entities, this is as easy as writing to the `doc.ents` attribute. For the parser, the best approach is to create an array with the head offsets you want, and then import the data with `doc.from_array()`.
Status: Issue closed
|
aurelia/ui-virtualization | 188768990 | Title: Rebinding List to New Array, List Still Shows Some of the Old Items
Question:
username_0: **I'm submitting a bug report**
* **Library Version:**
major.minor.patch-pre
1.0.0-beta.3.0.1
**Please tell us about your environment:**
* **Operating System:**
Windows 7
* **Node Version:**
6.9.0
* **NPM Version:**
3.10.8
* **JSPM
JSPM 0.16.45
* **Browser:**
Chrome
* **Language:**
ES6
**Current behavior:**
When you rebind the virtual list to a new array,
e.g. this.myArray = getNewData()
Then the list rebinds, but some of the items from the old array are still present.
* **What is the expected behavior?**
Not that! List should only contain new items after a rebind.
* **What is the motivation / use case for changing the behavior?**
I'm using the virtual-repeat.for list to display search results. If they enter a new search, I want the old results to go away, and the new results to display.
Answers:
username_1: Could you set up a basic reproducible gist for us to look at? Definitely seems strange. We're currently replacing our list in a similar manner for searching (although it's a bit different), and we don't have a problem. Might be a specific way you're doing it... But I'm not sure.
username_0: I'll have to do it this weekend. Thanks for the quick reply!
Aaron
username_0: I will say that we have a bunch of Compose elements in there. Don't know if that's relevant.
Thanks again.
Aaron
username_1: No problem, I just so happened to have my browser open and saw this issue fly in :)
Compose elements _might_ be a contributing factor.... But I wouldn't bet on it. I'd really just need to see what's going on to know for sure. I'm usually out over the weekend, so if nobody else gets a chance to look at it, I'll try to take a peek on monday.
username_0: OK, I feel stupid. How do you add aurelia and this plugin to a gist? I've never used gist before, just JSFiddle. Can you point me to a good resource on gist?
Thanks.
username_1: Don't feel stupid, sometimes Gist gives me a run for my money too :) If you can't get it to happen in a gist, just make a repository that I can clone and try it on.
A gist example of getting a custom element to work is here: https://gist.github.com/username_1/3a20dc24ca5c32c4b29fe6990dd30156
I don't have an example with a plugin right now.
username_2: @username_0 here's a gist you can use as a starting point: https://gist.run/?id=7542e061bc940cde506b
username_1: Related to: #106
Working on resolving there.
username_1: Resolved (hopefully) by https://github.com/aurelia/ui-virtualization/pull/107
Status: Issue closed
|
modelleicher/FS19_simpleIC | 773971246 | Title: [Suggestion] Possibiliti to add external control panels from machines to tractors
Question:
username_0: Lot of machines in real life use control panels which show capacity and precisely doses fertilizer or herbicide. 4D Modding realise baler with computer (Hesston Bale Control) and you have to add it in i3d. I think Simple IC should give a possibility to change possition of this parts, now game don't give this option and many people add static models od this computers without working dashes and icons, in my opinion this small addon to script can add more realism and fun for players which wont to play like in real life. I have small script converted from old Farming Simulator but it based on old IC panel and doesn't working exactly as it should.
Answers:
username_1: Hi, do you mean SimpleIC should support the adding of those panels or the controlling of them? Because the latter wouldn't be a "small addon" much rather a huge new script..
username_0: I think about points from machine using inside tractor
username_1: With the addition of implementControl at some point a few machine functions will be controllable via tractor
username_0: But you can create your own points with animations in machine like door animation in tractor? |
EnoxSoftware/OpenCVForUnity | 268713474 | Title: MobileNetssdExample
Question:
username_0: I am trying to get object detection via caffee model but whenever i try to do my unity editor gets crash.
Answers:
username_1: Are you using the following files?
MobileNetSSD_deploy.prototxt
https://github.com/chuanqi305/MobileNet-SSD/blob/master/MobileNetSSD_deploy.prototxt
dnn/MobileNetSSD_deploy.caffemodel
https://drive.google.com/open?id=0B3gersZ2cHIxRm5PMWRoTkdHdHc
username_2: I'm having problems getting this to run, also.
I'm using the two files mentioned above.
3058607249(checksum) 25036(bytes) MobileNet_deploy.prototxt
3242836829(checksum) 17028869(bytes) MobileNet_deploy.caffemodel
it seems to be related to the prototxt file and/or format?
on android i get this error at runtime:
'
04-12 15:25:24.590 12522-12574/com.cr3d.mssd E/libprotobuf-native: [libprotobuf ERROR /Users/satoo/opencv/android/opencv/3rdparty/protobuf/src/google/protobuf/text_format.cc:288] Error parsing text-format opencv_caffe.NetParameter: 7:1: Expected identifier, got: <
04-12 15:25:24.590 12522-12574/com.cr3d.mssd E/cv::error(): OpenCV(3.4.1-dev) Error: Unspecified error (FAILED: ReadProtoFromTextFile(param_file, param). Failed to parse NetParameter file: /storage/emulated/0/Android/data/com.cr3d.mssd/files/opencvforunity/dnn/MobileNetSSD_deploy.prototxt) in void cv::dnn::ReadNetParamsFromTextFileOrDie(const char*, opencv_caffe::NetParameter*), file /Users/satoo/opencv/android/opencv/modules/dnn/src/caffe/caffe_io.cpp, line 1145
04-12 15:25:24.590 12522-12574/com.cr3d.mssd E/org.opencv.dnn: dnn::readNetFromCaffe_10() : OpenCV(3.4.1-dev) /Users/satoo/opencv/android/opencv/modules/dnn/src/caffe/caffe_io.cpp:1145: error: (-2) FAILED: ReadProtoFromTextFile(param_file, param). Failed to parse NetParameter file: /storage/emulated/0/Android/data/com.cr3d.mssd/files/opencvforunity/dnn/MobileNetSSD_deploy.prototxt in function void cv::dnn::ReadNetParamsFromTextFileOrDie(const char*, opencv_caffe::NetParameter*)
`
the first lines of the prototxt file seem correct, there is no "<"
`name: "MobileNet"
input: "data"
input_shape {
dim: 1
dim: 3
dim: 224
dim: 224
}
layer {
name: "conv0"
type: "Convolution"
bottom: "data"
`
have any suggestions?
thanks!
username_1: Model files can be downloaded using this script.
https://github.com/opencv/opencv_extra/blob/master/testdata/dnn/download_models.py
MobileNet_deploy.prototxt 29,353 bytes
MobileNet_deploy.caffemodel 23,147,564 bytes
username_2: Thanks! This all works for me now!
@username_1 I wanted to ask your advice/opinions of caffe or tensorflow with a webcam on android? your caffe example seems to perform the best of your object detection/localization demos.....
How should I ask this question? email? wiki? issue?
thanks!
Status: Issue closed
username_1: @username_2 Could you send a question to Technical Inquiry?
https://enoxsoftware.com/opencvforunity/contact/technical-inquiry/ |
google/material-design-icons | 120601213 | Title: give warning if injectTapEventPlugin() not called -> items not interacting with mouse
Question:
username_0: Hey, just wondering if it would be possible to give a console warning if you didn't call injectTapEventPlugin.
I just started a new project and forgot to do that, and at some point discovered that nothing responded to my mouse input. (perhaps because my laptop also supports touch?)
Took me a while to figure out this was the missing part.
Status: Issue closed
Answers:
username_1: Is this issue intended for the material-design-icons project?
I'm closing this as unrelated, but feel free to reply if I'm mistaken.
username_0: sorry, yes wrong project |
kszytko/object-oriented-programming | 648806274 | Title: cmake w SHM
Question:
username_0: - [ ] Użyjcie zmiennej ${PROJECT_NAME}
- [ ] Lista plików cpp w zmiennej
- [ ] Wszystko poza plikiem main.cpp powinno kompilować się do biblioteki statycznej
- [ ] Binarka (main.cpp) powinna linkować się z powyższą biblioteką.
- [ ] Napiszcie odpowiedni plik .github/workflows/main.yml który spowoduje, że GitHub będzie automatycznie uruchamiał kompilację projektu dla każdego nowego commita. |
CocoaPods/Xcodeproj | 210399502 | Title: Target has two dependencies, "each remove from project" two times?
Question:
username_0: # once only remove one dependency?
```
require 'xcodeproj'
# ruby removeDe.rb "path of project"
project_path = ARGV[0]
project = Xcodeproj::Project.open(project_path)
# puts project
project.targets.each do |target|
if target.name == 'MyTarget'
target.dependencies.each { |e| puts e.remove_from_project }
end
end
project.save
project_path = ARGV[0]
project = Xcodeproj::Project.open(project_path)
# puts project
project.targets.each do |target|
if target.name == 'MyTarget'
target.dependencies.each { |e| puts e.remove_from_project }
end
end
project.save
```
Answers:
username_0: How to delete all dependencies at once?
username_1: Hey there, this looks like a great example of a question that you should ask on [StackOverflow](http://stackoverflow.com) - there is an active tag [CocoaPods](http://stackoverflow.com/tags/cocoapods) where people can help out with questions like this. We try to keep the Xcodeproj issues focused only on bug reports for the tool, and for upcoming feature requests. Which this doesn't look like.
username_0: Thank you for your information
But I think is a bug
target.dependencies.each { |e| puts e.remove_from_project }
Should remove all dependencies
username_2: same here, have to run this twice to remove my watch target dependency: target.dependencies.each { |e| puts e.remove_from_project } |
billie66/hdsp | 127937529 | Title: http://7xnm4l.com1.z0.glb.clouddn.com/004-git-init.mp4
Question:
username_0: http://7xnm4l.com1.z0.glb.clouddn.com/004-git-init.mp4
Answers:
username_0: {
"title": "the first episode",
"video": "167-css-length.mp4"
}
{
"video": "168-youziku.mp4"
}
{
"video": "169-bye-rails.mp4"
}
{
"video": "170-spa.mp4"
}
{
"video": "171-h5.mp4"
}
{
"video": "172-github-signup.mp4"
}
{
"video": "173-github-new-repo.mp4"
}
{
"video": "174-h5-tags.mp4"
}
{
"video": "175-100-pages.mp4"
}
{
"video": "176-css3-start.mp4"
}
{
"video": "177-transition.mp4"
}
{
"video": "178-gulp.mp4"
}
{
"video": "179-npm-es6.mp4"
}
{
"video": "180-why-meteor-react.mp4"
}
Status: Issue closed
|
tensorflow/tensorflow | 245249629 | Title: tf.one_hot indices out of bound
Question:
username_0: array([ 0., 0., 0., 0.], dtype=float32)
```
Both overflow and underflow result in a zeroed array. Should this be caught, and errors be thrown? If not, why and when would this zeroed behaviour make sense? Or is it costly to check and throw assertions?
Status: Issue closed
Answers:
username_1: This is indeed intentional. See the examples on the [C++ op documentation](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/core/ops/array_ops.cc#L4494). We cannot change this without breaking backwards comptability. |
vcu-swim-lab/stack-intheflow | 182337721 | Title: Figure out how to rank query results based on relevance
Question:
username_0: This is an issue of research, and is specifically about how to rank results from a query, not how that query is generated (that's another issue: https://github.com/vcu-swim-lab/stack-intheflow/issues/6). We've talked about this a few times and have brought up several approaches including:
- The algorithm in the Prompter paper
- Using Stack Overflow's order of results, most likely as a starting point
- Using some machine learning technique to learn ranking order based on implicit feedback (i.e. if people tend to click the 5th result most often, that should be the highest ranking result).
To be clear, once we've agreed on a general algorithm/approach to the problem, this issue will be complete.<issue_closed>
Status: Issue closed |
luvit/luvit | 417242438 | Title: uv.spawn() might fail
Question:
username_0: [uv.spawn()](https://github.com/luvit/luvit/blob/master/deps/childprocess.lua#L133) might fail, e.g. because of `ENOMEM: not enough memory`. In this case the handle is `nil` and the `pid` has the value of the error message.
`uv.spawn()` not returning a valid `handle` should be handled as an error and reported to the user.
Answers:
username_1: Similar issue from when this affected `coro-spawn`: https://github.com/luvit/lit/pull/220
Status: Issue closed
username_0: Closed, as the errors from uv.spawn() are actually handled. Not the way I thought it should be, but after studying the sources, e.g. looking at [test-process.lua](https://github.com/luvit/luvit/blob/master/tests/test-process.lua), I understand how to get the error messages. |
InsertKoinIO/koin | 539779824 | Title: Restarting single instances in Koin
Question:
username_0: I'm trying to implement a session mechanism, where the user can logout and login at any moment with other account. In order to avoid having inconsistent information, I wanted to restart some instances within a module that I like this:
```
val sessionRepositories = module {
single { UserSessionRepository(androidApplication(), get(), get()) } bind AbstractSessionRepository::class
}
```
To this I tried to do:
```
unloadKoinModules(sessionRepositories)
loadKoinModules(sessionRepositories)
```
And for what I know, this will drop the previous UserSessionRepository instance and will create a new one, creating a leak with the previous UserSessionRepository. Therefore, I tried to do:
```
MyApplication.getInstance().stopKoinInstance() //calls GlobalContext.stop()
MyApplication.getInstance().startKoinInstance() //restarts all instances, just to see if there was something with load and unload
```
And this keeps leaking, is there a way, where I can really restart my instances
Answers:
username_1: I did a session management like what you want using a custom scope. It works pretty well. After user logins, you can start scope, and when user logouts, you can close the session.
https://doc.insert-koin.io/#/koin-core/scopes?id=using-a-scope
Status: Issue closed
|
rkotze/git-mob-vs-code | 497648032 | Title: Open git-coauthors file from sidebar doesn't work on Windows
Question:
username_0: <!--
Have you read our Code of Conduct?
By filing an Issue, you are expected to comply with it, including treating everyone with respect: https://github.com/username_1/git-mob-vs-code/blob/master/CODE_OF_CONDUCT.md
-->
### Prerequisites
- [X] [Checked that your issue isn't already filed][git-mob-vs-code issues]
- [X] Tried upgrading to the latest git-mob-vs-code and git-mob version (`npm i -g git-mob`)
- [X] Tried upgrading to the latest git version (`brew upgrade git` if installed with Homebrew/Linuxbrew)
[git-mob-vs-code issues]: https://github.com/username_1/git-mob-vs-code/issues?utf8=%E2%9C%93&q=is%3Aissue
### Description
Clicking the `Open git-coauthors file` button in the sidebar results in an error.
The replace regex in [git-mob-coauthors-file.js#L13](https://github.com/username_1/git-mob-vs-code/blob/893549652f6455396c3df7ca7099b2eca3c0a58c/src/git-mob-coauthors-file.js#L13) removes the drive letter which results in wrong path.
`file:///Users/r15ch/.git-coauthors` or `\Users\r15ch\.git-coauthors`
### Steps to Reproduce
1. On Windows 10 using VSCode Portable
2. Click `Open git-coauthors file` button
**Expected behaviour:** [What you expect to happen]
Should open `.git-coauthors` file
**Actual behaviour:** [What actually happens]

### Versions
- operating system and version: Windows 10
- git-mob-vs-code version (can be found in the Extensions tab): 1.1.1
- git-mob version (`git-mob --version`): 1.0.1
- git version (`git --version`): 2.20.1.windows.1
### Additional Information
<!-- Any additional information, configuration or data that might be necessary to reproduce the issue. -->
Answers:
username_1: Interesting, I specifically remove the drive letter to make ti work on Windows
- [Commit](https://github.com/username_1/git-mob-vs-code/commit/4bbc868daebe3bd2f9b4978330131092c67bf4fb)
- [Current test file](https://github.com/username_1/git-mob-vs-code/blob/master/src/git-mob-coauthors-file.spec.js)
I think the difference being you run the portable VS code version. I wonder if that has anything to do with it.
I'll try to find time to investigate it.
username_1: @username_0 I've setup Git Mob on a portable version of VS Code and was able to open the `.git-coauthors` file.
Where you able to follow the path in the error message to the file? I'm just wondering if the path is incorrect? By chance are you using the environment variable `GITMOB_COAUTHORS_PATH`?
username_0: The path itself is correct, except the missing drive letter. `GITMOB_COAUTHORS_PATH` is not set.
It works just fine if I remove `.replace(/^([a-z]:[\\]|[\/])/gi, "")` from
https://github.com/username_1/git-mob-vs-code/blob/f43599280b33b9e6e7b7cf52974f1866d2b4b22d/src/git-mob-coauthors-file.js#L18
username_0: Example output after adding test output.

```js
get path() {
vscode.window.showErrorMessage("without regex: " + path.join(os.homedir(), this.file));
vscode.window.showErrorMessage("with regex: " + path.join(os.homedir(), this.file).replace(/^([a-z]:[\\]|[\/])/gi, ""));
return path
.join(os.homedir(), this.file);
}
```
username_1: Interesting, I'm on windows 10 and used the latest version of git-mob, git-mob-vs-code and VS Code.
The only thing I don't know is what version of VS Code do you have?
As far as I understand Portable VS Code, it does **not** auto-update. Maybe that version does not support file paths in the same way as the latest.
username_0: 
Installed via [Scoop](https://github.com/lukesampson/scoop/) to `D:\scoop\apps\vscode-portable\current`
username_1: I see, the issue could be that is looking for .git-coauthors on your D drive but it's on your C drive.
This is easy to fix for windows, just need to think about Unix based systems.
username_0: Shouldn't be `os.homedir()` enough? The directory separators are correct in both cases.
```js
const os = require('os')
const path = require('path')
console.log(path.join(os.homedir(), '.git-coauthors'))
```

Test in Ubuntu WSL
username_1: I remember the issue was on a Mac. So I'll need to confirm the path on a Mac.
When I first did this, I thought `os.homedir()` was the solution.
username_1: I've tried it on MAC OS and works fine. So there is no need to remove the drive letter.
I'll get this fix out soon.
username_1: Will be part of release 1.3
Status: Issue closed
|
jabieniek/mlproject | 390478674 | Title: Summary
Question:
username_0: ## Rubric Score
### Criteria 1: Valid Python Code
* _Score Level:_ 4 (Meets Expectations)
* _Comment(s):_ All code is valid without errors.
### Criteria 2: Exploration of Data
* _Score Level:_ 3 (Meets Expectations)
* _Comment(s):_ Data is explored well but could use a but more varying methods of data visualization such as histograms and also a correlation table would have been very useful to you.
### Criteria 3: Machine Learning Techniques used correctly
* _Score Level:_ 3 (Meets Expectations)
* _Comment(s):_ ML Techniques are used for their correct purposes but results are not confirmed by other similar models and results are not explained very well. If a model produces results that are not meaningful you should hypothesize as to why that is and try to confirm it.
### Criteria 4: Report - Are conclusions clear and supported by data?
* _Score Level:_ 3
* _Comment(s):_ Conclusion is clear and good job hypothesizing about reasons your models were not very useful. Evidence with number would have been helpful as well.
### Criteria 5: Code formatting
* _Score Level:_ 3
* _Comment(s):_ Code is formatted well with sufficient comments. A couple extra comments explaining what is happening would help.
### Overall Score: 16/20
Overall good job! Even though your models did not provide meaningful results it is good you at least commented as to why. My only criticisms are that data could have been explored more efficiently and conclusions could have been presented with more solid evidence. Keep it up and happy coding! |
google/docsy | 993248563 | Title: Link in community_links.html partial should be relative
Question:
username_0: I think the `href` on line 17 of should not be absolute but rather relative so that it still works when there is a base URL.
I suggest it should be:
`<p>You can find out how to contribute to these docs in our <a href="../docs/contribution-guidelines/">Contribution Guidelines</a>.`
Rather than:
`<p>You can find out how to contribute to these docs in our <a href="/docs/contribution-guidelines/">Contribution Guidelines</a>.`
In it's current form it will link to `localhost:1313/docs/contribution-guidelines/` rather than `localhost:1313/<mybase-url>/docs/contribution-guidelines`
If you think this this correct I can make a pull request to correct this.
Answers:
username_1: Oh yes, I think you're right! Please do make a pull request.
username_0: Hi There, I just submitted the PR https://github.com/google/docsy/pull/687
username_1: Fixed in #687
Status: Issue closed
|
ikedaosushi/tech-news | 518003835 | Title: Docker再入門
Question:
username_0: Docker再入門<br>
Dockerにあまり触れる機会が少なく、たまに使うとコマンドとか仕組みをすぐに忘れてしまう。そんな自分へのナレッジ<br>
https://ift.tt/2JPMgFU |
SkyrimLL/SDPlus | 164692360 | Title: Set up basic quest for Alicia player start (1 quest stage)
Question:
username_0: - improved Alicia player start (only 1 quest stage for now, and possible enslavement right off the start if the sanguine quest alias is assigned at that point... I noticed sometimes it is still 'none' by the time the enslavement event kicks in).<issue_closed>
Status: Issue closed |
MelvorIdle/melvoridle.github.io | 823129570 | Title: Agility: Cliff balance impact on prayers
Question:
username_0: **Describe the bug**
- Agility "cliff balance" (5)
- Says "-2 flat prayer points" in red
- Prayers UI apply it as -2 (dmg reduction listed as 3 pts instead of 5) making it a buff?
- Does not seems right as it would not make it an obstacle very balanced (should be +2 for a total of 7?)
- Tested briefly, still cost 5 to activate said prayer
**Expected behavior**
TBD.
Probably to affect cost of prayer as an increase and to be reflected in both UI and activation cost
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Browser**
Chrome latest (88.0.4324.190)
**Console output**
N/A
**Are you using any scripts?**
None (Native)
Answers:
username_1: This has been hotfixed
Status: Issue closed
|
rmc47/CamLog | 365726672 | Title: If the user enters a QSO before the QRZ lookup is finished, the locator appears in the wrong location
Question:
username_0: While operating GB18NH I noticed that if I entered a QSO (hit return) before the QRZ.com lookup had finished, then the locator found in the QRZ.com lookup is inserted in the blank/new QSO log line.
For simplicity, let us assume there has been no QSOs entered in the log, so the first QSO entered would be serial 001;
1. The operator, with CamLog open in front of them, calls CQ and 'M1BXF' responds.
2. The operator enters 'M1BXF' into the callsign box for serial 001 and hits TAB.
3. CamLog looks up 'M1BXF' on QRZ.com to fetch the locator.
4. The Operator completes QSO 001 and enters it, by hitting return before the QRZ.com lookup is complete.
5. The locator found for 'M1BXF' on the QRZ.com lookup is entered into the locator box for serial 002.
I believe this will only be noticed while on a slow internet connection. It is also likely that the locator for QSO 001, which is entered into QSO 002, will be overwritten when the operator enters a new callsign into QSO 002 and hits TAB, which will trigger another QRZ.com lookup and will overwrite the locator. |
ynfatal/ynfatal.github.io | 627677597 | Title: SpringBoot2 | 第六篇:整合SpringDataJPA | Fatal's Blog
Question:
username_0: https://username_0.github.io/2018/09/30/SpringBoot2/SpringBoot2%E7%AC%AC%E5%85%AD%E7%AF%87%E6%95%B4%E5%90%88SpringDataJPA/
上一篇介绍了Spring JdbcTemplate的使用,对比原始的JDBC而言,它更加的简洁。但随着表的增加,重复的CRUD工作让我们苦不堪言,这时候Spring Data Jpa的作用就体现出来了….. [TOC] JPA JPA 是 Java Persistence API 的简称,中文名 Java持久层API,是JDK 5.0注解或XML描述对象-关系表的映射关系,并将运行 |
wojtekmaj/react-pdf | 1147855388 | Title: _interopRequireWildcard is not a function
Question:
username_0: ### Before you start - checklist
- [X] I followed instructions in documentation written for my React-PDF version
- [X] I have checked if this bug is not already reported
- [X] I have checked if an issue is not listed in [Known issues](https://github.com/wojtekmaj/react-pdf/wiki/Known-issues)
- [ ] If I have a problem with PDF rendering, I checked if my PDF renders properly in [PDF.js demo](https://mozilla.github.io/pdf.js/web/viewer.html)
### Description
So i've installed @react-pdf/renderer in my project and i've tried to test the basic example, like creating a basic document and trying to download it
And I got some incompatibility with the webpack ( 4.4.2) :
var R = _interopRequireWildcard(__webpack_require__(/*! ramda */ "../../../../modules/my-project/code/node_modules/ramda/es/index.js"));
I tried adding ramda manually but it wouldn't work
### Steps to reproduce
Install the module and try the basic example
### Expected behavior
Should see the pdf
### Actual behavior
Not working as expected
### Additional information
_No response_
### Environment
- **Browser (if applicable)**:
- **React-PDF version**:
- **React version**:
- **Webpack version (if applicable)**: |
Welfenlab/tutor-student | 113148631 | Title: Browserified files should contain absolute paths
Question:
username_0: Strings like `/home/maik/Gits/tutor-meta/tutor-student/node_modules/...` should not appear in any distributed files.
Answers:
username_0: Even better than configuring browserify not to include full paths: https://www.npmjs.com/package/unpathify
Status: Issue closed
username_1: Strings like `/home/maik/Gits/tutor-meta/tutor-student/node_modules/...` should not appear in any distributed files.
username_1: Did you check that your change didn't break the discify builder. It requires full paths to analyze the packages.
username_0: The plugin only works if `fullPaths: false`, which is only set if `NODE_ENV == "production"`, so discify will work.
Status: Issue closed
|
remisharrock/c-programming-with-linux-MOOC-issues-tracker | 365561533 | Title: Taskgrader
Question:
username_0: Which course: 2 Language Foundations What type of content: Taskgrader URL: https://courses.edx.org/courses/course-v1:Dartmouth_IMTx+DART.IMT.C.02+1T2018/courseware/a86febae1149479d971ae6edbd850c51/fc9008c149444ca0a623cc332fc01e06/?child=first Description: Some euros (for the children) remains in the problem statements -) other comments: your name: Isa screenshot: Codecast timing: Tasgrader code submitted: Operating system: which web browser? Status actuel:
Answers:
username_1: resolved
Status: Issue closed
|
jlippold/tweakCompatible | 362467602 | Title: `App Percent` working on iOS 9.3.3
Question:
username_0: ```
{
"packageId": "com.hackyouriphone.apppercent",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "com.hackyouriphone.apppercent",
"deviceId": "iPhone8,2",
"url": "http://cydia.saurik.com/package/com.hackyouriphone.apppercent/",
"iOSVersion": "9.3.3",
"packageVersionIndexed": false,
"packageName": "App Percent",
"category": "HYI - Tweaks",
"repository": "HackYouriPhone",
"name": "App Percent",
"packageIndexed": true,
"packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.",
"id": "com.hackyouriphone.apppercent",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.0.7",
"shortDescription": "AppStore Installation in Percent (%) and Device Notification when an app download is finished. Configure option from Settings!",
"latest": "2.0.0",
"author": "pxcex",
"packageStatus": "Unknown"
},
"base64": "<KEY>",
"chosenStatus": "working",
"notes": ""
}
``` |
explosion/spaCy | 1059953122 | Title: EntityRecognizer throws IndexError when used in pipeline with Transformer and custom span getter
Question:
username_0: EntityRecognizer throws IndexError when used in pipeline with Transformer and custom span getter:
```
File "/home/---/---/research_spacy_ru/.venv/lib/python3.8/site-packages/spacy/language.py", line 1122, in update
proc.update(examples, sgd=None, losses=losses, **component_cfg[name])
File "spacy/pipeline/transition_parser.pyx", line 416, in spacy.pipeline.transition_parser.Parser.update
File "spacy/ml/parser_model.pyx", line 293, in spacy.ml.parser_model.ParserStepModel.finish_steps
File "spacy/ml/parser_model.pyx", line 456, in spacy.ml.parser_model.precompute_hiddens.begin_update.backward
File "/home/---/---/research_spacy_ru/.venv/lib/python3.8/site-packages/spacy/ml/_precomputable_affine.py", line 49, in backward
Xf = X[ids]
IndexError: index 221 is out of bounds for axis 0 with size 221
```
## How to reproduce the behaviour
I created my custom span_getter: https://gist.github.com/username_0/06e53b108f764e7240ea7ae8e2e830fd
It adapts number of words to respective number of word pieces, to better fit into transformer window.
Pipeline works with this function, the exception is thrown only at some documents.
I plug it into simple transformer + ner pipeline like this: https://github.com/username_0/natasha-spacy/blob/transformer-pipeline/project/config_trf.cfg
(in my tests I disabled all but transformer and NER)
This error is emitted at the line https://github.com/explosion/spaCy/blob/master/spacy/ml/_precomputable_affine.py#L49
## Your Environment
- **spaCy version:** 3.2.0
- **Platform:** Linux-5.4.0-90-generic-x86_64-with-glibc2.29
- **Python version:** 3.8.10
- **spacy-transformers** = "^1.0.6"
- **torch** = "1.9.1"
* Operating System: Ubuntu 20.04
* Python Version Used: Python 3.8.10
* spaCy Version Used: 3.2.0
Answers:
username_1: Thanks for the report and sorry it's taken us a long time to follow up on this. Unfortunately, because the issue is happening deep in the spaCy internals and your custom code isn't very simple, it's hard to be sure what's going on here.
Can you create a small example we can run to reproduce the problem? A repo like the one you linked to with a project file would be great, but that repo's project file doesn't seem to work and doesn't use Transformers anyway.
username_0: Thanks for your reply.
I reproduced the behavior based on one of spaCy tutorials: https://github.com/username_0/tutorial_spacy_custom_span_getter
The only changes I do are:
- I add my span getter (I added more comments to make its algorithm more clear)
- I alter config to use my transformer of choise
And the error remains.
P.S. The repo I linked in my first message does use transformer config, in project file it's called by "train_trf" and not "train" - to be able to use both configs. |
fossasia/pslab-android | 398692327 | Title: Multimeter should show a default value of 0.00
Question:
username_0: **Actual Behaviour**
Currently, the multimeter shows blank by default
**Expected Behaviour**
It should show a default value of 0.00
**Steps to reproduce it**
Go to Multimeter Activity
**Would you like to work on the issue?**
Yes<issue_closed>
Status: Issue closed |
AtsushiSakai/PythonRobotics | 512934802 | Title: Informed RRT* incorrectly update cBest for ellipse
Question:
username_0: **Describe the bug**
In function [informed_rrt_star_search](https://github.com/username_1/PythonRobotics/blob/master/PathPlanning/InformedRRTStar/informed_rrt_star.py#L38), `cBest` and `pathLen` seems have same meanings. After finding a new path, `pathLen` is used to check whether `cBest` should be updated, but `pathLen` itself is not being updated.
The performance of the bug is that the ellipse may increase sometimes when you find a new path.
Also, if you print `cBest` everytime ploting the ellipse, you'll find that `cBest` may increase sometimes.
**Screenshots**
Here is an example:
There is already a path with length 11.808038783431691.

After a new path with longer length 12.487373607747500 is found, cBest is updated.

Answers:
username_0: Simply fix: #248
Status: Issue closed
|
hundertzehn/mocoapp-api-docs | 634724038 | Title: [FEATURE REQUEST] - Endpoint for sending invoices via email
Question:
username_0: Hey there!
I already had a nice chat with Thomas about this.
We would like to automate our invoicing. For that it would be necessary to send an invoice via API.
For now - as a workaround - we send the emails using an email service and attach the invoices manually using the [/invoices/{id}.pdf](https://github.com/hundertzehn/mocoapp-api-docs/blob/master/sections/invoices.md#get-invoicesidpdf) endpoint.
Greetings to Switzerland,
Corvin
Answers:
username_1: We'll have a look at this @username_0. But it might take some time. Pretty busy atm with new features to come :muscle:
username_2: This is now available: https://github.com/hundertzehn/mocoapp-api-docs/pull/98/commits/4c96fceb59bba003488f3fb5903b9eedcae409a8
Let us know if there are any problems, best
Status: Issue closed
|
lovelmh13/myBlog | 562639268 | Title: input从右到左显示文字,并且光标始重在右侧。text-align和direction的区别
Question:
username_0: 其实很简单:
```css
.input {
text-align: right
}
```
需要分清和`direction: rtl;`的区别
`text-align`是让文本靠右,而`direction: rtl;`是改变文本的方向
```html
<style type="text/css">
div.one {
text-align: right
}
div.two {
direction: rtl;
}
</style>
<div class="one">Some text. Right-to-left direction.</div>
<div class="two">Some text. Left-to-right direction.</div>
```
`.class`的显示效果只是让文字整体平移到右边
`.two`除了文字移到了右侧显示,文字也发生了改变。 会先成:`.Some text. Left-to-right direction`,看见区别了吗?句号位置变了。
经过我的测试,**如果只是改变文本内容**,`direction: rtl;`会改变符号和文字的位置,而文字如果也想从右往左写,只需要稍加改动,就可以把顺序完全颠倒,变成从右往左写:
```css
div.two {
direction: rtl;
unicode-bidi:bidi-override;
}
```
就会变成:
`Some text. Left-to-right direction.`
但是,**如果换成是input**,除了显示跟本文的显示方式一样以为,光标是不一样的。
`direction: rtl;` 因为是改文字方向,所以光标也改成了在左边。使用`backspace`和`delete`按键也会相反。
`text-align: right`是改变了文字的位置,而不是方向,光标是正常的,所以这个才是我们需要用的属性。 |
crossplane/crossplane | 1092803887 | Title: Cannot spin up multiple ec2 resources at once
Question:
username_0: <!--
Thank you for helping to improve Crossplane!
Please be sure to search for open issues before raising a new one. We use issues
for bug reports and feature requests. Please find us at https://slack.crossplane.io
for questions, support, and discussion.
-->
### What happened?
I'm trying to spin up multiple AWS EC2 resources at once with multiple defined blocks using Composition, but it's always just one EC2 resource spinning up (and it's always the last defined resource).
### How can we reproduce it?
I put the configuration in my public repo, please check
https://github.com/username_0/infra
### What environment did it happen in?
Crossplane version:
- Crossplane 0.22.0
- provider/aws
- kubectl version
client: Major:"1", Minor:"21", GitVersion:"v1.21.1"
server: Major:"1", Minor:"20", GitVersion:"v1.20.5"
Status: Issue closed
Answers:
username_0: Sorry, I realized where my mistake is. Close the issue now |
michaelforney/swc | 374808126 | Title: Missing major()/minor() defines on Fedora 29
Question:
username_0: These are defined in `<sys/sysmacros.h>`.
Patch:
```
diff --git a/launch/launch.c b/launch/launch.c
index b8cc292..b6f0aa9 100644
--- a/launch/launch.c
+++ b/launch/launch.c
@@ -40,6 +40,7 @@
#include <sys/stat.h>
#include <sys/wait.h>
#include <sys/ioctl.h>
+#include <sys/sysmacros.h>
#include <linux/input.h>
#include <linux/kd.h>
#include <linux/major.h>
diff --git a/libswc/drm.c b/libswc/drm.c
index 910c034..a9c3ef9 100644
--- a/libswc/drm.c
+++ b/libswc/drm.c
@@ -39,6 +39,7 @@
#include <unistd.h>
#include <drm.h>
#include <xf86drm.h>
+#include <sys/sysmacros.h>
#include <wld/wld.h>
#include <wld/drm.h>
#include <wayland-server.h>
```
Status: Issue closed
Answers:
username_1: Thanks for the report. Should be fixed now. |
ionic-team/ionic-cli | 233431901 | Title: Ionic v1: `ionic serve` tells user that app is available at 0.0.0.0:8100 instead of localhost:8100
Question:
username_0: Noticed several threads about this on the forum:
https://forum.ionicframework.com/t/ionic-serve-is-not-working/92937
https://forum.ionicframework.com/t/ionic-serve-points-to-0-0-0-0-8100/92902
https://forum.ionicframework.com/t/how-i-run-my-ionic-1-application-on-cli-3-using-the-serve/92455
https://forum.ionicframework.com/t/ionic-serve-lab-not-working-after-updating-to-latest-ionic-version-ionic-v1/91743
Answers:
username_1: @username_0 could you reproduce the problems?
username_2: Yeah, I know what this is about. It defaults to `0.0.0.0` for Ionic 1.
Status: Issue closed
|
treasure-data/digdag | 335157836 | Title: How about add a link to docs.digdag.io to README
Question:
username_0: I tried to search the way installing digdag on Windows10 around https://github.com/treasure-data/digdag and could not find the way for a while.
I think that the link to www.digdag.io in https://github.com/treasure-data/digdag#Documentation is ambiguous if you are beginner and you want to install digdag first time.
To link docs.digdag.io directly will make the instruction more clear.
Answers:
username_1: Hello, @username_0
Thank you for your feedback.
The Documentation text is the link for that document.
Is it hard to find?

username_0: Hi, @username_1
Thank you for your instruction!
Only to find the link is not so hard, I think.
But the title "Documentation" refers different urls than that's contents and it may cause a problem.
I misunderstood the link in "Please check digdag.io for installation & user manual." refers the same url with its title.
And link to www.link.io in the sentence is somehow unhelpful when you just look for installation or user manual because there are too much information than you need.
Status: Issue closed
|
construct/construct | 295163407 | Title: compiler upgrade, generate Cython code
Question:
username_0: This is a major upgrade to the compiler. Benefits: much better performance. Drawback: adds optional dependency to `cython` package, if user wants to use the compiler that is, and compiles the `.compile` method somewhat.
Answers:
username_1: I don't understand nothing in cython and python internals. Maybe @tmr232 won't mind doing some review to it. Though he is quite busy.
username_0: Related https://github.com/cython/cython/issues/2103
username_0: Finished. Docs at:
https://construct.readthedocs.io/en/latest/compilation.html |
scalikejdbc/scalikejdbc | 418229748 | Title: How to do select or insert query?
Question:
username_0: I have the following code I would to execute
```
sql"""WITH s AS (
SELECT id
FROM $tableName
WHERE $columnName = $value
), i AS (
INSERT INTO $tableName ($columnName)
SELECT $value
WHERE NOT EXISTS (
SELECT id
FROM $tableName
WHERE $columnName = $value
)
RETURNING id
)
SELECT id
FROM i
UNION ALL
SELECT id
FROM s
"""
```
If I use `update` or `updateAndRetrieveGeneratedKey` within a `localTx` I get `org.postgresql.util.PSQLException: A result was returned when none was expected.`. If I use `single` within `readOnly` I get something like `... cannot perform a insert in a read only query`.
What is the proper way to go about this?
Answers:
username_1: I've never tried yet but what happens if you run the query as blow?
```scala
// or DB.autoCommit {
DB.localTx { implicit s =>
selectOrInsertQuery.map(toObject).list.apply()
}
```
username_2: The problem is in SQL.scala:814 which always assumes AutoSession is using DB.readonly irregardless what the AutoSession.isReadOnly flag is saying.
This complicates things a lot when using AutoSession- that is, they can't be used at all if at any point in the transaction there happens to be an insert.
username_1: @username_2 Thanks. The line 814 you mentioned is this line: https://github.com/scalikejdbc/scalikejdbc/blob/3.3.5/scalikejdbc-core/src/main/scala/scalikejdbc/SQL.scala#L814
I understand your point. But, on the other hand, I don't have any ideas to eliminate your problems here except for removing `AutoSesison` from the default param for methods or not to use AutoSession as much as possible.
username_1: As I mentioned above, there is no other suggestion apart from using the right session type (read-only, auto-commit, read-only) for each query in this scenario. Let me lose this issue now.
Status: Issue closed
|
MicrosoftDocs/azure-docs | 318145618 | Title: Misleading URL
Question:
username_0: The URL for this page is 'https://docs.microsoft.com/en-us/azure/app-service/containers/quickstart-nodejs'
which suggests the page has information regarding deploying a container to Azure App Service, but that is not the case.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 116d6e28-471b-2cf4-3eb8-6249b4ef3016
* Version Independent ID: f199d3a2-0c11-e2d6-18f0-d081c76d511d
* Content: [Create a Node.js in Azure App Service on Linux](https://docs.microsoft.com/en-us/azure/app-service/containers/quickstart-nodejs)
* Content Source: [articles/app-service/containers/quickstart-nodejs.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/containers/quickstart-nodejs.md)
* Service: **app-service-web**
* GitHub Login: @username_2
* Microsoft Alias: **username_2**
Answers:
username_1: @username_0 Thanks for the feedback! I have assigned the issue to the content author to investigate further and update the document as appropriate.
username_2: @username_0 Hi thanks. Technically it's true because App Service is deploying a built-in Node.js docker image to run the Node.js app in a container. However, you're right that it's not running your custom container. For running a custom docker image, please see: https://docs.microsoft.com/en-us/azure/app-service/containers/tutorial-custom-docker-image.
This is our content decision to have built-in and customer image content both under the "container" path. #please-close
Status: Issue closed
username_3: @username_0 We will now proceed to close this thread. If there are further questions regarding this matter,
please reopen it and we will gladly continue the discussion. |
renatomrcosta/jeeves-smugglesworth | 409681179 | Title: ask if user is still merging or forgot to release the token
Question:
username_0: if you have the token for more then 3 hours let jeeves send you a message if you are still merging
Answers:
username_1: Defining hard-rules for implementation:
- Every hour, on the hour, Jeeves will ask whoever has a token for 3h or longer if they are still merging.
- The same user will be asked again and again, every hour.
Later I can add working hours (so Jeeves doesn't repeat itself), or a flag indicating that a given queue has been warned beforehand. Dunno yet
username_1: Shoddily done, adding even more clutter to the code IMHO. Oh well.
Status: Issue closed
|
facebook/flow | 236196420 | Title: Debugging Flow
Question:
username_0: I am using Visual Studio Code with [vscode-ocaml](https://github.com/hackwaly/vscode-ocaml) extension. `./bin/flow` is compiled using `make build-flow-debug` (see also: https://github.com/facebook/flow/issues/4177).
`.vscode/launch.json`:
```js
{
"version": "0.2.0",
"configurations": [
{
"name": "OCaml",
"type": "ocamldebug",
"request": "launch",
"program": "${workspaceRoot}/bin/flow",
"console": "internalConsole",
"stopOnEntry": false,
"arguments": [
"${workspaceRoot}/local-test"
]
}
]
}
```
`local-test/test.js`:
```js
// @flow
const a: number = 'z' + 'd';
```
Debugging `main` seems to work:
<img width="794" alt="screen shot 2017-06-15 at 15 28 09" src="https://user-images.githubusercontent.com/794591/27183582-4f28422c-51df-11e7-9a48-af9d5b69d253.png">
However If I set a breakpoint somewhere in `src/typings/`, for example `statement.ml`, inside `variable_decl` function, then the breakpoint is skipped.
I suspect this is because of the server-client architecture - actual type checking happens inside a different process.
Is there any good way around this?
Answers:
username_1: I faced the same problem and I didn't solve it.
For the reference, I tried `flow check` command.
It seems to be checking type at foreground, but when using `ocamlbuild` with settings like below,
debugger responds with `cannot connect to debugger at 127.0.0.1 error: No such file or directory`.
```javascript
{
"program": "${workspaceRoot}/bin/flow",
"arguments": ["check", "path/to/mycode.js"],
}
```
The error occurred at `spawn` function in `hack/utils/daemon.ml`(yes, the debugger worked correctory), and I guess that it related to spawning server process.
username_1: Still examining, but it seems to get some progress.
As above comment I guess, the problem caused by some process that created behind `flow`.
Then, I try out arguments "--max-workers" that described with `Maximum number of workers to create (capped by number of cores)`
I don't understand it in detail, but it seems related to some process behind `flow`.
So, I modified my settings like follows and executed `ocamldebug`.
```
{
"program": "${workspaceRoot}/bin/flow",
"arguments": ["check", "--max-workers", "0", "path/to/mycode.js"],
}
```
And the result looks as I expected.
At least it can execute `flow` according to `breakpoint` and process exit normally.
username_2: @username_0 @username_1 have you tried to get debugger working with [vscode-ocaml-debugger](https://github.com/hackwaly/vscode-ocaml-debugger)? I can get the debugger working with simple example using `ocamlc -g helloworld.ml` but with Flow pretty much nothing happens, only the loading sign appears below debug menu,
username_2: Looks like ocaml debugger runs roughly this kind of command: `ocamlrun bin/flow check my-own-test/test.js --max-workers 0`, when running the flow with `ocamlrun` the program throws
```
Fatal error: unknown C primitive `unix_waitpid'
```
Maybe this is the culprit?
username_3: I don't use the ocaml debugger, but flow has a few debugging commands that I use to debug type checker behavior. `flow check --verbose-indent` will show the full list of flows in `__flow`. The function that adds errors is called `add_output`, and it also logs when it is called with `--verbose-indent`.
username_2: Nice, this probably already helps a lot, thanks for being so helpful!
username_2: Ok, got debugger finally working in VScode!
VScode debugger config:
```
{
"version": "0.2.0",
"configurations": [
{
"name": "OCaml",
"type": "ocaml-debugger",
"request": "launch",
"program": "${workspaceFolder}/bin/flow",
"console": "internalConsole",
"stopOnEntry": false,
"arguments": ["check", "${workspaceFolder}/local-test/test.js", "--max-workers=0", "--verbose-indent"]
}
]
}
```
- `${workspaceFolder}/local-test/test.js` being the js file to check
- `--max-workers=0` apparently runs everything in single process
- `--verbose-indent` Debugger works also without, but still very useful
Other requirements:
- OCaml 4.05.0 <- later versions might work too, I only tested with 4.05.0
- VScode debugger plugin: https://marketplace.visualstudio.com/items?itemName=hackwaly.ocaml-debugger
- debugger adapter: https://github.com/hackwaly/ocamlearlybird/
- needs to be installed with opam: `opam install earlybird`
- NOTE: Needs to be at least 0.1.3 version, earlier versions does not work
I also recommend to install Merlin: https://github.com/ocaml/merlin
```
opam install merlin
opam user-setup install
```
This might be worth documenting to e.g Flow wiki or readme.
username_4: I would also suggest using `--merge-timeout=0` as well to avoid timeouts.
username_4: @username_2 thanks for the instructions. Being able to step through things is going to be a big help.
username_5: Thanks @username_2 for such a useful comment.
I was unable to get mine running until I added to my user settings.json:
```json
{
"reason.path.ocamlmerlin": "_opam/bin/ocamlmerlin",
"reason.diagnostics.tools": [
"merlin"
],
"editor.formatOnSave": true,
"reason.codelens.enabled": true,
}
```
and to my debug configuration:
```json
/* start ocamlearlybird --server --port 4711 before running */
"debugServer": 4711
```
I then run `$ ocamlearlybird --server --port 4711` in the Flow directory before debugging. |
podium-lib/issues | 475677629 | Title: Remove deprecated events
Question:
username_0: Remove the deprecated [change](https://github.com/podium-lib/client/blob/9f0ff38d1d2d8ffe8f4c817758eea8570cb35c4d/lib/client.js#L102-L104) and [dispose](https://github.com/podium-lib/client/blob/9f0ff38d1d2d8ffe8f4c817758eea8570cb35c4d/lib/client.js#L110-L112) events in the @podium/client.<issue_closed>
Status: Issue closed |
forceworkbench/forceworkbench | 10351637 | Title: Serializable Field Mappings
Question:
username_0: _Original author: <EMAIL> (July 14, 2011 15:06:36)_
<b>1. Summary of the feature request:</b>
Ability to save mappings like DataLoader
<b>2. Typical scenerio where this feature would be used:</b>
Any insert, update, upsert scenario
<b>3. Reason current features do not meet needs:</b>
May have to repeat something many times and mapping each time is time consuming
<b>4. How soon would you like to see this feature implemented?</b>
Next API Release
<b>5. Would you like to contribute to the development of this feature?</b>
I would help
_Original issue: http://code.google.com/p/forceworkbench/issues/detail?id=480_<issue_closed>
Status: Issue closed |
Rydra/bgm-for-es | 771719545 | Title: Music doesn't stop with non-RA cores
Question:
username_0: I love this script but anytime I run a non-Retroarch core, like Daphne or MESS the music just keeps playing. Is there any way to stop the music when these cores load up?
Answers:
username_1: We just need to add the emulator name to the list that is in `MusicStateMachine.py`.
Mine does it on the `drastic` emulator for DS... I'm not very linux savvy or github savvy or else I would fork my own version and build it locally with all the emulator names I use.
username_2: Where is the location of MusicStateMachine.py?
username_0: /usr/lib/python2.7/dist-packages/bgm/
username_3: I have bumped to the new version of the Background music for EmulationStation. This new release is a complete revamp, with a new installation script. Please follow the following steps:
1. Uninstall the old script if you had it:
```
sudo apt remove python-es-bgm
```
2. Go to the main page of the project and reinstall the script, following the instructions of the documentation:
```
https://github.com/username_3/bgm-for-es
```
The new config file now allows you to update/customize the processes and emulators where the music should stop, without having to fiddle with code. Feedback always appreciated!
Status: Issue closed
|
kalkih/mini-media-player | 631533520 | Title: Get rid of box shadow
Question:
username_0: Hi Karl,
first of all thanks for the great plug-in. I integrated it to my mobile UI and it looks so much better than the normal media-control.
https://community.home-assistant.io/t/google-inspired-home-assistant-on-mobile/202191
I have two questions regarding your plug-in:
- Is there a way to get rid of the box-shadow?
- is there a way to change the color of the playbar (currently orange)?
i tried to integrate card-mod by thomas loven with the following code:
entity: media_player.sonos_wohnzimmer
show_header_toggle: false
icon: 'mdi:play-box-outline'
name: Sonos Play One Wohnzimmer
artwork: full-cover
style: |
ha-card {
box-shadow: none;
webkit-box-shadow: none;
}
type: 'custom:mini-media-player'
but without success.
Kind regards,
Philipp
Status: Issue closed
Answers:
username_0: reinstallation of card-mod fixed the issue |
milvus-io/milvus-sdk-go | 1076934154 | Title: [Bug]: panic: runtime error: slice bounds out of range
Question:
username_0: ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
Exceptions occur when searching multiple sets of vectors "panic: runtime error: slice bounds out of range "
### Expected Behavior
1、
```
func main() {
// Milvus instance proxy address, may verify in your env/settings
milvusAddr := `xxxxxxxxxxxx`
// setup context for client creation, use 2 seconds here
ctx := context.Background()
ctx, cancel := context.WithTimeout(ctx, 5*time.Second)
defer cancel()
c, err := client.NewGrpcClient(ctx, milvusAddr)
vector1 := entity.FloatVector([]float32{1,2,3,4,5,6,7,8})
vector2 := entity.FloatVector([]float32{1.1,1.2,1.3,1.4,1.5,1.6,1.7,1.8})
vector3 := entity.FloatVector([]float32{2.1,2.2,2.3,2.4,2.5,2.6,2.7,2.8})
vector4 := entity.FloatVector([]float32{3.1,3.2,3.3,3.4,3.5,3.6,3.7,3.8})
entitys := []entity.Vector{vector1,vector2,vector3,vector4}
sp, _ := entity.NewIndexFlatSearchParam(50)
sr, err := c.Search(ctx, "gosdk_insert_example", []string{}, "Year > 1990", []string{"ID"}, entitys, "Vector",
entity.L2, 10, sp)
fmt.Println(sr)
if err != nil {
log.Fatal("fail to search collection:", err.Error())
}
}
```
2、

### Steps To Reproduce
_No response_
### Environment
_No response_
### Anything else?
_No response_ |
jfrog/charts | 475500831 | Title: Providing custom binarystore.xml to the chart
Question:
username_0: <!-- Thanks for filing an issue! Before hitting the button, please answer these questions.
It's helpful to search the existing GitHub issues first. It's likely that another user
has already reported the issue you're facing, or it's a known issue that we're already aware of-->
**Is this a request for help?**:
---
**Is this a BUG REPORT or FEATURE REQUEST?** (choose one):
FEATURE REQUEST
<!--
If this is a BUG REPORT, please:
- Fill in as much of the template below as you can. If you leave out
information, we can't help you as well.
If this is a FEATURE REQUEST, please:
- Describe *in detail* the feature/behavior/change you'd like to see.
In both cases, be ready for followup questions, and please respond in a timely
manner. If we can't reproduce a bug or think a feature already exists, we
might close your issue. If we're wrong, PLEASE feel free to reopen it and
explain why.
-->
**Version of Helm and Kubernetes**:
all versions
**Which chart**:
artifactory
**What happened**:
As of writing this issue, the artifactory chart creates a general binarystore.xml based on persistence storage type. We use cachefs with filesystem and right now the filesystem type creates a default binarystore without cachefs.
**What you expected to happen**:
option to provide the chart a custom binarystore.xml to use
Answers:
username_1: @username_0 - This is a valid request. We'll provide a pull request in the near future.
Status: Issue closed
username_1: <!-- Thanks for filing an issue! Before hitting the button, please answer these questions.
It's helpful to search the existing GitHub issues first. It's likely that another user
has already reported the issue you're facing, or it's a known issue that we're already aware of-->
---
**Is this a BUG REPORT or FEATURE REQUEST?** (choose one):
FEATURE REQUEST
<!--
If this is a BUG REPORT, please:
- Fill in as much of the template below as you can. If you leave out
information, we can't help you as well.
If this is a FEATURE REQUEST, please:
- Describe *in detail* the feature/behavior/change you'd like to see.
In both cases, be ready for followup questions, and please respond in a timely
manner. If we can't reproduce a bug or think a feature already exists, we
might close your issue. If we're wrong, PLEASE feel free to reopen it and
explain why.
-->
**Version of Helm and Kubernetes**:
all versions
**Which chart**:
artifactory
**What happened**:
As of writing this issue, the artifactory chart creates a general binarystore.xml based on persistence storage type. We use cachefs with filesystem and right now the filesystem type creates a default binarystore without cachefs.
**What you expected to happen**:
option to provide the chart a custom binarystore.xml to use
username_1: Will close once `artifactory-ha` is also updated with the change.
Status: Issue closed
|
SiciliaHub/nonportaleregionesicilia | 160409658 | Title: Aggiungere testo descrittivo nella pagina "Cerca nei siti"
Question:
username_0: @username_1 aggiungerei [qui](https://github.com/SiciliaHub/nonportaleregionesicilia/blob/gh-pages/cerca.html#L92) un testo descrittivo brevissimo sul cosa è questa pagina.
Se sei d'accordo però pensaci tu, che ne sei l'ideatore
Answers:
username_1: @username_0 testo aggiunto ;) vedi se ti sembra che va bene e se no, modificalo pure o dimmi se devo aggiungere altre info.
username_0: Ho fatto una piccola modifica. Metterei in link alla tabella che contiene l'elenco degli URL mappati. Che ne dici?
username_1: Ottimo! Sono d'accordo per il link (all'inizio l'idea era quella di creare la pagina dei siti regionali, vedi #35). L'unico timore era che poi veniva simile a quella dei dataset e questo poteva confondere. Mi sembra che con il link questo pericolo non ci sia...
username_0: @username_1 dove trovo una tabella con l'elenco di link? Mi avrai già inviato un gdrive, ma non riesco a recuperarlo. Grazie
username_1: @username_0 è nel quarto foglio del google sheet del non-portale https://docs.google.com/spreadsheets/d/15cFU9v2x3ACb1IPcVE9Xg3wc5fuay8NAYDXQnoMcOFw/edit#gid=2118198704
Status: Issue closed
username_0: Ho aggiunto il link all'elenco dei siti mappati dal motore |
apache/pulsar | 550823488 | Title: Subscription metrics name error
Question:
username_0: The metric name of the document is pulse_subscription_unacked_messages but Prometheus print is pulsar_subscription_unacked_massages , word messages spelling mistakes
Answers:
username_1: Thanks @username_0 can you help fix the typo issue? |
BevoTechPack/BevoTechPack | 205702780 | Title: Bevo Tech Pack Server/Singleplayer World Loading Problems
Question:
username_0: Failure to give adequate information on issues/bugs will be subject to immediate possible closure/spankings.
Bevo Tech Pack Version : 2.21
### Description of issue/bug
At first sorry for my bad english.
Well.... i made a bevo tech pack server with my friends.
At first everything was fine and we could play on this server with no problems.
But since 3 days ago i couldn´t start the server.
### Steps to reproduce the issue/bug
I´ve already find out, that the problem is the world. I´ve removed the "world" folder, started the server, he created a new world and started with no problems.
I´ve tried to start the world in singleplayer and my Minecraft crashed...
I hope you can help me.
Greetings from Germany
### http://pastebin.com/5LNsKRv8
Questions:
Did you add any additional mods? - Just "fast craft"
Do you use optifine? - No
Did the problem appear in single player or on a server? - Both
Answers:
username_1: So you have 2 choices, lets start with the easy one and see if it fixes the world.
Edit config/forge.cfg on the server and change
B:removeErroringEntities=false
B:removeErroringTileEntities=false
both to true, load up your world, it may still crash, load it up again.
If it fixes your world change those 2 back to false (don't play your game with them set to true).
If it fixes the issue close this if it doesn't let me know.
username_0: No..... it haven´t worked.
First start: http://pastebin.com/D1fsRSWp
Second start: http://pastebin.com/wqU7CtAS
username_1: Ok so you are left with world editing out block located at -345,75,79 the issue you have is the server crashes so you have to do this manually using world edit gui. It's not something I can talk you through as it's complicated and a wrong move could ruin your world completely. Do you have someone with World Edit knowledge to do that?
username_0: Only some single basics...
I´ve tried to remove the block and i did it.
But now the server starts without mods..(I did an backup from the world before i did it)
Status: Issue closed
username_0: OH MY GOD
Thank you very much.
After trying to handle MC Edit, i did it. I removed the block and everything works fine now.
After the first try the whole world were reseted, but the mods were working. After the second try the mods didn´t worked, but the world was correct. And after the third try everything is working.
Thank you very much. |
rterrabh/LFApp | 164886875 | Title: Salvar a gramática testada quando sair e entrar do APP.
Question:
username_0: Procedimento: Após inserir uma gramática, funciona tudo de acordo. Porém, quando muda de Activity no Android, ou seja, vai pro Whatsapp ou Facebook, a gramática se perde.
Tem uma configuração que acredito que seja savedInstanceState, que tem como salvar o que foi feito anteriormente. Assim quando a pessoa sair e entrar, os registros estarão salvos.<issue_closed>
Status: Issue closed |
mockito/mockito | 351773017 | Title: Exception not thrown when mocking final method
Question:
username_0: Take the code snippet below MissingMethodInvocationException expected in both test case. But only the second fails.
Tested with Mockito 2.20 (found with mockito 2.8.47) using JDK 1.8.171, Win10x64.
```java
public class TestingMockito {
@Test
public void testShouldFailButPasses() {
// GIVEN
Foo mockFoo = mock(Foo.class);
Aaaa a = new Aaaa();
// WHEN
a.setSomething(mockFoo);
// THEN - next lines do not throw exception
Foo mockFoo2 = mock(Foo.class);
when(mockFoo2.getFoo()).thenReturn(null);
}
@Test
public void testFailAsExpected() {
Foo mockFoo2 = mock(Foo.class);
when(mockFoo2.getFoo()).thenReturn(null);
}
public static class Aaaa {
public void setSomething(Object something) {
System.out.println("This is " + something);
}
}
public static class Foo {
public final String getFoo() {
return "foo";
}
}
}
```
Answers:
username_1: @username_0
I tested your sample with v2.21.0, both tests turn green with plain mockito and mockito-inline. Why do you expect the when call to fail?
username_0: I expect both tests to fail with "MissingMethodInvocationException". Because the getFoo() method is **final** thus cannot be mocked by Mockito. I wonder what was the result in your environment? Did getFoo() returns <null> or "foo"?
```
org.mockito.exceptions.misusing.MissingMethodInvocationException:
when() requires an argument which has to be 'a method call on a mock'.
For example:
when(mock.getArticles()).thenReturn(articles);
Also, this error might show up because:
1. you stub either of: final/private/equals()/hashCode() methods.
Those methods *cannot* be stubbed/verified.
Mocking methods declared on non-public parent classes is not supported.
2. inside when() you don't call method on mock but on some other object.
at TestingMockito.testFailAsExpected(TestingMockito.java:29)
....
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86)
at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
```
username_1: When mockito-inline-mocker is not used, mockito can't intercept the final method. Therefore it can't detect when it is called.
One the other hand mockito could detect that no mock-method was called, when the `when` is called.
username_2: I have tested this on both Mockito 2.20 and the latest 2.23. I have observed that the object `a` is being intercepted by MockitoInline. Is that expected ? The test is passing only when some method is being called on object `a`. This looks like a bug. |
styled-components/styled-components | 452481751 | Title: Can I get the final className before render
Question:
username_0: ````
const Wrapper = styled.div`
border: 1px solid blue;
&__btn {
color: red;
}
`
class extends React.PureComponent<Props> {
public render() {
const wrapperClsName: string = (Wrapper as any).componentStyle.lastClassName
const cls = bemClsFactor(wrapperClsName)
return(
<Wrapper>
<button className={cls('btn')}>Hello World</button>
</Wrapper>
)
}
}
````
LIke the code before, I want to using styled-components and bem together, I don't want every component inside the Wrapper styled. But the first time render I can't get lastClassName.
is there any other way to do this?
Answers:
username_1: `Wrapper.styledComponentId` will give you the static className of the component, but I don't really recommend using styled-components in this way.
What exactly are you trying to accomplish here? A reproduction in a codesandbox would be helpful.
Status: Issue closed
|
bootstrapworld/curriculum | 1081342594 | Title: Switching back to default textbook alignment shows up empty
Question:
username_0: In lesson defining-values, the default textbook alignment (`IM.Alg.1`) shows `IM.Alg1.4.2 - Function Notation`. However, if you choose a different textbook in the dropdown and go back to `IM.Alg.1`, nothing is shown under it.
Answers:
username_1: @username_0 the value in the `option` elt is "textbook-IM_Alg.1" (with a dot). But the loop that compares the selection to the standards tables is seeing the value "textbook-IM_Alg_1". I couldn't find where this string munging happens, but somewhere along the lines the dot is turning into an underscore.
Status: Issue closed
|
keybase/client | 213959904 | Title: Backup software thinks /keybase is a folder that should be backed up, and then fails when it can't be accessed by root.
Question:
username_0: Keybase GUI Version: 1.0.20-20170309192151+e30557a
When trying to back up my computer with any backup agent that uses administrative privileges to perform backups, they universally report filesystem errors trying to access /keybase/ to back it up.
This is probably because 'root' has no keybase configuration. It may make sense on Mac to handle this specific case more effectively, so that backup clients either back up nothing (which is fine) or ignore the mount altogether (because you tune it to look like a network mount, or something, which is also fine) or some other manner altogether.
Answers:
username_1: Hi @username_0 -- it should already be configured to look like a network mount. You might just have to list if explicitly in your software's exclude list. Not sure what else to do on our end. It's true that `root` does not have permissions to read `/keybase`, only the user you mount it as.
username_0: For 'root', it would be invaluable if /keybase was an otherwise-empty
directory with a README explaining what's going on. That would help people
who do things as root on OS X (sudo or otherwise) understand why keybase
isn't accessible, when eventually they get around to trying that, and would
also give backup software something other than an disk read error.
username_1: Thanks for giving this some thought @username_0, interesting idea. Unfortunately, the way Finder works in macOS means that we can't present `root` with an empty folder. Finder often drops into `root` mode to look up attributes about files -- why it does that, I couldn't tell you. But the file system kernel module we use (osxfuse) has many fancy exceptions to let those Finder-related calls through, but not other types of root access like listing directories, etc.
Do the errors you're seeing actually break the backup software? Or does it just log the error and move onto the next directory? Also, are you willing to name the software so we can test with it? Thanks!
username_0: Ah, that explains why it's all getting confused. That's crazy! I had no
idea! I'm so sorry! :(
I should have noticed you were using FUSE. It means this is all basically a
FUSE problem and not a Keybase problem. They don't break it, it just logs
it and moves on.
Arq Backup is the one I use everyday, and is available to the general
public.
- R. |
supervisorphp/supervisor | 126661344 | Title: Failed to connect to localhost port 9001
Question:
username_0: I can't connect to my supervisor service give me that error every time i trie to get process
An error occurred when fetching the URI "http://localhost:9001/RPC2" with the adapter "guzzle6" ("cURL error 7: Failed to connect to localhost port 9001: Connection refused (see http://curl.haxx.se/libcurl/c/libcurl-errors.html)").
Answers:
username_1: Hey @username_0
What is in you supervisord.conf file?
username_0: ; supervisor config file
[unix_http_server]
file=/var/run/supervisor.sock ; (the path to the socket file)
chmod=0700 ; sockef file mode (default 0700)
[supervisord]
logfile=/var/log/supervisor/supervisord.log ; (main log file;default $CWD/supervisord.log)
pidfile=/var/run/supervisord.pid ; (supervisord pidfile;default supervisord.pid)
childlogdir=/var/log/supervisor ; ('AUTO' child log dir, default $TEMP)
; the below section must remain in the config file for RPC
; (supervisorctl/web interface) to work, additional interfaces may be
; added by defining them in separate rpcinterface: sections
[rpcinterface:supervisor]
supervisor.rpcinterface_factory = supervisor.rpcinterface:make_main_rpcinterface
[supervisorctl]
serverurl=unix:///var/run/supervisor.sock ; use a unix:// URL for a unix socket
; The [include] section can just contain the "files" setting. This
; setting can list multiple files (separated by whitespace or
; newlines). It can also contain wildcards. The filenames are
; interpreted as relative to this file. Included files *cannot*
; include files themselves.
[include]
files = /etc/supervisor/conf.d/*.conf
username_1: You didn't configure an inet section, only the unix socket one. Either use that one (not compatible with guzzle6) or configure inet:
http://supervisord.org/configuration.html#inet-http-server-section-settings
username_0: Thanks, for unix socket wich client i need to use ?
username_1: Assuming you are using Ivory Http Adapter, you can try this client:
username_0: Thanks i already configure the inet and now it works
Status: Issue closed
|
swimlane/ngx-charts | 529882448 | Title: ngcc error with Angular 9.0.0-rc.4
Question:
username_0: When upgraded angular from 9.0.0-rc.3 to 9.0.0-rc.4, ngcc fails with following error:
```
$ ngcc --properties es2015 browser module main --first-only --create-ivy-entry-point
Compiling @swimlane/ngx-charts : module as esm5
Error: Error on worker #5: Error: Failed to compile entry-point @swimlane/ngx-charts due to compilation errors:
node_modules/@angular/common/common.d.ts(115,22): error TS-996002: Appears in the NgModule.imports of TooltipModule, but could not be resolved to an NgModule class
node_modules/@angular/common/common.d.ts(115,22): error TS-996002: Appears in the NgModule.imports of AxesModule, but could not be resolved to an NgModule class
node_modules/@angular/common/common.d.ts(115,22): error TS-996002: Appears in the NgModule.imports of ChartCommonModule, but could not be resolved to an NgModule class
node_modules/@swimlane/ngx-charts/release/esm.js(2832,5): error TS-996002: Appears in the NgModule.imports of ChartCommonModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(1062,5): error TS-996002: Appears in the NgModule.imports of ChartCommonModule, but itself has errors
node_modules/@angular/common/common.d.ts(115,22): error TS-996003: Appears in the NgModule.exports of ChartCommonModule, but could not be resolved to an NgModule, Component, Directive, or Pipe class
node_modules/@swimlane/ngx-charts/release/esm.js(2832,5): error TS-996003: Appears in the NgModule.exports of ChartCommonModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(1062,5): error TS-996003: Appears in the NgModule.exports of ChartCommonModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of AreaChartModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of BarChartModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of BubbleChartModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of ForceDirectedGraphModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of HeatMapModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of LineChartModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of PieChartModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of PolarChartModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(13531,5): error TS-996002: Appears in the NgModule.imports of PolarChartModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(11717,5): error TS-996002: Appears in the NgModule.imports of PolarChartModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of NumberCardModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of TreeMapModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of GaugeModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(13531,5): error TS-996002: Appears in the NgModule.imports of GaugeModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(9549,5): error TS-996002: Appears in the NgModule.imports of GaugeModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(5816,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(9549,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(10196,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(10412,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(11000,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(11717,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(13560,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(13974,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(13531,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(14329,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors
node_modules/@swimlane/ngx-charts/release/esm.js(15029,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors
at ClusterWorker.compile (/charts-project/node_modules/@angular/compiler-cli/ngcc/src/main.js:170:27)
at Worker.<anonymous> (/charts-project/node_modules/@angular/compiler-cli/ngcc/src/execution/cluster/worker.js:41:42)
at Worker.emit (events.js:210:5)
at process.<anonymous> (internal/cluster/worker.js:34:12)
at process.emit (events.js:210:5)
at emit (internal/child_process.js:878:12)
at processTicksAndRejections (internal/process/task_queues.js:81:21)
at ClusterMaster.onWorkerMessage (/charts-project/node_modules/@angular/compiler-cli/ngcc/src/execution/cluster/master.js:158:27)
at /charts-project/node_modules/@angular/compiler-cli/ngcc/src/execution/cluster/master.js:46:95
at ClusterMaster.<anonymous> (/charts-project/node_modules/@angular/compiler-cli/ngcc/src/execution/cluster/master.js:238:57)
at step (/charts-project/node_modules/tslib/tslib.js:136:27)
at Object.next (/charts-project/node_modules/tslib/tslib.js:117:57)
at /charts-project/node_modules/tslib/tslib.js:110:75
at new Promise (<anonymous>)
at Object.__awaiter (/charts-project/node_modules/tslib/tslib.js:106:16)
at EventEmitter.<anonymous> (/charts-project/node_modules/@angular/compiler-cli/ngcc/src/execution/cluster/master.js:232:32)
at EventEmitter.emit (events.js:210:5)
```
Status: Issue closed
Answers:
username_1: Hi there, please refer to this comment: https://github.com/swimlane/ngx-charts/issues/1294#issuecomment-560380598
Closing this one. |
adjust/hashtypes | 484049958 | Title: Add binary I/O to the SHA types.
Question:
username_0: I made an [implementation of binary I/O for the various SHA types](https://github.com/username_0/hashtypes/tree/binary).
For now, I'm not making it a pull request because my work depends on the Windows port, [which is already a pull request](https://github.com/adjust/hashtypes/pull/17).
I bumped the version to 0.1.6 and made an upgrade script based on help from the PostgreSQL General mailing list.
I did not touch the `.in` files, and copied `hashtypes--0.1.5.sql` directly. I did not touch `sha.sql.type` because I don't know how it's used to generate the install script. |
Alexander-Barth/NCDatasets.jl | 407476515 | Title: ncgen does not preserve unlimited dimensions
Question:
username_0: `ncgen` does not preserve unlimited dimensions in the ncdf files it produces. And I could not find a function to set a dimension to unlimited. Is there one?
Answers:
username_1: It should work as it is part of the test-suite:
https://github.com/username_1/NCDatasets.jl/blob/master/test/test_ncgen.jl
The resulting file is:
```
netcdf juliauJUfOc {
dimensions:
lon = 3 ;
unlimited = UNLIMITED ; // (0 currently)
variables:
float lon(lon) ;
lon:string = "degrees_east" ;
lon:float32 = 1.f ;
lon:float64 = 1. ;
lon:float32_vector = 1.f, 2.f, 3.f ;
lon:float64_vector = 1., 2., 3. ;
lon:int32_vector = 1, 2, 3 ;
// global attributes:
:dollar = "a dollar $ stop" ;
:backslash = "a backslash \\ stop" ;
:doublequote = "a doublequote \" stop" ;
}
```
Can you provide a minimal netcdf file and julia script that show the error that you are seeing?
username_0: Here is an example using the file used in the documentation:
```> julia
_
_ _ _(_)_ | Documentation: https://docs.julialang.org
(_) | (_) (_) |
_ _ _| |_ __ _ | Type "?" for help, "]?" for Pkg help.
| | | | | | |/ _` | |
| | |_| | | | (_| | | Version 1.0.3 (2018-12-18)
_/ |\__'_|_|_|\__'_| | Official https://julialang.org/ release
|__/ |
julia> using NCDatasets
julia> ncgen("sresa1b_ncar_ccsm3-example.nc", "ncgen-sresa1b_ncar_ccsm3-example.jl")
julia> include("ncgen-sresa1b_ncar_ccsm3-example.jl")
julia> run(pipeline(`ncdump -h sresa1b_ncar_ccsm3-example.nc`, `head`));
netcdf sresa1b_ncar_ccsm3-example {
dimensions:
lat = 128 ;
lon = 256 ;
bnds = 2 ;
plev = 17 ;
time = UNLIMITED ; // (1 currently)
variables:
float area(lat, lon) ;
area:long_name = "Surface area" ;
julia> run(pipeline(`ncdump -h filename.nc`, `head`));
netcdf filename {
dimensions:
lat = 128 ;
lon = 256 ;
bnds = 2 ;
plev = 17 ;
time = 1 ;
variables:
float area(lat, lon) ;
area:long_name = "Surface area" ;
julia>
```
Note that the time dimension is unlimited in the original file but not in the `ncgen` file.
username_1: Thanks a lot, I could reproduce the issue and commit a fix to the match branch. Does it work for you?
username_0: Yes, that works. Thanks!
username_1: Thanks for confirming. I just made a new release with the fix.
Status: Issue closed
|
aciidb0mb3r/SPM-Cpp-Fixture | 187633730 | Title: Linker error
Question:
username_0: Tried on macOS with Xcode 8.0, 8.1 as well as Swift 3.0 on Linux and all get the following linker errors with this project:
```
$ swift build
Compile cpplib cpplib.cpp
Linking cpplib
Compile cwrapper cwapper.cpp
Compile cpp-exec main.cpp
Linking cwrapper
Compile Swift Module 'swift_exec' (1 sources)
Linking ./.build/debug/swift-exec
Linking cpp-exec
Undefined symbols for architecture x86_64:
"std::__1::basic_ostream<char, std::__1::char_traits<char> >::operator<<(int)", referenced from:
_main in main.cpp.o
"std::__1::cout", referenced from:
_main in main.cpp.o
ld: symbol(s) not found for architecture x86_64
clang: error: linker command failed with exit code 1 (use -v to see invocation)
<unknown>:0: error: build had 1 command failures
```
Could this be a regression in SwiftPM since this project was originally created?
Thanks for making this!
Answers:
username_1: Looks like it, can you report on bugs.swift.org
Thanks!
username_0: Thanks for confirming my suspicions! Filed [SR-3152](https://bugs.swift.org/browse/SR-3152) |
ContinuumIO/anaconda-issues | 267876473 | Title: Navigator Error
Question:
username_0: ## Main error
An unexpected error occurred on Navigator start-up<br>psutil.AccessDenied (pid=5840)
## Traceback
```
Traceback (most recent call last):
File "D:\Anaconda2\lib\site-packages\anaconda_navigator\exceptions.py", line 75, in exception_handler
return_value = func(*args, **kwargs)
File "D:\Anaconda2\lib\site-packages\anaconda_navigator\app\start.py", line 108, in start_app
if misc.load_pid() is None: # A stale lock might be around
File "D:\Anaconda2\lib\site-packages\anaconda_navigator\utils\misc.py", line 384, in load_pid
cmds = process.cmdline()
File "D:\Anaconda2\lib\site-packages\psutil\__init__.py", line 701, in cmdline
return self._proc.cmdline()
File "D:\Anaconda2\lib\site-packages\psutil\_pswindows.py", line 623, in wrapper
raise AccessDenied(self.pid, self._name)
AccessDenied: psutil.AccessDenied (pid=5840)
```
## System information
```
platform: win-64
version: 1.6.2
conda: 4.3.21
qt: 5.6.2
language: zh_CN
python: 2.7.13
os: Windows;7;6.1.7601;AMD64;Intel64 Family 6 Model 94 Stepping 3, GenuineIntel
pyqt: 5.6.0
```
Status: Issue closed
Answers:
username_1: **See Issue #1984 for more information on how to fix this.**
---
Closing as duplicate of #1984
---
Please remember to update to the latest version of Navigator to include
the latest fixes.
Open a terminal (on Linux or Mac) or the Anaconda Command Prompt (on windows)
and type:
```
$ conda update anaconda-navigator
$ conda update navigator-updater
``` |
MicrosoftDocs/azure-docs | 394329947 | Title: MQTT-over-WebSockets needs to explain how to set up the WS connection
Question:
username_0: Just like the document explains how to set up a raw MQTT connection when not using one of the Azure SDKs, it also needs to explain how to set up the MQTT-over-WebSockets connection when not using the SDKs. Specifically these two pieces of information are missing:
1. The path of the URL of the request must be `/$iothub/websocket`.
1. The request must contain the `sec-websocket-protocol` header with value `mqtt`
Thus an example handshake request would look like:
```
GET /$iothub/websocket HTTP/1.1
Host: foo.azure-devices.net
<other standard handshake headers>
sec-websocket-protocol: mqtt
```
Not doing both of these things will cause the server to reject the connection with HTTP 400.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: bcf2a86a-1de0-2786-db87-58279ceffb22
* Version Independent ID: 7e43624f-a87d-1c4d-94e6-e38c1d42686f
* Content: [Understand Azure IoT Hub MQTT support](https://docs.microsoft.com/en-us/azure/iot-hub/iot-hub-mqtt-support#retrieving-a-device-twins-properties)
* Content Source: [articles/iot-hub/iot-hub-mqtt-support.md](https://github.com/Microsoft/azure-docs/blob/master/articles/iot-hub/iot-hub-mqtt-support.md)
* Service: **iot-hub**
* GitHub Login: @username_2
* Microsoft Alias: **rezas**
Answers:
username_1: Thanks for the feedback! I have assigned the issue to the content author to evaluate and update as appropriate.
username_2: @username_0, sure, I'll add this in future revisions of this doc. In the meantime, if you have working samples to share feel free to send me here or via email (rezas).
username_1: We will now proceed to close this thread. If there are further questions regarding this matter, please tag me in your reply. We will gladly continue the discussion and we will reopen the issue.
Status: Issue closed
username_1: Just like the document explains how to set up a raw MQTT connection when not using one of the Azure SDKs, it also needs to explain how to set up the MQTT-over-WebSockets connection when not using the SDKs. Specifically these two pieces of information are missing:
1. The path of the URL of the request must be `/$iothub/websocket`.
1. The request must contain the `sec-websocket-protocol` header with value `mqtt`
Thus an example handshake request would look like:
```
GET /$iothub/websocket HTTP/1.1
Host: foo.azure-devices.net
<other standard handshake headers>
sec-websocket-protocol: mqtt
```
Not doing both of these things will cause the server to reject the connection with HTTP 400.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: bcf2a86a-1de0-2786-db87-58279ceffb22
* Version Independent ID: 7e43624f-a87d-1c4d-94e6-e38c1d42686f
* Content: [Understand Azure IoT Hub MQTT support](https://docs.microsoft.com/en-us/azure/iot-hub/iot-hub-mqtt-support#retrieving-a-device-twins-properties)
* Content Source: [articles/iot-hub/iot-hub-mqtt-support.md](https://github.com/Microsoft/azure-docs/blob/master/articles/iot-hub/iot-hub-mqtt-support.md)
* Service: **iot-hub**
* GitHub Login: @username_2
* Microsoft Alias: **rezas**
Status: Issue closed
username_1: We will now proceed to close this thread. If there are further questions regarding this matter, please tag me in your reply. We will gladly continue the discussion and we will reopen the issue.
username_3: In case someone stumbles on this - you need to use `/$iothub/websocket?iothub-no-client-cert=true` - otherwise the connection gets dropped. |
cake-tech/cake_wallet | 1090232099 | Title: [Bug] All nodes fail tests and show as down
Question:
username_0: used Cake Wallet v4.27 for android
All nodes fail tests and show as down(red), including cake's own node. For those nodes that require authentication, even if I fill in the correct username and password, the test still fails.
More than one person has reported this bug, hope to fix it in next version.
In addition, please consider updating the monero core to 0.17.3.0
Answers:
username_1: Thanks for your feedback. Unfortunately I'm not sure that we will fix it in next version, but I will put it in our list. For now we are focused on monero only app by cake wallet.
username_2: Main issue is dup of #204
Unsure about 0.17.3.0
Status: Issue closed
|
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.