repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
Monty15/Capstone
284806249
Title: Viewport meta tag Question: username_0: https://github.com/Monty15/Capstone/blob/master/index.html#L4-L9 Remember to include the `viewport` meta tag to properly control your website's viewport width. ```html <meta name="viewport" content="width=device-width, initial-scale=1.0"> ``` 💡 This is particularly important for any mobile responsive website.
vladbalmos/mitzasql
741618854
Title: Database names with dashes cause error 1064 (42000) Question: username_0: **Describe the bug** When selecting a database with a dash "-" in the name the following error is displayed. 1064 (42000): You have an error in your SQL syntax; check the manual that corresponds to your MySQL server version for the right syntax to use near '-test' at line 1 **To Reproduce** Steps to reproduce the behavior: 1. Create a database with name "wordpress-test" 2. Open Mitzasql and navigate to this database 3. When trying to enter the db you can see the error. 4. See error **Expected behavior** Tables inside the database would be displayed. **Screenshots** https://i.imgur.com/LsCfzAQ.png **Desktop (please complete the following information):** - OS: Debian Buster - Python version 3.7.3 Answers: username_1: Good find, thanks! Status: Issue closed username_0: Works like a charm. Thanks!
utPLSQL/utPLSQL
446432444
Title: Add JSON object comparision Question: username_0: Since JSON is gaining a popularity and since Oracle 12.2 we got a proper support of JSON in database I thought it would b a good idea to implement a JSON comparison matcher. I've did some initial work for cursor equality and will be adding a more methods.<issue_closed> Status: Issue closed
zalando/patroni
778628705
Title: Available implementations: """ + ', '.join(sorted(set(available_implementations)))) patroni.exceptions.PatroniFatalException: 'Can not find suitable configuration of distributed configuration store\nAvailable implementations: kubernetes, raft' Question: username_0: Hi ,i get same error like this: 2021-01-05 13:15:11,134 INFO: Failed to import patroni.dcs.consul 2021-01-05 13:15:11,139 INFO: Failed to import patroni.dcs.etcd 2021-01-05 13:15:11,142 INFO: Failed to import patroni.dcs.etcd3 2021-01-05 13:15:11,149 INFO: Failed to import patroni.dcs.exhibitor 2021-01-05 13:15:11,162 INFO: Failed to import patroni.dcs.zookeeper Traceback (most recent call last): File "patroni.py", line 6, in main() File "/home/kingbase/patroni/patroni/init.py", line 170, in main return patroni_main() File "/home/kingbase/patroni/patroni/init.py", line 138, in patroni_main abstract_main(Patroni, schema) File "/home/kingbase/patroni/patroni/daemon.py", line 98, in abstract_main controller = cls(config) File "/home/kingbase/patroni/patroni/init.py", line 29, in init self.dcs = get_dcs(self.config) File "/home/kingbase/patroni/patroni/dcs/init.py", line 107, in get_dcs Available implementations: """ + ', '.join(sorted(set(available_implementations)))) patroni.exceptions.PatroniFatalException: 'Can not find suitable configuration of distributed configuration store\nAvailable implementations: kubernetes, raft' My enviroment like this: Python 3.6 pip3 20.3.3 pip3 list Package Version boto 2.49.0 certifi 2020.12.5 chardet 4.0.0 click 7.1.2 dnspython 2.0.0 flake8 3.8.4 idna 2.10 importlib-metadata 3.3.0 kazoo 2.8.0 mccabe 0.6.1 patroni 2.5 pip 20.3.3 prettytable 2.0.0 psutil 5.8.0 psycopg2 2.5.4 pycodestyle 2.6.0 pyflakes 2.2.0 pysyncobj 0.3.7 python-consul 1.1.0 python-dateutil 2.8.1 python-etcd 0.4.5 PyYAML 5.3.1 requests 2.25.1 setuptools 28.8.0 six 1.15.0 typing-extensions 3.7.4.3 urllib3 1.26.2 wcwidth 0.2.5 wheel 0.36.2 ydiff 1.2 [Truncated] #timezone: 'PRC' #lc_messages: 'C' #lc_monetary: 'C' #lc_numeric: 'C' #lc_time: 'C' #full_page_writes: on #synchronous_commit: on #wal_log_hints: on #synchronous_standby_names: '' #max_replication_slots: 10 create_replica_methods: - basebackup basebackup: max-rate: '100M' tags: nofailover: false noloadbalance: false clonefrom: false nosync: false Answers: username_1: **python-etcd** module in order to use Etcd as DCS username_0: May i run it by source code? If i user python2.7 ,this is ok! username_1: It doesn't matter how you run it, `python-etcd` module must be installed if you want to run Patroni with etcd. username_0: python-etcd 0.4.5 in the pip3 list . I am sure i install it . username_1: Magic doesn't exists. username_0: python-etcd 0.4.5 in the pip3 list . I am sure i install it. It in the list: Package Version boto 2.49.0 certifi 2020.12.5 chardet 4.0.0 click 7.1.2 dnspython 2.0.0 flake8 3.8.4 idna 2.10 importlib-metadata 3.3.0 kazoo 2.8.0 mccabe 0.6.1 patroni 2.5 pip 20.3.3 prettytable 2.0.0 psutil 5.8.0 psycopg2 2.5.4 pycodestyle 2.6.0 pyflakes 2.2.0 pysyncobj 0.3.7 python-consul 1.1.0 python-dateutil 2.8.1 **### python-etcd 0.4.5** PyYAML 5.3.1 requests 2.25.1 setuptools 28.8.0 six 1.15.0 typing-extensions 3.7.4.3 urllib3 1.26.2 wcwidth 0.2.5 wheel 0.36.2 ydiff 1.2 zipp 3.4.0 username_1: I totally don't undestand what you are doing. The only purpose of release.sh is publishing Patroni on PyPI. username_0: I need change some code for our database ,the database base on postgresql, so i need build the package to publish some users. username_1: Perfect, then you should be qualified enough to figure out what is wrong with _your system_ and why python can't find/load certain modules. The recommended way of installing Patroni either from PyPI (pip install patroni[$EXTRAS]) or via PGDG packages (apt-get/yum install). We can't be responsible for something that you do on your own. Status: Issue closed
chargebee/chargebee-dotnet
422252320
Title: Upgrade from 2.6.7 to 2.6.9 -> Timestamp errors Question: username_0: Hi, I've updated from version 2.6.7 -> 2.6.9. Everything compiled fine and everything seemed fined after a few tests. Then I received a webhook from ChargeBee and everything went down... When creating a new webhook event from an io stream, I get this error: `OccurredAt = 'WebhookEvent.OccurredAt' threw an exception of type 'System.TypeInitializationException'` This is how I create the webhookEvent object: ` Public Sub ProcessRequest(context As HttpContext) Implements IHttpHandler.ProcessRequest Dim Response As String = Nothing Dim ResponseCode As HttpStatusCode = HttpStatusCode.OK Try ' Input Using StreamReader As New IO.StreamReader(context.Request.InputStream) InputData = StreamReader.ReadToEnd() End Using If String.IsNullOrWhiteSpace(InputData) Then Throw New ArgumentNullException("The data cannot be null.") ' Create the webhook event WebhookEvent = New ChargeBee.Models.Event(InputData) [...] Catch ex as Exception End Try [...] End Sub ` I checked and the timestamp is fine in the IO stream (1552915864). All the modules are up to date, everything else is working fine, I just can't make the webhook object creation works. If I use 2.6.7, this works flawlessly, in fact it's been working flawlessly all the way back with version 1.x.x. Please fix for the next release Answers: username_1: Stumbled upon same issue. Using v2.8.0 and it still fails for any date. username_2: Still broken in 2.8.5 username_3: Still broken in 2.8.7 username_4: **Update** I've developed a small fix for the conversion between Unix TimeStamp and UTC DateTime. Couldn't find how provide a solution to creators, so I pushed the fix to my personal space. https://github.com/username_4/chargebee-dotnet/tree/bugfix/patch-timestamp-datetime-conversion Hope it helps username_4: **Update** I've managed to fix this and now on my way to fork and provide the solution I've found to work. username_4: ### **[SOLVED]** **TL;DR;** Be sure to call **ChargeBee.Api.ApiConfig.Configure** before accessing DateTime or Timestamp fields to avoid unreasonable Exceptions, even when unit testing or offline. **Summary:** Accessing DateTime and Timestamp fields fail because the timeout for an HttpClientAdapter can't be zero. Adding one sentence to @username_3 [comment ](https://github.com/chargebee/chargebee-dotnet/issues/18#issuecomment-737283847) makes it work. ``` ChargeBee.Api.ApiConfig.Configure("fake-sitename", "fake-apikey"); var cbEvent = new Event(jsonProvidedAsString); var termEnd = cbEvent.Content.Subscription.CurrentTermEnd; ``` ChargeBee Api needs to be initialised before accessing its model. Otherwise it will throw exception when creating **HttpClient** on **ApiUtil** constructor. The same class contains **ConvertFromTimestamp** and **ConverttoTimestamp** methods triggering unexpected errors https://github.com/chargebee/chargebee-dotnet/blob/1d6934b645c31e9b3e17cc0fbbbf34d582334f0c/ChargeBee/Api/ApiUtil.cs#L21 ApiConfig.ConnectionTimeout is zero until `ChargeBee.Api.ApiConfig(<string>,<string>)` is run, being first assigned here: https://github.com/chargebee/chargebee-dotnet/blob/1d6934b645c31e9b3e17cc0fbbbf34d582334f0c/ChargeBee/Api/ApiConfig.cs#L45 Thereby, if **ApiConfig.Configure** is never invoked it would always raise an exception because the timeout for an HttpClientAdapter can't be zero. Thank you @username_3 on how to reproduce. and @username_0, @username_1, @username_2 for also reporting username_4: #27 addresses this. username_5: This has been addressed in v2.10.0 Status: Issue closed
bazelbuild/bazel
166594096
Title: Easy to mistakenly install bazel without specifying "--user" Question: username_0: When installing Bazel for the first time, I made the fairly obvious mistake of omitting the "--user" flag. The error I got back suggested I should use "sudo," which I did. Since I don't seem to be the only person to make this mistake (cf the title of #962), perhaps if the installer is about to error out due to insufficient permissions, it would be better to say something like: `If you are just installing this for yourself, use --user, but if you would like it to be accessible to everyone on the machine, ensure you have access to...` Answers: username_1: The error I get does not mention sudo, it says: ``` The Bazel installer must have write access to /usr/local/bin! Usage: bazel-0.3.0-jdk7-installer-linux-x86_64.sh [options] ``` and then prints the usage string. What version of the installer are you using? username_0: My mistake, I simply interpreted the lack of permissions as "I need to use sudo." I then avoided looking a few lines down to see the help text that referenced "--user". Again, it's a bit of an embarrassing mistake, I only filed the issue because I found someone else who made the same mistake. So I'm just suggesting a slight change to the message. Status: Issue closed
xmake-io/xmake
796521346
Title: Rebuild local Qt-dependent package with project Qt version? Question: username_0: I have Qt installed on Ubuntu via apt, as well as an older version installed via their web installer. I also have a Qt-dependent package in a local xrepo repo. We can set the version of Qt to build the project with: ``` xmake f --qt=~/Qt/5.11.3 ``` But the package is unaware of the change so continues to use the system-installed Qt. This causes version mismatch errors. Is there or could there be a solution to this? Some way of using the currently set `qmake` in the package? Answers: username_1: What build system does this package use? xmake? cmake or autoconf? username_0: It just uses `qmake` and `make`, here's what it looks like at the moment: ```lua package("qscintilla") set_homepage("https://www.riverbankcomputing.com/software/qscintilla/intro") set_description("QScintilla is a port to Qt of <NAME>'s Scintilla C++ editor control.") set_urls("https://www.riverbankcomputing.com/static/Downloads/QScintilla/$(version)/QScintilla-$(version).zip") add_versions("2.11.6", "ddd0945d90bbf9394e0d4a41cfeb5bd7c1a6b918c827aa90d4396ea3da0be9a9") on_install("linux", function(package) os.execv("qmake", { }, { curdir = "./Qt4Qt5"}) os.execv("make", { "-j8" }, { curdir = "./Qt4Qt5" }) os.cp("./Qt4Qt5/libqscintilla2_qt5.so.15.0.0", path.join(package:installdir("lib"), "libqscintilla2_qt5.so")) os.cp("./Qt4Qt5/libqscintilla2_qt5.so.15.0.0", path.join(package:installdir("lib"), "libqscintilla2_qt5.so.15.0.0")) os.cp("./Qt4Qt5/libqscintilla2_qt5.so.15.0.0", path.join(package:installdir("lib"), "libqscintilla2_qt5.so.15")) os.cp("./Qt4Qt5/Qsci/*.h", package:installdir("include/Qsci")) end) ``` username_1: ```lua on_install("linux", function(package) local qtdir = get_config("qt") end) ``` username_0: Thanks, that resolves the correct path, but `os.execv` doesn't like it: ```lua local qtdir = get_config("qt") os.execv(qtdir .. "/gcc_64/bin/qmake", { }, { curdir = "./Qt4Qt5"}) ``` `xmake f -cvD --qt=~/Qt/5.11.3` error: ```bash error: @programdir/core/sandbox/modules/os.lua:387: execv(~/Qt/5.11.3/gcc_64/bin/qmake ) failed(255) stack traceback: [C]: in function 'error' [@programdir/core/base/os.lua:787]: in function 'raise' [@programdir/core/sandbox/modules/os.lua:387]: in function 'execv' [./xmake-pkgs/packages/q/qscintilla/xmake.lua:12]: in function 'script' [...gramdir/actions/require/impl/actions/../utils/filter.lua:125]: in function 'call' [@programdir/actions/require/impl/actions/install.lua:168]: [C]: in function 'trycall' [@programdir/core/sandbox/modules/try.lua:121]: in function 'try' [@programdir/actions/require/impl/actions/install.lua:127]: in function 'action_install' [@programdir/actions/require/impl/package.lua:831]: in function 'jobfunc' [@programdir/modules/private/async/runjobs.lua:193]: [C]: in function 'trycall' [@programdir/core/sandbox/modules/try.lua:121]: in function 'try' [@programdir/modules/private/async/runjobs.lua:186]: in function 'cotask' [@programdir/core/base/scheduler.lua:317]: => install qscintilla 2.11.6 .. failed error: @programdir/modules/private/async/runjobs.lua:217: @programdir/actions/require/impl/actions/install.lua:256: install failed! stack traceback: [C]: in function 'error' [@programdir/core/base/os.lua:787]: in function 'raise' [@programdir/actions/require/impl/actions/install.lua:256]: in function 'catch' [@programdir/core/sandbox/modules/try.lua:127]: in function 'try' [@programdir/actions/require/impl/actions/install.lua:127]: in function 'action_install' [@programdir/actions/require/impl/package.lua:831]: in function 'jobfunc' [@programdir/modules/private/async/runjobs.lua:193]: [C]: in function 'trycall' [@programdir/core/sandbox/modules/try.lua:121]: in function 'try' [@programdir/modules/private/async/runjobs.lua:186]: in function 'cotask' [@programdir/core/base/scheduler.lua:317]: stack traceback: [C]: in function 'error' @programdir/core/base/os.lua:787: in function 'raise' @programdir/modules/private/async/runjobs.lua:217: in function 'catch' @programdir/core/sandbox/modules/try.lua:127: in function 'try' @programdir/modules/private/async/runjobs.lua:186: in function 'cotask' @programdir/core/base/scheduler.lua:317: in function <@programdir/core/base/scheduler.lua:315> ``` username_1: `~/xx`, you need use path.translate or path.absolute for qmake path. username_0: Yay, that's it - thanks! Status: Issue closed
nornir-automation/nornir
336520004
Title: Closing NAPALM connections automatically Question: username_0: When using NAPALM plugins connections stay open and can cause scripts to hang while waiting for the connections to timeout. Forcing the connections to close using a task helps with this. ```python task.host.connections["napalm"].close() ``` Maybe this could be done automatically? It'll require to find a way to know which connections are not useful anymore. Answers: username_1: How would we know that you didn't want to use the connection any more (i.e. in a subsequent task)? The process is probably to have the code writer to explicitly close the connection in some way. username_0: Yep that's complicated. I opened the issue because @username_2 wanted to track it. I don't see any simple way to do this. Maybe proposing a task closing all NAPALM connections could be an answer to this. username_2: I thought we could look into creating a “close” task so it’s easier for the user to explicitly close the connections or maybe do it automatically via the garbage collector by implementing some code in the __del__ magic method. username_3: I think the close task, or perhaps something like a context manager for Nornir could make sense. There might be scenarios where users don't want the connections to be terminated. (I've seen Huawei switches where a local script running on them stops running if the ssh connection dies). We could have some task which just iterates all of the hosts' connection dictionaries and calls a disconnect function. The main issue is perhaps that all of the connection plugins be consistent and support the same close method. username_2: For reference, some work that will enable this is being done in #189 username_4: @username_2 I think this is fixed with #195 and could be closed?
freedomofpress/securedrop-ux
408375462
Title: Online/Offline UI Differences Question: username_0: ### Problem When a journalist views the Client in offline mode, what does that look like? ### Considerations - How will parts of the UI requiring network connectivity be shown to communicate their non-functionality in Offline Mode? - Will something outright go away, or be greyed-out? - Where might there be suggestions or other affordances made to nudge a user to Sign In? - When accessing the client for the first time in Offline mode, what might that experience look like? ### Acceptance Criteria - Clickable wireframe demonstrating interactivity - All state changes clearly itemized/outlined - Spec'd wireframe, or multiple comments here? - Get on the needs testing punchlist --- This is a sub-task within #18 and #31 and #17 Answers: username_0: First Rev (submitted on Gitter, 07 Feb): CLICK ON TOP BAR to toggle back and forth between offline and online. Delete functionality for a Source Account also goes away, as does the drop-down menu on the Messages pane. https://invis.io/X4QEP5Y9EAT#/345671069_0_New_-_DR_On-_OFFLINE username_0: ### Feedback (cut-and-pasted from Gitter): <NAME> @eloquence Feb 07 17:12 I'm a fan! I like avoiding the standard "grayed out UI controls" pattern which can be very frustrating for users, and this also takes up less space <NAME> Alter @username_0 Feb 07 17:15 Oh—fwiw, clarification with the above—in the wireframe, the paper airplane dealie is the "Send" button. I believe the existing Client has a proper button with the word "Send" on it @eloquence Yay! Ya... the "omg, why won't this friggin' work?!" frustration quite sux. <NAME> @username_1 Feb 07 17:20 hey @username_0 this looks awesome <NAME> @username_0 Feb 07 17:24 Coolio! <NAME> @username_1 Feb 07 17:24 trying to find the image of what will be on the ... menu <NAME> @username_0 Feb 07 17:33 @username_1 You mean the content in that menu? I think @eloquence will eventually do that menu as its own Issue/Story "Export All Messages & Files" and "Delete Source" are the only two items therein, for now. ^ ...or probz "Export Source." We never did much testing on that menu. username_0: (crap, sorry that @'d everyone!) username_1: @username_0 - Hey just saw this and noticed that the conversation in gitter was left open-ended. I noticed while toggling back and forth between offline and online mode that you get a `. . .` options menu next to the date when online, so I was wondering if you had created any wireframes for this menu. I assume it'll include a way to logout? Is there anything else? username_0: @username_1 Nope, logging-out is handled from the user avatar having a dropdown-ish type thing. The dot-dot-dot menu you cite currently only has "Export All Submissions" and "Delete Account" on it, as of now. Older wireframes should show more functionality on it; I also don't really know what users may want in it, so could easily pack it with a dozen account-level options like "Export all files, delete all local files" etc... but wanna keep it to just those two account-level things, for now. Anything pertaining to the Journalist's authentication or connectivity happens in the top bar. username_0: I'd forgotten this ticket still existed! **Explorations**, here: [» Invision Mox (semi-interactive) «](https://invis.io/BSRX8TUN8K3) **Final direction**, in Zeplin (ignore "Empty" pane content): [Public Screen](https://scene.zeplin.io/project/5c807ea562f734bd2756b243/screen/5cd35ec0df6a8967aedc0d95) Status: Issue closed username_0: Closing, cuz this is done? @eloquence feel free to reopen if you disagree. Wishing this had gone into a review, but there is only so much team bandwidth and time... :)
DoctorVanGogh/ExtendedStorage
247234384
Title: Add toggle for user/real storage settings Question: username_0: implemented in [branch](/username_0/ExtendedStorage/tree/feature/settings-debug) ![es-user-store](https://user-images.githubusercontent.com/324067/28962695-c8b03cdc-7906-11e7-9e74-cbd85d2ff884.jpg) Status: Issue closed Answers: username_0: implemented in [branch](/username_0/ExtendedStorage/tree/feature/settings-debug) ![es-user-store](https://user-images.githubusercontent.com/324067/28962695-c8b03cdc-7906-11e7-9e74-cbd85d2ff884.jpg) Status: Issue closed
moby/moby
758436900
Title: IPv6 options refinement: --ipv6 behaving better out of the box (concerns both options and default behavior regarding IPv6 in custom networks) Question: username_0: I am just an outsider and not really a code contributor to docker/moby myself, so I just hope to get the discussion going. Basically I'm saying, just take my suggestions as a vague input but I shouldn't be the person to make any of the final calls, in the end I'm just an end-user as many others. Hopefully, this ticket will help the right people to come to the right conclusions! Answers: username_1: I have two related suggestions: - In addition to setting `--fixed-cidr-ipv6` to a private subnet when using ipv6 NAT, I suspect that `default-address-pools` should default to having a private subnet available too (maybe even *without* using ipv6 NAT), so you can create a custom network with `--ipv6` and have it assigned an address without having to manually manage this, just like with ipv4. - It might be useful if some option would be added (maybe even enabled by default) to get ipv6 enabled on user-created networks by default. It seems that you need to manually specify `--ipv6` to `network create` right now, it would be good if this could be made the default (distribution default or system default). username_0: I like those ideas! I agree in particular with it being a good suggestion that the IPv6 being enabled by default with a new option to override that when it's enabled for the daemon. I think the point of having this finally pretty "auto-magical" IPv6 NAT that works like almost IPv4 should be to have everyone do less manual tweaks to get it to "just work" on existing setups, including those with custom networks, so that seems like a really good idea to me. username_0: Any updates? I'm a bit concerned this might take another five years to get looked at, at which point people will probably be stung badly if these options are changed. Now seems like the best time if any to still integrate this. username_2: Just to add to the quorum. Having `--ip6tables` as a non-experimental flag would certainly help me sell the feature to my company. We are currently using `docker-ipv6nat` https://github.com/robbertkl/docker-ipv6nat/issues/65. To add something to this discussion (I don't know if this should be a separate ticket or not), docker-compose default created network should also be IPv6 enabled when `--ipv6` and `--ip6tables` are set. As it stands the difference of behaviour between docker and docker-compose is not very intuitive. A docker container using the default network is IPv6 enabled, but a container started by docker-compose isn't. username_3: I was recently wondering about this, myself. I'd like to do IPv6 NAT on my Docker swarms, without hacking [robertkl/docker-ipv6nat to work](https://github.com/robbertkl/docker-ipv6nat/issues/12) -- if that hack still works at all
juju/python-libjuju
1118809392
Title: TypeError: unsupported format string passed to NoneType.__format__ Question: username_0: Function "_print_status_apps" in file juju/status.py, below piece of code should check if app.workload_version or app.charm_channel is None. https://github.com/juju/python-libjuju/blob/cbeae7063f8a19dc1e1ed3638f8025735f6d9bc0/juju/status.py#L116 If a charm is deployed from local rather than from charmstore, workload_version and app.charm_channel may be None Traceback (most recent call last): File "./lib/test_lib_juju.py", line 26, in <module> main() File "./lib/test_lib_juju.py", line 7, in main juju_status = juju_get_status("bcache", keep_relations_info=True, jsfy=False, format_status=True) File "~lib/lib_juju.py", line 57, in juju_get_status juju_status = loop.run(get_juju_status(model_name, keep_relations_info, jsfy, format_status)) File "~venv/lib/python3.8/site-packages/juju/jasyncio.py", line 118, in run raise task.exception() File "~lib/lib_juju.py", line 108, in get_juju_status juju_status = await formatted_status(model) File "~venv/lib/python3.8/site-packages/juju/status.py", line 61, in formatted_status result_str += _print_status_apps(result_status) File "~venv/lib/python3.8/site-packages/juju/status.py", line 116, in _print_status_apps result_str += limits.format( TypeError: unsupported format string passed to NoneType.__format__ Task was destroyed but it is pending! Answers: username_1: Changes from #622 should fix this. Please reopen if needed. Status: Issue closed
CosmicMind/Material
144433209
Title: Depth Does Not Change on Pulse Question: username_0: When a view pulses (and scale), the shadow remains the same and does not present a change of depth. Scale I presume is there to simulate the view getting closer to the fingertip. Status: Issue closed Answers: username_1: The shadow slightly spreads, due to the scale effect. I believe you could add an animation by subclassing the pulse view and adding animations to the touch handlers that spreads the shadow, and contracts it. I am not going to make this a priority right now, but we are planning some nice additions with 3D touch and such, and that is when this issue will be considered at a high priority. Feel free to continue the dialog on this, though for cleanup reasons, I will close the issue.
PistonDevelopers/conrod
35726002
Title: Widget Ideas Question: username_0: Here is a space to collect Widget Type ideas. This list is taken directly from the [ofxUI] (https://github.com/rezaali/ofxUI) readme as a basis for ideas. Any more ideas are welcome! - Buttons (push, state, toggle, image, label) - Button Matrices - Dropdown Menus - Labels - Sliders (rotary, range, vertical, horizontal) - Number Dials (aka Spinners) - 2D Pads - Text Input Areas - Image Sliders - Image Buttons - Image Color Sampler - Value Plotters - Moving Graphs - Waveform & Spectrum Graphs - Radio Toggles - Text Areas - Sortable List Status: Issue closed Answers: username_0: Closing in favour of posting unique issues for drafting/discussing/requesting each widget.
google/yapf
1084716435
Title: [Question]yapfignore not working in pyproject.toml Question: username_0: Thank you for reading this issue. I'm currently working on a Django project and I want to use `yapf` for formatting. However, as I am trying to use `pyproject.toml` for excluding files, it does not works. ```toml [tool.yapfignore] ignore_patterns = [ '**/migrations]*.py', 'manage.py', 'venv/*', ] ``` On the other hand, what makes it interesting is that the command line behaves differently and successfully ignores the files I want. ```sh yapf --recursive --exclude '**/migrations/*.py' --exclude 'manage.py' --exclude 'venv/*' . ``` I still cannot figure out what would be the problem with this discrepancy. At the moment, a part my dependencies is like the following: ```txt pylint==2.12.2 pylint-django==2.4.4 pylint-plugin-utils==0.6 yapf==0.31.0 ``` and my current Python version is `3.9.9`. If there is any problem with the way I make this issue post, please don't hesitate to point it out! Thank you in advance! Answers: username_0: It still doesn't work, so I had a workaround, `.yapfignore`. ``` **/migrations/*.py manage.py venv/* ``` I have to admit that the project would be a bit more complex, but I think `pyproject.toml` is currently not working for excluding files and directories. Any suggestions? username_1: Hi @username_0 Related to https://github.com/google/yapf/issues/955 , I think yapf version is the problem. Latest version (0.32.0) of yapf released 26 Dec (ref: https://pypi.org/project/yapf/#history). In my local PC, yapf ignore feature correctly worked. Please check. Status: Issue closed username_0: @username_1 Thank you for your reply! I appreciate your kindness. Oh, I have missed the issue https://github.com/google/yapf/issues/955 I have checked and now it worked. Now this issue is resolved so I close it.
JuliaTime/TimeZones.jl
733844499
Title: Parsing a datetime string including "GMT" stopped working in v1.5.0 Question: username_0: `TimeZones.ZonedDateTime` used to work prior to release 1.5.0 when GMT was included. An error referring to legacy timezones is now returned. However, I'm not aware that such legacy timezones can be specified when parsing a string. ```julia julia> dt_string_1 = "Sat, 31 Oct 2020 11:31:54 GMT" "Sat, 31 Oct 2020 11:31:54 GMT" julia> TimeZones.ZonedDateTime(dt_string_1, "e, d u Y H:M:S Z") ERROR: ArgumentError: Unable to parse string "Sat, 31 Oct 2020 11:31:54 GMT" using format dateformat"e, d u Y H:M:S Z". The time zone "GMT" is of class `TimeZones.Class(:LEGACY)` which is currently not allowed by the mask: `TimeZones.Class(:FIXED) | TimeZones.Class(:STANDARD)` Stacktrace: [1] TimeZone(::SubString{String}, ::TimeZones.Class) at C:\Users\jerem\.julia\packages\TimeZones\fr1IP\src\types\timezone.jl:65 [2] TimeZone at C:\Users\jerem\.julia\packages\TimeZones\fr1IP\src\types\timezone.jl:46 [inlined] [3] ZonedDateTime(::Int64, ::Int64, ::Int64, ::Int64, ::Int64, ::Int64, ::Int64, ::SubString{String}) at C:\Users\jerem\.julia\packages\TimeZones\fr1IP\src\types\zoneddatetime.jl:132 [4] parse at C:\Users\jerem\AppData\Local\Programs\Julia-1.5.2\share\julia\stdlib\v1.5\Dates\src\parse.jl:285 [inlined] [5] ZonedDateTime(::String, ::DateFormat{Symbol("e, d u Y H:M:S Z"),Tuple{Dates.DatePart{'e'},Dates.Delim{String,2},Dates.DatePart{'d'},Dates.Delim{Char,1},Dates.DatePart{'u'},Dates.Delim{Char,1},Dates.DatePart{'Y'},Dates.Delim{Char,1},Dates.DatePart{'H'},Dates.Delim{Char,1},Dates.DatePart{'M'},Dates.Delim{Char,1},Dates.DatePart{'S'},Dates.Delim{Char,1},Dates.DatePart{'Z'}}}) at C:\Users\jerem\.julia\packages\TimeZones\fr1IP\src\parse.jl:88 [6] ZonedDateTime(::String, ::String; locale::String) at C:\Users\jerem\.julia\packages\TimeZones\fr1IP\src\parse.jl:101 [7] ZonedDateTime(::String, ::String) at C:\Users\jerem\.julia\packages\TimeZones\fr1IP\src\parse.jl:101 [8] top-level scope at REPL[96]:1 ``` Base function works when specifying the presence of Z, although the notion of timezone is ignored. This seems a viable approach for my use case. I'm just wondering whether it wouldn't be more convenient to keep supporting the parsing of GMT in `ZonedDateTime`? ``` julia> Dates.DateTime(dt_string_1, "e, d u Y H:M:S Z") 2020-10-31T11:31:54 ```<issue_closed> Status: Issue closed
Tunous/Dawn
654393250
Title: Highlighted text permanently shown Question: username_0: After long pressing and dragging to select text, a magnification bubble pops up and remains there until I restart the app. ![Screenshot_2020-07-02-23-06-24-217_me thanel dank](https://user-images.githubusercontent.com/50382414/87098151-ef821380-c246-11ea-937b-f78a2e7d3da3.jpg) Answers: username_1: This might be rom related, too. Can you post exact rom version you are on? username_0: My ROM is stable MIUI 12. MIUI Global 12.0.1 Stable 172.16.58.3(QFKEUXM)
tensorflow/tensorflow
387369750
Title: [ppc64le] //tensorflow/lite/experimental/micro unit test fail. Question: username_0: You can assign this issue to me, as I'm about ready to submit a PR to fix this. **System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): N/A - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux ppc64le Ubuntu 16.04 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A - TensorFlow installed from (source or binary): source - TensorFlow version (use command below): commit <PASSWORD> from Dec 4th, 2018 - Python version: 2.7 - Bazel version (if compiling from source): 0.15.0 - GCC/Compiler version (if compiling from source): gcc version 5.4.0 20160609 (Ubuntu/IBM 5.4.0-6ubuntu1~16.04.10) - CUDA/cuDNN version: N/A - GPU model and memory: N/A **Describe the current behavior** All tensorflow/lite/experimental/micro unit test fail See: https://powerci.osuosl.org/job/TensorFlow_PPC64LE_CPU_Build_and_Test/31/testReport/ **Describe the expected behavior** All unit test pass **Code to reproduce the issue** Unit test invoked by Jenkins: `./tensorflow/tools/ci_build/ci_build.sh cpu --dockerfile tensorflow/tools/ci_build/Dockerfile.cpu.ppc64le ./tensorflow/tools/ci_build/linux/ppc64le/cpu/run_py2.sh` `./tensorflow/tools/ci_build/linux/ppc64le/cpu/run_py2.sh` can be modified to just run `//tensorflow/lite/experimental/micro/...` Also the getting started section here recreates two of the issues: https://github.com/tensorflow/tensorflow/tree/master/tensorflow/lite/experimental/micro **Other info / logs** Running the getting started example: make -f tensorflow/lite/experimental/micro/tools/make/Makefile test_micro_speech This warning below is flagged and run the test runs it seg faults. As I understand the warning is because your not allowed to pass a string constant in this case. (It works however on x86) ``` g++ -O3 -DNDEBUG --std=c++11 -g -DTF_LITE_STATIC_MEMORY -I. -Itensorflow/lite/experimental/micro/tools/make/../../../../../ -Itensorflow/lite/experimental/micro/tools/make/../../../../../../ -Itensorflow/lite/experimental/micro/tools/make/downloads/ -Itensorflow/lite/experimental/micro/tools/make/downloads/gemmlowp -Itensorflow/lite/experimental/micro/tools/make/downloads/flatbuffers/include -I -I/usr/local/include -c tensorflow/lite/experimental/micro/examples/micro_speech/micro_speech_test.cc -o tensorflow/lite/experimental/micro/tools/make/gen/linux_ppc64le/obj/tensorflow/lite/experimental/micro/examples/micro_speech/micro_speech_test.o In file included from tensorflow/lite/experimental/micro/examples/micro_speech/micro_speech_test.cc:22:0: tensorflow/lite/experimental/micro/examples/micro_speech/micro_speech_test.cc: In function 'int main(int, char**)': ./tensorflow/lite/experimental/micro/testing/micro_test.h:95:51: warning: ISO C++ forbids converting a string constant to 'va_list {aka char*}' [-Wwrite-strings] micro_test::reporter->Report("Testing %s", #name); \ ^ tensorflow/lite/experimental/micro/examples/micro_speech/micro_speech_test.cc:28:1: note: in expansion of macro 'TF_LITE_MICRO_TEST' TF_LITE_MICRO_TEST(TestInvoke) { ^~~~~~~~~~~~~~~~~~ ``` Fixing it and running the example: make -f tensorflow/lite/experimental/micro/tools/make/Makefile test identifies the same problem in another place: ``` g++ -O3 -DNDEBUG --std=c++11 -g -DTF_LITE_STATIC_MEMORY -I. -Itensorflow/lite/experimental/micro/tools/make/../../../../../ -Itensorflow/lite/experimental/micro/tools/make/../../../../../../ -Itensorflow/lite/experimental/micro/tools/make/downloads/ -Itensorflow/lite/experimental/micro/tools/make/downloads/gemmlowp -Itensorflow/lite/experimental/micro/tools/make/downloads/flatbuffers/include -I -I/usr/local/include -c tensorflow/lite/experimental/micro/micro_error_reporter_test.cc -o tensorflow/lite/experimental/micro/tools/make/gen/linux_ppc64le/obj/tensorflow/lite/experimental/micro/micro_error_reporter_test.o tensorflow/lite/experimental/micro/micro_error_reporter_test.cc: In function 'int main(int, char**)': tensorflow/lite/experimental/micro/micro_error_reporter_test.cc:24:56: warning: ISO C++ forbids converting a string constant to 'va_list {aka char*}' [-Wwrite-strings] error_reporter->Report("~~~%s~~~", "ALL TESTS PASSED"); ^ [Truncated] LD_LIBRARY_PATH=/usr/local/cuda/extras/CUPTI/lib64:/usr/local/nvidia/lib:/usr/local/nvidia/lib64 \ OMP_NUM_THREADS=1 \ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/snap/bin \ PWD=/proc/self/cwd \ PYTHON_BIN_PATH=/usr/bin/python2 \ PYTHON_LIB_PATH=/usr/local/lib/python2.7/dist-packages \ TF_CUDA_CLANG=0 \ TF_CUDA_COMPUTE_CAPABILITIES=3.7 \ TF_CUDA_VERSION=9.2 \ TF_CUDNN_VERSION=7 \ TF_NCCL_VERSION='' \ TF_NEED_CUDA=1 \ TF_NEED_OPENCL_SYCL=0 \ TF_NEED_ROCM=0 \ external/local_config_cuda/crosstool/clang/bin/crosstool_wrapper_driver_is_not_gcc -o bazel-out/ppc-opt/bin/tensorflow/lite/experimental/micro/examples/micro_speech/feature_provider_test_binary -Wl,-no-as-needed -pie -Wl,-z,relro,-z,now '-Wl,--build-id=md5' '-Wl,--hash-style=gnu' -no-canonical-prefixes -fno-canonical-system-headers -B/usr/bin -Wl,--gc-sections -Wl,@bazel-out/ppc-opt/bin/tensorflow/lite/experimental/micro/examples/micro_speech/feature_provider_test_binary-2.params) /usr/bin/ld: bazel-out/ppc-opt/bin/tensorflow/lite/experimental/micro/examples/micro_speech/libpreprocessor_reference.a(preprocessor.o): undefined reference to symbol 'cos@@GLIBC_2.17' //lib/powerpc64le-linux-gnu/libm.so.6: error adding symbols: DSO missing from command line ``` Passing the link option "-lm" resolves this error. The makefile already does this. Answers: username_1: Thanks for your PR on this (and sorry for the slow response). username_0: closing old issues I opened. This issue is resolved. Status: Issue closed
jenkinsci/office-365-connector-plugin
637152225
Title: "View Build" option is broken in Microsoft Teams via Jenkins Question: username_0: Hi Team, Could you please help is resolving following issue? In Jenkins pipeline, we have added Notification Webhook for "Office 365 Connector". When build is notified to Microsoft teams, there is "View Build" button which is having below URL: http://<JENKINS_URL>/jenkins/job/<JOB_NAME>/job/<JOB>/50/display/redirect which is redirecting to: http://<JENKINS_URL>/jenkins/blue/organizations/jenkins/<JOB_NAME>%2F<JOB>/detail/<JOB>/50/ and throwing 400 error but actual URL is having "pipeline" in the end. http://<JENKINS_URL>/jenkins/blue/organizations/jenkins/<JOB_NAME>%2F<JOB>/detail/<JOB>/50/pipeline How to resolve this? Where to modify this URL? Please help, Thanks and Regards, <NAME> Answers: username_1: I don't use Blue ocean and so the plugin may not work with this. If you provide all details so I can reproduce the problem maybe I would be able to help or fix it username_0: Thanks for reply. This issue resolved. I disabled one plugin "Display URL for Blue Ocean" and result is coming fine. I think that plugin is causing some confusion with Teams URL. Thanks once again. We can close this username_2: Ideally, that link should be customizable, with the default being what it is now. We're currently getting hit by this too. username_1: Not sure what do you mean username_3: We experience this issue as well: when clicking 'View Build' the url 'https://teams.microsoft.com/null' is opened instead of the Build detail page within jenkins. username_4: Problem noticed yesterday. Yesterday system status: ``` The connector configuration was old (outlook in URL). jdk: 8 Windows ``` I thought we should update the JDK version and connector. So the current status: ``` Jenkins: 2.277.4 Plugin: 4.15.0 jdk: OpenJDK 11.0.2 Windows All other plugins are updated. The connector URL looks like: https://[name].webhook.office.com/webhookb2/[some GUID]/JenkinsCI/[some GUID] ``` But the problem still exists. username_5: We have a similar issue with the "View Pull Request" Button. It points to "_https://teams.microsoft.com/null_". Can I set, the url manually or is this a known bug? Thanks! username_6: I have URL's created manually that have the same behavior. Teams on my mobile works, but not desktop. Also, not everybody experienced the issue at once. I'd assume it is a Teams issue username_5: Thanks for your reply, that's interesting. Indeed it also works for my mobile version of Teams... username_1: Have you tried to regenerate hook url? I have done it and some problems have been solved username_7: We have the same problem too. I've regenerate hook url but the same problem still occurs. username_8: Yes. Multiple times. Issue persists. The webhook seems to be fine since, as others have mentioned, if you click the button on the mobile Teams app it seems to work fine. username_9: Just jumping in the conversation. Same problem here (MSTeams Connector) I also have a HTTP Post Request that uses a custom Payload and had to use markdown with a static URL and a variable since the default option didn't work either (custom Connector) username_1: Then maybe you have regenerated different hook? For regenerated url problems have gone. For old one I see problem on desktop and mobile username_1: OK, I found one more interesting thing: when updating connection configuration (without updating url) desktop Teams is fixed, mobile on team card is fixed but activity card on mobile still displays that message. username_9: No URL was changed, I've even tried to make a new one with the same configuration, just to see if there was any mistake made in the past, but no luck. Same result. I actually don't have any error on the webhook configuration. username_10: same problem too, but mobile no problem. username_1: Guys, have you follow this instruction https://docs.microsoft.com/pl-pl/microsoftteams/office-365-custom-connectors ? This provides guideline that solves problem for me username_11: yeah, same problem too. but mobile is work. -> https://teams.microsoft.com/null username_12: I'm wondering if this problem is caused by connector plugin, or not. I have tested some URL patterns against section "Jenkins URL" in Jenkins configuration page. Results: URL | Reachable :--|:-: `http://192.168.2.181:8484/` | No `http://dd7.local:8484/` | Yes `http://dd7:8484/` | No ## Pattern 1: `http://192.168.2.181:8484/` ![image](https://user-images.githubusercontent.com/5955540/125878215-864fc7b6-104b-4411-b9d6-b0dd0c583993.png) build something. ![2021-07-16_10h32_21](https://user-images.githubusercontent.com/5955540/125878225-e4c2dd03-c02d-496a-9bea-7a9f5c2cd95d.png) `View Build` leads to `https://teams.microsoft.com/null` ![2021-07-16_10h37_57](https://user-images.githubusercontent.com/5955540/125878583-6d752165-8723-45a6-b726-3a02492ff5a5.png) ## Pattern 2: `http://dd7.local:8484/` ![image](https://user-images.githubusercontent.com/5955540/125878299-04eaa412-873b-4b71-ac1e-c90b84263c63.png) build something. ![2021-07-16_10h33_49](https://user-images.githubusercontent.com/5955540/125878358-733e2274-6d89-4ff0-9b26-ca5bff0590d3.png) `View Build` leads to `http://dd7.local:8484/job/TeamsTest/2/display/redirect` ![2021-07-16_10h36_23](https://user-images.githubusercontent.com/5955540/125878475-56c6efa2-07f7-4f55-9572-df12ca9c2aeb.png) ## Pattern 3: `http://dd7:8484/` ![image](https://user-images.githubusercontent.com/5955540/125878256-a6bb93f9-e7ed-40ad-a301-b0d09f798190.png) build something. ![2021-07-16_10h34_39](https://user-images.githubusercontent.com/5955540/125878337-b0380fc2-df8f-4f64-ae1c-0bd7ee848746.png) `View Build` leads to `https://teams.microsoft.com/null` ![2021-07-16_10h38_22](https://user-images.githubusercontent.com/5955540/125878606-9e25f322-7877-4579-b49e-7b4b3c888303.png) username_11: I think I found a way to solve this problem. If it's not a secure website (e.g. https), it seems to go to a teams.microsoft.com/null page. Therefore, I think IP should be set up as an external network and apply SSL. username_8: I have SSL setup, https://<IP Address>:8443 How do I setup the IP as an external network?
getsentry/sentry-javascript
403459096
Title: Morgan and @sentry/node with the Express Handler not working well together Question: username_0: <!-- Requirements: please go through this checklist before opening a new issue --> - [x] Review the documentation: https://docs.sentry.io/ - [x] Search for existing issues: https://github.com/getsentry/sentry-javascript/issues - [x] Use the latest release: https://github.com/getsentry/sentry-javascript/releases ## Package + Version - [x] `@sentry/node` ### Version: ``` 4.5.3 ``` ## Description When using the morgan logging middleware, alongside `app.use(Sentry.Handlers.errorHandler() as express.ErrorRequestHandler)`, whenever there is an async controller that throws an Error, no futher requests can be made, and the following stack trace is shown: ``` GET /api/grin/stats 500 54.544 ms - - Error: connect ECONNREFUSED 172.16.17.32:8080 at Object._errnoException (util.js:1022:11) at _exceptionWithHostPort (util.js:1044:20) at TCPConnectWrap.afterConnect [as oncomplete] (net.js:1198:14) TypeError: this.app.get is not a function at IncomingMessage.ip (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/request.js:350:24) at getip (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/morgan/index.js:466:14) at logger (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/morgan/index.js:107:26) at Layer.handle [as handle_request] (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/router/layer.js:95:5) at trim_prefix (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/router/index.js:317:13) at /Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/router/index.js:284:7 at Function.process_params (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/router/index.js:335:12) at next (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/router/index.js:275:10) at jsonParser (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/body-parser/lib/types/json.js:110:7) at Layer.handle [as handle_request] (/Users/username_0/js/src/github.com/LuxorLabs/luxor-mining-v3/node_modules/express/lib/router/layer.js:95:5) ``` This does not happen if I comment out `app.use(morgan('dev'))`, or if I comment out `app.use(Sentry.Handlers.errorHandler() as express.ErrorRequestHandler)`... I just can't use both at the same time, and I'd like to do that. Appreciate any guidance here. Answers: username_1: We are experiencing weird things after updating from 4.5.0 -> 4.5.3, the context is different but the underlying `this.app.get` error is the same. username_0: @username_1 does downgrading to 4.5.0 fix the issue? username_1: @username_0 yeah it looks like 4.5.0 works as expected username_2: Seeing the same thing with bunyan (2.0.2) and express-bunyan-logger (1.3.2). Downgrading to 4.5.0 fixes it. username_3: Also seeing this. https://github.com/getsentry/sentry-javascript/issues/1859 looks like the same issue username_4: Same issue with Feathers and Winston https://github.com/feathersjs/feathers/issues/1183 username_5: I'm not using anything like Winston or Morgan, but I'm also experiencing this issue. Related bits: ``` Sentry.init({ beforeSend(event, hint) { const { originalException, } = hint const { message, } = originalException if (message.match(/NotAuthenticated/i)) { return null } return event }, dsn: env.SENTRY_DSN, }) // ... app.use(Sentry.Handlers.requestHandler()) // ... app.use((error, request, response, next) => { if (!error.stack.match(/NotAuthenticated/i)) { err(error.stack) } next(error) }) app.use(Sentry.Handlers.errorHandler()) app.use(express.errorHandler()) ``` Will check if downgrading fixes the issue. username_6: Fixed in `4.5.4` – https://github.com/getsentry/sentry-javascript/releases/tag/4.5.4 Sorry for the inconvenience :) Status: Issue closed username_7: The issue still exists in `"@sentry/node": "^4.6.2"` using as Express middleware ```TypeError: req.emit is not a function at Socket.socketOnTimeout (_http_server.js:427:48) at Socket.emit (events.js:197:13) at Socket.EventEmitter.emit (domain.js:446:20) at Socket._onTimeout (net.js:447:8) at listOnTimeout (timers.js:327:15) at processTimers (timers.js:271:5)``` username_6: Really?... can you provide a repro please? Every scenario I tested worked just fine. username_7: I think sentry tries to attach something to res object asynchronously, but the connection is already closed which crashes the api. ``` const api = express(); sentry.init({ dsn: process.env.SENTRY, environment: process.env.DOMAIN }); api.use(sentry.Handlers.requestHandler()); .... api.use(sentry.Handlers.errorHandler()); api.use((err, req, res, next) => { let { message, statusCode } = err; if (err.name === 'UnauthorizedError') { statusCode = 401; message = 'Token is invalid.'; } res.status(statusCode || 400).json({ error: message }); }); ``` username_6: Works just fine with your repro ¯\_(ツ)_/¯ ![image](https://user-images.githubusercontent.com/1523305/53088717-da4dfd00-350a-11e9-993c-dc0f3a08cf66.png) username_7: not sure, that was the logic before we removed it. Might be something related to our own controllers.
lipis/flag-icons
1165239735
Title: Reorder flags with the Unknown flag first on the list Question: username_0: My suggestion is that you should reorder the flag classes by placing the unknown flag (`xx`) before all other flags. In terms of CSS precedence, if `.fi-fr` is placed on the same element as `.fi-xx`, the unknown flag will override the French flag due to the fact that the French flag is declared earlier than the Unknown flag. By declaring the Unknown flag first, means that every other flag will override it if they reside on the same element. ### So why would you want to declare both flags on the same element? In my use case, I have a system that stores the 2-letter country code against items in a database. This allows me to display the relevant flag-icon on the page by adding the flag class to the span element (forgive me, I am using Vue): ```vue <span class="fi" :class="[`fi-${countryCode}`]"></span> ``` Unfortunately, the data may contain invalid 2-letter country codes. (eg. `vb`). To combat this, I would like it to display the "Unknown" flag as a fallback flag. ```vue <span class="fi fi-xx" :class="[`fi-${countryCode}`]"></span> ``` For a country code of `vb`, this would be translated to `class="fi fi-xx fi-vb"`. Since `.fi-vb` doesn't exist, this gets ignored and `.fi-xx` will produce the "Unknown" flag. For a country code of `fr`, this would be translated to `class="fi fi-xx fi-fr"`. If `.fi-fr` was declared after `.fi-xx` then the French flag would override the "Unknown" flag and page would display the French flag. Unfortunately, with `.fi-xx` declared last, the above example will show the "Unknown" flag instead. Please accept this proposal as I think this would be a benefit for a lot of people using your library.
Sage-Bionetworks/ChallengeWorkflowTemplates
580913878
Title: get_submission.cwl makes assumptions about output from challengeutils downloadsubmission Question: username_0: This line: https://github.com/Sage-Bionetworks/ChallengeWorkflowTemplates/blob/master/get_submission.cwl#L34 The workflow assumes that this is exactly where the file will be written and how it will be named. But `challengeutils downloadsubmission` already provides this information as part of its output. I think this fix is as simple as replacing that line with something like: ``` $(JSON.parse(self[0].contents)['file_path']) ``` Answers: username_1: @username_0, Thanks for reporting. Did you get an error when running your workflow? This is because `challengeutils` itself renames the submission file: https://github.com/Sage-Bionetworks/challengeutils/blob/master/challengeutils/__main__.py#L86 username_0: Interesting, lines like that seem to work otherwise: https://github.com/Sage-Bionetworks/ChallengeWorkflowTemplates/blob/master/get_submission.cwl#L41 username_1: Ahh, i forgot that CWL requires the File output to be in a dictionary format like so: ``` "results": { "location": "file:///Users/thomasyu/sage_projects/dream/ChallengeWorkflowTemplates/results.json", "basename": "results.json", "class": "File", "checksum": "sha1$af6758eac5c5bf4eefbd1ac836982ffa8b621051", "size": 222, "path": "/Users/thomasyu/sage_projects/dream/ChallengeWorkflowTemplates/results.json" } ``` Therefore the line in `outputEval` would have to mimic that structure above. username_1: Closing as I don't think there is a way to create the correct output blob to match what is required from a `File` type output Status: Issue closed
vcd94xt10z/zionphp
412513484
Title: Modulo test Question: username_0: Módulo para testar serviços web, funcionalidades: - cadastrar um fluxo de chamadas de URL com ou sem dados, status http esperado etc; - fazer botão simples para rodar os testes e exibir o resultado; Exemplo POST http://teste.com/usuario/salvar Status 201 sem response body PUT http://teste.com/usuario/1 Status 204 sem response body GET http://teste.com/usuario/1 Status 200 com response body DELETE http://teste.com/usuario/1 Status 202 sem response body<issue_closed> Status: Issue closed
godotengine/godot
354752076
Title: Tile selector: add option to resize tile entries Question: username_0: It is weird that even with a `list` mode, images are still horribly big. I mean tiles like in gridmap object, we should be able to resize it, for example: ![imgur-2018_08_28-13 33 55](https://user-images.githubusercontent.com/15695377/44720699-dd21e500-aac7-11e8-9320-7cca68373d01.png) Option to disable icons to get more space would be useful too. Answers: username_1: maybe an EditorSetting could be added to allow chose the maximum icon size, but it will scale the texture uniformly, you example looks horrible imo :P username_2: example looks terrible, please don't Status: Issue closed
PlantGenIE/PlantGenIE
1101178537
Title: Display expression values Question: username_0: What value should be used in expression tools? is it TPM+2? or TPM+1? Currently we have TPM+1 in all our tools to avoid special cases as follows. ``` log2(TPM+0.9)=-0.15 ∴ TPM=0 log2(TPM+1)=0 ∴ TPM=0 log2(TPM+2)=1 ∴ TPM=0 ``` Answers: username_0: yes, we display log2(TPM + 1) for all our expression tools. exHeatmap allow users to download expression values but that is also the same log2(TPM + 1) referred as absolute expression value.
mapgears/ol3-google-maps
154990007
Title: Problem with getStroke Question: username_0: Hello, when i load for the first time my olgm map, appear in console this error "ol3gm.js:244 Uncaught TypeError: d.getStroke is not a function" and the dots of polygon on layer are very bigger than normal. What is the problem? How can i fix it? thank you ![screenshot-localhost 2016-05-16 11-10-42](https://cloud.githubusercontent.com/assets/7781896/15285352/e23af732-1b56-11e6-826c-fbd10e10f274.png) Answers: username_1: the same problem username_2: Would you please provide a small demo featuring the issue, maybe with JSFiddle ? username_1: the problem in style function.When i use style object problem solved .Why don't i use style function? username_2: Style functions are not supported, as documented in the limitations: https://github.com/mapgears/ol3-google-maps/blob/master/LIMITATIONS.md#style-functions There's currently no plan to support them either. username_0: Sorry but i don't use a style object neither manipulate style. The problem is present in this function but i don't understand why (ol3gm.js - row 239). function fa(a, b) { var c = null, d = C(a); if (d) { var c = {}, e = d.getStroke(); if (e) { var f = e.getColor(); f && (c.strokeColor = x(f), f = B(f), null !== f && (c.strokeOpacity = f)); (e = e.getWidth()) && (c.strokeWeight = e) } if (e = d.getFill()) if (e = e.getColor()) c.fillColor = x(e), e = B(e), null !== e && (c.fillOpacity = e); if (f = d.getImage()) { d = {}; e = {}; if (f instanceof ol.style.Circle) { e.path = google.maps.SymbolPath.CIRCLE; var h = f.getStroke(); if (h) { var k = h.getColor(); k && (e.strokeColor = x(k)); e.strokeWeight = h.getWidth() } if (h = f.getFill()) if (h = h.getColor()) e.fillColor = x(h), h = B(h), e.fillOpacity = null !== h ? h : 1; (f = f.getRadius()) && (e.scale = f) } else f instanceof ol.style.Icon && ((h = f.getSrc()) && (e.url = h), h = f.getScale(), (k = f.getAnchor()) && (e.anchor = void 0 !== h ? new google.maps.Point(k[0] * h, k[1] * h) : new google.maps.Point(k[0], k[1])), (k = f.getOrigin()) && (e.origin = new google.maps.Point(k[0], k[1])), f = f.getSize()) && (e.size = new google.maps.Size(f[0], f[1]), void 0 !== h && (e.scaledSize = new google.maps.Size(f[0] * h, f[1] * h))); Object.keys(d).length ? c.icon = d : Object.keys(e).length && (c.icon = e) } 0 === Object.keys(c).length ? c.visible = !1 : void 0 !== b && (c.zIndex = 2 * b) } return c }; username_2: As I said, you can't use style functions with OLGM. It's not currently suported to do so. See the limitations. username_0: Ok, thank tou for your answer but this is a bug because the code is inside ol3gm. Can you suggest me a workaround, please? thank you username_2: This is not a bug. You must not use StyleFunctions with OLGM as it is stated in the limitations. There is no other workaround than "don't use StyleFunctions" with OLGM. I'm sorry that I can't be of more help regarding this for now. If you absolutely need this, then you're welcome to create a Pull Request adding this support or you can contact info at mapgears dot com if you wish to fund the development of this. To my knowledge, I think it's impossible to support style functions with OLGM. Status: Issue closed username_2: I'm closing this. See the limitations: https://github.com/mapgears/ol3-google-maps/blob/master/LIMITATIONS.md#style-functions
RasaHQ/rasa
604214539
Title: Makefile:43: recipe for target 'install' failed Question: username_0: While trying to install install dependencies using the make install command, I am getting the following error. ``` kamaldeep@kamaldeepsingh:~/git/testing/rasa1$ curl -sSL https://raw.githubusercontent.com/python-poetry/poetry/master/get-poetry.py | python Retrieving Poetry metadata Latest version already installed. kamaldeep@kamaldeepsingh:~/git/testing/rasa1$ make install poetry run python -m pip install -U 'pip<20' The virtual environment found in /home/kamaldeep/.cache/pypoetry/virtualenvs/rasa-LHgLSZoI-py3.6 seems to be broken. Recreating virtualenv rasa-LHgLSZoI-py3.6 in /home/kamaldeep/.cache/pypoetry/virtualenvs/rasa-LHgLSZoI-py3.6 [CalledProcessError] Command '['/home/kamaldeep/.cache/pypoetry/virtualenvs/rasa-LHgLSZoI-py3.6/bin/python', '-Im', 'ensurepip', '--upgrade', '--default-pip']' returned non-zero exit status 1. Makefile:43: recipe for target 'install' failed make: *** [install] Error 1 ``` Seems like there is some issue in the creation of the Virtual environment. Please help with the same. Status: Issue closed Answers: username_0: While trying to install install dependencies using the make install command, I am getting the following error. ``` kamaldeep@kamaldeepsingh:~/git/testing/rasa1$ curl -sSL https://raw.githubusercontent.com/python-poetry/poetry/master/get-poetry.py | python Retrieving Poetry metadata Latest version already installed. kamaldeep@kamaldeepsingh:~/git/testing/rasa1$ make install poetry run python -m pip install -U 'pip<20' The virtual environment found in /home/kamaldeep/.cache/pypoetry/virtualenvs/rasa-LHgLSZoI-py3.6 seems to be broken. Recreating virtualenv rasa-LHgLSZoI-py3.6 in /home/kamaldeep/.cache/pypoetry/virtualenvs/rasa-LHgLSZoI-py3.6 [CalledProcessError] Command '['/home/kamaldeep/.cache/pypoetry/virtualenvs/rasa-LHgLSZoI-py3.6/bin/python', '-Im', 'ensurepip', '--upgrade', '--default-pip']' returned non-zero exit status 1. Makefile:43: recipe for target 'install' failed make: *** [install] Error 1 ``` Seems like there is some issue in the creation of the Virtual environment. Please help with the same. Status: Issue closed
dotse/bbk
587993286
Title: Don't report 99 MBit on a 100-200Mbit connection as 'BAD'. Question: username_0: There need to be some sort of limit based in percentage of the connection to determine the speed, not just some hard limits. Answers: username_1: Have you checked the Readme? https://frontend.bredbandskollen.se/download/README.txt 'If the subscription can be identified, the download result will be evaluated as "GOOD", "ACCEPTABLE", or "BAD". The result is considered BAD if the speed is below the lower end of the expected speed range as stated by the service provider. It is considered ACCEPTABLE if it is above the lower end, but still clearly below the middle, of the expected speed range. If the result is better, it is considered GOOD. E.g. if the expected range is 150-250 Mbit/s, a result below 150 is BAD; a result between 150 and 190 is ACCEPTABLE, and a result above 190 is GOOD.' The "limit based in percentage" you're looking for is the 40% target that decides what is needed to get a GOOD result: 250 - 150 = 100 Mbit. 40% of 100 Mbit is 40 Mbit. 150 + 40 = 190 Mbit. username_0: Exactly and that's wrong in so many ways. 99,9% is NEVER a "bad" speed which implies it's unacceptable and hence makes people call customer support. username_2: We are getting the connection from the ISP - if we have the possibility. If you ISP says you connection should be 100-200 MB and you don't get up to the lowest limit we do consider that bad. username_0: And that is flawed since it's usually "up to" in the agreements and there might be other circumstances that lowers the speed. To actually know the connection quality you would have to have iperf server instances at the ISPs. This creates confusion which creates unnecessary support calls for those working in the field. username_3: I just randomly found this issue, and I'd just like to say this makes no sense. If the ISP considers 99 Mbit/s an acceptable speed on the connection that they sell, they should then sell the connection as a 99-200 Mbit/s connection, not a 100-200 Mbit/s connection. If the ISP promises 100-200 Mbit/s, then anything below 100 Mbit/s is unacceptable. (Assuming, for now, that the measurement server is well enough connected to the ISP's network to be a valid measurement of the connection speed itself.) Imagine, if I went to the supermarket, and I bought a box of milk that is specified to be 100-200 cl, and then I come home and see I only got 99 cl. In this case, the amount of milk that I received is out of spec, even if it's only 1% out of spec.
febobo/JS-EveryDay-Question
425260959
Title: JS每日一题:前端优化手段有哪些 Question: username_0: 20190118问: 前端优化手段有哪些? - 静态资源合并压缩(js,css, images) - 请求数量优化 - Gzip压缩 - 带宽优化 - CDN - 就近节点,减少DNS查找 - 按需加载 - lazyload - 减少请求 - 骨架屏 - 优化白屏 - web缓存 - 缓存ajax数据 - 减少重绘和重排 - 批量更新DOM样式 - 页面结构 - 将样式表放在顶部,将脚本放在底部,尽早刷新文档的输出
openenclave/openenclave
843626477
Title: Ability to make an ECALL from kernel mode Question: username_0: Currently another team in Microsoft is using OP-TEE directly, rather than building on Open Enclave SDK, because OE does not currently support the ability to make an ECALL from kernel mode. This forces them to jump through lots of extra hoops to write the enclave that OE would have solved for them, and limits their ability to work on other TEEs in the future. Currently in OE, one would have to first call up into user mode and have user mode make the ECALL, which loses extra perf (on Arm at least) and causes extra complexity and this approach isn't obvious either. Hence this is a feature request to support making an ECALL directly from kernel mode, which I am filing on the other team's behalf. Answers: username_0: This is peripherally related to issue #3804, but that is just for attestation, and that request came from a different team. username_1: On SGX, user mode context is required to make an ecall. username_1: Capture scenario in design doc.
AlvaroLarumbe/espalet-android
202401501
Title: When clicking on the refresh button there is no feedback to the user Question: username_0: It will be good to show spinner icon or at least make images blink once. Answers: username_1: Look at this: https://github.com/username_1/espalet-android/commit/b1030608674f4b0bbdb25932a19c5c86215feae3#diff-350204edff7826e85e2fac268b3e263cR61 username_0: Enough Status: Issue closed
invenia/Intervals.jl
524508865
Title: Range indexing fails Question: username_0: ``` julia> using Dates, TimeZones, Intervals julia> st = HE(ZonedDateTime(2014,1,1, tz"America/Winnipeg")) AnchoredInterval{-1 hour,ZonedDateTime}(ZonedDateTime(2014, 1, 1, tz"America/Winnipeg"), Inclusivity(false, true)) julia> r = st:Hour(1):(st + Day(1)) AnchoredInterval{-1 hour,ZonedDateTime}(ZonedDateTime(2014, 1, 1, tz"America/Winnipeg"), Inclusivity(false, true)):1 hour:AnchoredInterval{-1 hour,ZonedDateTime}(ZonedDateTime(2014, 1, 2, tz"America/Winnipeg"), Inclusivity(false, true)) julia> collect(r)[1:2] 2-element Array{AnchoredInterval{-1 hour,ZonedDateTime},1}: AnchoredInterval{-1 hour,ZonedDateTime}(ZonedDateTime(2014, 1, 1, tz"America/Winnipeg"), Inclusivity(false, true)) AnchoredInterval{-1 hour,ZonedDateTime}(ZonedDateTime(2014, 1, 1, 1, tz"America/Winnipeg"), Inclusivity(false, true)) julia> r[1:2] Error showing value of type StepRangeLen{AnchoredInterval{-1 hour,ZonedDateTime},AnchoredInterval{-1 hour,ZonedDateTime},Hour}: ERROR: MethodError: Cannot `convert` an object of type Hour to an object of type ZonedDateTime Closest candidates are: convert(::Type{T}, ::Interval{T}) where T at /Users/ericdavies/.julia/packages/Intervals/52k6l/src/interval.jl:124 convert(::Type{T}, ::AnchoredInterval{P,T}) where {P, T} at /Users/ericdavies/.julia/packages/Intervals/52k6l/src/anchoredinterval.jl:174 convert(::Type{T}, ::T) where T at essentials.jl:167 ... Stacktrace: [1] AnchoredInterval{-1 hour,ZonedDateTime}(::Hour) at /Users/ericdavies/.julia/packages/Intervals/52k6l/src/anchoredinterval.jl:79 [2] step(::StepRangeLen{AnchoredInterval{-1 hour,ZonedDateTime},AnchoredInterval{-1 hour,ZonedDateTime},Hour}) at ./range.jl:501 [3] show(::IOContext{REPL.Terminals.TTYTerminal}, ::StepRangeLen{AnchoredInterval{-1 hour,ZonedDateTime},AnchoredInterval{-1 hour,ZonedDateTime},Hour}) at ./range.jl:712 [4] show(::IOContext{REPL.Terminals.TTYTerminal}, ::MIME{Symbol("text/plain")}, ::StepRangeLen{AnchoredInterval{-1 hour,ZonedDateTime},AnchoredInterval{-1 hour,ZonedDateTime},Hour}) at ./show.jl:7 [5] display(::REPL.REPLDisplay{REPL.LineEditREPL}, ::MIME{Symbol("text/plain")}, ::StepRangeLen{AnchoredInterval{-1 hour,ZonedDateTime},AnchoredInterval{-1 hour,ZonedDateTime},Hour}) at /Users/ericdavies/.julia/packages/OhMyREPL/GFHgr/src/output_prompt_overwrite.jl:6 [6] display(::REPL.REPLDisplay, ::Any) at /Users/ericdavies/repos/julia1p2/usr/share/julia/stdlib/v1.2/REPL/src/REPL.jl:136 [7] display(::Any) at ./multimedia.jl:323 ``` Range indexing appears to be taking a `StepRange` and making a `StepRangeLen`, which we don't handle effectively. I think this might be a bug in `step(::StepRangeLen)`. Answers: username_0: I think that bug is there, but we can also avoid it by never making a `StepRangeLen`. We can prevent that by defining the traits `OrderStyle` and `ArithmeticStyle` to be the same as `AbstractTime`. We should also add `RangeStepStyle` while we're at it. username_0: See https://github.com/JuliaLang/julia/blob/master/base/range.jl and https://github.com/JuliaLang/julia/blob/master/base/traits.jl username_1: As of [Julia 1.5.0-DEV.207](https://github.com/JuliaLang/julia/pull/34563) (7d92a3aaed) this issue appears fixed: ```julia julia> r[1:2] AnchoredInterval{Hour(-1),ZonedDateTime}(ZonedDateTime(2014, 1, 1, tz"America/Winnipeg"), Inclusivity(false, true)):Hour(1):AnchoredInterval{Hour(-1),ZonedDateTime}(ZonedDateTime(2014, 1, 1, 1, tz"America/Winnipeg"), Inclusivity(false, true)) ``` Do we want to try and fix this on earlier versions of Julia? username_0: We should figure out the correct traits. If that fixes it for older versions of Julia, that would be a bonus. username_1: Another example of this issue but on Intervals 1.3 you just see a deprecation warning: ```julia julia> using Intervals julia> x = AnchoredInterval{-1}(1):1:AnchoredInterval{-1}(5) AnchoredInterval{-1,Int64,Open,Closed}(1):1:AnchoredInterval{-1,Int64,Open,Closed}(5) julia> y = x[1:4]; julia> y ┌ Warning: `convert(T, interval::AnchoredInterval{P,T})` is deprecated for intervals which are not closed with coinciding endpoints. Use `anchor(interval)` instead. │ caller = AnchoredInterval at anchoredinterval.jl:95 [inlined] └ @ Core ~/.julia/dev/Intervals/src/anchoredinterval.jl:95 ┌ Warning: `convert(T, interval::AnchoredInterval{P,T})` is deprecated for intervals which are not closed with coinciding endpoints. Use `anchor(interval)` instead. │ caller = AnchoredInterval at anchoredinterval.jl:95 [inlined] └ @ Core ~/.julia/dev/Intervals/src/anchoredinterval.jl:95 AnchoredInterval{-1,Int64,Open,Closed}(1):AnchoredInterval{-1,Int64,Open,Closed}(1):AnchoredInterval{-1,Int64,Open,Closed}(4) ``` username_1: I attempted to use traits to fix this problem but that didn't actually seem to be the source of the failure. I instead implemented the fix in #132 and made a [new issue for implementing traits](#133). Status: Issue closed
reuixiy/hugo-theme-meme
1025410447
Title: 关于shortcodes的扩展支持 Question: username_0: https://hugoloveit.com/zh-cn/theme-documentation-extended-shortcodes/ LoveIt主题的admonition很漂亮,如果不难的话希望加上这个feature XD Answers: username_1: 自己写了一些:https://github.com/username_1/shortcodes 演示:https://oi-io.me/tech/hugo-shortcodes/#note username_0: @username_1 非常酷! 等下班了试一下,感谢! 要是能PR到meme就好拉。 @reuixiy
geocollections/turvas
577375294
Title: Vaatluspunkti kaardivaade Question: username_0: Vaatluspunkti kaardivaates kuvatakse ala punkte, see on mugav ala uurimisel, aga vajalik oleks konkreetse valitud punkti eristamine kas suurema markeriga vms moel ning lisaks püsiv label kaardil. Püsiv label võiks olla alati kui on ainult üks aktiivne punkt kaardil - st ka proovi detailvaates.<issue_closed> Status: Issue closed
NVIDIA/VideoProcessingFramework
639325937
Title: Single GPU decoding some video Question: username_0: Hi, @username_1 : I looked at examples: SampleDecodeMultiThread.py, and tested our GPU(Single v100) decoding to the FPS 500. But I want a single GPU to decode multiple videos simultaneously. I found that they were serial-running decoded: one video, decode 6s ; two video , decode 12s How to make two video decoding process can call GPU decoding in parallel? Answers: username_1: Hi @username_0 To my best knowledge, this happens because of GIL. You can run multiple sessions in parallel threads and / or sub processes but you don't see much (or any) of a performance gain. Please look at the similar issue here: https://github.com/NVIDIA/VideoProcessingFramework/issues/62#issuecomment-641767610 I'm running out of depth here (don't have too much of a Python development expertise) so if you can try mentioned approach and come back as you get results that would be really great. username_0: thx @username_1 the #62 give me perfect solution! Status: Issue closed username_2: @username_0 Hi, I just adopt the multuprocessing method and the result is just the same as multi thread method. Have you achieve the improvements? could you just share the snip code for that? username_0: @username_2 sry,I have given up the research. username_2: Thanks for your reply. And I just want to know ,did you achieve idea twice improvement by using multiprocessing rather than multithread? username_0: no,only multithread
Azure/blobxfer
309790896
Title: AttributeError: 'NoneType' object has no attribute 'primary_endpoint' Question: username_0: Hi, I'm trying to use the High Level API to upload files to Azure Blob and I started from [this](https://blobxfer.readthedocs.io/en/latest/80-blobxfer-python-library/) from the documentation. I think I'm missing something because I don't understand how come I'm getting this kind of error, which is thrown by `create_destination_id` Here is the complete trace: ``` Traceback (most recent call last): File ".../upload.py", line 61, in <module> uploader.start() File "...\venv\lib\site-packages\blobxfer\operations\upload.py", line 1167, in start self._run() File "...\venv\lib\site-packages\blobxfer\operations\upload.py", line 1079, in _run create_destination_id(ase._client, ase.container, ase.name) File "...\venv\lib\site-packages\blobxfer\operations\upload.py", line 185, in create_destination_id return ';'.join((client.primary_endpoint, container, name)) AttributeError: 'NoneType' object has no attribute 'primary_endpoint' ``` These are the blobxfer parameters printed before the execution: ``` ============================================ Azure blobxfer parameters ============================================ blobxfer version: 1.1.1 platform: Windows-10-10.0.16299-SP0 components: CPython=3.6.4 azstor.blob=1.0.0 azstor.file=1.0.0 crypt=2.2.2 req=2.18.4 transfer direction: local -> Azure workers: disk=1 xfer=1 md5=1 crypto=0 log file: blobxfer.log resume file: None timeout: connect=30.0 read=30.0 mode: block skip on: fs_match=False lmt_ge=False md5=True delete extraneous: False overwrite: True recursive: True rename single: False access tier: hot chunk size bytes: 4194304 one shot bytes: 33554432 strip components: 0 store properties: attr=True md5=True rsa public key: None local source paths: ..\sourcefolder ============================================ ``` The code is like it follows, whereas config is a yaml file structured similarly to the sample configuration yaml [here](https://blobxfer.readthedocs.io/en/latest/sample_config.yaml). ``` upload_options = config['upload'][0]['options'] concurrency_dict = config['concurrency'] skip_on_options = upload_options.pop('skip_on', None) store_file_properties = upload_options.pop('store_file_properties', None) endpoint = "core.windows.net" concurrency = blobxfer.api.ConcurrencyOptions(**concurrency_dict) [Truncated] endpoint) source = blobxfer.api.LocalSourcePath() source.add_path('../sourcefolder') destination = blobxfer.api.AzureDestinationPath() destination.add_path_with_storage_account('container', 'storageaccountname') spec = blobxfer.api.UploadSpecification(upload_options, skip_on_options, source) spec.add_azure_destination_path(destination) uploader = blobxfer.api.Uploader(general_options, creds, spec) uploader.start() ``` I feel like I'm missing something trivial. Any help is appreciated. Thank you Answers: username_1: This call `destination.add_path_with_storage_account('container', 'storageaccountname')` must have a `creds` entry for `storageaccountname`. Please double check that is the case. You can also test getting the associated storage account with: ``` creds.get_storage_account(destination[0].lookup_storage_account('container')) ``` username_0: Hi @username_1, from what I see in the implementation of add_path_with_storage_account in `bloxfer/operations/azure/__init__.py`: ``` def add_path_with_storage_account(self, remote_path, storage_account): # type: (SourcePath, str, str) -> None """Add a path with an associated storage account :param SourcePath self: this :param str remote_path: remote path :param str storage_account: storage account to associate with path """ if len(self._path_map) >= 1: raise RuntimeError( 'cannot add multiple remote paths to SourcePath objects') rpath = blobxfer.util.normalize_azure_path(remote_path) self.add_path(rpath) self._path_map[rpath] = storage_account ``` both the parameters should be strings? That's why I used the storage account string. If I try to use `creds.get_storage_account(destination[0].lookup_storage_account('container'))` to check for the associated storage account, I get the following error: ``` Traceback (most recent call last): File ".../upload.py", line 55, in <module> print(creds.get_storage_account(destination.lookup_storage_account('container'))) File "...\venv\lib\site-packages\blobxfer\operations\azure\__init__.py", line 84, in get_storage_account return self._storage_accounts[name] KeyError: <blobxfer.operations.azure.StorageAccount object at 0x0000022CB4C85C88> ``` username_1: That's correct, it should be a string type. Sorry, my check was off, try: `creds.get_storage_account(spec.destinations[0].lookup_storage_account('container'))` username_0: Ok, I checked and the output is correct: the lookup of the storage account allows me to access the credentials I've set (name, endpoint, key, clients and so on) ``` sa = creds.get_storage_account(spec.destinations[0].lookup_storage_account('container')) print(sa.name, sa.endpoint, sa.key, sa.block_blob_client.primary_endpoint) ``` outputs correctly: `storageaccountname core.windows.net storageaccountkey storageaccountname.blob.core.windows.net` So, looking for the reason as to why ase._client would be None, I found the ase._client yielded by `_generate_destination_for_source` in `blobxfer/operations/upload.py` was None, because `populate_from_local` didn't assign the correct client to ase._client, as the mode passed "auto" (string) wasn't recognized as `StorageModes.Auto`. So, in the end, it was a really silly oversight on my part and used `mode=StorageModes.Auto` when creating the UploadOptions. Thank you, @username_1, for your help. Status: Issue closed
badcamp/camp_distro
311469327
Title: Text for Volunteer Page Question: username_0: ## Got Skills? Join Us + Volunteer! **Yes, you! BADCamp** would not be possible without love + support from volunteers! ### Volunteer opportunities include: Marketing + communications Day-of-coordination Registration desk help Store assistants + managers Room monitors + coordinators Stuff movers + badge checkers Room setup + maintenance AV, Internet + Photographers Signage distributors Venue setup + teardown ---WebForm ---
curiouslearning/workshop_drag_into_place_1
229487646
Title: User testing reflection (due Wednesday 5/31) Question: username_0: After you have completed user testing, please reflect on the experience and complete the assignment as [outlined here](https://docs.google.com/document/d/1uOWlF4pZqYn4c9yVkLotlZ3LFPSHrzdLvJbpeWvgfUQ/edit#heading=h.iyktjiy2ixr2) using the [following document as a template](https://docs.google.com/document/d/1VozsQ_IJCJ9cNHszWFiNQBSkpJ8jgd7sgB-_xkXGy94/edit?usp=sharing). When you have finished your write-up, you can [submit it as a Google Doc into this folder](https://drive.google.com/drive/folders/0B1Q2Y5s6EhXLMkNtRjBwcmVta1E?usp=sharing). Please note, unlike previous task cards, this will be due midweek on Wednesday 5/31 at 11:59PM as we wanted to give everyone more time to focus on completing the prototype and running a user test.
spring-cloud/spring-cloud-config
129642893
Title: Support placeholders for searchPaths on remote repos Question: username_0: To support a more flexible Git repo structure it would be good if the {placeholder} support was supported in the searchPaths option. Ideally we should be able to support having a single repo with config store under application (and profile) structure. Current implementation requires change to the searchPaths entry on the config server whenever a new application is added. ``` spring: cloud: config: server: git: uri: https://xxxx/config-repo.git searchPaths: {application} ``` Status: Issue closed Answers: username_0: I've tried this with 1.1.0.M5 but doesn't seem to work with either https or ssh git repo's. (application.yml) ``` spring: cloud: config: server: git: #uri: ssh://xxxxx:7999/mic/cloudconfig.git uri: https://xxxxxx/scm/mic/cloudconfig.git cloneOnStart: true searchPaths: {application} ``` ``` (cloudconfig repo) ├── application.properties ├── application.yml └── demo-app ├── demo-app-dev.yml ├── demo-app-prod.yml ├── demo-app.yml ├── devcloud.js └── simple.conf ``` Test - use URL: http://localhost:8888/demo-app/dev Expected result: 3 resource files returned (application.properties, application.yml, demo-app-dev.yml) Actual result: 2 resource files returned (application.properties, application.yml) username_1: Works for me. You need to quote the search paths value in YAML ("{" is not a valid literal). Maybe that's the problem? username_0: quoting in the yaml file fixed my issue. Thanks. ``` spring: cloud: config: server: git: #uri: ssh://xxxxx:7999/mic/cloudconfig.git uri: https://xxxxxx/scm/mic/cloudconfig.git cloneOnStart: true searchPaths: "{application}" ``` username_2: The solution with quotes is currently not working for me username_1: Probably a different issue then. This one is closed (you might have noticed?). It would be best if you can create a sample that reproduces the issue you see and then open another issue. username_3: I was also making same mistake, after quoting in doublesquotes it worked for me
flutter/flutter
349883943
Title: flutter run is error! Question: username_0: I tried to initialize the project is correct, I use the vscode editor, but, i Excuting an order `flutter run -v`, prompt : Connection timed out ! **I tried running it a dozen times, still the same error, which made me unable to understand how to use flutter!** @ianloic @paulcbetts @sgraham @sethladd @kevmoo ![image](https://user-images.githubusercontent.com/28696399/44010736-ba30f6b8-9ee6-11e8-8f74-e86c1a7e1ee3.png) ![tim 20180813122400](https://user-images.githubusercontent.com/28696399/44012768-03d4a3c0-9ef4-11e8-8334-c345c7634510.png) Answers: username_1: Please add the output of `flutter doctor -v`. username_1: How did you create the project? Does this also happen if you do ``` flutter create my_project cd my_project flutter run ``` in a terminal window? username_0: I have found the reason for this problem, because the network problem causes the dependency to be downloaded. thanks Status: Issue closed username_1: Thanks for the update! Sounds like this can be closed then.
explosion/spaCy
296125234
Title: How to convert simple NER format to spacy json Question: username_0: Currently I have data in the simple training data / offset format as shown in the docs: ``` TRAIN_DATA = [ ('Who is <NAME>?', { 'entities': [(7, 17, 'PERSON')] }), ('I like London and Berlin.', { 'entities': [(7, 13, 'LOC'), (18, 24, 'LOC')] }) ] ``` In order to utilize the CLI trainer, this data needs to be in spacys internal json training data format. to do this I am currently converting the simple offset format to conllner, then using the spacy cli converter. Is there a better way to directly get offset / simple format directly to spacy json format? Using the cli trainer seems to be able to train much faster than using the method specified [here](spacy/examples/training/train_ner.py) ## Your Environment * **spaCy version:** 2.0.5 * **Platform:** Linux-4.4.0-1049-aws-x86_64-with-debian-stretch-sid * **Python version:** 3.6.4 * **Models:** en_core_web_lg, en Answers: username_1: Yes, there's a [`gold.biluo_tags_from_offsets`](https://spacy.io/api/goldparse#biluo_tags_from_offsets) function that converts the entity offsets to a list of per-token BILUO tags: ```python from spacy.gold import biluo_tags_from_offsets doc = nlp(u'I like London.') entities = [(7, 13, 'LOC')] tags = biluo_tags_from_offsets(doc, entities) assert tags == ['O', 'O', 'U-LOC', 'O'] ``` username_0: Is there a way to save the bilou tags in the json format used by cli trainer? username_2: I think we currently lack a writer that does exactly that. I've been finding the current data-format situation frustrating too. Needs improvement. username_3: I have also the same problem in training NER. @username_2 @username_0 Is there any json sample training data for NER? username_1: You can find an the structure and an example of the training JSON format here: https://spacy.io/api/annotation#json-input So you should be able to write a script that produces data in this format, using the BILUO tags. I agree that spaCy should make this easier (and, as @honnial mentioned above, we're also thinking about overhauling the training format in general to make it easier to work with). username_3: Thanks @username_1 That is a very good idea, I'm looking forward to it. username_4: Same question that I posted on the chat channel today. I want my program to point to training data in a separate file rather than having it in the code. I currently have as follows based on the examples on the website <img width="1102" alt="screen shot 2018-04-03 at 2 44 06 pm" src="https://user-images.githubusercontent.com/4360712/38277517-838f5b46-374d-11e8-9b27-0b8331a3ad9b.png"> username_5: @username_4, recently I was looking for similar need. I figured a way to put the content to a separate file instead of in code. `inpt_text = open("text_file","r")` `TRAIN_DATA = inpt_text.read().replace('\n ', '')` `TRAIN_DATA = ast.literal_eval(' '.join(TRAIN_DATA.split()))` **text_file content format:** [ ('Who is <NAME>?', { 'entities': [(7, 17, 'PERSON')] }), ('I like London and Berlin.', { 'entities': [(7, 13, 'LOC'), (18, 24, 'LOC')] }) ] Plz share if there are any other approaches to maintain TRAIN_DATA in a separate file instead of in code. username_1: @username_4 @username_5 How you solve this depends on which file formats you like to work with. `ast.literal_eval` isn't *bad* – but it's also a very uncommon way to pass data around. And the transformations (replacing, splitting, joining) you have to do to read in the file can very easily lead to bugs. A simpler way might be working with JSON. Here's a an example of a file ```data.json```: ```json [ ["Who is <NAME>?", {"entities": [[7, 17, "PERSON"]]}], ["I like London and Berlin", {"entities": [[7, 13, "LOC"], [18, 24, "LOC"]]}] ] ``` JSON doesn't have tuples, so you need to use lists everywhere. In Python, you can then do the following: ```python import json import io with io.open('/path/to/data.json', encoding='utf8') as f: train_data = json.load(f) ``` Also remember that you can always re-write the training script to format your data differently. If you're working with large training sets, we also recommend to use the `spacy train` command instead – this will take care of batching up your examples correctly for optimal performance, and it'll also give you much more useful command line output. username_6: I am also about to write a script to convert from inline training format to CLI training format. However I train the model on sentences. So each paragraph entry will contain exactly one sentence instead. Is that fine? Since the NER model trains on context which are usually no more than sentences, I do not see a reason for paragraphs to be input with their raw text and then sentences. I think paragraphs are there for Dependency Parsing. Is that right? username_6: I am pasting an answer derived from [stackoverflow](https://stackoverflow.com/questions/48916768/how-do-i-convert-simple-training-style-data-to-spacys-command-line-json-format) which is not showing up easily in searches. Ofcourse you will have to improvise a bit if you have multiple paragraphs. ``` def getcharoffsetsfromwordoffsets(doc,entities): charoffsets = [] for entity in entities: span = doc[entity[0]:entity[1]] charoffsetentitytuple = (span.start_char, span.end_char, entity[2]) charoffsets.append(charoffsetentitytuple) return charoffsets def convertspacyapiformattocliformat(nlp, TRAIN_DATA): docnum = 1 documents = [] for t in TRAIN_DATA: doc = nlp(t[0]) charoffsetstuple = getcharoffsetsfromwordoffsets(doc,t[1]['entities']) tags = biluo_tags_from_offsets(doc, charoffsetstuple) ner_info = list(zip(doc, tags)) tokens = [] sentences = [] for n, i in enumerate(ner_info): token = {"head" : 0, "dep" : "", "tag" : "", "orth" : i[0].string, "ner" : i[1], "id" : n} tokens.append(token) sentences.append({'tokens' : tokens}) document = {} document['id'] = docnum docnum+=1 document['paragraphs'] = [] paragraph = {'raw': doc.text,"sentences" : sentences} document['paragraphs'] = [paragraph] documents.append(document) return documents ``` username_7: @username_1 is "spacy convert" command the solution for this issue? The reason why I am asking this is I don't see a converter for BILOU. Is that something currently mot supported? username_8: Hello everyone, I have the same question as @username_7 about the input data format which I have to pass to convert function in order to train my NER for my custom entities. Right now, my data looks like simple NER format with about 10 labels, when I try to pass it to the converter function it give me this error: Unknown format Can't find converter for json. Could you please help me: what am I doing wrong? username_2: In v2.1.0a1, this is now finally be easier. Thanks for your patience with the annoying format, which has a few extra levels of nesting so that it can accomodate whole document annotations, along with paragraph structure. In v2.1.0a1, there's now a new function `spacy.gold.docs_to_json`. This function takes a list of `Doc` objects and outputs properly formed json data, for use in `spacy train`. You just have to set up the `Doc` objects to have the annotations you want. For named entities, this is as easy as writing to the `doc.ents` attribute. For the parser, the best approach is to create an array with the head offsets you want, and then import the data with `doc.from_array()`. Status: Issue closed
aurelia/ui-virtualization
188768990
Title: Rebinding List to New Array, List Still Shows Some of the Old Items Question: username_0: **I'm submitting a bug report** * **Library Version:** major.minor.patch-pre 1.0.0-beta.3.0.1 **Please tell us about your environment:** * **Operating System:** Windows 7 * **Node Version:** 6.9.0 * **NPM Version:** 3.10.8 * **JSPM JSPM 0.16.45 * **Browser:** Chrome * **Language:** ES6 **Current behavior:** When you rebind the virtual list to a new array, e.g. this.myArray = getNewData() Then the list rebinds, but some of the items from the old array are still present. * **What is the expected behavior?** Not that! List should only contain new items after a rebind. * **What is the motivation / use case for changing the behavior?** I'm using the virtual-repeat.for list to display search results. If they enter a new search, I want the old results to go away, and the new results to display. Answers: username_1: Could you set up a basic reproducible gist for us to look at? Definitely seems strange. We're currently replacing our list in a similar manner for searching (although it's a bit different), and we don't have a problem. Might be a specific way you're doing it... But I'm not sure. username_0: I'll have to do it this weekend. Thanks for the quick reply! Aaron username_0: I will say that we have a bunch of Compose elements in there. Don't know if that's relevant. Thanks again. Aaron username_1: No problem, I just so happened to have my browser open and saw this issue fly in :) Compose elements _might_ be a contributing factor.... But I wouldn't bet on it. I'd really just need to see what's going on to know for sure. I'm usually out over the weekend, so if nobody else gets a chance to look at it, I'll try to take a peek on monday. username_0: OK, I feel stupid. How do you add aurelia and this plugin to a gist? I've never used gist before, just JSFiddle. Can you point me to a good resource on gist? Thanks. username_1: Don't feel stupid, sometimes Gist gives me a run for my money too :) If you can't get it to happen in a gist, just make a repository that I can clone and try it on. A gist example of getting a custom element to work is here: https://gist.github.com/username_1/3a20dc24ca5c32c4b29fe6990dd30156 I don't have an example with a plugin right now. username_2: @username_0 here's a gist you can use as a starting point: https://gist.run/?id=7542e061bc940cde506b username_1: Related to: #106 Working on resolving there. username_1: Resolved (hopefully) by https://github.com/aurelia/ui-virtualization/pull/107 Status: Issue closed
modelleicher/FS19_simpleIC
773971246
Title: [Suggestion] Possibiliti to add external control panels from machines to tractors Question: username_0: Lot of machines in real life use control panels which show capacity and precisely doses fertilizer or herbicide. 4D Modding realise baler with computer (Hesston Bale Control) and you have to add it in i3d. I think Simple IC should give a possibility to change possition of this parts, now game don't give this option and many people add static models od this computers without working dashes and icons, in my opinion this small addon to script can add more realism and fun for players which wont to play like in real life. I have small script converted from old Farming Simulator but it based on old IC panel and doesn't working exactly as it should. Answers: username_1: Hi, do you mean SimpleIC should support the adding of those panels or the controlling of them? Because the latter wouldn't be a "small addon" much rather a huge new script.. username_0: I think about points from machine using inside tractor username_1: With the addition of implementControl at some point a few machine functions will be controllable via tractor username_0: But you can create your own points with animations in machine like door animation in tractor?
EnoxSoftware/OpenCVForUnity
268713474
Title: MobileNetssdExample Question: username_0: I am trying to get object detection via caffee model but whenever i try to do my unity editor gets crash. Answers: username_1: Are you using the following files? MobileNetSSD_deploy.prototxt https://github.com/chuanqi305/MobileNet-SSD/blob/master/MobileNetSSD_deploy.prototxt dnn/MobileNetSSD_deploy.caffemodel https://drive.google.com/open?id=0B3gersZ2cHIxRm5PMWRoTkdHdHc username_2: I'm having problems getting this to run, also. I'm using the two files mentioned above. 3058607249(checksum) 25036(bytes) MobileNet_deploy.prototxt 3242836829(checksum) 17028869(bytes) MobileNet_deploy.caffemodel it seems to be related to the prototxt file and/or format? on android i get this error at runtime: ' 04-12 15:25:24.590 12522-12574/com.cr3d.mssd E/libprotobuf-native: [libprotobuf ERROR /Users/satoo/opencv/android/opencv/3rdparty/protobuf/src/google/protobuf/text_format.cc:288] Error parsing text-format opencv_caffe.NetParameter: 7:1: Expected identifier, got: < 04-12 15:25:24.590 12522-12574/com.cr3d.mssd E/cv::error(): OpenCV(3.4.1-dev) Error: Unspecified error (FAILED: ReadProtoFromTextFile(param_file, param). Failed to parse NetParameter file: /storage/emulated/0/Android/data/com.cr3d.mssd/files/opencvforunity/dnn/MobileNetSSD_deploy.prototxt) in void cv::dnn::ReadNetParamsFromTextFileOrDie(const char*, opencv_caffe::NetParameter*), file /Users/satoo/opencv/android/opencv/modules/dnn/src/caffe/caffe_io.cpp, line 1145 04-12 15:25:24.590 12522-12574/com.cr3d.mssd E/org.opencv.dnn: dnn::readNetFromCaffe_10() : OpenCV(3.4.1-dev) /Users/satoo/opencv/android/opencv/modules/dnn/src/caffe/caffe_io.cpp:1145: error: (-2) FAILED: ReadProtoFromTextFile(param_file, param). Failed to parse NetParameter file: /storage/emulated/0/Android/data/com.cr3d.mssd/files/opencvforunity/dnn/MobileNetSSD_deploy.prototxt in function void cv::dnn::ReadNetParamsFromTextFileOrDie(const char*, opencv_caffe::NetParameter*) ` the first lines of the prototxt file seem correct, there is no "<" `name: "MobileNet" input: "data" input_shape { dim: 1 dim: 3 dim: 224 dim: 224 } layer { name: "conv0" type: "Convolution" bottom: "data" ` have any suggestions? thanks! username_1: Model files can be downloaded using this script. https://github.com/opencv/opencv_extra/blob/master/testdata/dnn/download_models.py MobileNet_deploy.prototxt 29,353 bytes MobileNet_deploy.caffemodel 23,147,564 bytes username_2: Thanks! This all works for me now! @username_1 I wanted to ask your advice/opinions of caffe or tensorflow with a webcam on android? your caffe example seems to perform the best of your object detection/localization demos..... How should I ask this question? email? wiki? issue? thanks! Status: Issue closed username_1: @username_2 Could you send a question to Technical Inquiry? https://enoxsoftware.com/opencvforunity/contact/technical-inquiry/
google/material-design-icons
120601213
Title: give warning if injectTapEventPlugin() not called -> items not interacting with mouse Question: username_0: Hey, just wondering if it would be possible to give a console warning if you didn't call injectTapEventPlugin. I just started a new project and forgot to do that, and at some point discovered that nothing responded to my mouse input. (perhaps because my laptop also supports touch?) Took me a while to figure out this was the missing part. Status: Issue closed Answers: username_1: Is this issue intended for the material-design-icons project? I'm closing this as unrelated, but feel free to reply if I'm mistaken. username_0: sorry, yes wrong project
kszytko/object-oriented-programming
648806274
Title: cmake w SHM Question: username_0: - [ ] Użyjcie zmiennej ${PROJECT_NAME} - [ ] Lista plików cpp w zmiennej - [ ] Wszystko poza plikiem main.cpp powinno kompilować się do biblioteki statycznej - [ ] Binarka (main.cpp) powinna linkować się z powyższą biblioteką. - [ ] Napiszcie odpowiedni plik .github/workflows/main.yml który spowoduje, że GitHub będzie automatycznie uruchamiał kompilację projektu dla każdego nowego commita.
CocoaPods/Xcodeproj
210399502
Title: Target has two dependencies, "each remove from project" two times? Question: username_0: # once only remove one dependency? ``` require 'xcodeproj' # ruby removeDe.rb "path of project" project_path = ARGV[0] project = Xcodeproj::Project.open(project_path) # puts project project.targets.each do |target| if target.name == 'MyTarget' target.dependencies.each { |e| puts e.remove_from_project } end end project.save project_path = ARGV[0] project = Xcodeproj::Project.open(project_path) # puts project project.targets.each do |target| if target.name == 'MyTarget' target.dependencies.each { |e| puts e.remove_from_project } end end project.save ``` Answers: username_0: How to delete all dependencies at once? username_1: Hey there, this looks like a great example of a question that you should ask on [StackOverflow](http://stackoverflow.com) - there is an active tag [CocoaPods](http://stackoverflow.com/tags/cocoapods) where people can help out with questions like this. We try to keep the Xcodeproj issues focused only on bug reports for the tool, and for upcoming feature requests. Which this doesn't look like. username_0: Thank you for your information But I think is a bug target.dependencies.each { |e| puts e.remove_from_project } Should remove all dependencies username_2: same here, have to run this twice to remove my watch target dependency: target.dependencies.each { |e| puts e.remove_from_project }
billie66/hdsp
127937529
Title: http://7xnm4l.com1.z0.glb.clouddn.com/004-git-init.mp4 Question: username_0: http://7xnm4l.com1.z0.glb.clouddn.com/004-git-init.mp4 Answers: username_0: { "title": "the first episode", "video": "167-css-length.mp4" } { "video": "168-youziku.mp4" } { "video": "169-bye-rails.mp4" } { "video": "170-spa.mp4" } { "video": "171-h5.mp4" } { "video": "172-github-signup.mp4" } { "video": "173-github-new-repo.mp4" } { "video": "174-h5-tags.mp4" } { "video": "175-100-pages.mp4" } { "video": "176-css3-start.mp4" } { "video": "177-transition.mp4" } { "video": "178-gulp.mp4" } { "video": "179-npm-es6.mp4" } { "video": "180-why-meteor-react.mp4" } Status: Issue closed
tensorflow/tensorflow
245249629
Title: tf.one_hot indices out of bound Question: username_0: array([ 0., 0., 0., 0.], dtype=float32) ``` Both overflow and underflow result in a zeroed array. Should this be caught, and errors be thrown? If not, why and when would this zeroed behaviour make sense? Or is it costly to check and throw assertions? Status: Issue closed Answers: username_1: This is indeed intentional. See the examples on the [C++ op documentation](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/core/ops/array_ops.cc#L4494). We cannot change this without breaking backwards comptability.
vcu-swim-lab/stack-intheflow
182337721
Title: Figure out how to rank query results based on relevance Question: username_0: This is an issue of research, and is specifically about how to rank results from a query, not how that query is generated (that's another issue: https://github.com/vcu-swim-lab/stack-intheflow/issues/6). We've talked about this a few times and have brought up several approaches including: - The algorithm in the Prompter paper - Using Stack Overflow's order of results, most likely as a starting point - Using some machine learning technique to learn ranking order based on implicit feedback (i.e. if people tend to click the 5th result most often, that should be the highest ranking result). To be clear, once we've agreed on a general algorithm/approach to the problem, this issue will be complete.<issue_closed> Status: Issue closed
luvit/luvit
417242438
Title: uv.spawn() might fail Question: username_0: [uv.spawn()](https://github.com/luvit/luvit/blob/master/deps/childprocess.lua#L133) might fail, e.g. because of `ENOMEM: not enough memory`. In this case the handle is `nil` and the `pid` has the value of the error message. `uv.spawn()` not returning a valid `handle` should be handled as an error and reported to the user. Answers: username_1: Similar issue from when this affected `coro-spawn`: https://github.com/luvit/lit/pull/220 Status: Issue closed username_0: Closed, as the errors from uv.spawn() are actually handled. Not the way I thought it should be, but after studying the sources, e.g. looking at [test-process.lua](https://github.com/luvit/luvit/blob/master/tests/test-process.lua), I understand how to get the error messages.
InsertKoinIO/koin
539779824
Title: Restarting single instances in Koin Question: username_0: I'm trying to implement a session mechanism, where the user can logout and login at any moment with other account. In order to avoid having inconsistent information, I wanted to restart some instances within a module that I like this: ``` val sessionRepositories = module { single { UserSessionRepository(androidApplication(), get(), get()) } bind AbstractSessionRepository::class } ``` To this I tried to do: ``` unloadKoinModules(sessionRepositories) loadKoinModules(sessionRepositories) ``` And for what I know, this will drop the previous UserSessionRepository instance and will create a new one, creating a leak with the previous UserSessionRepository. Therefore, I tried to do: ``` MyApplication.getInstance().stopKoinInstance() //calls GlobalContext.stop() MyApplication.getInstance().startKoinInstance() //restarts all instances, just to see if there was something with load and unload ``` And this keeps leaking, is there a way, where I can really restart my instances Answers: username_1: I did a session management like what you want using a custom scope. It works pretty well. After user logins, you can start scope, and when user logouts, you can close the session. https://doc.insert-koin.io/#/koin-core/scopes?id=using-a-scope Status: Issue closed
rkotze/git-mob-vs-code
497648032
Title: Open git-coauthors file from sidebar doesn't work on Windows Question: username_0: <!-- Have you read our Code of Conduct? By filing an Issue, you are expected to comply with it, including treating everyone with respect: https://github.com/username_1/git-mob-vs-code/blob/master/CODE_OF_CONDUCT.md --> ### Prerequisites - [X] [Checked that your issue isn't already filed][git-mob-vs-code issues] - [X] Tried upgrading to the latest git-mob-vs-code and git-mob version (`npm i -g git-mob`) - [X] Tried upgrading to the latest git version (`brew upgrade git` if installed with Homebrew/Linuxbrew) [git-mob-vs-code issues]: https://github.com/username_1/git-mob-vs-code/issues?utf8=%E2%9C%93&q=is%3Aissue ### Description Clicking the `Open git-coauthors file` button in the sidebar results in an error. The replace regex in [git-mob-coauthors-file.js#L13](https://github.com/username_1/git-mob-vs-code/blob/893549652f6455396c3df7ca7099b2eca3c0a58c/src/git-mob-coauthors-file.js#L13) removes the drive letter which results in wrong path. `file:///Users/r15ch/.git-coauthors` or `\Users\r15ch\.git-coauthors` ### Steps to Reproduce 1. On Windows 10 using VSCode Portable 2. Click `Open git-coauthors file` button **Expected behaviour:** [What you expect to happen] Should open `.git-coauthors` file **Actual behaviour:** [What actually happens] ![image](https://user-images.githubusercontent.com/432127/65510195-a689ea00-ded4-11e9-9dcb-78c482756a3e.png) ### Versions - operating system and version: Windows 10 - git-mob-vs-code version (can be found in the Extensions tab): 1.1.1 - git-mob version (`git-mob --version`): 1.0.1 - git version (`git --version`): 2.20.1.windows.1 ### Additional Information <!-- Any additional information, configuration or data that might be necessary to reproduce the issue. --> Answers: username_1: Interesting, I specifically remove the drive letter to make ti work on Windows - [Commit](https://github.com/username_1/git-mob-vs-code/commit/4bbc868daebe3bd2f9b4978330131092c67bf4fb) - [Current test file](https://github.com/username_1/git-mob-vs-code/blob/master/src/git-mob-coauthors-file.spec.js) I think the difference being you run the portable VS code version. I wonder if that has anything to do with it. I'll try to find time to investigate it. username_1: @username_0 I've setup Git Mob on a portable version of VS Code and was able to open the `.git-coauthors` file. Where you able to follow the path in the error message to the file? I'm just wondering if the path is incorrect? By chance are you using the environment variable `GITMOB_COAUTHORS_PATH`? username_0: The path itself is correct, except the missing drive letter. `GITMOB_COAUTHORS_PATH` is not set. It works just fine if I remove `.replace(/^([a-z]:[\\]|[\/])/gi, "")` from https://github.com/username_1/git-mob-vs-code/blob/f43599280b33b9e6e7b7cf52974f1866d2b4b22d/src/git-mob-coauthors-file.js#L18 username_0: Example output after adding test output. ![image](https://user-images.githubusercontent.com/432127/66426583-74b76e00-ea12-11e9-978b-96aea17ddb64.png) ```js get path() { vscode.window.showErrorMessage("without regex: " + path.join(os.homedir(), this.file)); vscode.window.showErrorMessage("with regex: " + path.join(os.homedir(), this.file).replace(/^([a-z]:[\\]|[\/])/gi, "")); return path .join(os.homedir(), this.file); } ``` username_1: Interesting, I'm on windows 10 and used the latest version of git-mob, git-mob-vs-code and VS Code. The only thing I don't know is what version of VS Code do you have? As far as I understand Portable VS Code, it does **not** auto-update. Maybe that version does not support file paths in the same way as the latest. username_0: ![image](https://user-images.githubusercontent.com/432127/66466775-8be27400-ea83-11e9-97a6-593f30e2c704.png) Installed via [Scoop](https://github.com/lukesampson/scoop/) to `D:\scoop\apps\vscode-portable\current` username_1: I see, the issue could be that is looking for .git-coauthors on your D drive but it's on your C drive. This is easy to fix for windows, just need to think about Unix based systems. username_0: Shouldn't be `os.homedir()` enough? The directory separators are correct in both cases. ```js const os = require('os') const path = require('path') console.log(path.join(os.homedir(), '.git-coauthors')) ``` ![image](https://user-images.githubusercontent.com/432127/66470170-5b9dd400-ea89-11e9-9406-766a45ffacb0.png) Test in Ubuntu WSL username_1: I remember the issue was on a Mac. So I'll need to confirm the path on a Mac. When I first did this, I thought `os.homedir()` was the solution. username_1: I've tried it on MAC OS and works fine. So there is no need to remove the drive letter. I'll get this fix out soon. username_1: Will be part of release 1.3 Status: Issue closed
jabieniek/mlproject
390478674
Title: Summary Question: username_0: ## Rubric Score ### Criteria 1: Valid Python Code * _Score Level:_ 4 (Meets Expectations) * _Comment(s):_ All code is valid without errors. ### Criteria 2: Exploration of Data * _Score Level:_ 3 (Meets Expectations) * _Comment(s):_ Data is explored well but could use a but more varying methods of data visualization such as histograms and also a correlation table would have been very useful to you. ### Criteria 3: Machine Learning Techniques used correctly * _Score Level:_ 3 (Meets Expectations) * _Comment(s):_ ML Techniques are used for their correct purposes but results are not confirmed by other similar models and results are not explained very well. If a model produces results that are not meaningful you should hypothesize as to why that is and try to confirm it. ### Criteria 4: Report - Are conclusions clear and supported by data? * _Score Level:_ 3 * _Comment(s):_ Conclusion is clear and good job hypothesizing about reasons your models were not very useful. Evidence with number would have been helpful as well. ### Criteria 5: Code formatting * _Score Level:_ 3 * _Comment(s):_ Code is formatted well with sufficient comments. A couple extra comments explaining what is happening would help. ### Overall Score: 16/20 Overall good job! Even though your models did not provide meaningful results it is good you at least commented as to why. My only criticisms are that data could have been explored more efficiently and conclusions could have been presented with more solid evidence. Keep it up and happy coding!
google/docsy
993248563
Title: Link in community_links.html partial should be relative Question: username_0: I think the `href` on line 17 of should not be absolute but rather relative so that it still works when there is a base URL. I suggest it should be: `<p>You can find out how to contribute to these docs in our <a href="../docs/contribution-guidelines/">Contribution Guidelines</a>.` Rather than: `<p>You can find out how to contribute to these docs in our <a href="/docs/contribution-guidelines/">Contribution Guidelines</a>.` In it's current form it will link to `localhost:1313/docs/contribution-guidelines/` rather than `localhost:1313/<mybase-url>/docs/contribution-guidelines` If you think this this correct I can make a pull request to correct this. Answers: username_1: Oh yes, I think you're right! Please do make a pull request. username_0: Hi There, I just submitted the PR https://github.com/google/docsy/pull/687 username_1: Fixed in #687 Status: Issue closed
ikedaosushi/tech-news
518003835
Title: Docker再入門 Question: username_0: Docker&#20877;&#20837;&#38272;<br> Docker&#12395;&#12354;&#12414;&#12426;&#35302;&#12428;&#12427;&#27231;&#20250;&#12364;&#23569;&#12394;&#12367;&#12289;&#12383;&#12414;&#12395;&#20351;&#12358;&#12392;&#12467;&#12510;&#12531;&#12489;&#12392;&#12363;&#20181;&#32068;&#12415;&#12434;&#12377;&#12368;&#12395;&#24536;&#12428;&#12390;&#12375;&#12414;&#12358;&#12290;&#12381;&#12435;&#12394;&#33258;&#20998;&#12408;&#12398;&#12490;&#12524;&#12483;&#12472;<br> https://ift.tt/2JPMgFU
SkyrimLL/SDPlus
164692360
Title: Set up basic quest for Alicia player start (1 quest stage) Question: username_0: - improved Alicia player start (only 1 quest stage for now, and possible enslavement right off the start if the sanguine quest alias is assigned at that point... I noticed sometimes it is still 'none' by the time the enslavement event kicks in).<issue_closed> Status: Issue closed
MelvorIdle/melvoridle.github.io
823129570
Title: Agility: Cliff balance impact on prayers Question: username_0: **Describe the bug** - Agility "cliff balance" (5) - Says "-2 flat prayer points" in red - Prayers UI apply it as -2 (dmg reduction listed as 3 pts instead of 5) making it a buff? - Does not seems right as it would not make it an obstacle very balanced (should be +2 for a total of 7?) - Tested briefly, still cost 5 to activate said prayer **Expected behavior** TBD. Probably to affect cost of prayer as an increase and to be reflected in both UI and activation cost **Screenshots** If applicable, add screenshots to help explain your problem. **Browser** Chrome latest (88.0.4324.190) **Console output** N/A **Are you using any scripts?** None (Native) Answers: username_1: This has been hotfixed Status: Issue closed
rmc47/CamLog
365726672
Title: If the user enters a QSO before the QRZ lookup is finished, the locator appears in the wrong location Question: username_0: While operating GB18NH I noticed that if I entered a QSO (hit return) before the QRZ.com lookup had finished, then the locator found in the QRZ.com lookup is inserted in the blank/new QSO log line. For simplicity, let us assume there has been no QSOs entered in the log, so the first QSO entered would be serial 001; 1. The operator, with CamLog open in front of them, calls CQ and 'M1BXF' responds. 2. The operator enters 'M1BXF' into the callsign box for serial 001 and hits TAB. 3. CamLog looks up 'M1BXF' on QRZ.com to fetch the locator. 4. The Operator completes QSO 001 and enters it, by hitting return before the QRZ.com lookup is complete. 5. The locator found for 'M1BXF' on the QRZ.com lookup is entered into the locator box for serial 002. I believe this will only be noticed while on a slow internet connection. It is also likely that the locator for QSO 001, which is entered into QSO 002, will be overwritten when the operator enters a new callsign into QSO 002 and hits TAB, which will trigger another QRZ.com lookup and will overwrite the locator.
ynfatal/ynfatal.github.io
627677597
Title: SpringBoot2 | 第六篇:整合SpringDataJPA | Fatal's Blog Question: username_0: https://username_0.github.io/2018/09/30/SpringBoot2/SpringBoot2%E7%AC%AC%E5%85%AD%E7%AF%87%E6%95%B4%E5%90%88SpringDataJPA/ 上一篇介绍了Spring JdbcTemplate的使用,对比原始的JDBC而言,它更加的简洁。但随着表的增加,重复的CRUD工作让我们苦不堪言,这时候Spring Data Jpa的作用就体现出来了….. [TOC] JPA​ JPA 是 Java Persistence API 的简称,中文名 Java持久层API,是JDK 5.0注解或XML描述对象-关系表的映射关系,并将运行
wojtekmaj/react-pdf
1147855388
Title: _interopRequireWildcard is not a function Question: username_0: ### Before you start - checklist - [X] I followed instructions in documentation written for my React-PDF version - [X] I have checked if this bug is not already reported - [X] I have checked if an issue is not listed in [Known issues](https://github.com/wojtekmaj/react-pdf/wiki/Known-issues) - [ ] If I have a problem with PDF rendering, I checked if my PDF renders properly in [PDF.js demo](https://mozilla.github.io/pdf.js/web/viewer.html) ### Description So i've installed @react-pdf/renderer in my project and i've tried to test the basic example, like creating a basic document and trying to download it And I got some incompatibility with the webpack ( 4.4.2) : var R = _interopRequireWildcard(__webpack_require__(/*! ramda */ "../../../../modules/my-project/code/node_modules/ramda/es/index.js")); I tried adding ramda manually but it wouldn't work ### Steps to reproduce Install the module and try the basic example ### Expected behavior Should see the pdf ### Actual behavior Not working as expected ### Additional information _No response_ ### Environment - **Browser (if applicable)**: - **React-PDF version**: - **React version**: - **Webpack version (if applicable)**:
Welfenlab/tutor-student
113148631
Title: Browserified files should contain absolute paths Question: username_0: Strings like `/home/maik/Gits/tutor-meta/tutor-student/node_modules/...` should not appear in any distributed files. Answers: username_0: Even better than configuring browserify not to include full paths: https://www.npmjs.com/package/unpathify Status: Issue closed username_1: Strings like `/home/maik/Gits/tutor-meta/tutor-student/node_modules/...` should not appear in any distributed files. username_1: Did you check that your change didn't break the discify builder. It requires full paths to analyze the packages. username_0: The plugin only works if `fullPaths: false`, which is only set if `NODE_ENV == "production"`, so discify will work. Status: Issue closed
remisharrock/c-programming-with-linux-MOOC-issues-tracker
365561533
Title: Taskgrader Question: username_0: Which course: 2 Language Foundations What type of content: Taskgrader URL: https://courses.edx.org/courses/course-v1:Dartmouth_IMTx+DART.IMT.C.02+1T2018/courseware/a86febae1149479d971ae6edbd850c51/fc9008c149444ca0a623cc332fc01e06/?child=first Description: Some euros (for the children) remains in the problem statements -) other comments: your name: Isa screenshot: Codecast timing: Tasgrader code submitted: Operating system: which web browser? Status actuel: Answers: username_1: resolved Status: Issue closed
jlippold/tweakCompatible
362467602
Title: `App Percent` working on iOS 9.3.3 Question: username_0: ``` { "packageId": "com.hackyouriphone.apppercent", "action": "working", "userInfo": { "arch32": false, "packageId": "com.hackyouriphone.apppercent", "deviceId": "iPhone8,2", "url": "http://cydia.saurik.com/package/com.hackyouriphone.apppercent/", "iOSVersion": "9.3.3", "packageVersionIndexed": false, "packageName": "App Percent", "category": "HYI - Tweaks", "repository": "HackYouriPhone", "name": "App Percent", "packageIndexed": true, "packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.", "id": "com.hackyouriphone.apppercent", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.0.7", "shortDescription": "AppStore Installation in Percent (%) and Device Notification when an app download is finished. Configure option from Settings!", "latest": "2.0.0", "author": "pxcex", "packageStatus": "Unknown" }, "base64": "<KEY>", "chosenStatus": "working", "notes": "" } ```
explosion/spaCy
1059953122
Title: EntityRecognizer throws IndexError when used in pipeline with Transformer and custom span getter Question: username_0: EntityRecognizer throws IndexError when used in pipeline with Transformer and custom span getter: ``` File "/home/---/---/research_spacy_ru/.venv/lib/python3.8/site-packages/spacy/language.py", line 1122, in update proc.update(examples, sgd=None, losses=losses, **component_cfg[name]) File "spacy/pipeline/transition_parser.pyx", line 416, in spacy.pipeline.transition_parser.Parser.update File "spacy/ml/parser_model.pyx", line 293, in spacy.ml.parser_model.ParserStepModel.finish_steps File "spacy/ml/parser_model.pyx", line 456, in spacy.ml.parser_model.precompute_hiddens.begin_update.backward File "/home/---/---/research_spacy_ru/.venv/lib/python3.8/site-packages/spacy/ml/_precomputable_affine.py", line 49, in backward Xf = X[ids] IndexError: index 221 is out of bounds for axis 0 with size 221 ``` ## How to reproduce the behaviour I created my custom span_getter: https://gist.github.com/username_0/06e53b108f764e7240ea7ae8e2e830fd It adapts number of words to respective number of word pieces, to better fit into transformer window. Pipeline works with this function, the exception is thrown only at some documents. I plug it into simple transformer + ner pipeline like this: https://github.com/username_0/natasha-spacy/blob/transformer-pipeline/project/config_trf.cfg (in my tests I disabled all but transformer and NER) This error is emitted at the line https://github.com/explosion/spaCy/blob/master/spacy/ml/_precomputable_affine.py#L49 ## Your Environment - **spaCy version:** 3.2.0 - **Platform:** Linux-5.4.0-90-generic-x86_64-with-glibc2.29 - **Python version:** 3.8.10 - **spacy-transformers** = "^1.0.6" - **torch** = "1.9.1" * Operating System: Ubuntu 20.04 * Python Version Used: Python 3.8.10 * spaCy Version Used: 3.2.0 Answers: username_1: Thanks for the report and sorry it's taken us a long time to follow up on this. Unfortunately, because the issue is happening deep in the spaCy internals and your custom code isn't very simple, it's hard to be sure what's going on here. Can you create a small example we can run to reproduce the problem? A repo like the one you linked to with a project file would be great, but that repo's project file doesn't seem to work and doesn't use Transformers anyway. username_0: Thanks for your reply. I reproduced the behavior based on one of spaCy tutorials: https://github.com/username_0/tutorial_spacy_custom_span_getter The only changes I do are: - I add my span getter (I added more comments to make its algorithm more clear) - I alter config to use my transformer of choise And the error remains. P.S. The repo I linked in my first message does use transformer config, in project file it's called by "train_trf" and not "train" - to be able to use both configs.
fossasia/pslab-android
398692327
Title: Multimeter should show a default value of 0.00 Question: username_0: **Actual Behaviour** Currently, the multimeter shows blank by default **Expected Behaviour** It should show a default value of 0.00 **Steps to reproduce it** Go to Multimeter Activity **Would you like to work on the issue?** Yes<issue_closed> Status: Issue closed
AtsushiSakai/PythonRobotics
512934802
Title: Informed RRT* incorrectly update cBest for ellipse Question: username_0: **Describe the bug** In function [informed_rrt_star_search](https://github.com/username_1/PythonRobotics/blob/master/PathPlanning/InformedRRTStar/informed_rrt_star.py#L38), `cBest` and `pathLen` seems have same meanings. After finding a new path, `pathLen` is used to check whether `cBest` should be updated, but `pathLen` itself is not being updated. The performance of the bug is that the ellipse may increase sometimes when you find a new path. Also, if you print `cBest` everytime ploting the ellipse, you'll find that `cBest` may increase sometimes. **Screenshots** Here is an example: There is already a path with length 11.808038783431691. ![11-808038783431691](https://user-images.githubusercontent.com/11571437/67631568-31467600-f8d3-11e9-96d5-4f0c9e1c2102.png) After a new path with longer length 12.487373607747500 is found, cBest is updated. ![12-487373607747500](https://user-images.githubusercontent.com/11571437/67631570-373c5700-f8d3-11e9-917f-07d4840018ca.png) Answers: username_0: Simply fix: #248 Status: Issue closed
hundertzehn/mocoapp-api-docs
634724038
Title: [FEATURE REQUEST] - Endpoint for sending invoices via email Question: username_0: Hey there! I already had a nice chat with Thomas about this. We would like to automate our invoicing. For that it would be necessary to send an invoice via API. For now - as a workaround - we send the emails using an email service and attach the invoices manually using the [/invoices/{id}.pdf](https://github.com/hundertzehn/mocoapp-api-docs/blob/master/sections/invoices.md#get-invoicesidpdf) endpoint. Greetings to Switzerland, Corvin Answers: username_1: We'll have a look at this @username_0. But it might take some time. Pretty busy atm with new features to come :muscle: username_2: This is now available: https://github.com/hundertzehn/mocoapp-api-docs/pull/98/commits/4c96fceb59bba003488f3fb5903b9eedcae409a8 Let us know if there are any problems, best Status: Issue closed
lovelmh13/myBlog
562639268
Title: input从右到左显示文字,并且光标始重在右侧。text-align和direction的区别 Question: username_0: 其实很简单: ```css .input { text-align: right } ``` 需要分清和`direction: rtl;`的区别 `text-align`是让文本靠右,而`direction: rtl;`是改变文本的方向 ```html <style type="text/css"> div.one { text-align: right } div.two { direction: rtl; } </style> <div class="one">Some text. Right-to-left direction.</div> <div class="two">Some text. Left-to-right direction.</div> ``` `.class`的显示效果只是让文字整体平移到右边 `.two`除了文字移到了右侧显示,文字也发生了改变。 会先成:`.Some text. Left-to-right direction`,看见区别了吗?句号位置变了。 经过我的测试,**如果只是改变文本内容**,`direction: rtl;`会改变符号和文字的位置,而文字如果也想从右往左写,只需要稍加改动,就可以把顺序完全颠倒,变成从右往左写: ```css div.two { direction: rtl; unicode-bidi:bidi-override; } ``` 就会变成: `Some text. Left-to-right direction.` 但是,**如果换成是input**,除了显示跟本文的显示方式一样以为,光标是不一样的。 `direction: rtl;` 因为是改文字方向,所以光标也改成了在左边。使用`backspace`和`delete`按键也会相反。 `text-align: right`是改变了文字的位置,而不是方向,光标是正常的,所以这个才是我们需要用的属性。
crossplane/crossplane
1092803887
Title: Cannot spin up multiple ec2 resources at once Question: username_0: <!-- Thank you for helping to improve Crossplane! Please be sure to search for open issues before raising a new one. We use issues for bug reports and feature requests. Please find us at https://slack.crossplane.io for questions, support, and discussion. --> ### What happened? I'm trying to spin up multiple AWS EC2 resources at once with multiple defined blocks using Composition, but it's always just one EC2 resource spinning up (and it's always the last defined resource). ### How can we reproduce it? I put the configuration in my public repo, please check https://github.com/username_0/infra ### What environment did it happen in? Crossplane version: - Crossplane 0.22.0 - provider/aws - kubectl version client: Major:"1", Minor:"21", GitVersion:"v1.21.1" server: Major:"1", Minor:"20", GitVersion:"v1.20.5" Status: Issue closed Answers: username_0: Sorry, I realized where my mistake is. Close the issue now
michaelforney/swc
374808126
Title: Missing major()/minor() defines on Fedora 29 Question: username_0: These are defined in `<sys/sysmacros.h>`. Patch: ``` diff --git a/launch/launch.c b/launch/launch.c index b8cc292..b6f0aa9 100644 --- a/launch/launch.c +++ b/launch/launch.c @@ -40,6 +40,7 @@ #include <sys/stat.h> #include <sys/wait.h> #include <sys/ioctl.h> +#include <sys/sysmacros.h> #include <linux/input.h> #include <linux/kd.h> #include <linux/major.h> diff --git a/libswc/drm.c b/libswc/drm.c index 910c034..a9c3ef9 100644 --- a/libswc/drm.c +++ b/libswc/drm.c @@ -39,6 +39,7 @@ #include <unistd.h> #include <drm.h> #include <xf86drm.h> +#include <sys/sysmacros.h> #include <wld/wld.h> #include <wld/drm.h> #include <wayland-server.h> ``` Status: Issue closed Answers: username_1: Thanks for the report. Should be fixed now.
ionic-team/ionic-cli
233431901
Title: Ionic v1: `ionic serve` tells user that app is available at 0.0.0.0:8100 instead of localhost:8100 Question: username_0: Noticed several threads about this on the forum: https://forum.ionicframework.com/t/ionic-serve-is-not-working/92937 https://forum.ionicframework.com/t/ionic-serve-points-to-0-0-0-0-8100/92902 https://forum.ionicframework.com/t/how-i-run-my-ionic-1-application-on-cli-3-using-the-serve/92455 https://forum.ionicframework.com/t/ionic-serve-lab-not-working-after-updating-to-latest-ionic-version-ionic-v1/91743 Answers: username_1: @username_0 could you reproduce the problems? username_2: Yeah, I know what this is about. It defaults to `0.0.0.0` for Ionic 1. Status: Issue closed
treasure-data/digdag
335157836
Title: How about add a link to docs.digdag.io to README Question: username_0: I tried to search the way installing digdag on Windows10 around https://github.com/treasure-data/digdag and could not find the way for a while. I think that the link to www.digdag.io in https://github.com/treasure-data/digdag#Documentation is ambiguous if you are beginner and you want to install digdag first time. To link docs.digdag.io directly will make the instruction more clear. Answers: username_1: Hello, @username_0 Thank you for your feedback. The Documentation text is the link for that document. Is it hard to find? ![digdag](https://user-images.githubusercontent.com/767650/41825866-b3310086-785f-11e8-9a20-3c44cce782d0.png) username_0: Hi, @username_1 Thank you for your instruction! Only to find the link is not so hard, I think. But the title "Documentation" refers different urls than that's contents and it may cause a problem. I misunderstood the link in "Please check digdag.io for installation & user manual." refers the same url with its title. And link to www.link.io in the sentence is somehow unhelpful when you just look for installation or user manual because there are too much information than you need. Status: Issue closed
construct/construct
295163407
Title: compiler upgrade, generate Cython code Question: username_0: This is a major upgrade to the compiler. Benefits: much better performance. Drawback: adds optional dependency to `cython` package, if user wants to use the compiler that is, and compiles the `.compile` method somewhat. Answers: username_1: I don't understand nothing in cython and python internals. Maybe @tmr232 won't mind doing some review to it. Though he is quite busy. username_0: Related https://github.com/cython/cython/issues/2103 username_0: Finished. Docs at: https://construct.readthedocs.io/en/latest/compilation.html
scalikejdbc/scalikejdbc
418229748
Title: How to do select or insert query? Question: username_0: I have the following code I would to execute ``` sql"""WITH s AS ( SELECT id FROM $tableName WHERE $columnName = $value ), i AS ( INSERT INTO $tableName ($columnName) SELECT $value WHERE NOT EXISTS ( SELECT id FROM $tableName WHERE $columnName = $value ) RETURNING id ) SELECT id FROM i UNION ALL SELECT id FROM s """ ``` If I use `update` or `updateAndRetrieveGeneratedKey` within a `localTx` I get `org.postgresql.util.PSQLException: A result was returned when none was expected.`. If I use `single` within `readOnly` I get something like `... cannot perform a insert in a read only query`. What is the proper way to go about this? Answers: username_1: I've never tried yet but what happens if you run the query as blow? ```scala // or DB.autoCommit { DB.localTx { implicit s => selectOrInsertQuery.map(toObject).list.apply() } ``` username_2: The problem is in SQL.scala:814 which always assumes AutoSession is using DB.readonly irregardless what the AutoSession.isReadOnly flag is saying. This complicates things a lot when using AutoSession- that is, they can't be used at all if at any point in the transaction there happens to be an insert. username_1: @username_2 Thanks. The line 814 you mentioned is this line: https://github.com/scalikejdbc/scalikejdbc/blob/3.3.5/scalikejdbc-core/src/main/scala/scalikejdbc/SQL.scala#L814 I understand your point. But, on the other hand, I don't have any ideas to eliminate your problems here except for removing `AutoSesison` from the default param for methods or not to use AutoSession as much as possible. username_1: As I mentioned above, there is no other suggestion apart from using the right session type (read-only, auto-commit, read-only) for each query in this scenario. Let me lose this issue now. Status: Issue closed
MicrosoftDocs/azure-docs
318145618
Title: Misleading URL Question: username_0: The URL for this page is 'https://docs.microsoft.com/en-us/azure/app-service/containers/quickstart-nodejs' which suggests the page has information regarding deploying a container to Azure App Service, but that is not the case. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 116d6e28-471b-2cf4-3eb8-6249b4ef3016 * Version Independent ID: f199d3a2-0c11-e2d6-18f0-d081c76d511d * Content: [Create a Node.js in Azure App Service on Linux](https://docs.microsoft.com/en-us/azure/app-service/containers/quickstart-nodejs) * Content Source: [articles/app-service/containers/quickstart-nodejs.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/containers/quickstart-nodejs.md) * Service: **app-service-web** * GitHub Login: @username_2 * Microsoft Alias: **username_2** Answers: username_1: @username_0 Thanks for the feedback! I have assigned the issue to the content author to investigate further and update the document as appropriate. username_2: @username_0 Hi thanks. Technically it's true because App Service is deploying a built-in Node.js docker image to run the Node.js app in a container. However, you're right that it's not running your custom container. For running a custom docker image, please see: https://docs.microsoft.com/en-us/azure/app-service/containers/tutorial-custom-docker-image. This is our content decision to have built-in and customer image content both under the "container" path. #please-close Status: Issue closed username_3: @username_0 We will now proceed to close this thread. If there are further questions regarding this matter, please reopen it and we will gladly continue the discussion.
renatomrcosta/jeeves-smugglesworth
409681179
Title: ask if user is still merging or forgot to release the token Question: username_0: if you have the token for more then 3 hours let jeeves send you a message if you are still merging Answers: username_1: Defining hard-rules for implementation: - Every hour, on the hour, Jeeves will ask whoever has a token for 3h or longer if they are still merging. - The same user will be asked again and again, every hour. Later I can add working hours (so Jeeves doesn't repeat itself), or a flag indicating that a given queue has been warned beforehand. Dunno yet username_1: Shoddily done, adding even more clutter to the code IMHO. Oh well. Status: Issue closed
facebook/flow
236196420
Title: Debugging Flow Question: username_0: I am using Visual Studio Code with [vscode-ocaml](https://github.com/hackwaly/vscode-ocaml) extension. `./bin/flow` is compiled using `make build-flow-debug` (see also: https://github.com/facebook/flow/issues/4177). `.vscode/launch.json`: ```js { "version": "0.2.0", "configurations": [ { "name": "OCaml", "type": "ocamldebug", "request": "launch", "program": "${workspaceRoot}/bin/flow", "console": "internalConsole", "stopOnEntry": false, "arguments": [ "${workspaceRoot}/local-test" ] } ] } ``` `local-test/test.js`: ```js // @flow const a: number = 'z' + 'd'; ``` Debugging `main` seems to work: <img width="794" alt="screen shot 2017-06-15 at 15 28 09" src="https://user-images.githubusercontent.com/794591/27183582-4f28422c-51df-11e7-9a48-af9d5b69d253.png"> However If I set a breakpoint somewhere in `src/typings/`, for example `statement.ml`, inside `variable_decl` function, then the breakpoint is skipped. I suspect this is because of the server-client architecture - actual type checking happens inside a different process. Is there any good way around this? Answers: username_1: I faced the same problem and I didn't solve it. For the reference, I tried `flow check` command. It seems to be checking type at foreground, but when using `ocamlbuild` with settings like below, debugger responds with `cannot connect to debugger at 127.0.0.1 error: No such file or directory`. ```javascript { "program": "${workspaceRoot}/bin/flow", "arguments": ["check", "path/to/mycode.js"], } ``` The error occurred at `spawn` function in `hack/utils/daemon.ml`(yes, the debugger worked correctory), and I guess that it related to spawning server process. username_1: Still examining, but it seems to get some progress. As above comment I guess, the problem caused by some process that created behind `flow`. Then, I try out arguments "--max-workers" that described with `Maximum number of workers to create (capped by number of cores)` I don't understand it in detail, but it seems related to some process behind `flow`. So, I modified my settings like follows and executed `ocamldebug`. ``` { "program": "${workspaceRoot}/bin/flow", "arguments": ["check", "--max-workers", "0", "path/to/mycode.js"], } ``` And the result looks as I expected. At least it can execute `flow` according to `breakpoint` and process exit normally. username_2: @username_0 @username_1 have you tried to get debugger working with [vscode-ocaml-debugger](https://github.com/hackwaly/vscode-ocaml-debugger)? I can get the debugger working with simple example using `ocamlc -g helloworld.ml` but with Flow pretty much nothing happens, only the loading sign appears below debug menu, username_2: Looks like ocaml debugger runs roughly this kind of command: `ocamlrun bin/flow check my-own-test/test.js --max-workers 0`, when running the flow with `ocamlrun` the program throws ``` Fatal error: unknown C primitive `unix_waitpid' ``` Maybe this is the culprit? username_3: I don't use the ocaml debugger, but flow has a few debugging commands that I use to debug type checker behavior. `flow check --verbose-indent` will show the full list of flows in `__flow`. The function that adds errors is called `add_output`, and it also logs when it is called with `--verbose-indent`. username_2: Nice, this probably already helps a lot, thanks for being so helpful! username_2: Ok, got debugger finally working in VScode! VScode debugger config: ``` { "version": "0.2.0", "configurations": [ { "name": "OCaml", "type": "ocaml-debugger", "request": "launch", "program": "${workspaceFolder}/bin/flow", "console": "internalConsole", "stopOnEntry": false, "arguments": ["check", "${workspaceFolder}/local-test/test.js", "--max-workers=0", "--verbose-indent"] } ] } ``` - `${workspaceFolder}/local-test/test.js` being the js file to check - `--max-workers=0` apparently runs everything in single process - `--verbose-indent` Debugger works also without, but still very useful Other requirements: - OCaml 4.05.0 <- later versions might work too, I only tested with 4.05.0 - VScode debugger plugin: https://marketplace.visualstudio.com/items?itemName=hackwaly.ocaml-debugger - debugger adapter: https://github.com/hackwaly/ocamlearlybird/ - needs to be installed with opam: `opam install earlybird` - NOTE: Needs to be at least 0.1.3 version, earlier versions does not work I also recommend to install Merlin: https://github.com/ocaml/merlin ``` opam install merlin opam user-setup install ``` This might be worth documenting to e.g Flow wiki or readme. username_4: I would also suggest using `--merge-timeout=0` as well to avoid timeouts. username_4: @username_2 thanks for the instructions. Being able to step through things is going to be a big help. username_5: Thanks @username_2 for such a useful comment. I was unable to get mine running until I added to my user settings.json: ```json { "reason.path.ocamlmerlin": "_opam/bin/ocamlmerlin", "reason.diagnostics.tools": [ "merlin" ], "editor.formatOnSave": true, "reason.codelens.enabled": true, } ``` and to my debug configuration: ```json /* start ocamlearlybird --server --port 4711 before running */ "debugServer": 4711 ``` I then run `$ ocamlearlybird --server --port 4711` in the Flow directory before debugging.
podium-lib/issues
475677629
Title: Remove deprecated events Question: username_0: Remove the deprecated [change](https://github.com/podium-lib/client/blob/9f0ff38d1d2d8ffe8f4c817758eea8570cb35c4d/lib/client.js#L102-L104) and [dispose](https://github.com/podium-lib/client/blob/9f0ff38d1d2d8ffe8f4c817758eea8570cb35c4d/lib/client.js#L110-L112) events in the @podium/client.<issue_closed> Status: Issue closed
forceworkbench/forceworkbench
10351637
Title: Serializable Field Mappings Question: username_0: _Original author: <EMAIL> (July 14, 2011 15:06:36)_ <b>1. Summary of the feature request:</b> Ability to save mappings like DataLoader <b>2. Typical scenerio where this feature would be used:</b> Any insert, update, upsert scenario <b>3. Reason current features do not meet needs:</b> May have to repeat something many times and mapping each time is time consuming <b>4. How soon would you like to see this feature implemented?</b> Next API Release <b>5. Would you like to contribute to the development of this feature?</b> I would help _Original issue: http://code.google.com/p/forceworkbench/issues/detail?id=480_<issue_closed> Status: Issue closed
Rydra/bgm-for-es
771719545
Title: Music doesn't stop with non-RA cores Question: username_0: I love this script but anytime I run a non-Retroarch core, like Daphne or MESS the music just keeps playing. Is there any way to stop the music when these cores load up? Answers: username_1: We just need to add the emulator name to the list that is in `MusicStateMachine.py`. Mine does it on the `drastic` emulator for DS... I'm not very linux savvy or github savvy or else I would fork my own version and build it locally with all the emulator names I use. username_2: Where is the location of MusicStateMachine.py? username_0: /usr/lib/python2.7/dist-packages/bgm/ username_3: I have bumped to the new version of the Background music for EmulationStation. This new release is a complete revamp, with a new installation script. Please follow the following steps: 1. Uninstall the old script if you had it: ``` sudo apt remove python-es-bgm ``` 2. Go to the main page of the project and reinstall the script, following the instructions of the documentation: ``` https://github.com/username_3/bgm-for-es ``` The new config file now allows you to update/customize the processes and emulators where the music should stop, without having to fiddle with code. Feedback always appreciated! Status: Issue closed
kalkih/mini-media-player
631533520
Title: Get rid of box shadow Question: username_0: Hi Karl, first of all thanks for the great plug-in. I integrated it to my mobile UI and it looks so much better than the normal media-control. https://community.home-assistant.io/t/google-inspired-home-assistant-on-mobile/202191 I have two questions regarding your plug-in: - Is there a way to get rid of the box-shadow? - is there a way to change the color of the playbar (currently orange)? i tried to integrate card-mod by thomas loven with the following code: entity: media_player.sonos_wohnzimmer show_header_toggle: false icon: 'mdi:play-box-outline' name: Sonos Play One Wohnzimmer artwork: full-cover style: | ha-card { box-shadow: none; webkit-box-shadow: none; } type: 'custom:mini-media-player' but without success. Kind regards, Philipp Status: Issue closed Answers: username_0: reinstallation of card-mod fixed the issue
milvus-io/milvus-sdk-go
1076934154
Title: [Bug]: panic: runtime error: slice bounds out of range Question: username_0: ### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior Exceptions occur when searching multiple sets of vectors "panic: runtime error: slice bounds out of range " ### Expected Behavior 1、 ``` func main() { // Milvus instance proxy address, may verify in your env/settings milvusAddr := `xxxxxxxxxxxx` // setup context for client creation, use 2 seconds here ctx := context.Background() ctx, cancel := context.WithTimeout(ctx, 5*time.Second) defer cancel() c, err := client.NewGrpcClient(ctx, milvusAddr) vector1 := entity.FloatVector([]float32{1,2,3,4,5,6,7,8}) vector2 := entity.FloatVector([]float32{1.1,1.2,1.3,1.4,1.5,1.6,1.7,1.8}) vector3 := entity.FloatVector([]float32{2.1,2.2,2.3,2.4,2.5,2.6,2.7,2.8}) vector4 := entity.FloatVector([]float32{3.1,3.2,3.3,3.4,3.5,3.6,3.7,3.8}) entitys := []entity.Vector{vector1,vector2,vector3,vector4} sp, _ := entity.NewIndexFlatSearchParam(50) sr, err := c.Search(ctx, "gosdk_insert_example", []string{}, "Year > 1990", []string{"ID"}, entitys, "Vector", entity.L2, 10, sp) fmt.Println(sr) if err != nil { log.Fatal("fail to search collection:", err.Error()) } } ``` 2、 ![image](https://user-images.githubusercontent.com/6262877/145593910-68257ed6-71c1-4d0f-bafd-ece8e31e3e6d.png) ### Steps To Reproduce _No response_ ### Environment _No response_ ### Anything else? _No response_
jfrog/charts
475500831
Title: Providing custom binarystore.xml to the chart Question: username_0: <!-- Thanks for filing an issue! Before hitting the button, please answer these questions. It's helpful to search the existing GitHub issues first. It's likely that another user has already reported the issue you're facing, or it's a known issue that we're already aware of--> **Is this a request for help?**: --- **Is this a BUG REPORT or FEATURE REQUEST?** (choose one): FEATURE REQUEST <!-- If this is a BUG REPORT, please: - Fill in as much of the template below as you can. If you leave out information, we can't help you as well. If this is a FEATURE REQUEST, please: - Describe *in detail* the feature/behavior/change you'd like to see. In both cases, be ready for followup questions, and please respond in a timely manner. If we can't reproduce a bug or think a feature already exists, we might close your issue. If we're wrong, PLEASE feel free to reopen it and explain why. --> **Version of Helm and Kubernetes**: all versions **Which chart**: artifactory **What happened**: As of writing this issue, the artifactory chart creates a general binarystore.xml based on persistence storage type. We use cachefs with filesystem and right now the filesystem type creates a default binarystore without cachefs. **What you expected to happen**: option to provide the chart a custom binarystore.xml to use Answers: username_1: @username_0 - This is a valid request. We'll provide a pull request in the near future. Status: Issue closed username_1: <!-- Thanks for filing an issue! Before hitting the button, please answer these questions. It's helpful to search the existing GitHub issues first. It's likely that another user has already reported the issue you're facing, or it's a known issue that we're already aware of--> --- **Is this a BUG REPORT or FEATURE REQUEST?** (choose one): FEATURE REQUEST <!-- If this is a BUG REPORT, please: - Fill in as much of the template below as you can. If you leave out information, we can't help you as well. If this is a FEATURE REQUEST, please: - Describe *in detail* the feature/behavior/change you'd like to see. In both cases, be ready for followup questions, and please respond in a timely manner. If we can't reproduce a bug or think a feature already exists, we might close your issue. If we're wrong, PLEASE feel free to reopen it and explain why. --> **Version of Helm and Kubernetes**: all versions **Which chart**: artifactory **What happened**: As of writing this issue, the artifactory chart creates a general binarystore.xml based on persistence storage type. We use cachefs with filesystem and right now the filesystem type creates a default binarystore without cachefs. **What you expected to happen**: option to provide the chart a custom binarystore.xml to use username_1: Will close once `artifactory-ha` is also updated with the change. Status: Issue closed
SiciliaHub/nonportaleregionesicilia
160409658
Title: Aggiungere testo descrittivo nella pagina "Cerca nei siti" Question: username_0: @username_1 aggiungerei [qui](https://github.com/SiciliaHub/nonportaleregionesicilia/blob/gh-pages/cerca.html#L92) un testo descrittivo brevissimo sul cosa è questa pagina. Se sei d'accordo però pensaci tu, che ne sei l'ideatore Answers: username_1: @username_0 testo aggiunto ;) vedi se ti sembra che va bene e se no, modificalo pure o dimmi se devo aggiungere altre info. username_0: Ho fatto una piccola modifica. Metterei in link alla tabella che contiene l'elenco degli URL mappati. Che ne dici? username_1: Ottimo! Sono d'accordo per il link (all'inizio l'idea era quella di creare la pagina dei siti regionali, vedi #35). L'unico timore era che poi veniva simile a quella dei dataset e questo poteva confondere. Mi sembra che con il link questo pericolo non ci sia... username_0: @username_1 dove trovo una tabella con l'elenco di link? Mi avrai già inviato un gdrive, ma non riesco a recuperarlo. Grazie username_1: @username_0 è nel quarto foglio del google sheet del non-portale https://docs.google.com/spreadsheets/d/15cFU9v2x3ACb1IPcVE9Xg3wc5fuay8NAYDXQnoMcOFw/edit#gid=2118198704 Status: Issue closed username_0: Ho aggiunto il link all'elenco dei siti mappati dal motore
apache/pulsar
550823488
Title: Subscription metrics name error Question: username_0: The metric name of the document is pulse_subscription_unacked_messages but Prometheus print is pulsar_subscription_unacked_massages , word messages spelling mistakes Answers: username_1: Thanks @username_0 can you help fix the typo issue?
BevoTechPack/BevoTechPack
205702780
Title: Bevo Tech Pack Server/Singleplayer World Loading Problems Question: username_0: Failure to give adequate information on issues/bugs will be subject to immediate possible closure/spankings. Bevo Tech Pack Version : 2.21 ### Description of issue/bug At first sorry for my bad english. Well.... i made a bevo tech pack server with my friends. At first everything was fine and we could play on this server with no problems. But since 3 days ago i couldn´t start the server. ### Steps to reproduce the issue/bug I´ve already find out, that the problem is the world. I´ve removed the "world" folder, started the server, he created a new world and started with no problems. I´ve tried to start the world in singleplayer and my Minecraft crashed... I hope you can help me. Greetings from Germany ### http://pastebin.com/5LNsKRv8 Questions: Did you add any additional mods? - Just "fast craft" Do you use optifine? - No Did the problem appear in single player or on a server? - Both Answers: username_1: So you have 2 choices, lets start with the easy one and see if it fixes the world. Edit config/forge.cfg on the server and change B:removeErroringEntities=false B:removeErroringTileEntities=false both to true, load up your world, it may still crash, load it up again. If it fixes your world change those 2 back to false (don't play your game with them set to true). If it fixes the issue close this if it doesn't let me know. username_0: No..... it haven´t worked. First start: http://pastebin.com/D1fsRSWp Second start: http://pastebin.com/wqU7CtAS username_1: Ok so you are left with world editing out block located at -345,75,79 the issue you have is the server crashes so you have to do this manually using world edit gui. It's not something I can talk you through as it's complicated and a wrong move could ruin your world completely. Do you have someone with World Edit knowledge to do that? username_0: Only some single basics... I´ve tried to remove the block and i did it. But now the server starts without mods..(I did an backup from the world before i did it) Status: Issue closed username_0: OH MY GOD Thank you very much. After trying to handle MC Edit, i did it. I removed the block and everything works fine now. After the first try the whole world were reseted, but the mods were working. After the second try the mods didn´t worked, but the world was correct. And after the third try everything is working. Thank you very much.
rterrabh/LFApp
164886875
Title: Salvar a gramática testada quando sair e entrar do APP. Question: username_0: Procedimento: Após inserir uma gramática, funciona tudo de acordo. Porém, quando muda de Activity no Android, ou seja, vai pro Whatsapp ou Facebook, a gramática se perde. Tem uma configuração que acredito que seja savedInstanceState, que tem como salvar o que foi feito anteriormente. Assim quando a pessoa sair e entrar, os registros estarão salvos.<issue_closed> Status: Issue closed
mockito/mockito
351773017
Title: Exception not thrown when mocking final method Question: username_0: Take the code snippet below MissingMethodInvocationException expected in both test case. But only the second fails. Tested with Mockito 2.20 (found with mockito 2.8.47) using JDK 1.8.171, Win10x64. ```java public class TestingMockito { @Test public void testShouldFailButPasses() { // GIVEN Foo mockFoo = mock(Foo.class); Aaaa a = new Aaaa(); // WHEN a.setSomething(mockFoo); // THEN - next lines do not throw exception Foo mockFoo2 = mock(Foo.class); when(mockFoo2.getFoo()).thenReturn(null); } @Test public void testFailAsExpected() { Foo mockFoo2 = mock(Foo.class); when(mockFoo2.getFoo()).thenReturn(null); } public static class Aaaa { public void setSomething(Object something) { System.out.println("This is " + something); } } public static class Foo { public final String getFoo() { return "foo"; } } } ``` Answers: username_1: @username_0 I tested your sample with v2.21.0, both tests turn green with plain mockito and mockito-inline. Why do you expect the when call to fail? username_0: I expect both tests to fail with "MissingMethodInvocationException". Because the getFoo() method is **final** thus cannot be mocked by Mockito. I wonder what was the result in your environment? Did getFoo() returns <null> or "foo"? ``` org.mockito.exceptions.misusing.MissingMethodInvocationException: when() requires an argument which has to be 'a method call on a mock'. For example: when(mock.getArticles()).thenReturn(articles); Also, this error might show up because: 1. you stub either of: final/private/equals()/hashCode() methods. Those methods *cannot* be stubbed/verified. Mocking methods declared on non-public parent classes is not supported. 2. inside when() you don't call method on mock but on some other object. at TestingMockito.testFailAsExpected(TestingMockito.java:29) .... at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86) at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38) ``` username_1: When mockito-inline-mocker is not used, mockito can't intercept the final method. Therefore it can't detect when it is called. One the other hand mockito could detect that no mock-method was called, when the `when` is called. username_2: I have tested this on both Mockito 2.20 and the latest 2.23. I have observed that the object `a` is being intercepted by MockitoInline. Is that expected ? The test is passing only when some method is being called on object `a`. This looks like a bug.
styled-components/styled-components
452481751
Title: Can I get the final className before render Question: username_0: ```` const Wrapper = styled.div` border: 1px solid blue; &__btn { color: red; } ` class extends React.PureComponent<Props> { public render() { const wrapperClsName: string = (Wrapper as any).componentStyle.lastClassName const cls = bemClsFactor(wrapperClsName) return( <Wrapper> <button className={cls('btn')}>Hello World</button> </Wrapper> ) } } ```` LIke the code before, I want to using styled-components and bem together, I don't want every component inside the Wrapper styled. But the first time render I can't get lastClassName. is there any other way to do this? Answers: username_1: `Wrapper.styledComponentId` will give you the static className of the component, but I don't really recommend using styled-components in this way. What exactly are you trying to accomplish here? A reproduction in a codesandbox would be helpful. Status: Issue closed
bootstrapworld/curriculum
1081342594
Title: Switching back to default textbook alignment shows up empty Question: username_0: In lesson defining-values, the default textbook alignment (`IM.Alg.1`) shows `IM.Alg1.4.2 - Function Notation`. However, if you choose a different textbook in the dropdown and go back to `IM.Alg.1`, nothing is shown under it. Answers: username_1: @username_0 the value in the `option` elt is "textbook-IM_Alg.1" (with a dot). But the loop that compares the selection to the standards tables is seeing the value "textbook-IM_Alg_1". I couldn't find where this string munging happens, but somewhere along the lines the dot is turning into an underscore. Status: Issue closed
keybase/client
213959904
Title: Backup software thinks /keybase is a folder that should be backed up, and then fails when it can't be accessed by root. Question: username_0: Keybase GUI Version: 1.0.20-20170309192151+e30557a When trying to back up my computer with any backup agent that uses administrative privileges to perform backups, they universally report filesystem errors trying to access /keybase/ to back it up. This is probably because 'root' has no keybase configuration. It may make sense on Mac to handle this specific case more effectively, so that backup clients either back up nothing (which is fine) or ignore the mount altogether (because you tune it to look like a network mount, or something, which is also fine) or some other manner altogether. Answers: username_1: Hi @username_0 -- it should already be configured to look like a network mount. You might just have to list if explicitly in your software's exclude list. Not sure what else to do on our end. It's true that `root` does not have permissions to read `/keybase`, only the user you mount it as. username_0: For 'root', it would be invaluable if /keybase was an otherwise-empty directory with a README explaining what's going on. That would help people who do things as root on OS X (sudo or otherwise) understand why keybase isn't accessible, when eventually they get around to trying that, and would also give backup software something other than an disk read error. username_1: Thanks for giving this some thought @username_0, interesting idea. Unfortunately, the way Finder works in macOS means that we can't present `root` with an empty folder. Finder often drops into `root` mode to look up attributes about files -- why it does that, I couldn't tell you. But the file system kernel module we use (osxfuse) has many fancy exceptions to let those Finder-related calls through, but not other types of root access like listing directories, etc. Do the errors you're seeing actually break the backup software? Or does it just log the error and move onto the next directory? Also, are you willing to name the software so we can test with it? Thanks! username_0: Ah, that explains why it's all getting confused. That's crazy! I had no idea! I'm so sorry! :( I should have noticed you were using FUSE. It means this is all basically a FUSE problem and not a Keybase problem. They don't break it, it just logs it and moves on. Arq Backup is the one I use everyday, and is available to the general public. - R.
supervisorphp/supervisor
126661344
Title: Failed to connect to localhost port 9001 Question: username_0: I can't connect to my supervisor service give me that error every time i trie to get process An error occurred when fetching the URI "http://localhost:9001/RPC2" with the adapter "guzzle6" ("cURL error 7: Failed to connect to localhost port 9001: Connection refused (see http://curl.haxx.se/libcurl/c/libcurl-errors.html)"). Answers: username_1: Hey @username_0 What is in you supervisord.conf file? username_0: ; supervisor config file [unix_http_server] file=/var/run/supervisor.sock ; (the path to the socket file) chmod=0700 ; sockef file mode (default 0700) [supervisord] logfile=/var/log/supervisor/supervisord.log ; (main log file;default $CWD/supervisord.log) pidfile=/var/run/supervisord.pid ; (supervisord pidfile;default supervisord.pid) childlogdir=/var/log/supervisor ; ('AUTO' child log dir, default $TEMP) ; the below section must remain in the config file for RPC ; (supervisorctl/web interface) to work, additional interfaces may be ; added by defining them in separate rpcinterface: sections [rpcinterface:supervisor] supervisor.rpcinterface_factory = supervisor.rpcinterface:make_main_rpcinterface [supervisorctl] serverurl=unix:///var/run/supervisor.sock ; use a unix:// URL for a unix socket ; The [include] section can just contain the "files" setting. This ; setting can list multiple files (separated by whitespace or ; newlines). It can also contain wildcards. The filenames are ; interpreted as relative to this file. Included files *cannot* ; include files themselves. [include] files = /etc/supervisor/conf.d/*.conf username_1: You didn't configure an inet section, only the unix socket one. Either use that one (not compatible with guzzle6) or configure inet: http://supervisord.org/configuration.html#inet-http-server-section-settings username_0: Thanks, for unix socket wich client i need to use ? username_1: Assuming you are using Ivory Http Adapter, you can try this client: username_0: Thanks i already configure the inet and now it works Status: Issue closed
swimlane/ngx-charts
529882448
Title: ngcc error with Angular 9.0.0-rc.4 Question: username_0: When upgraded angular from 9.0.0-rc.3 to 9.0.0-rc.4, ngcc fails with following error: ``` $ ngcc --properties es2015 browser module main --first-only --create-ivy-entry-point Compiling @swimlane/ngx-charts : module as esm5 Error: Error on worker #5: Error: Failed to compile entry-point @swimlane/ngx-charts due to compilation errors: node_modules/@angular/common/common.d.ts(115,22): error TS-996002: Appears in the NgModule.imports of TooltipModule, but could not be resolved to an NgModule class node_modules/@angular/common/common.d.ts(115,22): error TS-996002: Appears in the NgModule.imports of AxesModule, but could not be resolved to an NgModule class node_modules/@angular/common/common.d.ts(115,22): error TS-996002: Appears in the NgModule.imports of ChartCommonModule, but could not be resolved to an NgModule class node_modules/@swimlane/ngx-charts/release/esm.js(2832,5): error TS-996002: Appears in the NgModule.imports of ChartCommonModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(1062,5): error TS-996002: Appears in the NgModule.imports of ChartCommonModule, but itself has errors node_modules/@angular/common/common.d.ts(115,22): error TS-996003: Appears in the NgModule.exports of ChartCommonModule, but could not be resolved to an NgModule, Component, Directive, or Pipe class node_modules/@swimlane/ngx-charts/release/esm.js(2832,5): error TS-996003: Appears in the NgModule.exports of ChartCommonModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(1062,5): error TS-996003: Appears in the NgModule.exports of ChartCommonModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of AreaChartModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of BarChartModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of BubbleChartModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of ForceDirectedGraphModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of HeatMapModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of LineChartModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of PieChartModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of PolarChartModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(13531,5): error TS-996002: Appears in the NgModule.imports of PolarChartModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(11717,5): error TS-996002: Appears in the NgModule.imports of PolarChartModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of NumberCardModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of TreeMapModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996002: Appears in the NgModule.imports of GaugeModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(13531,5): error TS-996002: Appears in the NgModule.imports of GaugeModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(9549,5): error TS-996002: Appears in the NgModule.imports of GaugeModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(4282,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(5816,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(9549,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(10196,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(10412,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(11000,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(11717,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(13560,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(13974,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(13531,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(14329,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors node_modules/@swimlane/ngx-charts/release/esm.js(15029,5): error TS-996003: Appears in the NgModule.exports of NgxChartsModule, but itself has errors at ClusterWorker.compile (/charts-project/node_modules/@angular/compiler-cli/ngcc/src/main.js:170:27) at Worker.<anonymous> (/charts-project/node_modules/@angular/compiler-cli/ngcc/src/execution/cluster/worker.js:41:42) at Worker.emit (events.js:210:5) at process.<anonymous> (internal/cluster/worker.js:34:12) at process.emit (events.js:210:5) at emit (internal/child_process.js:878:12) at processTicksAndRejections (internal/process/task_queues.js:81:21) at ClusterMaster.onWorkerMessage (/charts-project/node_modules/@angular/compiler-cli/ngcc/src/execution/cluster/master.js:158:27) at /charts-project/node_modules/@angular/compiler-cli/ngcc/src/execution/cluster/master.js:46:95 at ClusterMaster.<anonymous> (/charts-project/node_modules/@angular/compiler-cli/ngcc/src/execution/cluster/master.js:238:57) at step (/charts-project/node_modules/tslib/tslib.js:136:27) at Object.next (/charts-project/node_modules/tslib/tslib.js:117:57) at /charts-project/node_modules/tslib/tslib.js:110:75 at new Promise (<anonymous>) at Object.__awaiter (/charts-project/node_modules/tslib/tslib.js:106:16) at EventEmitter.<anonymous> (/charts-project/node_modules/@angular/compiler-cli/ngcc/src/execution/cluster/master.js:232:32) at EventEmitter.emit (events.js:210:5) ``` Status: Issue closed Answers: username_1: Hi there, please refer to this comment: https://github.com/swimlane/ngx-charts/issues/1294#issuecomment-560380598 Closing this one.
adjust/hashtypes
484049958
Title: Add binary I/O to the SHA types. Question: username_0: I made an [implementation of binary I/O for the various SHA types](https://github.com/username_0/hashtypes/tree/binary). For now, I'm not making it a pull request because my work depends on the Windows port, [which is already a pull request](https://github.com/adjust/hashtypes/pull/17). I bumped the version to 0.1.6 and made an upgrade script based on help from the PostgreSQL General mailing list. I did not touch the `.in` files, and copied `hashtypes--0.1.5.sql` directly. I did not touch `sha.sql.type` because I don't know how it's used to generate the install script.
Alexander-Barth/NCDatasets.jl
407476515
Title: ncgen does not preserve unlimited dimensions Question: username_0: `ncgen` does not preserve unlimited dimensions in the ncdf files it produces. And I could not find a function to set a dimension to unlimited. Is there one? Answers: username_1: It should work as it is part of the test-suite: https://github.com/username_1/NCDatasets.jl/blob/master/test/test_ncgen.jl The resulting file is: ``` netcdf juliauJUfOc { dimensions: lon = 3 ; unlimited = UNLIMITED ; // (0 currently) variables: float lon(lon) ; lon:string = "degrees_east" ; lon:float32 = 1.f ; lon:float64 = 1. ; lon:float32_vector = 1.f, 2.f, 3.f ; lon:float64_vector = 1., 2., 3. ; lon:int32_vector = 1, 2, 3 ; // global attributes: :dollar = "a dollar $ stop" ; :backslash = "a backslash \\ stop" ; :doublequote = "a doublequote \" stop" ; } ``` Can you provide a minimal netcdf file and julia script that show the error that you are seeing? username_0: Here is an example using the file used in the documentation: ```> julia _ _ _ _(_)_ | Documentation: https://docs.julialang.org (_) | (_) (_) | _ _ _| |_ __ _ | Type "?" for help, "]?" for Pkg help. | | | | | | |/ _` | | | | |_| | | | (_| | | Version 1.0.3 (2018-12-18) _/ |\__'_|_|_|\__'_| | Official https://julialang.org/ release |__/ | julia> using NCDatasets julia> ncgen("sresa1b_ncar_ccsm3-example.nc", "ncgen-sresa1b_ncar_ccsm3-example.jl") julia> include("ncgen-sresa1b_ncar_ccsm3-example.jl") julia> run(pipeline(`ncdump -h sresa1b_ncar_ccsm3-example.nc`, `head`)); netcdf sresa1b_ncar_ccsm3-example { dimensions: lat = 128 ; lon = 256 ; bnds = 2 ; plev = 17 ; time = UNLIMITED ; // (1 currently) variables: float area(lat, lon) ; area:long_name = "Surface area" ; julia> run(pipeline(`ncdump -h filename.nc`, `head`)); netcdf filename { dimensions: lat = 128 ; lon = 256 ; bnds = 2 ; plev = 17 ; time = 1 ; variables: float area(lat, lon) ; area:long_name = "Surface area" ; julia> ``` Note that the time dimension is unlimited in the original file but not in the `ncgen` file. username_1: Thanks a lot, I could reproduce the issue and commit a fix to the match branch. Does it work for you? username_0: Yes, that works. Thanks! username_1: Thanks for confirming. I just made a new release with the fix. Status: Issue closed
aciidb0mb3r/SPM-Cpp-Fixture
187633730
Title: Linker error Question: username_0: Tried on macOS with Xcode 8.0, 8.1 as well as Swift 3.0 on Linux and all get the following linker errors with this project: ``` $ swift build Compile cpplib cpplib.cpp Linking cpplib Compile cwrapper cwapper.cpp Compile cpp-exec main.cpp Linking cwrapper Compile Swift Module 'swift_exec' (1 sources) Linking ./.build/debug/swift-exec Linking cpp-exec Undefined symbols for architecture x86_64: "std::__1::basic_ostream<char, std::__1::char_traits<char> >::operator<<(int)", referenced from: _main in main.cpp.o "std::__1::cout", referenced from: _main in main.cpp.o ld: symbol(s) not found for architecture x86_64 clang: error: linker command failed with exit code 1 (use -v to see invocation) <unknown>:0: error: build had 1 command failures ``` Could this be a regression in SwiftPM since this project was originally created? Thanks for making this! Answers: username_1: Looks like it, can you report on bugs.swift.org Thanks! username_0: Thanks for confirming my suspicions! Filed [SR-3152](https://bugs.swift.org/browse/SR-3152)
ContinuumIO/anaconda-issues
267876473
Title: Navigator Error Question: username_0: ## Main error An unexpected error occurred on Navigator start-up<br>psutil.AccessDenied (pid=5840) ## Traceback ``` Traceback (most recent call last): File "D:\Anaconda2\lib\site-packages\anaconda_navigator\exceptions.py", line 75, in exception_handler return_value = func(*args, **kwargs) File "D:\Anaconda2\lib\site-packages\anaconda_navigator\app\start.py", line 108, in start_app if misc.load_pid() is None: # A stale lock might be around File "D:\Anaconda2\lib\site-packages\anaconda_navigator\utils\misc.py", line 384, in load_pid cmds = process.cmdline() File "D:\Anaconda2\lib\site-packages\psutil\__init__.py", line 701, in cmdline return self._proc.cmdline() File "D:\Anaconda2\lib\site-packages\psutil\_pswindows.py", line 623, in wrapper raise AccessDenied(self.pid, self._name) AccessDenied: psutil.AccessDenied (pid=5840) ``` ## System information ``` platform: win-64 version: 1.6.2 conda: 4.3.21 qt: 5.6.2 language: zh_CN python: 2.7.13 os: Windows;7;6.1.7601;AMD64;Intel64 Family 6 Model 94 Stepping 3, GenuineIntel pyqt: 5.6.0 ``` Status: Issue closed Answers: username_1: **See Issue #1984 for more information on how to fix this.** --- Closing as duplicate of #1984 --- Please remember to update to the latest version of Navigator to include the latest fixes. Open a terminal (on Linux or Mac) or the Anaconda Command Prompt (on windows) and type: ``` $ conda update anaconda-navigator $ conda update navigator-updater ```
MicrosoftDocs/azure-docs
394329947
Title: MQTT-over-WebSockets needs to explain how to set up the WS connection Question: username_0: Just like the document explains how to set up a raw MQTT connection when not using one of the Azure SDKs, it also needs to explain how to set up the MQTT-over-WebSockets connection when not using the SDKs. Specifically these two pieces of information are missing: 1. The path of the URL of the request must be `/$iothub/websocket`. 1. The request must contain the `sec-websocket-protocol` header with value `mqtt` Thus an example handshake request would look like: ``` GET /$iothub/websocket HTTP/1.1 Host: foo.azure-devices.net <other standard handshake headers> sec-websocket-protocol: mqtt ``` Not doing both of these things will cause the server to reject the connection with HTTP 400. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: bcf2a86a-1de0-2786-db87-58279ceffb22 * Version Independent ID: 7e43624f-a87d-1c4d-94e6-e38c1d42686f * Content: [Understand Azure IoT Hub MQTT support](https://docs.microsoft.com/en-us/azure/iot-hub/iot-hub-mqtt-support#retrieving-a-device-twins-properties) * Content Source: [articles/iot-hub/iot-hub-mqtt-support.md](https://github.com/Microsoft/azure-docs/blob/master/articles/iot-hub/iot-hub-mqtt-support.md) * Service: **iot-hub** * GitHub Login: @username_2 * Microsoft Alias: **rezas** Answers: username_1: Thanks for the feedback! I have assigned the issue to the content author to evaluate and update as appropriate. username_2: @username_0, sure, I'll add this in future revisions of this doc. In the meantime, if you have working samples to share feel free to send me here or via email (rezas). username_1: We will now proceed to close this thread. If there are further questions regarding this matter, please tag me in your reply. We will gladly continue the discussion and we will reopen the issue. Status: Issue closed username_1: Just like the document explains how to set up a raw MQTT connection when not using one of the Azure SDKs, it also needs to explain how to set up the MQTT-over-WebSockets connection when not using the SDKs. Specifically these two pieces of information are missing: 1. The path of the URL of the request must be `/$iothub/websocket`. 1. The request must contain the `sec-websocket-protocol` header with value `mqtt` Thus an example handshake request would look like: ``` GET /$iothub/websocket HTTP/1.1 Host: foo.azure-devices.net <other standard handshake headers> sec-websocket-protocol: mqtt ``` Not doing both of these things will cause the server to reject the connection with HTTP 400. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: bcf2a86a-1de0-2786-db87-58279ceffb22 * Version Independent ID: 7e43624f-a87d-1c4d-94e6-e38c1d42686f * Content: [Understand Azure IoT Hub MQTT support](https://docs.microsoft.com/en-us/azure/iot-hub/iot-hub-mqtt-support#retrieving-a-device-twins-properties) * Content Source: [articles/iot-hub/iot-hub-mqtt-support.md](https://github.com/Microsoft/azure-docs/blob/master/articles/iot-hub/iot-hub-mqtt-support.md) * Service: **iot-hub** * GitHub Login: @username_2 * Microsoft Alias: **rezas** Status: Issue closed username_1: We will now proceed to close this thread. If there are further questions regarding this matter, please tag me in your reply. We will gladly continue the discussion and we will reopen the issue. username_3: In case someone stumbles on this - you need to use `/$iothub/websocket?iothub-no-client-cert=true` - otherwise the connection gets dropped.
cake-tech/cake_wallet
1090232099
Title: [Bug] All nodes fail tests and show as down Question: username_0: used Cake Wallet v4.27 for android All nodes fail tests and show as down(red), including cake's own node. For those nodes that require authentication, even if I fill in the correct username and password, the test still fails. More than one person has reported this bug, hope to fix it in next version. In addition, please consider updating the monero core to 0.17.3.0 Answers: username_1: Thanks for your feedback. Unfortunately I'm not sure that we will fix it in next version, but I will put it in our list. For now we are focused on monero only app by cake wallet. username_2: Main issue is dup of #204 Unsure about 0.17.3.0 Status: Issue closed