repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
lmrodriguezr/nonpareil
166852840
Title: nonpareil alignment version can't use fastq files Question: username_0: There is a problem with nonpareil alignment version. It can't use fastq files. It will go through the whole code. But, it will output files with all zeros. [SRR096387_1.npo.txt](https://github.com/lmrodriguezr/nonpareil/files/376416/SRR096387_1.npo.txt)
ornladios/ADIOS2
860223425
Title: Question: ADIOS2 VTU example Question: username_0: Is there a working example of creating the vtu file to be read into VTK with data on multiple ranks? I was able to get the vtx reader to work with all data on a single rank, but not with distributed data (in local arrays). The documentation implies that this should work without issue as long as local arrays are used. Answers: username_1: @username_0 this is integrated with a MFEM example that does exactly that, see [here](https://github.com/mfem/mfem/blob/master/examples/ex9p.cpp#L507). I don't have a bp dataset stored, but generating it should be a matter of following MFEM docs (it shouldn't be difficult). Hope it helps. Let me know if you have any questions. username_0: This was not as simple as one would hope. A few issues came up: 1. Only adios 2.5.0 works with mfem. `[email protected]` writes a `bp` file that cannot be read back if the `Save` function is called more than once. `[email protected]` cannot be used with `mfem` at all because some of the functions that are used in the mfem code are missing 2. The following patch was needed for the `mfem` makefile [0001-fix-adios2-build.patch.txt](https://github.com/ornladios/ADIOS2/files/6330600/0001-fix-adios2-build.patch.txt). Once the patch is applied the `ADIOS2_OPTS` and `ADIOS2_LIB` options need to be specified during configure. So, I guess it seems like the `adios2` vtk unstructured grid support is not widely used/tested? Anyways based on the concepts from `mfem` I will try to get a simple example working with `[email protected]`. username_0: Based on the outputs from MFEM I was able to get something working. The vtk/ParaView reader seems to be a bit of a fickle beast that just crashes ParaView on any errors (I understand that there are a lot of moving parts I'm not trying to complain just help anyone else who ends up stuck on these issues). I think it would be helpful if the document the following since it differs from how the data arrays are defined in the traditional xml based readers: 1. the connectivity array must be uint64_t or the treader will crash 2. the connectivity array contains the information that is normally stored as connectivity and offsets, so each entry includes the number of element nodes then the connectivity for that element. @username_1 do you know if it is possible to use an arrays that are not called "verticies" and "connectivity"? I am doing a multiscale analysis where I have one macroscale mesh distributed across many processors and many microscale meshes on a single processor using on node parallel. Since I have to store the microscale mesh data in my restart file (which I'm considering using adios2 for). It would be really cool if I could somehow point the verticies/connectivity array to one of the microscale data sets to visualize it. username_1: @username_0 apologies for the late reply, it's been a busy time. Your assessment is correct there are too many moving parts. The current bottleneck is the integration of more recent versions of adios2 2.5.0 with ParaView superbuild due to a VTK pending issue see [here](https://gitlab.kitware.com/vtk/vtk/-/issues/18004). Also, getting things into ParaView/VTK can take a very long time. For now there is no alternative to using "vertices" and "connectivity", these are implemented in VTK BTW. Hope it helps and thanks for the patch!
silknow/converter
446671131
Title: Use perio.do from representing time periods Question: username_0: www.perio.do Answers: username_0: @AvgustinKastelic Could you give me a status update for the handover? username_0: a ecrm:E12_Production ; ecrm:P108_has_produced <http://data.silknow.org/object/d8bdaaa5-b7fd-302f-aba2-737c4f396556> ; ecrm:P126_employed <http://data.silknow.org/vocabulary/368> ; ecrm:P32_used_general_technique "weaving technique"@en , "trama llançada"@en ; ecrm:P4_has_time-span <http://n2t.net/ark:/99152/p086kj97q3t> , <http://n2t.net/ark:/99152/p086kj9t437> ; username_1: As we have discussed, this approach will ultimately not be suitable in numerous cases, since, in your concrete example, the URIs http://n2t.net/ark:/99152/p086kj97q3t and http://n2t.net/ark:/99152/p086kj9t437 refer to respectively the 16th and 17th century in South Carolina! perio.do seems to be lacking encompassing periods, since it always tries to bind one with a geographical area, which makes less sense when describing centuries. Hence, we should rather rely on Getty AAT that provides such notions of generic period. For example, 16th century will be http://vocab.getty.edu/aat/300404510 username_0: We made some progress here as it now (almost) works out of the box with Getty AAT, @username_2 works on an update of string2vocabulary for that. However, time periods are exclusively English on Getty AAT. Our solution is to create a local copy and modify it to include Italian, Spanish and French translations of the most common periods as skos:altLabels for now. In the long-term we should suggest these labels to Getty to include them officially maybe? username_1: Yes, this is an excellent idea! To progress on the translation, you should output a CSV containing all the English period names found in the data. Then, load this CSV in a google spreadsheet so our experts from UVEG, UNIPA and CNRS can help in the translation. username_0: We need to consider other options to parse time expressions instead of matching them with an vocabulary. username_0: Just an internal reminder to extend the labels to more centuries (e.g. 13th century) username_0: Overhaul with Wikidata Reconciliation necessary: https://wdreconcile.toolforge.org/en/api username_1: See also https://github.com/schemaorg/suggestions-questions-brainstorming/issues/219 and around for handling uncertain dates username_2: And how we manage uncertainty on DOREMUS: http://data.doremus.org/artist/e8e2694f-ead1-3d37-8895-2473a25f9858/birth/interval username_1: Closing since we have now #66. Status: Issue closed
opencontainers/oci-conformance
581784732
Title: Opt-in for registries to appear on public test dashboard Question: username_0: Regarding the conformance dashboard located here: https://oci.bloodorange.io I think the [source](https://github.com/bloodorangeio/oci-distribution-conformance-results) for that github pages site + CI should be migrated to this repo, and some domain setup such as `conformance.opencontainers.org`. However, it's not clear which results should be included. What if, as part of submission for certification via PR, the submitter can set a field in `PRODUCT.yaml`: ``` ci: true ``` or something similar, which indicates that there is an associated CI definition located in this repo, which will be built on some timer, updating the dashboard results page. As part of this, commercial vendors would need to supply secrets/credentials with access to some test registry that we can use for the tests. Answers: username_0: This is now possible with new process for providing badges Status: Issue closed
felangel/bloc
638169665
Title: How can I `yield` a state from the recurring callback of a Timer? Question: username_0: I have an uninitialized Timer object in my class and when user hits start button, it is initialized, and inside it's callback I need to call `yield ...`. I'm new to dart and this is the code I have so far. Where should I modify? ``` dart if (event is CounterETimerStart) { timer = Timer.periodic(Duration(seconds: 1), (timer) async* { yield CounterNewSecond(++m.passedTime); }); } ``` Answers: username_1: Hi @username_0 👋 You can add an event and react to it. ``` if (event is CounterETimerStart) { timer = Timer.periodic(Duration(seconds: 1), (timer) { add(TimerTicked(payload)); }); } if (event is TimerTicked) { yield CounterNewSecond(event.payload); } ``` username_0: Hey @username_1, Thanks. This is very good, but it just doesn't feel that elegant! Is it the only way to do this? username_1: I assume you're after something like ``` yield* Timer.periodic(Duration(seconds: 1), (timer) async* { yield something; }); ``` but that's not possible. Alternatively you could use: ``` Stream<int> get stream async* { yield* Stream.periodic(const Duration(seconds: 1), (index) => index); } ``` username_0: ``` dart stream = Stream.periodic(const Duration(seconds: 1), (index) => CounterNewSecond(++m.passedTime)); yield* stream; ``` I wrote this and now `CounterNewSecond` is yielding, but the bloc now can't listen to any event. I believe I've not written the code correctly. Could you provide a more complete code, because I don't really know what to do with the code in the alternate way you offered. username_1: This assumes `CounterNewSecond` is a state class and not an event. username_0: I realize that. What I'm saying is when I'm listening to `start` event, I want to keep yielding the `CounterNewSecond` state, and the code above does that perfectly. The problem with it is that when the above code starts running, the bloc wont receive any other events that I send to it from the screen code. username_1: You're adding an event per second, there's plenty of time for other events to be processed. The events are processed in the same order they are added, e.g.: `CounterNewSecond`, `SomeEventFromUI`, `CounterNewSecond`. If you think there's some issues you can't overcome feel free to share a minimal reproduction gist/repo and I can have a look 👍 username_0: ... It seems like when `Stream.periodic(...)` starts working, the thread is occupied by it and the bloc wont react to any other event coming in. username_1: Your model `m` should not be on the bloc but on the state itself. Here's a [gist](https://gist.github.com/username_1/41c284e6b36dc716b50a15c4ea081418) which does what I assume you want. My bad, the approach was not working because the default bloc event processing operator `asyncExpand` pauses the event stream in order to process it, event by event and `Stream.periodic` never completes so other events won't get the chance of being processed, unless for some rx magic. `switchMap` converts each incoming event into a new `Stream` that will complete after being processed so now other events will get the chance of being handled. Hope that solves your issue and sorry for missing this crucial aspect ✌ username_0: No problem. Sorry for late replies. About the model not being on the bloc, why? If bloc is supposed to house all the business logic, and the logic does some work on the model and modifies it and later saves it, wouldn't the best place for it be the bloc? As I understand and read about it, bloc pattern is like MVVM in native android. The ViewModel class holds all the data, and it's then passed down to View so it can show stuff based on it, and sometimes directly access the data through the `viewModel` object, but the view never directly holds it or does operations on it. Have I understood the BLoC pattern wrong? And thanks for the explanation, but could I trouble you to use the `switchMap` in a code snippet? I'm not well versed enough in rx yet! username_1: My previous reply already contains a link to a complete rework of your code to use `switchMap` so if you click on the blue 'gist' you'll get to it. 👍 username_0: Totally overlooked the gist link. Thanks for both links, much appreciated. Status: Issue closed
kasper/phoenix
263149340
Title: Add support for Open Scripting Architecture Question: username_0: Hello, Apple supports Javascript in [OSA](https://developer.apple.com/library/content/documentation/AppleScript/Conceptual/AppleScriptX/Concepts/osa.html) That would open a whole new world for Phoenix. Answers: username_1: You can already run OSA by running it as an external task. Check the Task API, maybe that could help you achieve what you are aiming to do. username_0: Hmm, I meant to expose OSA APIs to phoenix.js. It'll make possible something like this right in the config: ```js const itunes = App.get ( 'iTunes' ); const selection = itunes.selection(); itunes.play(selection[0]); ``` Status: Issue closed username_1: Hi! Thanks for the feedback, but this is not in the context of this app. Use macOS AppleScript/JavaScript automation and run those as tasks from Phoenix.
jeremybarbet/react-native-modalize
946853338
Title: Add a hook to access Modalize itself inside it's children Question: username_0: **Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] A modal will be opened outside and will be closed after it finishes its job. The modal should be closed by itself after its jobs are done. Consider we have a `SingleSelectBottomSheetModal` component, which displays a bottom sheet with a flatlist and a close button inside the header, which let the user chose an option. the modal should be closed when the user selects an option or hits the close button. But i think the outer components should only care about the option of the user select, the modal should be closed inside the `SingleSelectBottomSheetModal ` **Describe the solution you'd like** A clear and concise description of what you want to happen. Add a hook named `useModalize` or something, it includes some `Modalize` methods such as `close()`. When the user taps on the close button, we can close the modalize inside the modalize, we don't need to add any callback to outer components. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. // TODO // BTW, Modalize is a great library~ Answers: username_1: Recently he has merged one implementation of this hook, but idk how to use username_0: https://jeremybarbet.github.io/react-native-modalize/#/HOOKS Maybe this? But this is not what i want, i need a hook that accesses the modalize **INSIDE** itself, which allows me to close the modal inside itself.
jupyterlab/jupyterlab
364436823
Title: Move the `share` command into its own extension to allow it to be overridden. Question: username_0: In order to allow custom building of URLs in the file browsers "Copy Shareable Link" functionality, we should move the `'filebrowser:share-main'` command's implementation and addition to the command registry into a small extension that can be disabled in specialized environments. cf. https://gitter.im/jupyterlab/jupyterlab?at=5bacc22e56877c463a90c648 cc: @username_1 Answers: username_1: Thanks! cross-ref https://github.com/jupyterhub/jupyterhub/issues/2017 where it would be handy to make 'shareable link' copy the `/hub/user-redirect` link instead of the local-only url username_2: Would it be best to have this as a separate extension, or a separate plugin only (keeping it in the original extension)? username_0: Sorry for being imprecise @username_2, I meant a `JupyterLabPlugin` and should have called it a plugin. I updated the issue. Status: Issue closed
WICG/ua-client-hints
551156379
Title: The server hosting the page should be able to set Client Hint restrictions that apply to anything that runs on hosted pages Question: username_0: I believe that server owners, should be able to add extra restrictions beyond the baseline on what scripts on the page are able to access, this would allow them to define a level of privacy for their users in line with my interests and allows those sites interested in increasing the level of privacy for their users to do so explicitly and potentially in a way the browser can signal to users. It could also become the basis of an additional plug in to the proposals around the [privacy budget](https://github.com/bslassey/privacy-budget). Let's take an example in which I am the owner of a server hosting a privacy conscious website. If I don't believe anything on my site would need access to the users' Architecture via a Client Hint, I should be able to not accept it in my server headers and also define to the browser that anything acting on the page to request that information is doing so against my wishes, thus restricting the access to the on-page object. If I want to run ads, but make this information unavailable to any ad that potentially seeks to violate users privacy out of my control, this seems an ideal mechanism. I would think that the ideal mechanism for this would be for the browser to read `Accept-CH` and translate that policy to the `navigator` object. I see there is a discussion of potentially using Feature Policy for this, and that might be an acceptable (if potentially repetitive approach), but it would mean that it would need equal detail for being able to accept or restrict specific Client Hints. In that situation it would be especially useful if Feature Policy could set Client Hint allowances in detail and have it filter down to default settings for iFrames, and to have that level of detailed control available to place on iFrames themselves (so a user could set Allow on UA at a page level, but disallow it at a particular iframe level) Answers: username_1: Feature Policy [allows for that](https://w3c.github.io/webappsec-feature-policy/#iframe-allow-attribute). Marking this as a "feature request", since this doesn't have to be a core part of UA-CA. username_2: I'd say that it makes sense to have feature-/privacy-parity between the headers and JavaScript API. Access to the information should be governed the same way, by Feature-Policies. Arguments for this solution include: - Same opt-in mechanism for same information - Enables control over both passive- _and active_ fingerprinting. username_3: In terms of headers, you can set a Feature Policy to "none" to disallow use even from yourself. So for example if you send a response with the header `Feature Policy: ch-arch 'none'` then no first or third party subresource will get the sec-ch-ua-arch header, even if the send an equivalent accept-ch response. I do agree though that these restrictions should have parity between JS and CH. username_1: I suspect that barring UA information from 3P iframes without explicit 1P delegation would not add much from a privacy perspective (as that usage is accounted for), but will result in more developer pain when we'd want 3P developers to adopt the new APIs. username_2: Maybe som "pain" will lead to a more "educated" use of the new APIs. Looking at the history that made the User-Agent develop to its current state, the reckless use of poor regular expressions in server side code could have been avoided if it was made clear that the User-Agent was a part of a bigger policy system. Today the server side code relying on the User-Agent has pretty much settled on scalable solutions, but it is not hard to find ugly JavaScript code making use of string matches on `navigator.UserAgent`. I can easily see the same thing happen for `getUserAgent()`. Restricting usage of `getUserAgent() ` with feature policies will not solve that problem, but definitely lead to more educated use IMO. username_2: I'd like to give this issue a friendly bump. What is the reason that the JS API is not protected either by Feature-Policy/Permission-Policy nor any user permission, while hints in transmitted through headers are gated both by opt-on and Feature-Policy/Permission-Policy? If the intention is to ship Chrome with the JS API completely open (even for 3rd parties!), I think it will undermine the initial motivation for this whole project; Privacy. High-entropy information is suddenly available to anyone... As a minimum I'd expect Feature-Policy/Permission-Policy to explicitly delegate access to specific hints and a user permission dialog (like we have for the geolocation API) to allow the user to be in charge of his own privacy. username_1: High entropy information is already available to everyone that's able to run active content. The goal of this project was to limit *passive* entropy from being sent to everyone by default. Clamping down on active entropy (e.g. through a [Privacy Budget](https://github.com/bslassey/privacy-budget)) would be a separate project. username_2: I do understand the challenge of aligning all related projects in the team, but my concern still stands: Why address passive fingerprinting by adding more active un-gated entropy? Just seems strange. I think it's naive to think that the information provided by the new APIs will not be used in functionality that normally raise privacy concerns. Even if there is a long term idea of Privacy Budgets etc., I can easily imagine the community embracing the new API making it difficult to gate at a later stage, and we'll have another "user-agent freeze situation". username_3: To be clear: all of the information from these APIs are currently available in most browsers' User-Agent (which is [available in JS and unrestricted](https://developer.mozilla.org/en-US/docs/Web/API/NavigatorID/userAgent)) as well, with the expectation that information that isn't normally available being left blank. There's no new entropy being revealed that wasn't freely available before username_2: Totally get the message @username_3, but that is only true until the [user-agent freeze](https://groups.google.com/a/chromium.org/g/blink-dev/c/-2JIRNMWJ7s/m/yHe4tQNLCgAJ). So this just feels like a lost opportunity to improve privacy. username_4: All new features that can be detected will provide some level of new entropy, even if the entropy is duplicate information, and with the userAgent freeze feature, there is another level of entropy introduced by comparing the 2 features. `if (!navigator.userAgentData) // user doesn't have this feature else // detect if userAgent is forzen by comparing user's HighEntropyValues with Navigator platform/userAgent ` I don't think there is anyway of escaping that this introduces a new and unique entropy. But, why introduce if it is already available? Doesn't this now provides a new compare fingerprint and isn't it unnavoidable that some navigator.userAgent strings will match and some won't? For example, I'm on beta and userAgent returns 192.168.3.11, but uaFullVersion returns 84.0.4147.21. So, if a script is comparing these, it should be inferred that my userAgent is frozen or spoofed.
MicrosoftDocs/appcenter-docs
486351941
Title: Missing blue chat button Question: username_0: "To get help on any issue you may experience, use the blue chat button at the bottom-right corner of every App Center portal page" Yet on every app center page I can see, there is no blue chat button. Why is this? I even tried turning off all my ad and content blockers but I still don't see any blue chat button. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 45dd71a0-a249-a87b-e4eb-e4da0a20472f * Version Independent ID: 450c26f5-ab13-8df5-d45a-6086f5391b23 * Content: [App Center Support - Visual Studio App Center](https://docs.microsoft.com/en-us/appcenter/general/support-center#feedback) * Content Source: [docs/general/support-center.md](https://github.com/MicrosoftDocs/appcenter-docs/blob/live/docs/general/support-center.md) * Service: **vs-appcenter** * GitHub Login: @username_1 * Microsoft Alias: **username_1** Answers: username_1: Hello @username_0 ! This could be due to an ad-blocker. Can you try disabling those? Thanks! username_0: @username_1 That was my first thought as well, which is why as I indicated in my first message, I tried turning all of them off and refreshing the page. username_1: Oh sorry I totally missed that part! Let me chat with the team and get back to you as soon as I have an update for you. username_1: Could you please confirm your OS version + Browser version? Thank you! username_0: Mac os 10.13.6 Safari 11.1.2. I forget but I may also have tried in chrome. —<NAME> username_1: Hello! Thanks for sending that over. It looks like we are unable to reproduce your issue. The team would like to ask you a few additional questions. Would you mind sending us an email at <EMAIL> with a link to this issue? Thank you! Status: Issue closed username_1: I'll close this issue here, but feel free to re-open if you don't get this solved via the support email. username_2: I have this problem now. tested with chrome / firefox on my mac. and firefox on my pc. A few days ago I had no problem with the chat. username_1: Hi @username_2 , thanks for reporting this. We are having a hard time reproducing this issue, if you could reach out to us via <EMAIL> we will to look into it further. Thanks!
Automattic/mongoose
140428834
Title: Connection returned from mongoose.connect has undefined connection.db property in 4.4.7 Question: username_0: We're connecting to our MongoDB instance with `mongoose.connect`. Up till the release of Mongoose `4.4.7`, the returned Connection's `connection.db` property gave us a `dropDatabase` method. After upgrading to `4.4.7`, the `connection.db` property is undefined. We connect this way: ```javascript var db = mongoose.connect(config.db.uri, config.db.options, function (err) { // Log Error if (err) { console.error(chalk.red('Could not connect to MongoDB!')); console.log(err); } else { // Enabling mongoose debug mode if required mongoose.set('debug', config.db.debug); // Call callback FN if (cb) cb(db); } }); ``` And we have a Grunt task to drop the database: ```javascript grunt.task.registerTask('dropdb', 'drop the database', function () { // async mode var done = this.async(); // Use mongoose configuration var mongoose = require('./config/lib/mongoose.js'); mongoose.connect(function (db) { db.connection.db.dropDatabase(function (err) { if (err) { console.log(err); } else { console.log('Successfully dropped db: ', db.connection.db.databaseName); } db.connection.db.close(done); }); }); }); ``` I reverted back to Mongoose `4.4.6` and this issue went away. Answers: username_0: Upon further investigation, it appears the connection isn't being made. When I change the above example that connects to the database, to use `createConnection` instead of `connect`, I get the following error thrown. ![gh-1264-mongoose-createconnection-error](https://cloud.githubusercontent.com/assets/5613734/13730475/8af2e30c-e90d-11e5-8a5e-c89c77d66198.PNG) Perhaps, I'm complicating the matter by changing the connection method to use `createConnection` (for the purposes of debugging) but I find it interesting that no error is thrown when attempting the connection by using `connect`; but it's not making the connection. Any ideas? username_1: Same situation, same problem... username_2: I am having connection issues as well in 4.4.7 Status: Issue closed username_0: @username_3 My issue went away after upgrading to 4.4.8. Thanks for the prompt fix!
cpriego/valet-linux
220404469
Title: Valet not pointing to 127.0.0.1 Question: username_0: but still, it's pointing to 127.0.53.53 I'm using `Ubuntu 16.04` Answers: username_1: 127.0.53.53 is an alert meaning "the TLD you used has another master DNS", basically your computer is hitting a public DNS server instead of dnsmasq. You can check which dns is responding using the command: ``` dig test.dev ``` Ironically that address still "points" to localhost so it should work anyway. username_0: ![screenshot from 2017-04-12 10-14-54](https://cloud.githubusercontent.com/assets/20656462/24938318/fac8bc1e-1f68-11e7-830a-d48fe2ae3f39.png) I'm not really good at network stuff. But as I've understand `test.dev` is using the `127.0.53.53` ip. Am I right? And for some reason I still can't access my `local-directory.dev` site ![screenshot from 2017-04-12 10-22-55](https://cloud.githubusercontent.com/assets/20656462/24938483/0eed3e44-1f6a-11e7-96ce-7cb2ba400f22.png) username_2: Same thing here, using Arch Linux username_2: Problem solved for me. i added `[main] plugins=keyfile dns=dnsmasq` to my NetworkManager configuration file, then reinstalled nginx adn started the nginx.service again. In fact, there was an error starting the service, that's why i couldn't my local-directory.dev site. @username_1 check if you can manualy start the nginx.service, in arch this would be: `systemctl start nginx.service` username_0: @username_2 thanks for the tip. I'll try that when I get home. username_0: @username_2 @username_1 @username_3 i think I've found something that concludes ( base on my understanding ) why I can't access my `local-directory.dev` site. ![screenshot from 2017-04-13 19-51-27](https://cloud.githubusercontent.com/assets/20656462/25003611/d95088f8-2082-11e7-9062-aa811d228d0c.png) tho I can't find out why nginx can't access port 80 when in fact nginx itself uses the port. ![screenshot from 2017-04-13 19-56-09](https://cloud.githubusercontent.com/assets/20656462/25003721/64af302a-2083-11e7-84f0-8c5c34f464c5.png) username_3: That last image was really helpful. For some reason nginx is running as www-data. With Valet, nginx should be run with **your** username, otherwise it is unable to access your files. Could you please show the contents of your `/etc/nginx/nginx.conf` file? username_4: @username_0 I'm using 17.04 and ran into a similar issue. As a temporary workaround I added `nameserver 127.0.1.1` to /etc/resolv.conf. Still looking for a permanent solution myself. username_0: @username_3 here my `nginx.conf` file settings ``` user www-data; worker_processes auto; pid /run/nginx.pid; events { worker_connections 768; # multi_accept on; } http { ## # Basic Settings ## sendfile on; tcp_nopush on; tcp_nodelay on; keepalive_timeout 65; types_hash_max_size 2048; # server_tokens off; # server_names_hash_bucket_size 64; # server_name_in_redirect off; include /etc/nginx/mime.types; default_type application/octet-stream; ## # SSL Settings ## ssl_protocols TLSv1 TLSv1.1 TLSv1.2; # Dropping SSLv3, ref: POODLE ssl_prefer_server_ciphers on; ## # Logging Settings ## access_log /var/log/nginx/access.log; error_log /var/log/nginx/error.log; ## # Gzip Settings ## gzip on; gzip_disable "msie6"; # gzip_vary on; # gzip_proxied any; # gzip_comp_level 6; # gzip_buffers 16 8k; # gzip_http_version 1.1; # gzip_types text/plain text/css application/json application/javascript text/xml application/xml application/xml+rss text/javascript; ## [Truncated] # # See sample authentication script at: # # http://wiki.nginx.org/ImapAuthenticateWithApachePhpScript # # # auth_http localhost/auth.php; # # pop3_capabilities "TOP" "USER"; # # imap_capabilities "IMAP4rev1" "UIDPLUS"; # # server { # listen localhost:110; # protocol pop3; # proxy on; # } # # server { # listen localhost:143; # protocol imap; # proxy on; # } #} ``` username_0: @username_3 after updating my dnsmasq on my OS, `*.dev` sites are now pointing to `127.0.0.1` but wasn't accessing the right directories when using `valet park` and thus using the `/var/www` directory in every `*.dev` site. username_3: Yes, you need to reinstall valet with `valet install` to correct the `/etc/nginx/nignx.conf` misconfiguration. The file was overwritten when you reinstalled nginx. username_0: wow.. works like magic. thanks @username_3. Status: Issue closed username_5: I found a workaround. Today is the second day I am doing something related to Laravel so please bare with me if this does not work. :) Run command `ping anything.test` if it pings to 127.0.0.1 then we are good to go. Now just run `valet link` inside the directory you are willing to place your websites in and now you should be able to access your sites on browser using `any_project.test`.
pachyderm/pachyderm
261161198
Title: auth admin username typo = locked out of cluster Question: username_0: I was activating auth and didn't realize my GH username has capitalization (because past joey was a moron). The real issue here arises because now there are no admins in my cluster even though I'm logged in. And there's no way to fix this without deleting my cluster and data. <img width="1709" alt="screen shot 2017-09-27 at 5 38 17 pm" src="https://user-images.githubusercontent.com/1795008/30944447-aa8247e8-a3ac-11e7-9220-7b8ed9b7473b.png"> @username_1 proposed a fix where activation and auth happen at the same time making whoever activates auth an admin.<issue_closed> Status: Issue closed
oracle/weblogic-kubernetes-operator
409431405
Title: Logstash container (based on image logstash:5) inside the weblogic-operator pod doesn't start due to a bad permission on /var/lib/logstash directory Question: username_0: When installing weblogic-operator using Helm, the value.yaml file used as a template to create (in my case) my-operator-values.yaml, contains the attribute logStashImage with value logstash:5. Unfortunately this container does not start due to the error: Path "/var/lib/logstash" must be a writable directory. It is not writable To reproduce the error just run, after having created your version of values.yaml (in my case my-operator-values.yaml) the command: helm install kubernetes/charts/weblogic-operator \ --name weblogic-operator \ --namespace weblogic \ --values my-operator-values.yaml \ --wait Answers: username_0: Just opened the PR "Issue #876 fix" that, in may case, solved the issue username_0: ``` $ minishift version minishift v1.30.0+186b034 $ minishift ssh docker version Client: Version: 1.13.1 API version: 1.26 Package version: docker-1.13.1-88.git07f3374.el7.centos.x86_64 Go version: go1.9.4 Git commit: 07f3374/1.13.1 Built: Fri Dec 7 16:13:51 2018 OS/Arch: linux/amd64 Server: Version: 1.13.1 API version: 1.26 (minimum version 1.12) Package version: docker-1.13.1-88.git07f3374.el7.centos.x86_64 Go version: go1.9.4 Git commit: 07f3374/1.13.1 Built: Fri Dec 7 16:13:51 2018 OS/Arch: linux/amd64 Experimental: false ``` Status: Issue closed
tpai/tpai.github.io
317942189
Title: Could not use sequelize.js with mysql:8 Question: username_0: **Situation** Use `sequelize.js` with `mysql:8` will cause an error: ``` SequelizeConnectionError: Client does not support authentication protocol requested by server ``` No matter you upgrade the `sequelize` or `mysql2` package version, it still happen. **Workaround** Downgrade mysql to `5.6`. Answers: username_1: In your SQL Workbench, execute `ALTER USER 'root'@'localhost' IDENTIFIED WITH mysql_native_password BY '<PASSWORD>';` Replace '<PASSWORD>' with your new password Ctrl + C existing npm start and do a npm start again.
gemfire/node-examples
673976172
Title: Minor doc update Question: username_0: When I was running and following the `book-service` example with `pcc`, found that if we can improve that last section on `README.md` to also consider deleting the `service-key`. 1. If the Cloud Cache service instance is no longer needed, delete the service-key and then the service: Deleting service-key ``` $ cf delete-service-key PCC-TLS PCC-TLS-service-key Really delete the service key PCC-TLS-service-key?> y Deleting key PCC-TLS-service-key for service instance PCC-TLS as admin... OK ``` Now delete the service ``` $ cf delete-service PCC-TLS Really delete the service PCC-TLS?> y Deleting service PCC-TLS in org test_org / space test_space as admin... OK Delete in progress. Use 'cf services' or 'cf service PCC-TLS' to check operation status. ``` Answers: username_1: Please review the pull request I submitted to incorporate your improvement: https://github.com/gemfire/node-examples/pull/5 Cheers, <NAME>
Jimmy-sha256/deribit_websocket_v2
664990617
Title: what is the api variable value in function loop(self, api, request) Question: username_0: Hello James I am trying to use your deribit ws api to retrieve market data. I am struggling what is the value of the api variable in the loop function? Maybe you have also a quick example to retrieve market data? Thanks for your assistance. Greetings, Peter Answers: username_1: Hi the api variable determines whether we are connecting to the public or private endpoint. The function `async def private_api(self, request):` returns the variable containing the data for a private api connection and `async def public_api(self, request):` returns the variable containing the data for a public api connection I hope this makes sense, if you bear with me I will put an example together showing how to retrive market data. username_1: A quick example of a single script that will fetch the index prices of test net eth / btc, hope this helps. Let me know if your still struggling or want an example doing for a particular end point. import asyncio import websockets import json client_id = 'enter_client_id_here' client_secret = 'enter_client_secret_here' client_url = 'wss://test.deribit.com/ws/api/v2' # create a websocket object class WS_Client(object): def __init__(self, client_id=None, client_secret=None): self.client_id = client_id self.client_secret = client_secret self.client_url = client_url self.json = { "jsonrpc" : "2.0", "id" : 1, "method" : None, } # send a public method request async def public_api(self, request): async with websockets.connect(self.client_url) as websocket: await websocket.send(request) response = await websocket.recv() response = json.loads(response) return response # create an asyncio event loop def loop(self, api, request): response = asyncio.get_event_loop().run_until_complete( api(json.dumps(request))) return response # example get index function https://docs.deribit.com/?python#public-get_index def index(self, currency): options = {"currency" : currency} self.json["method"] = "public/get_index" self.json["params"] = options return self.loop(self.public_api, self.json) client = WS_Client(client_id, client_secret) btc_index = client.index('BTC') eth_index = client.index('ETH') print(btc_index) print(eth_index)
preciofishbone/Omnia-Foundation
208856001
Title: Support Regions in Distributed and Memory Cache Question: username_0: Add support for specifying regions where cached objects can be stored. Examples of API below **Adding object to cache using region** ` cache.AddOrUpdateDistributedCache("test", "pelle", DateTimeOffset.Now.AddHours(1), region: "customregion"); ` **Getting item from a specific region** ` cache.GetFromDistributedCache<string>("test", region: "customregion"); ` **Getting items in a region** ` var cachedItems = cache.GetRegionFromDistributedCache( "customregion"); `<issue_closed> Status: Issue closed
rust-lang/rust
339167994
Title: [Rust 2018] Inconsistent anonymous lifetimes Question: username_0: This is allowed by Rust 2018: ```rust impl<T: SomeTrait<'_>> SomeType<T> { ... } ``` while this is not allowed: ```rust fn foo<T: SomeTrait<'_>>() { ... } ``` Answers: username_0: I replicated this on the playground: https://play.rust-lang.org/?gist=b7f75c8e5df6945c2065892882b59764&version=nightly&mode=debug&edition=2018 Status: Issue closed username_1: Duplicate of https://github.com/rust-lang/rust/issues/45667
seek4science/seek
1138807741
Title: Harmonized registration process for all types, with upload, metadata extraction, and then form filling. Question: username_0: different types behave differently, and there are different implementations behind them with some duplication of code. Each type needs to follow the same set of steps, and similar UI, that handles 1. Provide files 2. Extract any metadata 3. Metadata form (including extracted metadata) 4. Submit with validation Answers: username_1: sorry, what is "type" in this context?
gocql/gocql
268231230
Title: Issue with LWT -- "if not exists" Question: username_0: Please answer these questions before submitting your issue. Thanks! ### What version of Cassandra are you using? [cqlsh 5.0.1 | Cassandra 2.1.15.1423 | CQL spec 3.2.1 | Native protocol v3] one Cassandra cluster with 3 nodes. ### What version of Gocql are you using? ### What did you do? We are using the cassandra feature of `IF NOT EXIST` clause during the insertion of the row. When multiple goroutines are trying to insert, we are observing that often multiple goroutines are succeeding at a time while ONLY one should have succeeded. Sample code and log are attached. [lwtsample.gz](https://github.com/gocql/gocql/files/1412878/lwtsample.gz) schema for lwt is as follows: ``` desc testlock CREATE KEYSPACE testlock WITH replication = {'class': 'NetworkTopologyStrategy', 'USC1': '3'} AND durable_writes = true; CREATE TABLE testlock.paas_lock ( tenant text, namespace text, pool text, created timestamp, owner text, uuid text, PRIMARY KEY ((tenant, namespace, pool))) ``` ### What did you expect to see? ### What did you see instead? Here is the log from sample code. you'd notice that goroutine 6 and 5 both got lock once it was released by 27. Only one should have succeeded. Have anyone faced this before; Is there anything amiss in code? : ``` go run main.go 2017/10/16 11:00:31 CassandraDB: endpoint(10.180.52.46:9042), keyspace(testlock) 2017/10/16 11:00:31 CassandraDB: timeout(600ms), retry(&{NumRetries:3}) 2017/10/16 11:00:38 lockobj: &{Tenant:paypal Ns:mkr Pool:PATCH_POLICY Owner: Uuid:27 Created:2017-10-16 11:00:35.13594583 -0700 PDT Success:false} 2017/10/16 11:00:38 lock: &{Tenant: Ns: Pool: Owner: Uuid: Created:0001-01-01 00:00:00 +0000 UTC Success:false} 2017/10/16 11:00:38 got lock 27 2017/10/16 11:00:38 released lock 27 2017/10/16 11:00:39 lockobj: &{Tenant:paypal Ns:mkr Pool:PATCH_POLICY Owner: Uuid:6 Created:2017-10-16 11:00:35.134512416 -0700 PDT Success:false} 2017/10/16 11:00:39 lock: &{Tenant: Ns: Pool: Owner: Uuid: Created:0001-01-01 00:00:00 +0000 UTC Success:false} 2017/10/16 11:00:39 got lock 6 2017/10/16 11:00:39 lockobj: &{Tenant:paypal Ns:mkr Pool:PATCH_POLICY Owner: Uuid:20 Created:2017-10-16 11:00:35.135152313 -0700 PDT Success:false} 2017/10/16 11:00:39 lock: &{Tenant: Ns: Pool: Owner: Uuid: Created:0001-01-01 00:00:00 +0000 UTC Success:false} 2017/10/16 11:00:39 lockobj: &{Tenant:paypal Ns:mkr Pool:PATCH_POLICY Owner: Uuid:5 Created:2017-10-16 11:00:35.13465921 -0700 PDT Success:false} 2017/10/16 11:00:39 lock: &{Tenant: Ns: Pool: Owner: Uuid: Created:0001-01-01 00:00:00 +0000 UTC Success:false} 2017/10/16 11:00:39 got lock 5 ``` If you are having connectivy related issues please share the following additional information ### Describe your Cassandra cluster please provide the following information - output of `nodetool status` - output of `SELECT peer, rpc_address FROM system.peers` - rebuild your application with the `gocql_debug` tag and post the output Answers: username_1: Is this the issue that is being discussed on #265 ? My understanding is that if Cassandra doesn't support client side timestamps with LWT then it should either ignore them or error out, this could be a Cassandra bug but we can work around it. username_0: @username_1 Not sure if this is same as #265 related to client side timestamp. Will try out #996 fix. Any debugging suggestions, in case if issue still exists? username_1: Are you running into something like http://datanerds.io/post/cassandra-no-row-consistency/? I verified locally that doing LWT works correctly from the driver standpoint and #996 does not cause or solve any issues. username_1: I can't replicate any issue by hammering different LWT queries, closing. Status: Issue closed
open-telemetry/opentelemetry-collector-contrib
895789475
Title: Metric tags are being dropped when using ObserverType: summary Question: username_0: Currently, the statsd receiver supports the conversion of timings to either gauge/summary. While using gauge we can see that the tags of the original incoming metric are correctly translated to output metric labels but this does not happen when using summary which is seen https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/main/receiver/statsdreceiver/protocol/metric_translator.go#L71 **Steps to reproduce** Use config: ``` statsd: endpoint: 0.0.0.0:<port> aggregation_interval: 60s enable_metric_type: false timer_histogram_mapping: - statsd_type: histogram observer_type: summary - statsd_type: timing observer_type: summary And send timing metrics with labels. Answers: username_1: Thanks! The summary part does not have tags. I will add. username_1: Are working on a fix, will have the PR tomorrow. username_1: Fixed by that PR 3440. username_1: Cam we close that? username_2: Done. Status: Issue closed
zengfr/easymodbus4j
747580670
Title: Java 11支持么? Question: username_0: 我用了modbus4j仅支持jdk8,使用jdk11,master可以发送数据,接收数据jvm报错。 Answers: username_1: 你好 错误发来看下 或截图 username_0: @username_1 ``` # # A fatal error has been detected by the Java Runtime Environment: # # EXCEPTION_ACCESS_VIOLATION (0xc0000005) at pc=0x0000000180005b00, pid=12208, tid=4156 # # JRE version: OpenJDK Runtime Environment (172.16.17.32+1) (build 172.16.17.32+1-LTS) # Java VM: OpenJDK 64-Bit Server VM (172.16.17.32+1-LTS, mixed mode, tiered, compressed oops, g1 gc, windows-amd64) # Problematic frame: # C [rxtxSerial.dll+0x5b00] # # No core dump will be written. Minidumps are not enabled by default on client versions of Windows # # If you would like to submit a bug report, please visit: # https://bugreport.java.com/bugreport/crash.jsp # The crash happened outside the Java Virtual Machine in native code. # See problematic frame for where to report the bug. # --------------- S U M M A R Y ------------ Command Line: -Djava.library.path=D:\Workspaces-Learning\sshx-ktzl\lib -javaagent:C:\Users\clayton\AppData\Local\JetBrains\Toolbox\apps\IDEA-U\ch-0\202.7660.26\lib\idea_rt.jar=58732:C:\Users\clayton\AppData\Local\JetBrains\Toolbox\apps\IDEA-U\ch-0\202.7660.26\bin -Dfile.encoding=UTF-8 com.mytest.serial.SerialPortTest Host: Intel(R) Core(TM) i5-7200U CPU @ 2.50GHz, 4 cores, 15G, Windows 10 , 64 bit Build 19041 (10.0.19041.546) Time: Fri Nov 20 15:49:19 2020 �й���׼ʱ�� elapsed time: 12.366819 seconds (0d 0h 0m 12s) ``` username_0: @username_1 I use `jssc-2.9.2` ,`rxtx-2.2pre2`, `modbus4j-3.0.3` solved the problem. But, I don't quite understand yet at the very start, I have problems with `rxtx` , `modbus4j` . ``` <dependency> <groupId>org.rxtx</groupId> <artifactId>rxtx</artifactId> <version>2.2pre2</version> </dependency> <dependency> <groupId>io.github.java-native</groupId> <artifactId>jssc</artifactId> <version>2.9.2</version> </dependency> <dependency> <groupId>com.infiniteautomation</groupId> <artifactId>modbus4j</artifactId> <version>3.0.5</version> </dependency> ``` username_1: 一、 [rxtxSerial.dll+0x5b00] EXCEPTION_ACCESS_VIOLATION win10下可能出现。和 easymodbus4j无关,尝试下以下方法: 1、win10系统补丁 以及 最新Java 11 以及确认都是64位。 2、dll 放置正确的目录位置 调用的是正确的64位版本 3、添加serialPort.addEventListener(eventListener);解决 二、easymodbus4j 只支持基于TcpIp的 tcp/rtu协议。 不支持直接接串口的RTU。串口的RTU 需硬件接 转tcp的RTU或TCP硬件模块(几十块钱 透明转换)。 Status: Issue closed
karol-f/vue-custom-element
669701520
Title: Possible to use actions and router using custom element? Question: username_0: I´m trying to reuse a component that makes calls to the router i.e: `this.$route.params;` and I get an undefined when packing it into a vue-custom-element same thing happens for actions, i´m getting errors like `TypeError: Cannot read property 'commit' of undefined` Any thoughts on this? Do i need to do something else in order to call $route in Vue, or define something when injecting Vuex context variables?
InsertKoinIO/koin
702361315
Title: kotlinx-coroutines-core:1.3.9 is dependent, InstanceCreationException occurs. Question: username_0: **Describe the bug** if `implementation "org.jetbrains.kotlinx:kotlinx-coroutines-core:1.3.9"` occurs `org.koin.core.error.InstanceCreationException` **Koin project used and used version (please complete the following information):** ```` ext.koinVersion = '2.1.6' implementation "org.koin:koin-core:$koinVersion" implementation "org.koin:koin-androidx-scope:$koinVersion" implementation "org.koin:koin-androidx-viewmodel:$koinVersion" ```` **Additional moduleDefinition** - **It doesn't happen on all devices.** - **It does not occur in apk file. (only aab file in playstore)** **My Application Code** ```` startKoin { androidLogger() androidContext(this@SangSikApplication) androidFileProperties() koin.loadModules(DiModule) koin.createRootScope() } val DiModule = listOf( retrofitPart, apiPart, prefPart, filePart, resourcePart, activityViewModelPart, fragmentViewModelPart ) ````
Clinical-Genomics/scout
769720088
Title: Bam files missing for mtDNA Question: username_0: I have noticed that BAM files for mtDNA are missing for several cases. I have now checked and all cases uploaded in Scout after 20th of November miss BAM files for mtDNA. Could you please check what has happened. Answers: username_1: Oh no! We'll check and fix ASAP! username_2: Its from the CG remake: the config value with the path to mt_bam is missing from the "new" API exported config files. Transferring there, but we will keep tracking it!
GreenWaves-Technologies/gap_sdk
876835277
Title: Bidirectional transport for NINA Question: username_0: I noticed in the nina_w10 transport API, the __nina_w10_receive_async function is not implemented. Is it planned to be implemented? https://github.com/GreenWaves-Technologies/gap_sdk/blob/master/rtos/pmsis/pmsis_bsp/transport/nina_w10/nina_w10.c I am using the AI deck on a Crazyflie and I would like to route communication between the GAP8 and the AI deck over the NINA so it is compatible with the lighthouse deck. I have it working in the GAP8->NINA->CF direction, but I found that the SPI transport functions in the opposite direction is not implemented. We can do CF->GAP8 communication over UART1, but with the caveat that the LED on the lighthouse has to be disabled and UART1 interferes with the hyperflash memory on the AI deck. Answers: username_1: Hi Sorry for that, it's really very low priority for us, so we don't have a date about when it will be implemented. I suggest you to implement it by yourself, we have lots of examples on using SPI and UART, and if you have any question, we are happy to help. username_0: Ok, that's good to know. I can look into implementing it myself and possibly submit a pull request. Status: Issue closed username_1: Thanks, please feel free to post your question and pull request.
brinckmann/montepython_public
388401919
Title: Best way to run a model? Question: username_0: What is the best way to run a model? Should i run once with high N and --update 50, Or repeat the same run for a model various times for increasing N say 10000, 100000, 1000000. Is there even any difference between the two ways? If so which way will theoretically get me to the highest resolution the fastest? thank you Answers: username_0: Thank you! So using -r i can continue old chains that were completed and increase N, so like having the best of both worlds. Status: Issue closed
MicrosoftDocs/powerbi-docs
377633904
Title: Continuous vs Categorical could use more detail Question: username_0: This page briefly mentions continuous and categorical, but it doesn't provide enough detail, or a link to more detail about them. Thanks --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ad604499-b845-44de-acc9-1bd9a7a97488 * Version Independent ID: f16079ec-98ee-e35d-f928-c5d493365051 * Content: [Customize X-axis and Y-axis properties - Power BI](https://docs.microsoft.com/en-us/power-bi/visuals/power-bi-visualization-customize-x-axis-and-y-axis) * Content Source: [powerbi-docs/visuals/power-bi-visualization-customize-x-axis-and-y-axis.md](https://github.com/MicrosoftDocs/powerbi-docs/blob/live/powerbi-docs/visuals/power-bi-visualization-customize-x-axis-and-y-axis.md) * Service: **powerbi** * GitHub Login: @username_1 * Microsoft Alias: **username_1** Status: Issue closed Answers: username_1: Hi @username_0 , I'll add more information to the document and republish. Look for an update by next week. Thanks for your feedback, Michele
morganericsson/msengcs
282733100
Title: Beskrivning av forskning inom DV Answers: username_1: Forskning inom datavetenskap Forskningen inom datavetenskap riktar in sig mot teknik för att bygga smartare mjukvarusystem. Aktiviteterna bedrivs i huvudsak inom fyra forskningsområden med flera beröringspunkter: - CPS – Cyber-fysikaliska system, - ISOVIS – Informations och mjukvaruvisualisering, - AdaptWise – Självanpassande mjukvarusystem och - Software Technology Labs. Forskningsgruppen för Cyber-fysikaliska Systems (CPS) studerar samverkan och återkoppling mellan de inbyggda mjukvarukomponenterna (cyber) och fysiska komponenterna, exempelvis, mekaniska komponenter, energisystem, mänskliga aktiviteter och omgivande miljö vilka utgör det sammanhållna systemet CPS-design tar hänsyn till flera faktorer som real-tids aspekter, energiförbrukning, tillförlitlighet, tillgänglighet och säkerhet. För att hantera denna komplexa verklighet krävs samverkan av experter från olika discipliner. Några av de konkreta utmaningarna finns inom modellering för analys, analyser, simuleringar, transformationstekniker för modell baserad utveckling, verifiering och realtids-analyser. Gruppen fokuserar på flera aspekter, (i) modellbaserad design, syntes och verifikation av CPS, (ii) CPS-tillförlitlighet, säkerhet och integritet, (iii) storskalig dataanalys för CPS, (iv) hierarkisk modellering och optimering för CPS, samt (v) applikationer av CPS i smarta energisystem och transportsystem. Det huvudsakliga målet för forskningen, smartare system, studeras interaktionen mellan mjukvarusystemen, som tillhandahåller "smartare" och de fysiska enheter som utgör "systemen". Informations- och mjukvaruvisualiseringsforskningen inom ISOVIS fokuserar på explorativ analys och visualisering av stora och komplexa informationsmängder, exempelvis från biovetenskap, humaniora eller mjukvaruutveckling. Vår vision är att attackera utmaningar kopplade till stora datamängder med en kombination av individ-centrerad dataanalys och interaktiv visualisering för förbättrat beslutsstöd. En interaktiv visualisering och visuell analys ger effektivare och mer tillförlitliga analyser av de stora och komplexa datamängder som samlas in från smarta system. Som ett resultat kan vi skapa en bättre förståelse för och följaktligen också skapa bättre förutsättningar för att utveckla system, exempelvis öka möjligheterna att förutse systemens prestanda och konstrueras för att förbättra möjligheterna att förutsäga och kontrollera prestanda och beteende. Forskargruppen AdaptWise studerar metoder och tekniker för självanpassande mjukvarusystem. Gruppen fokuserar på att skapa försäkringar för att systemen uppfyller de egenskaper som efterfrågas. Självanpassning är en princip för att hantera osäkerheter i driftsatta mjukvarusystem, exempelvis dynamisk resurstillgång och förändringar av kraven Att lägga till självtillpassningsfunktioner till programvarusystem gör det möjligt för dessa system att agera autonomt med en dynamik som är omöjlig att bestämma vid tidpunkten då systemen designas och implementeras. Därför kan dessa system ta över enformiga uppgifter som vanligtvis utförs av operatörer med kortare svarstider och med högre kvalitet, vilket ger smartare system. Software Technology Labs studerar teknologier för dataintensiva mjukvarusystem. Maskininlärning och artificiell intelligens använder (stora) datamängder för att automatisera uppgifter som resonemang, planering, beslutsfattande och prediktion. Skalbara teknologier möjliggör en hantering av stora datavolymer, exempelvis i stora real-tids dataströmmar. Gruppen studerar även möjligheten att se IT-system som informationskällor. Idén är att på detta vis skapa möjligheter för kontinuerlig analys och därmed förbättra kvalitetsarbetet. Teknologi för dataintensiva mjukvarusystem möjliggör att omvandling av data till information och information till kunskap vilket bidrar till en enklare implementation av smartare system och vissa fal även en mer eller mindre helautomatisk generering av komponenter för dessa system. Status: Issue closed
SlimeKnights/TinkersConstruct
70636484
Title: render issue Question: username_0: hello all. I have found a little render bug: ![2015-04-24_11 20 18](https://cloud.githubusercontent.com/assets/11927133/7316070/ebcf6b1c-ea73-11e4-98c3-bf404d951a6c.png) Answers: username_1: It is not a render bug. This happens because the cast texture doesn't have the exact same cut out as a shovel head. username_0: ok but, why that append only cobalt ? username_1: The shovel head must be a bit smaller with cobalt. Status: Issue closed
platformio/platformio-home
631593382
Title: Libraries: Could not update library Question: username_0: PIO Core Call Error: "Library Storage: C:\\Users\\peter\\.platformio\\lib\r\nUpdating ArduinoJson @ 6.15.1 [6.15.2]\r\nUninstalling ArduinoJson @ 6.15.1: \t[OK]\r\nLooking for ArduinoJson library in registry\r\nFound: https://platformio.org/lib/show/64/ArduinoJson\r\nLibraryManager: Installing id=64 @ 6.15.2\r\nArduinoJson @ 6.15.2 has been successfully installed!\r\n\n\nError: VCS: Could not receive an output from `['git', 'ls-remote']` command ({'out': '', 'err': \"remote: Repository not found.\\nfatal: repository 'https://github.com/Jason2866/Arduino.git/' not found\\n\", 'returncode': 128})"
marcus-ma/myBlog
1047177658
Title: 利用双数组树来压缩高效解决大量敏感词匹配问题 Question: username_0: #情景导入 ##敏感词识别 &emsp;&emsp;手头上有一份含有1万条违规词数据的词库文件,需要监控游戏中玩家所发的聊天文本,当识别出包含违规词的文本,把违规词以 `*`号替换。此外识别到拉人等相关违规词需要把玩家账号封禁。 <br> ##正则匹配 &emsp;&emsp;面对字符串匹配问题,最简单直接的方案就是使用正则。但是问题来了,若是单纯使用正则匹配,每段聊天文本都要进行1w次的匹配,且该功能具备实时性,但众所周知正则是很耗费性能的,一旦数据量翻个几十倍,这性能可想而知。 <br> ##trie树 &emsp;&emsp;`Trie树,又称单词查找树,是一种树形结构,是一种哈希树的变种。典型应用是用于统计,排序和保存大量的字符串(但不仅限于字符串),所以经常被搜索引擎系统用于文本词频统计。它的优点是:利用字符串的公共前缀来减少查询时间,最大限度地减少无谓的字符串比较,查询效率比哈希树高.` <br><br> &emsp;&emsp;以上是关于trie树的介绍,利用前缀字典树的这一数据结构(构建方法是利用map与指针),可以达到匹配一段文本最坏也就O(n+m)`n为文本长度,m为违规词长度`。这是一种以空间换时间的方法。 <br> ##AC自动机 &emsp;&emsp;Trie虽然能用于多模式匹配,但是每次匹配失败都需要进行回溯,如果模式串很长的话会很浪费时间。 <br> &emsp;&emsp;所以AC自动机应运而生,如同Manacher一样,AC自动机利用构建的fail指针操作阻止了模式串匹配阶段的回溯,将时间复杂度优化到了O ( n ) `n为文本长度`。详细概念可以参考一下文章:[AC自动机](https://zhuanlan.zhihu.com/p/80325757) <br> ##双数组树+AC自动机 &emsp;&emsp;上面所提到的trie树,虽然构建完后的查询匹配会很快,但相对的,其所占用的空间也大,map的底层使用hash函数也有性能的损耗,含有大量指针不利于GC。 <br> &emsp;&emsp;针对以上的情况,查了一下相关资料,找到了一种很好的数据压缩方案且不需要指针的数据结构:`Double Array trie` <br> &emsp;&emsp;`双数组Trie树(DoubleArrayTrie)是一种空间复杂度低的Trie树,应用于字符区间大的语言(如中文、日文等)分词领域。是Trie结构的压缩形式,仅用两个线性数组来表示Trie树,该结构有效结合了数字搜索树(Digital Search Tree)检索时间高效的特点和链式表示的Trie空间结构紧凑的特点。双数组Trie的本质是一个确定有限状态自动机(DFA),每个节点代表自动机的一个状态,根据变量不同,进行状态转移,当到达结束状态或无法转移时,完成一次查询操作。在双数组所有键中包含的字符之间的联系都是通过简单的数学加法运算表示,不仅提高了检索速度,而且省去了链式结构中使用的大量指针,节省了存储空间。` <br> &emsp;&emsp;详细概念与构建可以参考一下文章:[双数组Trie树(DoubleArrayTrie)Java实现](https://www.hankcs.com/program/java/%E5%8F%8C%E6%95%B0%E7%BB%84trie%E6%A0%91doublearraytriejava%E5%AE%9E%E7%8E%B0.html) <br> ##实战 &emsp;&emsp;接下来将参考上面那篇文章,把原JAVA版本的代码改用golang来实现。 <br> &emsp;&emsp;其构建的原理有2条:1)敏感词需要先排序 ,2)字符串的字符状态转移方程 ```golang base[t]+c = base[tc] check[tc] = t ``` <br> &emsp;&emsp;源码如下 ```go package main import ( "bufio" "encoding/gob" "errors" "fmt" "log" "os" "runtime" "sort" "strings" "time" ) const ( _ROOT_STATE = 0 [Truncated] var m runtime.MemStats runtime.ReadMemStats(&m) mb := 1024 * 1024.0 fmt.Printf("\nAlloc = %v\tTotalAlloc = %v\tSys = %v\t NumGC = %v\n\n", float64(m.Alloc)/mb, float64(m.TotalAlloc)/mb, float64(m.Sys)/mb, m.NumGC) } func main(){ start:= time.Now() d,err:=BuildFromFile("./dictionary.txt") if err!=nil { panic(err) } runtime.GC() res := d.SearchAndReplace("打秋风u打算盘") fmt.Println("查询成功:",time.Since(start).Nanoseconds()) fmt.Println(res) } ```
SAP/ui5-webcomponents
721598359
Title: ui5-link does not support aria-label Question: username_0: **Describe the bug** The ui5-link component does not pass any aria attributes down to the anchor tag inside it, so screenreaders only read the inner text of the link. **To reproduce** Steps to reproduce the behavior: 1. Go to https://sap.github.io/ui5-webcomponents/playground/components/Link/ in chrome 2. using dev tools, add an aria-label to a ui5-link 3. notice the anchor tag in the shadow root does not get the aria-label 4. using any screenreader, focus on the link and notice the aria-label is not read out. **Expected behavior** The aria-label and aria-labelledby attributes should be respected by ui5-link and passed to the inner anchor element to allow compliance with accessibility standards. **Screenshots** ![Screen Shot 2020-10-14 at 8 40 37 AM](https://user-images.githubusercontent.com/24360164/96016393-88539c80-0dfd-11eb-949d-f4bafca5f319.png) **Context** - UI5 Web Components version: any - screen reader: NVDA, JAWS, Mac VoiceOver **Affected components** ui5-link **Organization:** SuccessFactors **Priority:** Medium Thank you!<issue_closed> Status: Issue closed
porres/pd-else
551563347
Title: pd-lib-builder makefile does not use PDDIR or DESTDIR Question: username_0: In v1.0-beta26-128-ga1e469c, specifying PDDIR or DESTDIR does not affect the make process. These have to be defined in the pd-lib-builder makefile to have any effect. I did not check the rest of the environment variables to see if they're working Answers: username_1: I don't really know what you're talking about or what is the issue. Are you saying I should change https://github.com/username_1/pd-else/blob/master/Makefile , but why? username_1: or are you having issues with pd-lib-builder? In which case you should take it to https://github.com/pure-data/pd-lib-builder username_0: To build the else library, i had to specify the PDDIR and DESTDIR in the pd-lib-builder/Makefile.pdlibbuilder . Traditionally, for the other pure data libraries that use pd-lib-builder, doing this "make PDDIR=<path to puredata> DESTDIR=<install directory> build" works. But in this case it does not. It could be the current pd-lib-builder used in the else library is older. username_1: which one, for example? username_1: well, I just updated to the latest pdlubuilder (6.0), since this is strictly a pdlibbuilder issue, I'm closing this, we can take it to the pdlibbuilder repository's issue if somehow pdlibbuilder isn't working as expected for ELSE, cause I have literally no clue Status: Issue closed
llldddbbb/dbblog
451841219
Title: 为什么在运行sql文件的时候,category那里的rank字段会出现语法错误? Question: username_0: 我在搭数据库的时候,这个地方会中断,而且我没找到问题出在什么地方,日志里也没有说问题所在,只说rank字段创建语法有问题。 1064 - You have an error in your SQL syntax; check the manual that corresponds to your MySQL server version for the right syntax to use near 'rank int null comment '级别', parent_id int default '0' null comment '父主键' at line 6 Answers: username_0: 了解了,版本太高,导致rank字段不能创建了。。 Status: Issue closed
OrchardCMS/OrchardCore
554648875
Title: Unable to resolve service for type 'YesSql.ISession' while attempting to activate 'OrchardCore.Workflows.Http.Services.HttpRequestRouteActivator' Question: username_0: I am getting following error on OrchardCore CMS startup. Any ideas on how to fix it? ``` fail: OrchardCore.Media.Azure.Startup[0] Azure Media Storage is enabled but not active because ConnectionString is missing or empty in application configuration. fail: OrchardCore.Media.Azure.Startup[0] Azure Media Storage is enabled but not active because ContainerName is missing or empty in application configuration. fail: Microsoft.AspNetCore.Server.Kestrel[13] Connection id "0HLT0MKP5R3C9", Request id "0HLT0MKP5R3C9:00000001": An unhandled exception was thrown by the application. System.InvalidOperationException: Unable to resolve service for type 'YesSql.ISession' while attempting to activate 'OrchardCore.Workflows.Http.Services.HttpRequestRouteActivator'. at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteFactory.CreateArgumentCallSites(Type serviceType, Type implementationType, CallSiteChain callSiteChain, ParameterInfo[] parameters, Boolean throwIfCallSiteNotFound) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteFactory.CreateConstructorCallSite(Type serviceType, Type implementationType, CallSiteChain callSiteChain) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteFactory.TryCreateExact(ServiceDescriptor descriptor, Type serviceType, CallSiteChain callSiteChain) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteFactory.TryCreateEnumerable(Type serviceType, CallSiteChain callSiteChain) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteFactory.CreateCallSite(Type serviceType, CallSiteChain callSiteChain) at Microsoft.Extensions.DependencyInjection.ServiceLookup.ServiceProviderEngine.CreateServiceAccessor(Type serviceType) at System.Collections.Concurrent.ConcurrentDictionary`2.GetOrAdd(TKey key, Func`2 valueFactory) at Microsoft.Extensions.DependencyInjection.ServiceLookup.ServiceProviderEngine.GetService(Type serviceType, ServiceProviderEngineScope serviceProviderEngineScope) at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService(IServiceProvider provider, Type serviceType) at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService[T](IServiceProvider provider) at OrchardCore.Modules.ModularTenantContainerMiddleware.Invoke(HttpContext httpContext) in C:\projects\orchardcore\src\OrchardCore\OrchardCore\Modules\ModularTenantContainerMiddleware.cs:line 73 at Microsoft.AspNetCore.Server.Kestrel.Core.Internal.Http.HttpProtocol.ProcessRequests[TContext](IHttpApplication`1 application) fail: Microsoft.AspNetCore.Server.Kestrel[13] Connection id "0HLT0MKP5R3CA", Request id "0HLT0MKP5R3CA:00000001": An unhandled exception was thrown by the application. System.InvalidOperationException: Unable to resolve service for type 'YesSql.ISession' while attempting to activate 'OrchardCore.Workflows.Http.Services.HttpRequestRouteActivator'. at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteFactory.CreateArgumentCallSites(Type serviceType, Type implementationType, CallSiteChain callSiteChain, ParameterInfo[] parameters, Boolean throwIfCallSiteNotFound) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteFactory.CreateConstructorCallSite(Type serviceType, Type implementationType, CallSiteChain callSiteChain) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteFactory.TryCreateExact(ServiceDescriptor descriptor, Type serviceType, CallSiteChain callSiteChain) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteFactory.TryCreateEnumerable(Type serviceType, CallSiteChain callSiteChain) at Microsoft.Extensions.DependencyInjection.ServiceLookup.CallSiteFactory.CreateCallSite(Type serviceType, CallSiteChain callSiteChain) at Microsoft.Extensions.DependencyInjection.ServiceLookup.ServiceProviderEngine.CreateServiceAccessor(Type serviceType) at System.Collections.Concurrent.ConcurrentDictionary`2.GetOrAdd(TKey key, Func`2 valueFactory) at Microsoft.Extensions.DependencyInjection.ServiceLookup.ServiceProviderEngine.GetService(Type serviceType, ServiceProviderEngineScope serviceProviderEngineScope) at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService(IServiceProvider provider, Type serviceType) at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService[T](IServiceProvider provider) at OrchardCore.Modules.ModularTenantContainerMiddleware.Invoke(HttpContext httpContext) in C:\projects\orchardcore\src\OrchardCore\OrchardCore\Modules\ModularTenantContainerMiddleware.cs:line 73 at Microsoft.AspNetCore.Server.Kestrel.Core.Internal.Http.HttpProtocol.ProcessRequests[TContext](IHttpApplication`1 application) ``` Answers: username_1: How can we repro this issue? username_2: If you accidentally type AddOrchardCore() in the 'ConfigureServices' in Startup.cs instead of type AddOrchardCms() ,then you get this error username_0: Sorry for the long delay, but yeah, as @username_2 mentioned it was typo in Startup Status: Issue closed
cypress-io/cypress
738925877
Title: 5.5.0 breaks window.addListener('message', listener) Question: username_0: ### Current behavior Since upgrading from 5.4.0 to 5.5.0, this code doesn't execute anymore: ``` const listener = (event: MessageEvent) => { if (event.data.type === 'form-submit') { console.log('yay'); } }; window.addEventListener('message', listener); ``` ### Desired behavior When testing with Cypress 5.4.0, the event listener is called. ### Test code to reproduce ``` const listener = (event: MessageEvent) => { if (event.data.type === 'form-submit') { console.log('yay'); } }; window.addEventListener('message', listener); ``` This code should be called when a form is submitted (in my case I'm using typeform). ### Versions Failing: 5.5.0 Working: 5.4.0 OS: MacOS / Linux Env: Local / DroneCI Answers: username_1: Can you provide the test code to run against this app code that will show the error? username_0: Hello, sorry about the delay, it has been fixed with cypress 6.0.0 Status: Issue closed username_2: <EMAIL>​ username_3: Seeing the same issue in v9.1.1 For the time being, moved to test "implementation detail" instead of testing integration. FWIW, the issue appears when testing communication between a page and an iframe. The parent can send messages, but will not receive messages from the iframe via `window.addEventListener('message', listener);`
LeaVerou/awesomplete
58850746
Title: ∪∴bamWar∵【닷 컴】동탄오피↙양산오피 상주오피☜ Question: username_0: ∪∴bamWar∵【닷 컴】동탄오피↙양산오피 상주오피☜∪∴bamWar∵【닷 컴】동탄오피↙양산오피 상주오피☜∪∴bamWar∵【닷 컴】동탄오피↙양산오피 상주오피☜∪∴bamWar∵【닷 컴】동탄오피↙양산오피 상주오피☜∪∴bamWar∵【닷 컴】동탄오피↙양산오피 상주오피☜∪∴bamWar∵【닷 컴】동탄오피↙양산오피 상주오피☜∪∴bamWar∵【닷 컴】동탄오피↙양산오피 상주오피☜∪∴bamWar∵【닷 컴】동탄오피↙양산오피 상주오피☜∪∴bamWar∵【닷 컴】동탄오피↙양산오피 상주오피☜∪∴bamWar∵【닷 컴】동탄오피↙양산오피 상주오피☜
CS2113-AY1819S2-T09-1/main
430741525
Title: Improve `find` command to support most module attributes Question: username_0: The `find` command only support mainly the module's name, code and its credits. A user would expect most of the module's attributes to be searchable. We should consider including the remaining attributes. i.e. `tags` and `semesters` Answers: username_1: Closing in view of #206. Status: Issue closed username_0: The `find` command only support mainly the module's name, code and its credits. A user would expect most of the module's attributes to be searchable. We should consider including the remaining attributes. i.e. `tags` and `semesters` Status: Issue closed
UpCloudLtd/upcloud-ansible
716433840
Title: Timeout as argument for individual methods Question: username_0: Would be nice to have timeout as argument for individual methods, too, like set 2 min timeout for create server -call _Originally posted by @Darep in https://github.com/UpCloudLtd/upcloud-ansible/pull/23#issuecomment-704788052_ Answers: username_0: Won't do, since we will be deprecating this repo and moving to https://github.com/UpCloudLtd/upcloud-ansible-collection Status: Issue closed
FusionAuth/fusionauth-issues
364241917
Title: Feature : Flexible Social Login Question: username_0: ## Social Login ### Problem In order to reduce friction during user acquisition, allowing a potential user to authenticate or register using an existing set of credentials such as Facebook, Google or LinkedIn can increase user adoption. ### Solution To allow a flexible solution that supports using social login on the FusionAuth OAuth login, through a Login API, or an SPA using a JavaScript widget - a comprehensive solution is required. Proposed initial provider support: * Facebook * Google * LinkedIn Possible additional providers: * Twitter * Instagram * GitHub * Microsoft * Amazon * MySpace (ha, just kidding) ### How to vote Please give us a thumbs up or thumbs down as a reaction to help us prioritize this feature. Feel free to comment if you have a particular need or comment on how this feature should work. Answers: username_0: Feature is in nearing development complete and should be released yet this month. username_1: once logged in, does app have access to user's info to make API calls on their behalf? ie: log in with Google to fusion auth means now I can put an inbox inside my app where they can read Gmail and do other googley things?? username_0: Hi @username_1 If I understand your question correctly, once a user logs into FusionAuth using Google (for example) will you have access to the Google credentials to retrieve information or make additional requests from Google on behalf of the logged in user? The short answer is yes. Any long lived token we receive from the identity provider such as Google will be stored and accessible to the FusionAuth API caller. username_1: oh, wonderful. I was thinking it would be great to store user tokens along with their profiles and user settings. my plan requires a solution that makes everyone, including me, totally locked off from users information. user logs into the app, it becomes their encrypted pipeline from them to app, my app will check if authed, if so, go ahead and make calls, if not, use auth call, whichever call, it goes to KrakenD, which is going to call a lot of APIs, return tokens to app which immediately encrypts the token in the appropriate sub-table with all the auth info and preferences for each particular service. , basically, my branch directly to that is app>userDB>userID>Modules>ModuleID, where moduleID holds the auth info and settings for the g-mail module. that's about all that gets held in that database. I'm wanting the auth server/container to be three layers really. the top layer authenticates with the database that this person is, in fact, a registered user of the app, doing so with MFA, unless trusted device or browser on a 2 week last connected lease before asks them again..... that's for logging into the app... may add the option for "always trust this ip".... second layer underneath that, now that they are authenticated as an app user, they can open the app and request to add further services for individual modules, this is where the auth server gets hit hard if that particular user has like 50 different services that they added (oh yes, there will be a price for that tier of service)... The auth service (in this case, fusion), would pull all the tokens and forward each one appropriately to the individual module (aka service aka set of API calls).. third level is error handling, because good propbability that one or more of the modules may have different rules, require an app-password, or have it's own 2FA, so the auth service has to have a layer capable of handling whatever secondary action is required and return the info to those particular modules that it's auth is successful and it may now go ahead with it's api calls.... planning on doing this all withReact frontend and holding the auth info in an escrow cache for any 2FA, and display "additional steps required" in the module on their screen, while dropping the errored out logins and placing a "log in again" link in the module's display.... but all the successfule modules are independent and got their auth's, so they are already busy making their calls and displaying their returned data... make sense so far? overall thinking is ... So Google or facebook or wherever would have all auth info and the module number and prefs data in one chunk in that userID table,... but i mean.. first they have to maintain their authed connection from their react frontend to my backend middleware. meaning that the order of things is 1 app server container sends a "hey, i need you to auth me" to fusion auth 2. fusion does it's thing, lets the user into the app or denies them or sends them password reset emails, 3. once they are authed with the app, Fusion calls up all the services this user wants to use, asking for auth refreshes (or first time auths).... 4. fusion returns the yes/no/wait2FA/eror status to each module 5. the modules that passed (now holding logged in auth) are authed to make the API calls 6. Modules individually ask middleware "fetch.$servicename.authcode" commands to the middleware. 7. Middleware makes the relevant stack of API calls to KrakenD Containerized cloud app. 8. KrakenD calls Google with 4-5 apisto get the different info it all needs, then filters/sums responses 8. krakend fires back streamlined api replies to the middleware, 9. middleware sends the info to the cache on the device 10, react hot reloads the information from cache and display is updated for that module. ########## Confusing part about the auth.. do i need auth server continually involved, sending the auth to mix with the API request at middleware?.... or am i right in thinking that once the module is authed, fusion is out of the picture till a change in state at the client side.... but until then, the app can make it's own GET/PUT calls... THEN... on logout from app, the modules are no longer running, modules no longer exist, there are no open authed paths to that user's services because the only point authorized to is the module.... so the user is protected because at no point in time does the server hold an auth position.... fusion would process between the modules and the services as if the module was the user, that gives the modules authorization to retrieve personal info for the user, and each module makes encrypted API calls from the module to the middleware (over the open internet), middleware decrypts the shortcode, makes the full API call to KrakenD, without needing to talk to Fusion, KrakenD calls, gets the response, to hand off to middleware, middleware takes the already filtered and summed response from Krakend, and makes shortcode data packs to pass back to the module's cache, it hot reloads from seeing changed data, and that's the summary ... except that i would need settings to de-auth the user's app and shut it down if the user is either idle for 15 minutes, closes the app, or logs out... some sort of settings, and that's when Fusion would step back in, yes? so Fusion has a couple important jobs to do, but they only occur at 5 points, then it sits back and just waits to see useridle/disconnected/logoff, at which point it revokes App access, and doesn't need to do anything else. Fusion would then be responsible for.... 1. App > register/Auth app use with email confirm OR send e-mail/SMS 2fa returning user log in 2. Module > Requests first auth to a service by whichever method service specified -----> Fusion stores refresh token/app specific password/whatever... ?? into it's own userDB ?? 3. User passes app-auth, Fusion notifies middleware, which then loads app, Modules then sequentially request Fusion provide them auth. 4. Fusion makes auth requests for the modules, and handles the responses, Passing modules their authority to make calls, OR that there was Error/Waiting/MFA.. whatever status that wasn't a pass, 5. Waiting for the 2FA or requesting a new password, whatever it needs, asks for to get user auth'd 6. Noticing the disconnection, and logging user out of my app. this is all how i understand things to be. i suppose i just realized, Fusion will have it's own userID database, and that my userDB with the users preferences and status (free user, or premium user, what's their module limit, what is their subscription, what is the layout grid of their modules, what theme are they using)... these actually Should be on different databases.. so fusionauth will have its own DB with auth codes and refresh tokens and saved passwords and all that, so that it's SSO as far as the user sees it on a day to day normal use where fusion pulls the auth stuff all from it's own database, handling all the auth storage.... and i can make that DB connect to one thing in this world, being the fusion auth container server, then if i ever have to perform maintenance, i just remote desktop to Fusionauth container and can use tools on it to manage it's DB.. but at least that way user data will be really really protected and secure. only have one connection and filtering by i.p address, the ip address really being a load balancing router so that the DB can be distributed on a CDN, and fusion auth just connects to one IP address. can it do that? i guess is what i'm wondering all about... between the hastags here is the most important. ########## ANYWAY... as far as users go.... Their personal data, images, writing, personal info, e-mail, whatever.. anything that the API's call for never get stored. My database holds the framework and user profiles for them as app users, as well as the list of microservice modules and layouts and stuff... upon them opening the app, the app plops the mostly empty framework on their screen, using the module placement, basic username, and simple profile information... Fusion Auth does all the authy stuff App gets the information itself by calling the up KrakenD API gateway, which returns filtered minimized clean info direct to the app. planning on a Redis cache to hold the app framework objects, and not sure yet about which DB to use to hold the user profiles, my little diagram i'm sending along with this post is going to have to change how i connect it all, having realized i'll be keeping FusionAuth's database separated from the userdb for the app.. both contain sensitive information so going to tuck them two steps removed and impossible for anyone but admin to log onto the os that's allowed to remotely log into the DB.. and keep it encrypted TLS anytime anything gets sent anywhere along any line sorry, i guess i'm super chatty... beeen at the computer too long and haven't talked to anyone in a couple days, even online... but i'm working this out slowly moving towards the solution soon as i started typing it really got me to wonder if fusionauth works the way i think it does... going to attach a little flowchart that's oversimplified, but can share a more detailed one if needed, soi can figure out the right solution. i gotta know... does fusion do all i think it can? or do i just go with something like Feathersjs as middleware and write individual passport.js into each module, then that way each module is it's own app self-authorizing and making it's own calls, and my parent app just holds them... i suppose that way i could open source the whole module framework, borrowing from other open source libraries to make one library for module building, with generators for each type of auth, then anyone could make a module for their favorite service and it could get added to the roster... and because they are all individual apps on the dash being used, it'd be suuuuuuper easy to integrate something like integromat on it's own tab where people could not only have all their apps visible and usable on the same screen, but actually just hit a button, their dashboard does a 180 and you can see the back of it and draw connections in either direction from visual endpoints.. be just like a wiring patchbay ,.. anyway!!! i'll shutthe hell up with my ideas, because even though i suck at code, i'm wicked at system design and ideas. ![overview](https://user-images.githubusercontent.com/777603/47767424-2c87bb80-dc99-11e8-9713-414dd13fc8c0.png) username_0: Twitter, Facebook, Google and a generic OpenID Connect provider are available in version 1.1.0. https://fusionauth.io/docs/v1/tech/release-notes/ username_0: Follow on features you may vote on for future work. https://github.com/FusionAuth/fusionauth-issues/issues/33 https://github.com/FusionAuth/fusionauth-issues/issues/34 https://github.com/FusionAuth/fusionauth-issues/issues/35
OpenDSA/OpenDSA-LTI
205960445
Title: Code Workout exercises not shown in Plain books Question: username_0: When we compile a book in "plain" (non-Canvas) form, the Code Workout exercises do not appear, let alone be usable. I spoke with Ayaan. While support for this use case will need to be developed, he thinks that it should not be hard. So, we need to do something on our side to provide the view to the exercise. Related to this, even in Canvas there should be a way to compile the book so that Code Workout exercises have "no deadline", and are not tied to any particular class from Code Workout's point of view. This will be useful for our "demonstration" books.
facebook/flow
207343562
Title: $PropertyType equivalent for Arrays Question: username_0: Is there, or could there be added, an equivalent of $PropertyType that works on Arrays? In other words, `$ArrayElement<Array<T>> == T` I’m working with auto-generated Flow types from GraphQL queries (using apollo-codegen), so being able to define my own types in relation to parts of those auto-generated ones is valuable: ``` // auto-generated type ServicesLoadQuery = { services: Array<{ code: string, name: string, hasMetadata: boolean, }>, }; // my type export type Service = $ArrayElement<$PropertyType<ServicesLoadQuery, 'services'>>; ``` Forgive me if I’ve missed this already being available. Answers: username_1: Not sure if it's the simplest solution, but should work: ```js type _ArrayElement<El, Arr: Array<El>> = El; type ArrayElement<Arr> = _ArrayElement<*, Arr>; ```; username_0: Awesome! That seems to work. username_1: Great! Should it be closed then? :) Status: Issue closed username_0: Sure. Might be nice to have it built-in, but rolling my own is not a problem. username_2: @username_0 Could you tell me how you use it? I tried like this (and many others), but it didn't work: type Services = ArrayElement<$PropertyType<ServicesLoadQuery, 'services'>>; const myServices: Services = [{code:'2', name:'2', hasMetadata: true}] username_0: `$ArrayElement` gives you what’s inside the array. So my guess is what you’d want is: ``` type Service = $ArrayElement<$PropertyType<ServicesLoadQuery, 'services'>>; const myServices: Service[] = [{code:'2', name:'2', hasMetadata: true}] ``` username_2: thx, works nice with type Service = ArrayElement<$PropertyType<ServicesLoadQuery, 'services'>>; const myServices: Service[] = [{code:'2', name:'2', hasMetadata: true}] username_3: A native flow solution for this could be nice. But thanks for providing a workaround! username_3: @username_1 thanks so much! How did you come up with this solution? It clearly shows that there are some part of Flow I don't understand... can you explain how it actually works? Thanks! username_4: Thanks for providing this workaround. It would be really nice to have it as a utility type.
o2team/tech-weekly
734649213
Title: Introducing the New JSX Transform Question: username_0: https://reactjs.org/blog/2020/09/22/introducing-the-new-jsx-transform.html React 17 中引入了新的 JSX 语法转换机制,好处除了书写 JSX 时不再需要提前引入 `React` 对象外,还做了一些性能优化和概念简化的工作。如 `key` 独立于其它 `props` 而单独传递,不建议在函数式组件中使用 `defaultProps` 等。文章介绍了相关背景和大致的改动内容,以及升级与兼容的方法。关于此改动的动机和实现细节,可以浏览此 [RFC](https://github.com/reactjs/rfcs/blob/createlement-rfc/text/0000-create-element-changes.md#detailed-design)。 Answers: username_1: 感谢贡献 Status: Issue closed
dotnet/roslyn
487480908
Title: Object synchronization method was called from an unsynchronized block of code Question: username_0: **Version Used**: net core sdk version: 3.0.100-preview8-013656 os version: ubuntu 16.04.4 LTS x64 **Steps to Reproduce**: clone this [project](https://github.com/username_0/ENode-Infrastructure.git) go to folder: test/ENode.Kafka.Tests run dotnet build "ENode.Kafka.Tests.csproj" -c Release -o /app/build **Exception** ``` You are using a preview version of .NET Core. See: https://aka.ms/dotnet-core-preview /usr/share/dotnet/sdk/3.0.100-preview8-013656/Roslyn/Microsoft.CSharp.Core.targets(59,5): error MSB3883: Unexpected exception: [/src/enode/src/ENode.Kafka/ENode.Kafka.csproj] /usr/share/dotnet/sdk/3.0.100-preview8-013656/Roslyn/Microsoft.CSharp.Core.targets(59,5): error : System.AggregateException: One or more errors occurred. (Object synchronization method was called from an unsynchronized block of code.) [/src/enode/src/ENode.Kafka/ENode.Kafka.csproj] /usr/share/dotnet/sdk/3.0.100-preview8-013656/Roslyn/Microsoft.CSharp.Core.targets(59,5): error : ---> System.ApplicationException: Object synchronization method was called from an unsynchronized block of code. [/src/enode/src/ENode.Kafka/ENode.Kafka.csproj] /usr/share/dotnet/sdk/3.0.100-preview8-013656/Roslyn/Microsoft.CSharp.Core.targets(59,5): error : at System.Threading.Mutex.ReleaseMutex() [/src/enode/src/ENode.Kafka/ENode.Kafka.csproj] /usr/share/dotnet/sdk/3.0.100-preview8-013656/Roslyn/Microsoft.CSharp.Core.targets(59,5): error : at Microsoft.CodeAnalysis.CommandLine.ServerNamedMutex.Dispose() [/src/enode/src/ENode.Kafka/ENode.Kafka.csproj] /usr/share/dotnet/sdk/3.0.100-preview8-013656/Roslyn/Microsoft.CSharp.Core.targets(59,5): error : at Microsoft.CodeAnalysis.CommandLine.BuildServerConnection.RunServerCompilationCore(RequestLanguage language, List`1 arguments, BuildPathsAlt buildPaths, String pipeName, String keepAlive, String libEnvVariable, Nullable`1 timeoutOverride, Func`3 tryCreateServerFunc, CancellationToken cancellationToken) [/src/enode/src/ENode.Kafka/ENode.Kafka.csproj] /usr/share/dotnet/sdk/3.0.100-preview8-013656/Roslyn/Microsoft.CSharp.Core.targets(59,5): error : --- End of inner exception stack trace --- [/src/enode/src/ENode.Kafka/ENode.Kafka.csproj] /usr/share/dotnet/sdk/3.0.100-preview8-013656/Roslyn/Microsoft.CSharp.Core.targets(59,5): error : at System.Threading.Tasks.Task.Wait(Int32 millisecondsTimeout, CancellationToken cancellationToken) [/src/enode/src/ENode.Kafka/ENode.Kafka.csproj] /usr/share/dotnet/sdk/3.0.100-preview8-013656/Roslyn/Microsoft.CSharp.Core.targets(59,5): error : at System.Threading.Tasks.Task.Wait(CancellationToken cancellationToken) [/src/enode/src/ENode.Kafka/ENode.Kafka.csproj] /usr/share/dotnet/sdk/3.0.100-preview8-013656/Roslyn/Microsoft.CSharp.Core.targets(59,5): error : at Microsoft.CodeAnalysis.BuildTasks.ManagedCompiler.ExecuteTool(String pathToTool, String responseFileCommands, String commandLineCommands) [/src/enode/src/ENode.Kafka/ENode.Kafka.csproj] ``` Answers: username_1: Dup of https://github.com/dotnet/roslyn/issues/37974 Status: Issue closed
janEntikan/tmx2bam
733896568
Title: [Issue] Tiles don't match the ones used in the TMX file. Question: username_0: A quick comparison of the issue with a bucket fill TMX file. It appears as if tmx2bam is choosing a different tile then the one required for the slot. ![Gyazo](https://i.gyazo.com/f5e3045bb0adba53b669ebf2b9bc651c.png) Answers: username_1: Fixed and thanks so much :) Status: Issue closed
wix/react-native-navigation
233773820
Title: Modal disables Android Developer Menu Question: username_0: ### Issue Description When a modal is active on top screen, Android Developer Menu is blocked and cannot be viewed. Also for the same reason I suppose, modal also disables native components which uses `View` or `ReactView` ### Steps to Reproduce / Code Snippets / Screenshots Install example project and show modal and try to show developer menu by cmd+M. --- ### Environment * React Native Navigation version: 1.1.16 * React Native version: 0.43.4 * Platform(s) (iOS, Android, or both?): Android * Device info (Simulator/Device? OS version? Debug/Release?): Android Emulator SDK 23 Answers: username_1: +1 username_2: This also affects React Native Navigation v2.
facebook/react-native
506163487
Title: setTimeout not working when remote debugging is enabled Question: username_0: I am trying to load a second screen after some timeout. The second screen does not open up even after waiting for a long time. React Native version: 0.60.5 ## Steps To Reproduce 1. Suppose I am on a Screen A. 2. From A, I want to open Screen B after timeout of 100 ms. I used the following code in componentDidMount of Screen A. setTimeout(() => { navigation.navigate("ScreenB"); }, 100) But, ScreenB does not open up. Answers: username_1: If you could reproduce your issue in a separate git repo, people could help you much faster and better. username_2: Facing the same. On [that](https://github.com/facebook/react-native/issues/9030#issuecomment-235639617) comment they say It's related to the difference between device and computer clock. And another old issue related to that: #4470 username_3: I have exactly https://github.com/facebook/react-native/issues/4470 but the issue is old and closed. When debugger is activated the following code executes before the time has passed. Without debugging the code executes after the time has passed. ``` const timeout = 60 * 1000 const t1 = moment() global.setTimeout( () => { const t2 = moment() console.log('setTimeout', timeout, ', result:', t2 - t1) }, timeout ) ``` without debugging: ``` LOG setTimeout 60000 , result: 60012 // If I change it from 60 to 4: LOG setTimeout 4000 , result: 8263 ``` with debugging: ``` setTimeout 60000 , result: 42189 // If I change it from 60 to 4: setTimeout 4000 , result: 194 ``` If this is about device and host clocks being different, would it be better to just choose one clock and stick with that?
UbuntuKylin/youker-assistant
53703577
Title: 为什么需要python? Question: username_0: debhelper (>= 9), dh-translations, libphonon-dev, libqt4-dev, fcitx-libs-dev, pkg-config, python-dev, python (>= 2.6.6-3~), qt4-default, qtchooser, phonon-backend-gstreamer, python-dbus, python-psutil, python-mechanize, python-beautifulsoup, python-pygame Answers: username_1: 先問問您自己,為什麼不需要 Python?
jsforce/jsforce
715775903
Title: Cannot escape `-` for SOSL queries Question: username_0: I am trying to execute a SOSL query which includes `-` in the search keyword. According to the official SOSL documentation `-` is a reserved keyword and can be escaped using `\-`. But when using with jsforce, I couldn't seem to escape it. ## Screenshot ![image](https://user-images.githubusercontent.com/30751793/95221016-a7a66480-0814-11eb-85a8-52df2bcb64a7.png)
goto-bus-stop/awestruct
422278680
Title: write new docs Question: username_0: thanks for the feedback! I don't have much time today, so will just post two very quick points in response to 2/6/7 that I hope are relevant to you: structs can be nested like ```js const A = Struct([ ['someValue', Struct.types.int8] ]) const B = Struct([ ['key', A], // reads an `A` and puts it at property `b.key` A, // reads an `A` and merges its properties into the resulting `b` ]) assert.deepEqual( B(Buffer.from([1, 2])), { key: { someValue: 1 }, someValue: 2 } ) ``` `Struct.types.skip` allows you to jump around by a certain number of bytes, but there is no builtin type to find a sequence and jump to it at the moment. I made a bunch of modules to read binary file formats using this package, for example: https://github.com/genie-js/genie-dat/blob/master/src/object.js which extensively uses nesting. Answers: username_1: Thanks for writing **awestruct** - I like the neat way to describe parsing schema you're doing here! Regarding how to improve your documentation (it's actually quite decent already...): I saw [jsBinarySchemaParser](https://github.com/matt-way/jsBinarySchemaParser), but it had pretty much no documentation. One nice thing they did tho was give real world example or parsing a GIF file. [binary-parser](https://github.com/keichi/binary-parser) has nice documentation and does schema description in similar way (chaining functions rather than using array notation), if you'd want some inspiration. Few questions: 1) What would be the best way to write a loop in array notation? For example I have a section in binary file with readings from a sensor that are delimited by a string (ie `^reading#time20191013#value32145^reading...`) 2) (tl;dr: probably yes - rtfm!) Is it possible to define small Structs as a building block and reference them in another struct? For example, main struct parses the file and when it encounters a section for [person], it calls a Struct_Person to parse the section and output an object `userid: {name: xx, surname: yy, gender: F,dob: yymmdd}`. Or from question 1, outputs array of readings like `[{date: yymmddhhmmss, value: 32145},{date: yymmddhhmmss, value: 987123}, {...}]` . Hmm, looking at your Readme examples it looks like it should be possible, since you're declaring Structs in place. I'd say for complex files it might be better to declare Structs elsewhere and just reference them, following DRY principles. Ah, you've got `Struct.type` , yaay! I'll give it a try! 3) BigInt/int64 is with us! [Node 12](https://nodejs.org/dist/latest-v12.x/docs/api/buffer.html#buffer_buf_readbigint64be_offset) and [V8 engine](https://v8.dev/blog/bigint) I might give it a go writing a type for it at some point, but I'm just starting with buffers/typedArrays so not there yet to do it. 4) **Value Paths** is a huge thing that will help with creating a parsing script. Thanks a lot for this! 5) for [Custom types: Struct.Type(type)](https://github.com/username_0/awestruct#custom-types-structtypetype) when you write "_// always 1 byte, could also write as { size: 1 }_", do you mean that the return here should always be 1, or you mean that if we declare 1 it will be always 1? How would I dynamically declare this value when invoking this type during parsing? 6) Might be in the docs, but I didn't see it and at the moment not sure how I do this: Create a Struct/Type that scans the file searching for a particular string (text or raw bits/hexes) and skips the buffer/stream until it finds the trigger sequence. Something like (pseudocode) `LOOP (for buffer index i) if (Int8AtIndex(i) != 'AF0CBB7D') {skip(i), i++} else {foundIt => parseAtIndex(foundIt)}` . The reason is that I have a file where there are many sections for which I don't know what they are and I'm not interested in spending time to figure them out, but there are few sections that have data I want to extract. So I want to skip the junk, go to a section that starts with a keyword (i.e. sample_names), parse that section until it ends with a keyword and then jump to another section further down (i.e. reading_values), parse that section until its end keyword and finish. 7) Is it possible to use offsets to jump throughout the file from one location to another? Not for my current user case, but I know there are file formats where the structure is not linear but uses pointers instead. Or, actually, in my file I have a table of contents with names of entries and pointers to a later table with sections of values corresponding to entries in the table of contents. I was planning to parse the sections separately in sequence and then combine, but if there is a way to use pointers, that would sometimes help. Something like `[TOC: {Jane:c, Andy:a, Nick:b}] [......] [VALUES: {a:321, b:765, c: 984}]` Thanks for reading! username_0: thanks for the feedback! I don't have much time today, so will just post two very quick points in response to 2/6/7 that I hope are relevant to you: structs can be nested like ```js const A = Struct([ ['someValue', Struct.types.int8] ]) const B = Struct([ ['key', A], // reads an `A` and puts it at property `b.key` A, // reads an `A` and merges its properties into the resulting `b` ]) assert.deepEqual( B(Buffer.from([1, 2])), { key: { someValue: 1 }, someValue: 2 } ) ``` `Struct.types.skip` allows you to jump around by a certain number of bytes, but there is no builtin type to find a sequence and jump to it at the moment. I made a bunch of modules to read binary file formats using this package, for example: https://github.com/genie-js/genie-dat/blob/master/src/object.js which extensively uses nesting. username_1: Thanks a lot for fast reply, but no need to rush - my questions are for your consideration when convenient (I hope they will help other people as well). I'll play with the code and see how far I get but also got other work to do in between ^^ I'm happy tho as it looks like writing my parsing script will become that much easier. username_1: I see in [genie-dat/struct.js](https://github.com/genie-js/genie-dat/blob/master/src/struct.js ) that it's possible to use reference as `s.$parent`. Might be worth adding description of referencing this way to [Value Paths](https://github.com/username_0/awestruct#value-paths) readme section. username_2: An alternative to new docs could be to convert to TypeScript...
gabrielsaruhashi/RAMP
370820475
Title: Create 211 Endpoint with map + list toggle Question: username_0: https://projects.invisionapp.com/share/RQN6YTQUMB7#/screens/311032688 1) toggle between map and listview 2) design page according to mocks 3) create new screen with information about PLACE endpoint 4) add functionality to marker Answers: username_0: one commit per feature Status: Issue closed
rust-lang/rust
294220485
Title: `map` causes wasm target to return pointer to incorrect memory value Question: username_0: Hey everyone 👋 I'm new to Rust and I'm loving it so far! Nice work! 😄 I'm having an issue when compiling to WebAssembly with a pointer being returned to an incorrect value after using `map`: #[no_mangle] pub fn pointer_to_six() -> *const u32 { let _unused = vec![0].into_iter().map(|x| x).collect::<Vec<u32>>(); let v = vec![6]; v.as_ptr() } This returns a pointer to the value `3612` instead of the value `6`. If I remove the line with the `map` on it it returns `6` as expected. Note: that line doesn't reference any existing variables or set any variables that are used in the future. I created a [minimum reproducible failure repo here](https://github.com/username_0/wasm-map-bug). If there's anything else I can provide that would help track this down let me know! Thanks! rustc 1.25.0-nightly (def3269a7 2018-01-30) binary: rustc commit-hash: def3269a71be2e737cad27418a3dad9f5bd6cd32 commit-date: 2018-01-30 host: x86_64-apple-darwin release: 1.25.0-nightly LLVM version: 4.0 Answers: username_1: `v` is deallocated as soon as the function returns, the returned pointer is no longer valid when it is used (use after free). username_0: ah ha! For anyone wondering this how to fix it: use std::mem; #[no_mangle] pub fn pointer_to_six() -> *const u32 { let _unused = vec![0].into_iter().map(|x| x).collect::<Vec<u32>>(); let v = vec![6]; let ptr = v.as_ptr(); mem::forget(v); ptr } Thanks! Sorry for the noise 🙂 Status: Issue closed
elytra/GlassHearts
419747478
Title: Glass Hearts gives free self-sacrifice with Blood Magic Question: username_0: Minecraft Version: 1.12.2 Forge: 14.23.5.2815 Glass hearts: 1.12.1-1.0.66 Blood Magic: 1.12.2-2.4.0-102 Reproduction is easy, just pick up a Sacrificial Dagger and rightclick with it near a blood altar. You will see that the altar accumulates LP, but despite the appropriate sound effect, camera flinch, and even the health bar flashing, health is not removed. Also affects directly "using" an orb to directly sacrifice health into your personal LP network. Notably, overdrawing your LP network with a sigil (I tested with a lava sigil) subtracts health correctly.
emergenzeHack/covid19italia_segnalazioni
598094262
Title: Le professionalità dei Live al servizio del Paese #ChiamateNoi è una piattaforma di rappresentanza i Question: username_0: <pre><yamldata> Da_chi_offerta: chiamate noi Descrizione: "Le professionalità dei Live al servizio del Paese\n#ChiamateNoi è una\ \ piattaforma di rappresentanza indipendente partita dal basso che raccoglie professionisti\ \ ed aziende del comparto dell’organizzazione di eventi e che raduna in un unico\ \ elenco le alte professionaltà spendibili in settori diversi da quello dello spettacolo\n\ \n\"Siamo un insieme di professionisti dello spettacolo e degli eventi che ricopre\ \ a 360gradi tutto il mansionario necessario per la costruzione da zero di un evento\ \ e/o una struttura anche complessa, atta ad accogliere pubblico e grandi folle.\"\ \n\nIl Nostro obiettivo: dare una risposta immediata alla richiesta di specifiche\ \ professionalità, abituate ad operare in contesti difficili, sempre nuovi e con\ \ tempistiche molto contratte, ma garantendo sempre il miglior risultato possibile\ \ a prescindere dalle difficoltà incontrate.”\n\n#Chiamatenoi nel rispetto di tutti\ \ i soggetti e gli attori del mondo dello spettacolo teme che la crisi generata\ \ da Covid-19 creerà forti disagi economici ai propri colleghi e si prefigge senza\ \ scopo di lucro di:\n\n- essere ulteriore luogo di incontro tra domanda e offerta\ \ per mansioni utili ad ideare, montare, smontare, spostare, organizzare, strutture\ \ temporanee o permanenti atte a fronteggiare l’emergenza sanitaria \n\n- essere\ \ ulteriore luogo di incontro tra domanda e offerta in filiere produttive diverse\ \ da quello dello spettacolo, anche nei mesi futuri, con la speranza di dare la\ \ possibilità agli iscritti di trovare occasioni di ricollocamento temporaneo in\ \ altri settori." Destinatari: tutti Link: https://www.chiamatenoi.it/ Natura: solidale Tipo_di_soggetto: pubblico Titolo: I professionisti dello spettacolo a servizio dell'Emergenza </yamldata></pre>
openaddresses/machine
187815723
Title: Try to recover from S3 multipart upload errors Question: username_0: Saw one of these when doing a recent tile index: ``` Traceback (most recent call last): File "/usr/local/lib/python3.4/dist-packages/openaddr/ci/__init__.py", line 1115, in decorated_function return route_function(*args, **kwargs) File "/usr/local/lib/python3.4/dist-packages/openaddr/ci/tileindex.py", line 133, in main tile.publish(s3.bucket) File "/usr/local/lib/python3.4/dist-packages/openaddr/ci/tileindex.py", line 76, in publish collect.write_to_s3(s3_bucket, zipfile.filename, keyname) File "/usr/local/lib/python3.4/dist-packages/openaddr/ci/collect.py", line 227, in write_to_s3 mp.complete_upload() File "/usr/local/lib/python3.4/dist-packages/boto/s3/multipart.py", line 319, in complete_upload self.id, xml) File "/usr/local/lib/python3.4/dist-packages/boto/s3/bucket.py", line 1805, in complete_multipart_upload response.status, response.reason, body) boto.exception.S3ResponseError: S3ResponseError: 200 OK <?xml version="1.0" encoding="UTF-8"?> <Error><Code>InternalError</Code><Message>We encountered an internal error. Please try again.</Message><RequestId>39412F28EBB52A21</RequestId><HostId>Nf/irK7q5vrQnmQhG7YZUT6jqcXGCkqBQnphdMZkANSOrhPw2VpP7OE4VUujPlcz8ujLSsoOvRI=</HostId></Error> ```
co-IT/ngrx-ducks
542531986
Title: Create docs for new API Question: username_0: ## Contents - Getting started guide - createDuck - reducers - actions - selectors - bindSelectors (plain selectors, selectors with props) - pick - effects - testing - reducer - actions - effects - schematics Answers: username_0: Demo of new API can be found at: https://stackblitz.com/edit/ngrx-ducks-8-4 username_0: See https://co-it.gitbook.io/ngrx-ducks Status: Issue closed
ARM-DOE/pyart
104287661
Title: Additional examples in documentation Question: username_0: Additional examples should be added to Py-ART's documentation to describe new features which have been added to the package. This list will be updated as suggestions are made: * Dealias with region based algorithm. * Dealias with unwrap_phase algorithm. * Plot VPT data * GateFilter usage. * Create a Grid with the gates_to_grid algorithm Answers: username_1: I'm going to take a stab at this so can you assign #366 to me. -C username_1: Thanks. username_1: @username_0, Did you have notebooks with some of these already created that I could draw from? username_0: I think a few examples can be found in the notebooks from the [2015 AMS](https://github.com/openradar/AMS-Short-Course-on-Open-Source-Radar-Software) and [2015 ARM/ASR meeting](https://github.com/ARM-DOE/notebooks/tree/master/ASR_PI_2015) course materials. I know that both have examples of the region based dealiasing algorithm. username_2: Examples for most of these exist. I also added more notebooks for ingest and gridding etc. Closing as this is outdated. Can open a new issue if new example ideas arise. Status: Issue closed
wbuchanan/brewscheme
120288509
Title: Stata installation error Question: username_0: Dear William, I have tried to install using your instruction: net inst brewscheme, from("http://username_1.github.io/brewscheme/") replace But, I get the error back from Stata: file http://username_1.github.io/brewscheme/brewcolors.ado not found could not copy http://username_1.github.io/brewscheme/brewcolors.ado (no action taken) After that, I tried to install manually using your Github files, and was somewhat succesful installing the two foldes brewcolors and style under C:\ado\personal\ but I am not sure if the whole system operates as designed by you. Can you check the net install on your server (and should I re-install)? Best regards, Eric Answers: username_1: The installation issue should be resolved now. Installing this way is actually pulling the dev branch versions of the files. It shouldn't be an issue where the .ado files are located as long as they are somewhere on the ADOPATH. Status: Issue closed
accordproject/concerto-ui
429740163
Title: Component: Concerto (Model) Graphical Editor Question: username_0: - Viewer/editor for Concerto models as interactive UML style diagram, using https://github.com/projectstorm/react-diagrams Answers: username_1: Hello @username_2 and @username_3 . I want to work on Concerto Model Web Editor for GSOC. I have read the documentation. As per my understanding from the zoom call , we need to create a web editor where the user edits the graphical representation, and the system generates a model file (.cto) from it. This editor can be used by both developer and non-developer . I have made a small layout for this, Kindly have a look and guide me further for this project. ![concerto_model_web_editor](https://user-images.githubusercontent.com/29794748/76355868-09642600-633b-11ea-9e73-7e8dfc0b5f62.png) username_1: For generating diagrams, we will be using https://github.com/projectstorm/react-diagrams? username_2: Adding @username_4, @dselman, and @mttrbrts for feedback also. username_3: Hi @username_1, this library looks nice! That being said, we are not tied to a certain library. Best to research & see which one makes the most sense for this project. username_1: I will make a small prototype using this https://github.com/projectstorm/react-diagrams and share with you all for your feedback. Thanks! username_4: The UI mockup you made looks great for the purposes of a proposal. @username_1 If your proposal is selected, I can help you make a higher fidelity mockup. username_1: I will try my best in proposal. I am really excited to work with you @username_4 .
AtomLinter/linter-clojure
123961697
Title: ClojureScript Question: username_0: Would it be simple to modify this plugin to support ClojureScript as well? Answers: username_1: Assuming the clojure.jar file supports it with the same options and the same output format it should be as simple as adding that to the list of supported scopes. Have you tried that locally? username_0: @username_1 - can you explain a little better what you mean? I'm very new to Clojure / ClojureScript development, though an experienced programmer username_1: In [this block](https://github.com/AtomLinter/linter-clojure/blob/v1.1.1/lib/init.coffee#L22-L28) of code the command that gets executed is built up. First either `java` or the contents of `javaExecutablePath` is used as the command base. The the parameters get assembled. By default it looks like it runs something similar to this: ```ShellSession java -jar /path/to/closure-1.7.0.jar -i /path/to/file_currently_open.ext ``` So if you manually run a command like that on a ClojurScript file and get reasonable output, ClojustScript should work if it is added to the list of `grammarScopes` for this plugin. Having said all that, now that I've checked while writing that up I see that it _should_ already be working since `source.clojurescript` is in the list of allowed grammars, do you have an example file where it isn't working? What do you get when you run `Editor: Log Cursor Scope` (<kbd>Ctrl</kbd> + <kbd>Shift</kbd> + <kbd>Alt</kbd> + <kbd>P</kbd>) on a ClosureScript file? username_0: Thanks for working through this :) When I run log cursor scope this is what's returned: ``` Scopes at Cursor source.clojure meta.expression.clojure meta.definition.global.clojure meta.expression.clojure meta.expression.clojure meta.expression.clojure meta.expression.clojure ``` And here's an example of a basic linting error: <img width="517" alt="screen shot 2015-12-28 at 3 20 26 pm" src="https://cloud.githubusercontent.com/assets/236943/12027123/fcddb104-ad76-11e5-970f-73fa12f4312b.png"> username_1: Looks like it's working properly already then :stuck_out_tongue: Were you expecting something that it isn't doing, or had you just not tried it yet? username_0: Well what I mean about supporting ClojureScript is `js->clj` shouldn't be unresolved (its a common built in function in the global js namespace), and neither should `defproject` -- unless I am misunderstanding something? username_1: You'll have to ask somebody familiar with ClojureScript for that, @username_2? username_1: Marking this as closed since linter-clojure is already running on ClojureScript files, feel free to continue the discussion though :wink:. Status: Issue closed username_0: Sounds good. @username_2, any help would be appreciated. I know there is a `cljs.jar` file that is used to compile ClojureScript; wondering if we could leverage that somehow within the library along with the typical `clojure.jar` that is used for linting. username_1: Just to be clear, I have no experience with Clojure beyond knowing it's a language. If there are actual implementation details that need changing to get proper support in by all means please reopen this issue. username_0: I'm assuming there is, but I honestly have no clue what they could be (total noob here). Whatever it is I imagine its trivial, though, as there are linters available for other editors. username_2: I'll take a look. I've been ill lately but i'll look after the new year.
bitpopulous/alpha
251922807
Title: My Wallet - PPT vs Pokens Value / Visuals (Chart) Question: username_0: - Add a section that differentiates between your PPT value and your Token Value. - How many Pokens you have converted from the PPT and their value. - Is the value of Pokens valued at the time they are converted to Pokens from PPT, or does that value fluctuation as the value of PPT moves. - If its fixed at the time of transforming PPT to Pokens, you should have 2 boxes, showing PPT value and Pokens value. (I as a user might not convert all PPT over to Pokens all at once)
MicrosoftDocs/azure-docs
1004809392
Title: Helm chart command removed from helm 3.7 cli, update page to reflect on helm cli 3.7 release Question: username_0: To push a helm chart to ACR: ``` helm create test helm package test Successfully packaged chart and saved it to: /home/bap/test/test-0.1.0.tgz helm push /home/bap/test/test-0.1.0.tgz oci://contenthub.azurecr.io/helm/test ``` --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 59c4be11-ec02-7160-e65b-2218bfd70f68 * Version Independent ID: 19e08339-390f-ccfa-885c-2643722bd186 * Content: [Store Helm charts - Azure Container Registry](https://docs.microsoft.com/en-us/azure/container-registry/container-registry-helm-repos) * Content Source: [articles/container-registry/container-registry-helm-repos.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/container-registry/container-registry-helm-repos.md) * Service: **container-registry** * GitHub Login: @username_2 * Microsoft Alias: **danlep** Answers: username_1: Thanks for the feedback! I have assigned the issue to the content author to investigate further and update the document as appropriate. username_2: @username_0 - Thanks very much for raising this issue and noting the differences in Helm 3.7. We'll update the article soon. username_0: 👍 Question, can I do doc PR? username_2: @username_0 - Yes, you can make a PR and we'd be happy to review it! Instructions begin here: https://github.com/MicrosoftDocs/azure-docs/blob/master/CONTRIBUTING.md Thanks a lot username_2: The Helm article has been updated for Helm 3.7.1 commands, so I am closing this issue. Thanks again for contacting us. #please-close Status: Issue closed
tngan/samlify
514881703
Title: Error: ERR_FAILED_STATUS with top tier code: urn:oasis:names:tc:SAML:2.0:status:Requester, second tier code: urn:oasis:names:tc:SAML:2.0:status:RequestDenied Question: username_0: parseLoginResponse method give this error Error: ERR_FAILED_STATUS with top tier code: urn:oasis:names:tc:SAML:2.0:status:Requester, second tier code: urn:oasis:names:tc:SAML:2.0:status:RequestDenied Answers: username_0: how to get rid of this error username_1: @username_0 Looks like the NameIDFormat is not set properly. Checkout this https://samlify.js.org/#/sp-configuration?id=required-parameters for detail. username_0: can you give exaple for sp configuration without using metadata file username_1: Hi @username_0, https://github.com/username_1/samlify/blob/32948dfdf8083bef4106e1f8980ea29866461b22/test/issues.ts#L41-L76 username_1: Hi @username_0, any update on this issue ? username_1: @username_0 I will close this ticket since there is no update for couple weeks. Feel free to reopen the ticket if you have further question. Status: Issue closed
w8r/polygon-centroid
269620118
Title: Add support for self intersected polygons Question: username_0: Currently self intersected polygons are not supported. It produces wrong centroid in that case. It would be nice to calculate it also for self intersected polygons. http://kineme.net/forum/Discussion/General/JavascriptCentroid Probably it will require one additional step in algorithm.
htacg/tidy-html5
941529469
Title: warning: visibility attribute not supported in this configuration Question: username_0: I'm getting lots of ``` warning: visibility attribute not supported in this configuration; ignored [-Wattributes]``` messages when building tidy with gcc on cygwin (windows 10) All those warnings go away if I change src/forward.h to ``` #if defined(_WIN32) || defined(__CYGWIN__) <============ this line #define TY_PRIVATE #else #define TY_PRIVATE __attribute__((__visibility__("hidden"))) #endif ``` Cygwin isn't windows, so gcc doesn't have _WIN32 defined: ``` $ echo | gcc -dM -E -xc - | grep -i win32 $ echo | gcc -dM -E -xc - | grep -i cygwin #define __CYGWIN__ 1 ``` But apparently cygwin inherits windows limitations, so related to issue #743? Answers: username_1: Yup, it's _exactly_ related to #743. I'll push a fix within the next 24 hours, or sooner if you want to do a quick PR, because the review is pretty much already complete; I just don't have a Cygwin environment. I'd love to start testing Cygwin, but I'm not really interested in learning about how to do it _in depth_ in Windows, because although you say it's not Windows, it's still Windows ;-) Can you give me a quick and dirty primer to compiling this on Windows, assuming no existing development environment? Then I'll try to replicate it in Github Actions like I did for the mingw-w64 builds, and these types of screwups won't happen again! Thanks. username_1: So this was a perfect use of Windows 10 sandbox; getting this to compile under Cygwin was trivially easy with the provided fix, which I'll happily push, but by removing the attribute in Cygwin, are we exposing Cygwin uses to the same concerns as #743? Will soname become meaningless on Cygwin once again? Status: Issue closed
react-navigation/react-navigation
667139508
Title: Bottom tabs have a "weird" -4 pixels offset Question: username_0: The bottom tabs fall behind the native navigation of the phone. See screenshot (colors are to make it more obvious) I think this is introduced here: https://github.com/react-navigation/react-navigation/commit/849e04ab6a541fffb490ffdfa9819608b88494f4 `const paddingBottom = Math.max(insets.bottom - 4, 0);` 3 button navigation: ![image](https://user-images.githubusercontent.com/810327/88681790-3cdcfc80-d0f2-11ea-8a52-f6fed6c4702e.png) Gesture navigation: ![image](https://user-images.githubusercontent.com/810327/88681836-48302800-d0f2-11ea-82da-ac16ef5365cb.png) **Expected Behavior** Bottom tabs not to fall behind the native navigation. **How to reproduce** Use the bottom tabs, use android and have a translucent bottom navigation set. **Your Environment** | software | version | | ----------------------------- | ------- | | iOS or Android | Android (probably both, more visible on Android) | @react-navigation/native | 5.7.1 | @react-navigation/bottom-tabs | 5.7.2 | react-native-screens | 2.9.0 | react-native | 0.63.2 Answers: username_0: Verified. This is also an issue with those versions. Status: Issue closed
TorquilH/Repo3
635747117
Title: Strong Password test - Pass Question: username_0: Create user account and choose password "<PASSWORD>!!" Works great Answers: username_0: https://bitmoderntestdemo.testquality.com defect [D23](https://bitmoderntestdemo.testquality.com/project/9896/result/1929850/defect/6798). Do not remove or edit.
MichaelCurrin/twitterverse
247203318
Title: Make db summary run with missing tables Question: username_0: ``` python -m lib.database -s Getting table summary... Traceback (most recent call last): File "/usr/lib/python2.7/runpy.py", line 162, in _run_module_as_main "__main__", fname, loader, pkg_name) File "/usr/lib/python2.7/runpy.py", line 72, in _run_code exec code in run_globals File "/home/username_0/apps/twitterverse/app/lib/database.py", line 232, in <module> main(sys.argv[1:]) File "/home/username_0/apps/twitterverse/app/lib/database.py", line 227, in main tableCounts.showTableCounts() File "lib/query/schema/tableCounts.py", line 24, in showTableCounts count = tableClass.select().count() File "/home/username_0/.local/virtualenvs/twitterverse/lib/python2.7/site-packages/sqlobject/sresults.py", line 229, in count count = self.accumulate('COUNT(*)') File "/home/username_0/.local/virtualenvs/twitterverse/lib/python2.7/site-packages/sqlobject/sresults.py", line 211, in accumulate return conn.accumulateSelect(self, *exprs) File "/home/username_0/.local/virtualenvs/twitterverse/lib/python2.7/site-packages/sqlobject/dbconnection.py", line 507, in accumulateSe lect val = self.queryOne(q) File "/home/username_0/.local/virtualenvs/twitterverse/lib/python2.7/site-packages/sqlobject/dbconnection.py", line 482, in queryOne return self._runWithConnection(self._queryOne, s) File "/home/username_0/.local/virtualenvs/twitterverse/lib/python2.7/site-packages/sqlobject/dbconnection.py", line 348, in _runWithConn ection val = meth(conn, *args) File "/home/username_0/.local/virtualenvs/twitterverse/lib/python2.7/site-packages/sqlobject/dbconnection.py", line 475, in _queryOne self._executeRetry(conn, c, s) File "/home/username_0/.local/virtualenvs/twitterverse/lib/python2.7/site-packages/sqlobject/sqlite/sqliteconnection.py", line 222, in _ executeRetry raise dberrors.OperationalError(ErrorMessage(e)) sqlobject.dberrors.OperationalError: no such table: place ``` Answers: username_0: Fixed by printing missing table message in place of row count. ``` commit e9c91bf5f0b805221770d1be36968e40322ca708 Author: <NAME> <<EMAIL>> Date: Fri Sep 1 23:11:42 2017 +0200 Improve database docs and tableCounts query. ``` Status: Issue closed
inveniosoftware/invenio-app-rdm
1168485714
Title: Missing space between orcid icon and name in deposit form Question: username_0: Package version (if known): 8.0 ## Describe the bug Space missing between orcid icon and name: ![orcid_spacing_deposit_page](https://user-images.githubusercontent.com/1932651/158196576-0e49deed-f7e1-49f6-88b4-60e23196d208.png) Answers: username_1: According to [Orcid brand guidelines](https://info.orcid.org/brand-guidelines/#Inline_ORCID_iD), the icon must follow the name. WDYT of this. Other ideas? ![Screen Shot 2022-03-14 at 5 26 34 PM](https://user-images.githubusercontent.com/86330/158264699-247fb34a-721e-4931-999c-7320cd9928e1.png)
bergware/dynamix
157787054
Title: HDD activity not detected for disks behind Adaptec RAID controller - solution proposal Question: username_0: I have an Adaptec 51254 RAID controller allowing for 12 SATA ports. Problem: Your S3 sleep plugin won't allow bringing the system to S3 if HDD activity should be evaluated. HDD's cannot be spinned down via unraid UI or hdparm, the controller brings them back up immediately. Solution is to not use each disk as JBOD but instead use them as individual arrays consisting of just one disk. For arrays spin down times can be defined in the controller menu and that works fine. However - "hdparm -C" will never get the proper result, the disks are always "active" though they are off. I had created some code to handle this before but found a better way around that on the web. Instead of relying on the state of the drive, check if there has been any IO happening. The following code works perfectly - it finds there is no activity on the disks regardless of their activity state. I changed my code locally (repacked into txz and adjusted the MD5) and it works fine - should work well for all disks. The code I'm using now: HDD_activity() { result= if [[ $checkHDD == yes ]]; then # Create a file on the ramdisk and cycle it to test for disk activity ( if [ ! -f /dev/shm/1 ] ; then touch /dev/shm/1 /dev/shm/2; fi ; mv /dev/shm/1 /dev/shm/2; cat /proc/diskstats > /dev/shm/1 ) >/dev/null 2>&1 for dev in ${array[@]}; do # Check if drive has been non idle since last run if [ "$(diff /dev/shm/1 /dev/shm/2 | grep $dev )" = "" ]; then result= else result=1 break; fi done fi if [[ -n $result ]]; then log "Disk activity on going: $dev" echo $result fi } Answers: username_1: Thanks for your proposal. When time permits I make an update. Status: Issue closed username_1: New version is out, which has your proposed solution.
cli/cli
708416111
Title: <EMAIL> Question: username_0: ### Impact _What kind of vulnerability is it? Who is impacted?_ ### Patches _Has the problem been patched? What versions should users upgrade to?_ ### Workarounds _Is there a way for users to fix or remediate the vulnerability without upgrading?_ ### References _Are there any links users can visit to find out more?_ ### For more information If you have any questions or comments about this advisory: * Open an issue in [example link to repo](http://example.com) * Email us at [example email address](mailto:<EMAIL>)<issue_closed> Status: Issue closed
pyinvoke/invoke
505377646
Title: DeprecationWarning on usage of imp in loader.py Question: username_0: When running my test suite which uses invoke (via fabric) I see: ``` ======================= warnings summary ======================== venv/lib64/python3.7/site-packages/invoke/loader.py:3 venv/lib64/python3.7/site-packages/invoke/loader.py:3: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses import imp -- Docs: https://docs.pytest.org/en/latest/warnings.html ========== 6 passed, 1 warnings in 1651.00s (0:27:30) =========== ``` The used parts of the `imp` module have been deprecated since Python 3.3 and the replacement `importlib` has been available since Python 3.1. Answers: username_1: For what it's worth, there's a workaround that will suppress that one particular warning. I use it on pytest but I believe it's going to be mostly the same with other test frameworks: `-W ignore::DeprecationWarning:invoke.loader` username_2: I have also spotted this - the question is really: which versions of python is invoke committed to supporting? The docs say 2.7 and >=3.4. However, we have now passed 2020-01-01, so the sun has officially gone down on python 2. If I had a vote (which I don't obviously), it would be to drop support for 2.7 and switch this code to `importlib` Instead I'm adding a warning filter (https://docs.pytest.org/en/latest/warnings.html#pytest-mark-filterwarnings) and have logged an issue on our project. username_3: This issue was brought up again in #829. I would be happy to submit a PR if the team is looking for help. username_4: Hi, I'm starting to help out on issue triage. Invoke (and the Fabric and Paramiko family) are conservative with the supported versions of python. But yes, very soon Python2 will be dropped clearing out issues such as this. https://bitprophet.org/projects/#roadmap
rflament/loggedfs
435289913
Title: Implementation of open is unsafe Question: username_0: Closely related to #16: In [your implementation of `open`](https://github.com/rflament/loggedfs/blob/loggedfs-0.9/src/loggedfs.cpp#L566), you are using the `open` system call. You should be using the `openat` system call instead.
storybookjs/storybook
1172330885
Title: Interactions relying on focus lose focus on manual step-through Question: username_0: Hello 👋 **Describe the bug** Stories which rely on an element being focused do not work when manually stepped through, because the *Go forward* button in the `interactionsDebugger` steals focus. This causes subsequent keyboard events to not work. e.g. in: ```tsx await userEvent.tab(); await userEvent.keyboard("{Enter}"); ``` focus is lost before the keyboard event is called. **To Reproduce** https://github.com/username_0/sb-focus-bug-repro https://gilstroem.github.io/sb-focus-bug-repro/?path=/story/example-button--primary **System** Environment Info: System: OS: macOS 12.0 CPU: (8) arm64 Apple M1 Pro Binaries: Node: 17.0.1 - /opt/homebrew/bin/node Yarn: 3.2.0 - /opt/homebrew/bin/yarn npm: 8.1.0 - /opt/homebrew/bin/npm Browsers: Chrome: 99.0.4844.74 Safari: 15.0 npmPackages: @storybook/addon-actions: ^6.5.0-alpha.49 => 6.5.0-alpha.49 @storybook/addon-docs: ^6.5.0-alpha.49 => 6.5.0-alpha.49 @storybook/addon-essentials: ^6.5.0-alpha.49 => 6.5.0-alpha.49 @storybook/addon-interactions: ^6.4.19 => 6.4.19 @storybook/addon-links: ^6.5.0-alpha.49 => 6.5.0-alpha.49 @storybook/builder-webpack5: ^6.5.0-alpha.49 => 6.5.0-alpha.49 @storybook/jest: ^0.0.9 => 0.0.9 @storybook/manager-webpack5: ^6.5.0-alpha.49 => 6.5.0-alpha.49 @storybook/node-logger: ^6.5.0-alpha.49 => 6.5.0-alpha.49 @storybook/preset-create-react-app: ^4.1.0 => 4.1.0 @storybook/react: ^6.5.0-alpha.49 => 6.5.0-alpha.49 @storybook/test-runner: ^0.0.4 => 0.0.4 @storybook/testing-library: ^0.0.9 => 0.0.9 **Additional context** Might be related to: https://github.com/storybookjs/storybook/issues/16139 Please let me know if you need any more information
eramdam/BetterTweetDeck
174933228
Title: Port to Firefox Question: username_0: Now that Firefox 48 with support of WebExtensions dropped, it's time to work on the Firefox port. I've gotta do some research about how it's going to work but if someone has some experience with that, I'm more than happy to hear them out :) Answers: username_0: Here are the differences I noticed after quick testing, will complete later: - [ ] `chrome.storage.local` is not implemented in the same way so the `browserHelper.settings` methods need to be updated. I need to change the way I call `chrome.storage.local/sync.get` since I've been using for the same way for 3 years and it turns out it's completely wrong.. - [ ] As for Edge ( #64 ) I need to adapt the tooling so I can create browser specific builds username_0: Closing this in favor of smaller issues regrouped in a milestone https://github.com/username_0/BetterTweetDeck/milestone/4 Status: Issue closed
CrossLead/temporal-mongo
111668091
Title: aggregate() API does not match promised-mongo/mongo Question: username_0: Per PR #3, @username_1 caught fact that our `aggregate()` API is non-standard. @mcgrit can you take an initial look? Answers: username_1: Potential fix in #5 https://github.com/CrossLead/temporal-mongo/commit/1f59ec117d60ceb3afad35f984caba62d20a7d02 Status: Issue closed username_1: fixed in https://github.com/CrossLead/temporal-mongo/commit/9a59ed0ae132a10ad5d1269d550812e02d06da80
usnistgov/REFPROP-wrappers
496489363
Title: How to code refprop 10 with .net framework C# and VB.net languages Question: username_0: How to code refprop 10 with .net framework C# and VB.net languages? Any examples with calculation of all thermodynamic properties in VB.net and C# would be helpful. Status: Issue closed Answers: username_1: For VB.net, see: https://trc.nist.gov/refprop/LINKING/Linking.htm#NETApplications
chrishelgert/zazu-fkill
443092112
Title: Plugin failed to load Question: username_0: ``` ERROR: [19:04:27] [chrishelgert/zazu-fkill :Process] Plugin failed to load { "block": "Process", "error": "/Users/username_0/.zazu/plugins/chrishelgert/zazu-fkill/node_modules/ps-list/index.js:9\nconst windows = async () => {\n ^\nSyntaxError: Unexpected token (\n at Object.exports.runInThisContext (vm.js:76:16)\n at Module._compile (module.js:528:28)\n at Module._extensions..js (module.js:565:10)\n at Object.require.extensions.(anonymous function) [as .js] (/Applications/Zazu.app/Contents/Resources/app.asar/node_modules/babel-register/lib/node.js:152:7)\n at Module.load (module.js:473:32)\n at tryModuleLoad (module.js:432:12)\n at Function.Module._load (module.js:424:3)\n at Module.require (module.js:483:17)\n at require (internal/module.js:20:19)\n at Object.<anonymous> (/Users/username_0/.zazu/plugins/chrishelgert/zazu-fkill/src/processes.js:1:16)", "level": "error", "message": "Plugin failed to load", "plugin": "chrishelgert/zazu-fkill", "time": "2019-05-12T11:04:27.160Z" } ```
oaeproject/Hilary
220152757
Title: Re-enable users when the tenancy gets re-enabled Question: username_0: When the disabled tenant is re-started then the users are still marked as deleted. And if we want to re-create the same users, we cannot do so, because the error says that the username has already been taken. This is a follow-up fix to https://github.com/oaeproject/Hilary/issues/1304<issue_closed> Status: Issue closed
gogoout/nativescript-rater
570095837
Title: Feature on iOS - Add app country code parameter due to iTunes lookup fail. Question: username_0: Hello, There is an issue on iOS when the app is available in only one country. The iTunes lookup return an empty response. Related issue : https://github.com/takecian/SwiftRater/issues/30 SwiftRater documentation : `SwiftRater.countryCode = "fr"` Could you please add this parameter in the next version of this NS plugin ? Thank you. Answers: username_1: Hi, sorry for the late response. You can access to `SwiftRater` by `appRater.ios`. So I think `appRater.ios.countryCode = "fr"` will just work! Status: Issue closed
swagger-api/swagger-codegen
321475401
Title: [PHP] Handle non-required property as nullable Question: username_0: ##### Description When a property isn't required, it is not shown in Doc Block, that the getter can actually return "null". Also it is not shown in the setter, that "null" is a valid value ##### Swagger-codegen version 2.3 ##### Swagger declaration file content or url ```yaml /path post: parameters: - in: body required: true name: objectName schema: type: object required: - requiredPropertyName properties: requiredPropertyName: type: string notRequiredPropertyName: type: string ``` ##### Suggest a fix/enhancement Change returnType in Doc Block for getters to correctly show, that the returned value can be null (if the property is not required). Also shown in Doc Block for setters, that value can be null (if property is not required) Answers: username_1: How does `required` interact with `nullable`? ``` /path post: parameters: - in: body required: true name: objectName schema: type: object required: - requiredPropertyName - requiredNullablePropertyName properties: requiredPropertyName: type: string notRequiredPropertyName: type: string requiredNullablePropertyName: type: string nullable: true notRequiredNullablePropertyName: type: string nullable: true ``` Isn't `|null` supposed to be defined by the nullable property instead of required? username_0: @username_1 That sounds correct :) The issue though is, that we only have null to represent optional. If the field is optional, and it's not filled, then it's value will be "null".
Intel-Media-SDK/MediaSDK
531657285
Title: commit 444ecb4 caused regression in vp9 decoder Question: username_0: The output Shift is not set since commit commit 444ecb4, which results in regression when running a gst-msdk pipeline: https://github.com/Intel-Media-SDK/MediaSDK/commit/444ecb480a3918cc6faeb7662ca7492fd564f301#diff-d8bc9fb5bc14276fa6b733fff663ba7aL151-L153 gst-launch-1.0 filesrc location=input.10bit.ivf ! ivfparse ! msdkvp9dec ! fakesink Redistribute latency... 0:00:00.021322781 15413 0x557733521050 ERROR msdkdec gstmsdkdec.c:396:gst_msdkdec_init_decoder:<msdkvp9dec0> Init failed (invalid video parameters) 0:00:00.021337017 15413 0x557733521050 ERROR msdkdec gstmsdkdec.c:868:gst_msdkdec_negotiate:<msdkvp9dec0> Failed to re-negotiate Answers: username_0: BTW for HEVC decoder, the output Shift is still equal to the input Shift. I think both VP9 and HEVC encoder should keep consistency for 10bit decoding username_1: @username_0 : do you use MFX_VERSION>=MFX_VERSION_NEXT path to build mediasdk or you see regression when this condition is not true? username_0: @username_1 I see the regression no matter this condition is true or not. Status: Issue closed
lesuixin/Hearthstone-Hearthbuddy
1044676566
Title: 2021.11 update Question: username_0: 2021.11 游戏更新 迷你系列“死亡矿井” 2021.11 game update Mini-Set "Deadmine" Answers: username_0: 修复方法参考 ``` 首先你要找到一个没有壳,然后能用dnspy打开并能定位到class276的hb 然后参考那篇文章,把那些intptr0 + XXX的偏移地址对照exports-5-5.txt里的地址,找到函数名,然后再到exports-11-3.txt里找到新的偏移地址。 最后把最新的偏移地址更新下就好了 ``` username_0: 我获得一份压缩包,但需等等 ``` 有没有在使用阿里云盘的朋友?帮忙去bald9-hb,2.2 https://www.aliyundrive.com/s/MEZePQu4a6s/folder/6183af38dcc1382930644f45bf758a7020a03f2a 下载.exe,压缩包发上来 ``` username_1: [](url [hearthbuddy-bald9发行版(2.0).zip](https://github.com/username_0/Hearthstone-Hearthbuddy/files/7476250/hearthbuddy-bald9.2.0.zip) ) username_1: [](url [Hearthbuddy.zip](https://github.com/username_0/Hearthstone-Hearthbuddy/files/7476290/Hearthbuddy.zip) ) username_2: #22 然后参考那篇文章,把那些intptr0 + XXX的偏移地址对照exports-5-5.txt里的地址,找到函数名,然后再到exports-11-3.txt里找到新的偏移地址。 最后把最新的偏移地址更新下就好了 能细说一下是哪个文章么,新的memo2函数和地址都能看到,但是不知道是哪个,exports-X-X.txt这个是怎么得出来的,望指教! username_2: __ ![1636085773(1)](https://user-images.githubusercontent.com/28940843/140457618-61cde667-6ce7-4377-84dd-cb2ddb3357fa.jpg) 是这样么? username_2: ....最新的这些问题依旧修复了,我这个版本老了 ̄□ ̄|| username_0: [hearthbuddy-bald9发行版(2.2).zip](https://github.com/username_0/Hearthstone-Hearthbuddy/files/7481153/hearthbuddy-bald9.2.2.zip) 因为bald9觉得是小更新,只更新了hb.exe,组合成2.2版不带策略 程序hb: hearthbuddy-bald9发行版(2.0) + 2.2版 策略Routines: 磁石战士lwh 2021-9-13策略Routines + 自动识别新皮肤 auto distinguish new skins + 还缺迷你系列35 Mini-Set新卡说明CardDefs.xml +35 card sim ,需要自己替换组合 username_0: Oo孤独念旧人oO的 11.3更新修复方法 https://tieba.baidu.com/p/7602844533 ``` 首先使用CE获取新旧mono-2.0-bdwgc.dll的地址,这里我打包分享吧 aHR0cHM6Ly90cnlzdHVkeS5sYW56b3V3LmNvbS9pdGRrOHc1c3MxZQ== 然后把无壳兄弟用dnspy打开,找到ns27.Class276.Class276如图 ![找到ns27 Class276 Class276如图](https://user-images.githubusercontent.com/70421104/140632490-4fe7c981-a555-4c7d-b835-e8fc5d2f752f.png) 然后右键编辑IL指令,将下图所有0x后面的在旧地址txt中查找,找到对应名称,再在新地址txt中找到对应的,替换完点确定再保存就行了 ![将下图所有0x后面的在旧地址txt中查找,找到对应名称](https://user-images.githubusercontent.com/70421104/140632494-ba9d91ef-20d2-483e-ab7d-d070da0f4f90.png) ``` [新旧地址合集.zip](https://github.com/username_0/Hearthstone-Hearthbuddy/files/7491736/default.zip) username_3: Sim_DED_001a应当是暗礁德鲁伊的变形,但是文件夹的注释是推土壮汉,不正确,能不能把爬虫的程序放出来呢
scikit-learn/scikit-learn
668452870
Title: LabelSpreading predict_proba return nan? Question: username_0: <!-- **If your issue is a usage question, please submit it in one of these other channels instead:** - **StackOverflow with the scikit-learn tag: https://stackoverflow.com/questions/tagged/scikit-learn** - **Mailing List: https://mail.python.org/mailman/listinfo/scikit-learn** - **Gitter: https://gitter.im/scikit-learn/scikit-learn** - **For more information, see User Questions: http://scikit-learn.org/stable/support.html#user-question** The issue tracker is used only to report issues and feature requests. For questions, please use either of the above platforms. Most question issues are closed without an answer on this issue tracker. Thanks for your understanding. --> Answers: username_0: why use this api got some nans?but use predict can get the predict labels username_1: please provide a fully reproducible code snippet username_0: clf=LabelSpreading(max_iter=100,kernel='knn',n_neighbors=8) clf.fit(data, data_df['label']) labels = clf.predict_proba(data) there are some nans in labels. But we use the API predict, it will get the predict labels. username_2: We would need the data to be able to reproduce the issue. username_0: we split the features to 37 column. then put them in to the above model. columns = ['feature0','feature1','feature2','feature3','feature4','feature5'...,'feature36'] data = data_df[columns] [train.zip](https://github.com/scikit-learn/scikit-learn/files/5004696/train.zip) username_0: use this get the nans`data_df = pd.read_csv(file_name, header = 0, index_col = 0) for i in range(37): feature_name = 'feature' + str(i) data_df[feature_name] = data_df.apply(lambda x: x['features'].split(',')[i], axis=1) columns = ['feature0','feature1','feature2','feature3','feature4','feature5'...,'feature36'] data = data_df[columns] clf=LabelSpreading(max_iter=100,kernel='knn',n_neighbors=8) clf.fit(data, data_df['label']) labels = clf.predict_proba(data)` but when we change the api `labels = clf.predicta(data)`. we can get the predict labels. username_0: This problem also exists.Is it my method of operation wrong? username_3: Hi @username_0 , your code is not reproducible: I did my best in reconstructing it, ```py from sklearn.semi_supervised import LabelSpreading import pandas as pd data_df = pd.read_csv('train.csv', header = 0, index_col = 0) columns = [] for i in range(37): feature_name = 'feature' + str(i) data_df[feature_name] = data_df.apply(lambda x: x['Features'].split(',')[i], axis=1) columns.append(feature_name) data = data_df[columns] clf = LabelSpreading(max_iter=100,kernel='knn',n_neighbors=8) clf.fit(data, data_df['label']) labels = clf.predict_proba(data) ``` but 'label' is not a key in your dataframe. I'm closing this issue, as this is closer to a usage question than a bug report. If you manage to have a reproducible snippet of code feel free to join the community on [Stack Overflow](https://stackoverflow.com/questions/tagged/scikit-learn) or [the scikit-learn mailing list](https://mail.python.org/mailman/listinfo/scikit-learn). Thanks. Status: Issue closed username_0: I am sorry,the original size of the dataset is huge, so I sample it. But I forgot upload the label. from sklearn.semi_supervised import LabelSpreading import pandas as pd data_df = pd.read_csv('train.csv', header = 0, index_col = 0) columns = [] for i in range(37): feature_name = 'feature' + str(i) data_df[feature_name] = data_df.apply(lambda x: x['Features'].split(',')[i], axis=1) columns.append(feature_name) data_df['label'] = data_df.apply(lambda x: 0 if x['morel']==True else 1 if [x['less']==True](url [train.zip](https://github.com/scikit-learn/scikit-learn/files/5222137/train.zip) ) else -1, axis=1) data = data_df[columns] clf = LabelSpreading(max_iter=100,kernel='knn',n_neighbors=8) clf.fit(data, data_df['label']) labels = clf.predict_proba(data)
sp614x/optifine
823488309
Title: [Minecraft Crash] Eu uso o minecraft Shiginima Launcher e estou a fazer um modpack, mas quando adiciona da este crash Question: username_0: Os detalhes do Crash: Java HotSpot(TM) 64-Bit Server VM warning: Using incremental CMS is deprecated and will likely be removed in a future release [23:20:49] [main/INFO] [LaunchWrapper]: Loading tweak class name net.minecraftforge.fml.common.launcher.FMLTweaker [23:20:49] [main/INFO] [LaunchWrapper]: Using primary tweak class name net.minecraftforge.fml.common.launcher.FMLTweaker [23:20:49] [main/INFO] [LaunchWrapper]: Calling tweak class net.minecraftforge.fml.common.launcher.FMLTweaker [23:20:49] [main/INFO] [FML]: Forge Mod Loader version 14.23.5.2854 for Minecraft 1.12.2 loading [23:20:49] [main/INFO] [FML]: Java is Java HotSpot(TM) 64-Bit Server VM, version 1.8.0_261, running on Windows 10:amd64:10.0, installed at C:\Program Files\Java\jre1.8.0_261 [23:20:49] [main/INFO] [FML]: Searching C:\Users\HP\AppData\Roaming\.minecraft\mods for mods [23:20:49] [main/INFO] [FML]: Loading tweaker optifine.OptiFineForgeTweaker from .OptiFine_1.12.2_HD_U_F5.jar [23:20:49] [main/WARN] [FML]: The coremod Backpacked (com.mrcrayfish.backpacked.asm.BackpackedPlugin) is not signed! [23:20:49] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in cxlibrary-1.12.1-1.6.1.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [23:20:49] [main/WARN] [FML]: The coremod cubex2.cxlibrary.CoreModLoader does not have a MCVersion annotation, it may cause issues with this version of Minecraft [23:20:49] [main/WARN] [FML]: The coremod CXLibraryCore (cubex2.cxlibrary.CoreModLoader) is not signed! [23:20:49] [main/INFO] [FML]: Loading tweaker org.spongepowered.asm.launch.MixinTweaker from MalisisCore-1.12.2.jar [23:20:49] [main/WARN] [FML]: The coremod micdoodle8.mods.miccore.MicdoodlePlugin does not have a MCVersion annotation, it may cause issues with this version of Minecraft [23:20:49] [main/WARN] [FML]: The coremod MicdoodlePlugin (micdoodle8.mods.miccore.MicdoodlePlugin) is not signed! [23:20:49] [main/WARN] [FML]: The coremod ObfuscatePlugin (com.mrcrayfish.obfuscate.asm.ObfuscatePlugin) is not signed! [23:20:49] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in OpenModsLib-1.12.2-0.12.2.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [23:20:49] [main/WARN] [FML]: The coremod openmods.core.OpenModsCorePlugin does not have a MCVersion annotation, it may cause issues with this version of Minecraft [23:20:49] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in RandomThings-MC1.12.2-4.2.7.4.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [23:20:49] [main/WARN] [FML]: The coremod lumien.randomthings.asm.LoadingPlugin does not have a MCVersion annotation, it may cause issues with this version of Minecraft [23:20:49] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in redstonepaste-mc1.12-1.7.5.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [23:20:49] [main/WARN] [FML]: The coremod net.fybertech.nwr.NWRTweak does not have a MCVersion annotation, it may cause issues with this version of Minecraft [23:20:49] [main/WARN] [FML]: The coremod NWRTweak (net.fybertech.nwr.NWRTweak) is not signed! [23:20:49] [main/INFO] [FML]: Loading tweaker org.spongepowered.asm.launch.MixinTweaker from wolfarmor-1.12.2-3.7.5-universal-signed.jar [23:20:49] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in Xaeros_Minimap_21.4.1_Forge_1.12.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [23:20:49] [main/WARN] [FML]: The coremod XaeroMinimapPlugin (xaero.common.core.XaeroMinimapPlugin) is not signed! [23:20:49] [main/INFO] [FML]: Loading tweaker org.spongepowered.asm.launch.MixinTweaker from phosphor-1.12.2-0.2.6+build50.jar [23:20:49] [main/INFO] [LaunchWrapper]: Loading tweak class name net.minecraftforge.fml.common.launcher.FMLInjectionAndSortingTweaker [23:20:49] [main/INFO] [LaunchWrapper]: Loading tweak class name optifine.OptiFineForgeTweaker [23:20:49] [main/INFO] [LaunchWrapper]: Loading tweak class name org.spongepowered.asm.launch.MixinTweaker Exception in thread "main" [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: java.util.ServiceConfigurationError: org.spongepowered.asm.service.IMixinServiceBootstrap: Provider org.spongepowered.asm.service.modlauncher.MixinServiceModLauncherBootstrap could not be instantiated [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.util.ServiceLoader.fail(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.util.ServiceLoader.access$100(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.util.ServiceLoader$LazyIterator.nextService(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.util.ServiceLoader$LazyIterator.access$700(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.util.ServiceLoader$LazyIterator$2.run(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.security.AccessController.doPrivileged(Native Method) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.util.ServiceLoader$LazyIterator.next(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.util.ServiceLoader$1.next(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at org.spongepowered.asm.service.MixinService.runBootServices(MixinService.java:77) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at org.spongepowered.asm.service.MixinService.<init>(MixinService.java:72) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at org.spongepowered.asm.service.MixinService.getInstance(MixinService.java:92) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at org.spongepowered.asm.service.MixinService.boot(MixinService.java:102) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at org.spongepowered.asm.launch.MixinBootstrap.<clinit>(MixinBootstrap.java:76) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at org.spongepowered.asm.launch.MixinTweaker.<init>(MixinTweaker.java:44) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.lang.reflect.Constructor.newInstance(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.lang.Class.newInstance(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at net.minecraft.launchwrapper.Launch.launch(Launch.java:98) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at net.minecraft.launchwrapper.Launch.main(Launch.java:28) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: Caused by: java.lang.SecurityException: class "org.spongepowered.asm.service.ServiceInitialisationException"'s signer information does not match signer information of other classes in the same package [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.lang.ClassLoader.checkCerts(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.lang.ClassLoader.preDefineClass(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.lang.ClassLoader.defineClass(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.security.SecureClassLoader.defineClass(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.net.URLClassLoader.defineClass(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.net.URLClassLoader.access$100(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.net.URLClassLoader$1.run(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.net.URLClassLoader$1.run(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.security.AccessController.doPrivileged(Native Method) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.net.URLClassLoader.findClass(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.lang.ClassLoader.loadClass(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at sun.misc.Launcher$AppClassLoader.loadClass(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.lang.ClassLoader.loadClass(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.lang.Class.getDeclaredConstructors0(Native Method) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.lang.Class.privateGetDeclaredConstructors(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.lang.Class.getConstructor0(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: at java.lang.Class.newInstance(Unknown Source) [23:20:49] [main/INFO] [STDERR]: [java.lang.ThreadGroup:uncaughtException:-1]: ... 19 more Status: Issue closed Answers: username_1: We do not support hacked clients username_0: Isso não faz sentido, deviam ajudar todos os jogadores, só por eu não poder comprar o cliente original não dão ajuda? Deviam mudar essas regras!
umijs/umi
395546321
Title: 访问http://localhost:8001/时报TypeError: reactRouter.Router.computeRootMatch is not a function Question: username_0: bigfish工程,访问根目录http://localhost:8001/时直接抛错了 ![image](https://user-images.githubusercontent.com/2724034/50636222-4216ab80-0f90-11e9-8197-481547b78f8b.png) 查了下网上的说法是,react-router-config 4.4应与react-router4.4来匹配使用 https://github.com/ReactTraining/react-router/issues/6492 看umi-build-dev里的确是版本不匹配的,https://github.com/umijs/umi/blob/master/packages/umi-build-dev/package.json#L46 Answers: username_1: 给可复现的仓库? username_0: http://gitlab.alipay-inc.com/wealthprod/fund-research 我也去问下bigfish的同学 username_1: 重现了,配置了 exportStatic,没有 layout,没有 `/` 路由时会报这个错。 Status: Issue closed username_2: ![image](https://user-images.githubusercontent.com/13595509/57679061-0934c480-765d-11e9-9d44-09c9d122237a.png) UC 下重现了
FriendsOfFlarum/recaptcha
743102701
Title: Never works with invisible recaptcha Question: username_0: Beta 14, reCpatcha 0.2.0 Invislble ReCaptcha Error console. ``` Uncaught TypeError: this.recaptcha.execute is not a function exports index.js:37 e extend.js:61 handleEvent render.js:888 I render.js:910 T render.js:735 l render.js:730 l render.js:122 l render.js:63 u render.js:50 l render.js:134 l render.js:63 u render.js:50 l render.js:134 l render.js:63 l render.js:162 l render.js:66 f render.js:423 h render.js:302 f render.js:477 f render.js:416 f render.js:488 f render.js:419 f render.js:488 f render.js:419 h render.js:302 exports render.js:964 a mount-redraw.js:14 show ModalManagerState.js:35 onclick HeaderSecondary.js:77 handleEvent render.js:888 I render.js:910 T render.js:735 l render.js:730 l render.js:122 l render.js:63 l render.js:162 l render.js:66 u render.js:50 l render.js:134 l render.js:63 u render.js:50 l render.js:134 l render.js:63 l render.js:162 l render.js:66 u render.js:50 h render.js:277 exports render.js:964 mount mount-redraw.js:45 mount ForumApplication.js:128 boot Application.js:176 <anonymous> (index):160 ``` Answers: username_0: Issue is present also with checkbox. Users are unable to register. username_1: Yep I have the same issue with both checkbox and invisible recapcha on Beta 14.1. If I disable the extension, everything works fine username_2: Same issue on my side username_3: Hopefully, these problems will be fixed with the new release (will be tagged soon). Feel free to re-open the issue persists! Status: Issue closed username_0: Thanks for fixing this
google/kctf
576507598
Title: RBAC support Question: username_0: https://kubernetes.io/docs/reference/access-authn-authz/rbac/ Status: Issue closed Answers: username_0: this is already kindof supported in beta (I can give you access to modify "challenge" resources without giving you access to create pods/deployments directly) the security isolation here isn't really strong, though, because the challenge can just define itself privileged, which I think is probably necessary :( username_1: This should change though once the support for clean proc made it to stable k8s
luvit/lit
145216773
Title: custom luvi Question: username_0: Hello, I have run in a bit of trouble running in Alpine Linux, and musl. Since lit will download the supposedly right version of the precompiled binary from github it will not run on my small system (an ARM server from scaleway and my Raspberry). I already have *luvi*, *lit* and *luvit* installed in `/usr/local`. Additionally I had to replace the downloaded luvi with my compiled version `~/.litdb.git/cache/24339028e53ad0607172d8c0f0105a5814cd7bc9/luvi`. A quick inspection of the code revealed tha there is currently no way to set a preferred `luvi` executable. Is this something you would like to support? If you are positive I could give it a try and send a pull request. Answers: username_1: ``` And it will build the zip for `luvit/luvit`, and append it to the file at `/path/to/prefix` and store it in `/usr/local/bin/luvit` The undocumented feature is the 3rd argument to `lit make` is the custom luvi path. username_1: I would like to think of some tweaks to luvi that makes this custom luvi easier, maybe a config option? username_0: Thanks for the prompt response! A config option would be helpful to force a custom version of luvi instead of grabbing a prebuild binary username_1: Agreed, I'll leave this issue open as a reminder to add such an option when I'm back at my desk next week.
eregnier/vue2-gmap-custom-marker
1081081449
Title: Custom marker popup going off screen Question: username_0: <div v-if="pin" class="property-thumbnail"> ..... </div> <div class="detail"> ..... </div> </div> </div> ``` </gmap-custom-marker> ``` So lets say a marker is close to the edge of the map and click on it, the box will go off the screen (off the map). And I have to drag the map in order to see the box. Is it possible to check somehow when we are on the edges and prevent the box from going beyond it? Thanks for the help (and for the plugin) Answers: username_1: <div v-if="pin" class="thumbnail"> ..... </div> <div class="detail"> ..... </div> </div> </div> </template> <script> export default { mounted() { console.log('the map item', this.$parent.$map) // Do some computation with the map object } } </script> ``` Then it should be possible to compute your panel location depending on map information (this is google an underlaying map object you have to study) and compare it to your marker position then choose to move left or right some content of the custom marker (maybe making it larger and use relative / absolute positions) I never had to do this but I would do it this way (maybe I am naive) the first time. Is it clear enough or do you need something else ? Eric username_0: Thanks for your reply @username_1 What I did in case anyone is interested (I am not sure if this is a good solution but it works): I used a "ref" as username_1 mentioned above. And checked the position of the parent element to the child element Parent element = whole map (mapPos) Child element = marker popup box (popupPos) ``` toggleProperty (id) { // clearing the classes before re-adjusting this.$refs['marker_el_' + id][0].classList.remove('adjustTop') this.$refs['marker_el_' + id][0].classList.remove('adjustLeft') this.$refs['marker_el_' + id][0].classList.remove('adjustRight') // get the positions let mapPos = this.$refs['map'].getBoundingClientRect() let popupPos = this.$refs['marker_el_' + listingId][0].getBoundingClientRect() let position = {} position.top = popupPos.top - mapPos.top position.right = popupPos.right - mapPos.right position.left = popupPos.left - mapPos.left // set the classes (the numbers used below are what I wanted based on the position of my map and popups if (position.top <= 350) { this.$refs['marker_el_' + id][0].classList.add('adjustTop') } if (relativePos.left <= 150) { this.$refs['marker_el_' + id][0].classList.add('adjustLeft') } if (relativePos.right >= -150) { this.$refs['marker_el_' + id][0].classList.add('adjustRight') } } ``` And in css I set position absolute to the marker popup ``` .adjustTop { top: 0px !important; position: absolute; } .adjustLeft { left: 0px !important; position: absolute; } .adjustRight { right: 0px !important; position: absolute; } ``` Status: Issue closed
ember-learn/ember-website
512215177
Title: [redesign] Mascots Commission Pages - Update Markup Question: username_0: ## To Do: - [ ] Update the markup for the 'mascots/commission-sent' page - [ ] Update the markup for the 'mascots/commission' page - [ ] Update the markup for the 'mascots/faq' page - [ ] Update the markup for the 'mascots/payment-sent' page - [ ] Update the markup for the 'mascots/payment' page --- ### Notes (important) This issue is part of the [Ember Website Redesign RFC](https://github.com/emberjs/rfcs/blob/master/text/0425-website-redesign.md) and is one of the **last remaining issues** before we are ready to release. As we are moving very quickly on this project you should comment on this issue that you are taking it on so that we do not have multiple people working on the same issue. **Note:** please don't take on an issue if you do not think you can complete it within 24 hours 🙏 This issue is **strictly to only update the markup** of a page on the website and should be possible to do using only components and class helpers in the new [ember-styleguide](https://deploy-preview-145--ember-styleguide.netlify.com/). If you find that you need to create page-specific styles, complete the page as best you can and open a PR **without** adding any styles. Make a note in your PR what was missing from the styleguide that prevented you from completing the page. When creating your PR you should target the `redesign-rfc` branch. This way when your PR is merged it will not update `master` it will instead update the long-lived branch that is tracking the website redesign work. Answers: username_0: Fixed in #506 Status: Issue closed
BSData/horus-heresy
218741354
Title: Mechanicum: Missing HQ- Secutarii Axiarch Question: username_0: Hey guys, I have been waiting for a bit since book 7 came out to see if you were going to add the Axiarch HQ choice. Noticed that the rules for Secutarii Peltasts and Hoplites are in but still no Axiarch. Is there a reason it isn't in yet that I just am not aware of? Or maybe I have missed it somehow? Thanks, Teun<issue_closed> Status: Issue closed
revelrylabs/bn-mobile-react
424221522
Title: Deployment Screencast Question: username_0: - [ ] Do a deployment with a screencast Answers: username_1: The deployment process is less about knowing what to do (Bump version numbers, run deploy script) than it is about making sure you have the correct permissions for everything. Running the scripts the right way wont work if people dont have their keys and everything set up. username_0: Nothing stopping us from recording a screencast with cleared keys or even just better documenting a deploy process with screenshots. We at least need to update that ReadMe with Sentry stuff and better directions on what file paths to use where. Updated AC + Title username_2: I made some updates but putting back in up next to see if anyone has anything else to add username_1: Will post the current README in here, and work on it on github, so I can use github image uploading username_1: # bn-mobile-react BigNeon React Native Repo This is the react native Big Neon Mobile App repository, part of the Big Neon ticketing system. # Overall project architecture Big Neon is a multi-tiered micro-services architecture for selling and managing tickets. The software system is made up of multiple components. As such, the code is split across multiple repositories. To get an overall picture of how everything fits together have a look at the [docs repository](https://github.com/big-neon/docs.git) ## Project Setup - If you don't use the `asdf` version manager, set it up: https://github.com/asdf-vm/asdf - Run `./bin/setup` from the project directory. - Run `./bin/start` to start the app on your dev machine. ### Tech Stack #### Explanation This will be a React Native app for iOS and Android. #### React Native Standalone Debugger To use the RN Standalone Debugger with an iPhone simulator, first install it with: ``` brew update && brew cask install react-native-debugger ``` Then, make sure You disable Remote Debugging on any running simulators. (`Cmd+D -> Disable Remote JS Debugging`) Then run `npm run debug` After the standalone debugger loads up, you can then re-enable Remote JS Debugging, which should then open in the standalone app, and not in a browser window. ### Requirements ### Mobile React Native (https://facebook.github.io/react-native/) Unstated (https://github.com/jamiebuilds/unstated) React Navigation (https://reactnavigation.org/) ### Deploying There are staging and production deploy scripts, based on using Expo's release channels. These will build the apps with the appropriate staging or production config settings. You will need Apple and Google developer accounts. We're using [Expo Build releasing through Fastlane](https://blog.expo.io/automating-standalone-expo-app-builds-and-deployments-with-fastlane-exp-and-exptool-9b2f5ad0a2cd) as a guideline if you want to try to set up your own deploy. #### Set up your environment for deployment The `expo` CLI should be installed if you've run `npm install`. Login to expo using the BigNeon credentials: [Truncated] ``` You will also need a `google-deploy-key.json` file to upload to Google. Contact Brad at Revelry or Keith at Big Neon for this file, and save it in the .creds directory. Never commit it to github. #### Deploying First, make sure you bump the `version` and `android:versionCode` numbers in `app.json`. - `version` should follow semantic versioning: `major.minor.patch`. If the updates are minor, just bump the patch number by one. Moderate changes bump the minor number by 1. And large changes bump the major number by 1. - `versionNumber` must always be an integer, and can just be incremented by one. ``` $> ./deploy_staging ``` This will build the app with the staging variables, and deploy to the Google Play store's alpha track, and to Apple's App Store Connect. In the Apple App Store Connect, you may need to manually go to the build listing and fix the Missing Compliance issue. Also, make sure there is always a `user: <EMAIL>, pass: test` account so Apple can log in to the app to test it. username_1: (Making deployment its own DEPLOYMENT_README.md) ---- ### Deploying There are staging and production deploy scripts, based on using Expo's release channels. These will build the apps with the appropriate staging or production config settings. You will need Apple and Google developer accounts. We're using [Expo Build releasing through Fastlane](https://blog.expo.io/automating-standalone-expo-app-builds-and-deployments-with-fastlane-exp-and-exptool-9b2f5ad0a2cd) as a guideline if you want to try to set up your own deploy. #### Set up your environment for deployment The `expo` CLI should be installed if you've run `npm install`. Login to expo using the BigNeon credentials: ``` $ npm run expo -- login username/email: bigneon password: <<PASSWORD>> ``` In your `.bash_profile` (or whichever file hosts your bash/zsh/etc config), 2 ``` export LC_ALL=en_US.UTF-8 export LANG=en_US.UTF-8 export DELIVER_USERNAME=apple developer username export DELIVER_PASSWORD=<PASSWORD> export FASTLANE_ITC_TEAM_ID=119447135 export SECRET_SENTRY_AUTH_TOKEN=sentry auth token ``` The Team ID is the Big Neon Team ID in the App Store. To deploy, you will need to be added to the organization. [Based on the steps here](https://docs.expo.io/versions/latest/guides/building-standalone-apps.html#3-start-the-build), first you need to set up some keystores: _Android:_ ``` $> expo build:android ``` In the following text, select `2) I want to upload my own keystore!` ``` Would you like to upload a keystore or have us generate one for you? If you don't know what this means, let us handle it! :) 1) Let Expo handle the process! 2) I want to upload my own keystore! ``` You can get the keystore file from Brad (<EMAIL>), as well as the `keystpre password`, `key alias`, and `key password`, which will all be needed. _iOS:_ ``` $> expo build:ios ``` [Truncated] ``` You will also need a `google-deploy-key.json` file to upload to Google. Contact Brad at Revelry or Keith at Big Neon for this file, and save it in the .creds directory. Never commit it to github. #### Deploying First, make sure you bump the `version` and `android:versionCode` numbers in `app.json`. - `version` should follow semantic versioning: `major.minor.patch`. If the updates are minor, just bump the patch number by one. Moderate changes bump the minor number by 1. And large changes bump the major number by 1. - `versionNumber` must always be an integer, and can just be incremented by one. ``` $> ./deploy_staging ``` This will build the app with the staging variables, and deploy to the Google Play store's alpha track, and to Apple's App Store Connect. In the Apple App Store Connect, you may need to manually go to the build listing and fix the Missing Compliance issue. Also, make sure there is always a `user: <EMAIL>, pass: test` account so Apple can log in to the app to test it. username_2: With #709, is this needed anymore? username_3: Can we either get this over the finish line or close it out? username_2: Closing Status: Issue closed
Ordinastie/MalisisCore
251545966
Title: 1.12 Server Crash. Question: username_0: 1.12 Server with latest forge crashes on launch. Client with the same mods installed launches and runs fine. The server crashes with any 1.12 version installed (6.0.1, 6.0.2, 6.0.3 & 6.0.4). Any fix/help would be appreciated. ---- Minecraft Crash Report ---- WARNING: coremods are present: LoadingPlugin (Quark-r1.2-104.jar) BCModPlugin (backpacks 1.12 - 3.3.0.jar) Do not report to Forge! Remove FoamFixAPI (or replace with FoamFixAPI-Lawful) and try again. (foamfix-0.7.4-1.12-anarchy.jar) MalisisCorePlugin (malisiscore-1.12-6.0.4.jar) IELoadingPlugin (ImmersiveEngineering-0.12-70.jar) CTMCorePlugin (CTM-MC1.12-0.2.2.7.jar) Inventory Tweaks Coremod (InventoryTweaks-1.63+beta.107.jar) LoadingPlugin (RandomThings-MC1.12.1-4.0.5.jar) MalisisSwitchesPlugin (malisisswitches-1.12-5.0.0.jar) Contact their authors BEFORE contacting forge // Ouch. That hurt :( Time: 8/20/17 8:27 PM Description: Exception in server tick loop java.lang.NoSuchMethodError: net.malisis.core.block.IBlockComponent.getSubBlocks(Lnet/minecraft/block/Block;Lnet/minecraft/creativetab/CreativeTabs;Lnet/minecraft/util/NonNullList;)V at net.malisis.core.block.MalisisBlock.func_149666_a(MalisisBlock.java:400) at net.minecraft.item.ItemBlock.func_150895_a(ItemBlock.java:149) at crafttweaker.mc1120.item.MCItemUtils.createItemList(MCItemUtils.java:39) at crafttweaker.mc1120.brackets.BracketHandlerItem.rebuildItemRegistry(BracketHandlerItem.java:65) at crafttweaker.mc1120.events.CommonEventHandler.registerRecipes(CommonEventHandler.java:40) at net.minecraftforge.fml.common.eventhandler.ASMEventHandler_221_CommonEventHandler_registerRecipes_Register.invoke(.dynamic) at net.minecraftforge.fml.common.eventhandler.ASMEventHandler.invoke(ASMEventHandler.java:90) at net.minecraftforge.fml.common.eventhandler.EventBus$1.invoke(EventBus.java:143) at net.minecraftforge.fml.common.eventhandler.EventBus.post(EventBus.java:179) at net.minecraftforge.registries.GameData.fireRegistryEvents(GameData.java:733) at net.minecraftforge.common.crafting.CraftingHelper.loadRecipes(CraftingHelper.java:623) at net.minecraftforge.fml.common.Loader.initializeMods(Loader.java:779) at net.minecraftforge.fml.server.FMLServerHandler.finishServerLoading(FMLServerHandler.java:107) at net.minecraftforge.fml.common.FMLCommonHandler.onServerStarted(FMLCommonHandler.java:336) at net.minecraft.server.dedicated.DedicatedServer.func_71197_b(DedicatedServer.java:216) at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:483) at java.lang.Thread.run(Unknown Source) A detailed walkthrough of the error, its code path and all known details is as follows: --------------------------------------------------------------------------------------- -- System Details -- Details: Minecraft Version: 1.12 Operating System: Windows 10 (amd64) version 10.0 Java Version: 1.8.0_144, Oracle Corporation Java VM Version: Java HotSpot(TM) 64-Bit Server VM (mixed mode), Oracle Corporation Memory: 5776305296 bytes (5508 MB) / 8396472320 bytes (8007 MB) up to 8396472320 bytes (8007 MB) JVM Flags: 2 total; -Xmx8192M -Xms8192M IntCache: cache: 0, tcache: 0, allocated: 0, tallocated: 0 FML: MCP 9.40 Powered by Forge 14.21.1.2443 124 mods loaded, 124 mods active States: 'U' = Unloaded 'L' = Loaded 'C' = Constructed 'H' = Pre-initialized 'I' = Initialized 'J' = Post-initialized 'A' = Available 'D' = Disabled 'E' = Errored UCH minecraft{1.12} [Minecraft] (minecraft.jar) [Truncated] UCH voidcraft{0.24.4} [VoidCraft] (VoidCraft-1.12-0.24.4.jar) UCH waystones{4.0.15} [Waystones] (Waystones_1.12.1-4.0.15.jar) UCH xnet{1.4.0} [XNet] (xnet-1.12-1.4.0.jar) UCH xtones{1.12-1.0.5-8} [Xtones] (Xtones-1.12-1.0.5-8.jar) Loaded coremods (and transformers): LoadingPlugin (Quark-r1.2-104.jar) vazkii.quark.base.asm.ClassTransformer BCModPlugin (backpacks 1.12 - 3.3.0.jar) brad16840.common.asm.ClassTransformer Do not report to Forge! Remove FoamFixAPI (or replace with FoamFixAPI-Lawful) and try again. (foamfix-0.7.4-1.12-anarchy.jar) pl.asie.foamfix.coremod.FoamFixTransformer MalisisCorePlugin (malisiscore-1.12-6.0.4.jar) net.malisis.core.util.chunkcollision.ChunkCollisionTransformer net.malisis.core.util.chunkblock.ChunkBlockTransformer net.malisis.core.renderer.transformer.MalisisRendererTransformer net.malisis.core.renderer.icon.asm.TextureMapTransformer net.malisis.core.util.clientnotif.ClientNotifTransformer IELoadingPlugin (ImmersiveEngineering-0.12-70.jar) blusunrize.immersiveengineering.common.asm.IEClassTransformer CTMCorePlugin (CTM-MC1.12-0.2.2.7.jar) Answers: username_0: It seems to be a problem with Crafttweaker. It launches fine with that and Modtweaker removed and they aren't really necessary for my build. Status: Issue closed username_1: It seems that Crafttweaker uses a method that was previously only available on the client but was recently changed. The next update will have it fixed.
valentina2pucer/APPR-2015-16
140602856
Title: shiny Question: username_0: Potrebovala bi pomoč, želim namreč pri zemljevidu,ki je že narejen dodati aplikacijo, kjer se bodo z določeno barvo, obarvali določeni predeli glede na gostoto, ki si izbere da ga zanima.(gostota smrtnih žrtev). Answers: username_0: v Shiny.r bi rada naredila da mi na zemljevidu pokaže ( obarva ) področja, kjer določena gostota porazdelitve. Pripravila sem primerno tabelo, vendar nevem kako naj napišem podatke, da mi bo pri zemljevidu to upoštevalo.. Na internetu sem našla, da je glede na procente prebivalstva. in je prav funkcija za procente... js sem pa podatke že pripravila, vendar ne vem kam naj jih točno upišem da mi to upošteva username_1: Najprej se opravičujem - sem že napisal odgovor na prvotno sporočilo, a mi ga očitno ni poslalo. Za Shiny naredi tako: * premakni vse ukaze `library` iz `shiny.r` v `lib/libraries.r`, * premakni celoten ukaz `shinyUI` iz `shiny.r` v `shiny/ui.R`, * premakni celoten ukaz `shinyServer` iz `shiny.r` v `shiny/server.R` (pred njega lahko premakneš še preostalo kodo, ali pa jo daš v katerega od programov, ki ga vključiš s `source`), * pred `runApp("shiny")` dodaj vse ukaze `source` iz poročila, V `shiny.r` imaš tako le ukaze `source` in `runApp("shiny")` (stavek `if` lahko pobrišeš) - ko ga zaženeš, se torej zažene tvoj celoten program, nato se pa zažene še Shiny. Tako organizirano aplikacijo lahko potem še vključiš v poročilo, ne da bi bilo potrebno še enkrat zaganjati celotnega programa. Pazi še na to, da enako poimenuješ komponente v strežniškem in odjemalskem delu. Zemljevidu si v `server.R` dala ime `zem`, zato isto ime uporabi v `ui.R` pri `plotOutput`. Zemljevida ti sicer ne izriše, ker uporabljaš funkcijo `percent_map`, ki je nimaš v nobenem izmed uvoženih paketov. Da boš lahko podatke sploh risala na zemljevid, namesto razpredelnice `nesrece` izpelji `zem_tabela` iz razpredelnice, kjer imaš že razdelano razdelitev na države, npr. `tabela_1`. Potem lahko zemljevid narišeš na podoben način kot v `vizualizacija.r`: ```R ggplot() + geom_polygon(data = zem_tabela %>% group_by(Lokacija, Razvrstitev) %>% filter(Razvrstitev == input$var) %>% right_join(zem, by = c("Lokacija" = "STATE_NAME")), aes(x = long, y = lat, group = group, fill = Razvrstitev)) + scale_fill_manual(values = c("Redko" = "darkgreen", "Manj gosto" = "black", "Gosto" = "darkorange", "Najgosteje" = "darkviolet"), na.value = "darkgrey") ``` Funkcija `scale_fill_manual` omogoča določanje barv za posamezne kategorije. Tako uporaba stavkov `switch` ni potrebna. username_0: Zanima me, kaj mi še manjka, da bo sploh delovalo. + poročila mi noče prevesti (tudi brez shiny), ker je nekaj narobe z zadnjim grafom , ki sem ga dodala. username_1: V `shiny.r` kličeš program `helpers.R`, ki ga nimaš na repozitoriju, tako da to odstrani. Iz `server.R` odstrani še ukaze `source` in stavek `if` ter pri definiciji razpredelnice `zem_tabela` nadomesti `nesrece` s `tabela_1`. Potem lahko Shiny aplikacijo poženeš tako, da poženeš program `shiny.r`. Poročila ti ne prevede zaradi napake v `analiza.r` - v vrstici 21 premakni `+` v prejšnjo vrstico. Prav tako pri zadnjem grafu v poročilu manjka `+` na koncu tretje vrstice. Da se bo graf prikazal, pobriši presledek med oznako bloka `graf4` in vejico. username_0: sedaj sem popravila pomojem use, vendar še vedno ne deluje shiny .. mi ga ne požene in v poročilu ga tudi ne najde. username_1: V `shiny.r` pokliči še program `analiza/analiza.r`. Da ti bo prevedlo poročilo, spremeni prvi parameter pri `shinyAppDir` na `"shiny"` (to je ime mape, kjer se nahaja aplikacija). Status: Issue closed
Bouncyrock/TaleSpire-Alpha-Public-Issue-Tracker
400071721
Title: Visibility/ population issue Question: username_0: **Describe the bug** I was testing layers of the map in player mode. I switched back to GM and jumped out to a high view. The 2nd layer didnt populate the map in some areas until i went to the 2nd layer specifically **To Reproduce** Steps to reproduce the behavior: 1. build a multi level structure 3+ (large) 2. move to a lower level - removing higher levels from view 3. move camera away from part of the upper levels 4. jump to a high level [i went from 1 to 6] 5. move camera around map parts of lower levels do not populate correct. **Screenshots/Videos/Twitch Clips** ![screenshot 55](https://user-images.githubusercontent.com/46600569/51289094-ec4cf500-19cc-11e9-83f3-0c8fdda4ece5.png) (above) this is in 6th layer view ![screenshot 56](https://user-images.githubusercontent.com/46600569/51289101-eeaf4f00-19cc-11e9-862b-fac2c65826a9.png) (above) this is in 2nd layer view. **Additional context** I would like to not that part of the 3rd layer, was built in 2nd layer view... stacking blocks into third layer, But none of those blocks are on screen they are elsewhere in the map. Answers: username_0: (Below) from 6th layer ![screenshot 57](https://user-images.githubusercontent.com/46600569/51289373-e1469480-19cd-11e9-8ac7-6470e9053a49.png) (Below) 2nd layer ![screenshot 58](https://user-images.githubusercontent.com/46600569/51289378-e4da1b80-19cd-11e9-8e2a-289827c92f28.png) username_0: ![screenshot 69](https://user-images.githubusercontent.com/46600569/52670709-3ef3d100-2ee7-11e9-8e2f-80105e5c283c.png) Im dropping this in, idk why when loading the tiles load all over and then begin to load chunk by chunk if while looking at this code could get the initial loading to begin with the chunk by chunk approach we could avoid this weird thing that was on my screen for like 30 seconds. username_0: ALSO when looking at the screen shots click the pictures because those are hard to see until clicked then everything looks normal. username_1: There are a few things playing into this issue (one of which is what areas around the players position are updated) and I agree we can do a bunch more to mitigate it. We are changing how fog of war is placed in the near future which will affect some of this but we should definitely keep the loading icon up until the board is fully placed. username_1: I've got a small pile of fixes for tile reveal in the release for this weekend. I think it would be a good idea review what bugs in the above remain after that ships
serilog/serilog
234636126
Title: Best way to work with Azure Functions Question: username_0: I'm a total noob at this, but I am having some difficulty working with Serilog in an Azure Functions. I'm just getting started writing some sample things to make sure I understand and that it's going to work for this project. Here's the simple (awful) code. ILogger logger = new LoggerConfiguration() .WriteTo.AzureTableStorage(storage, Serilog.Events.LogEventLevel.Verbose, null, "TrustyFunctionsLog") .Enrich.FromLogContext() .WriteTo.LiterateConsole() .WriteTo.Console() .WriteTo.Trace() .MinimumLevel.Verbose() .CreateLogger(); Writing to: `AzureTableStorage` is fine. `Console` - nothing appears in the console window during local execution, or in the application logging windows in Azure Functions portal. `LiterateConsole` - same as above `Trace` - this seems to work locally to the debug window, but not in production. I know the AF docs suggest not writing to the console, but I was hoping to see it during development. I know I can abstract it away to write in multiple places myself, but I have feeling that I'm doing something really wrong... I know I'm missing something, I just don't know what. Answers: username_1: Hi @username_0, I have recently been looking at this area as well. I believe what you are looking for is to write to the `Microsoft.Azure.WebJobs.Script.InterceptingTraceWriter`. Details are over at [Microsoft.Azure.WebJobs.Script.InterceptingTraceWriter](https://github.com/Azure/azure-webjobs-sdk-script/blob/dev/src/WebJobs.Script/Diagnostics/InterceptingTraceWriter.cs). This is a class that I do not believe a sink currently exists for. I have started playing with some ideas [here](https://github.com/username_1/serilog-sinks-azure-functions) but nothing functioning. username_0: I saw your repo, but there wasn't anything in it. Time constraints for me are such that I can't start another project on top of what I'm doing. Wish I had the time to help out... GL with it. username_1: @username_0 sorry, it can't be of much help in the short term. Closing this out and we can pick things up at the other repo. Status: Issue closed username_2: @username_0 one option you might consider is to write to a network-based viewer (Seq, Splunk, Elasticsearch, ...) when doing local development. You can spin up Seq on localhost and get a pretty good "console-like" experience via the Seq sink. username_0: @username_2 Thanks, it's working great. I have some methods that I am also using on my Asp.Net site that I wanted to use for my new Azure Functions. They are looking for a `Microsoft.Extensions.Logging` instance and I wanted to stick with an Application Insights all MS solution. Looks like I'll have to replace all that stuff with Serilog. username_3: @username_0 Did you manage to get the Microsoft.Extensions.Logging working along with Serilog in your azure function apps? If so please provide the code snipset
Rycieos/factorio-leaflet-maps
602723340
Title: Map doesn't load with single date Question: username_0: If there is only a single entry in the array in dates.json, the map does not display. Answers: username_1: I'm not able to recreate this. A JSON file with only this: ```json [ "2020-04-11" ] ``` loads correctly. Are you sure you didn't have a comma (`,`) trailing the only value? JSON is picky about that. However, this has made me realize that we shouldn't show the date selection bar if there is only one value in the array. username_0: Went back and could not reproduce. I had checked that the json looked right, but must have missed something. Status: Issue closed