repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
w3c/ttml1
115104257
Title: region attribute unspecified Question: username_0: Another issue that has concerned me lately is what it means when the region attribute is unspecified. According to the spec, this simply means the sub-tree does not show up in the rendering, Which is fine and simple to understand and implement; however I'm wondering if it should perhaps instead have the special meaning of 'any region'. For example: <body> <div region="r1"> <p tts:color="red"begin="0s" end="10s">This text must be red.</p> </div> <div region="r2"> <p tts:color="blue"begin="0s" end="10s">This text must be blue.</p> </div> </body> In this case, clearly the intent is that div 1 shows up in r1 and div 2 shows up in r2, however without specifying <body region="r1 r2"> this won't happen. In complex scenarios I can imagine the amount of region attributes getting quite tedious. Also, if as we have discussed, we are going to allow the idea of an anonymous default region, which I think most of the test suite now relies on, we will need some sort of rule like this anyway in order to get content to be targeted at that anonymous region. We could still allow that an explicit declaration of no region <body region=""> means that it will show up in no region at all to retain the current semantics. (raised by <NAME> on 2008-12-11) From tracker issue http://www.w3.org/AudioVideo/TT/tracker/issues/42<issue_closed> Status: Issue closed
getgauge/gauge
217125426
Title: gauge --init java_maven is initializing with an old Gauge-java and giving a Protobuf exception Question: username_0: **Expected behavior** I should be able to run the tests without an error. ``` mvn clean test [INFO] Scanning for projects... [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Building art1 1.0-SNAPSHOT [INFO] ------------------------------------------------------------------------ [INFO] [INFO] --- maven-clean-plugin:2.5:clean (default-clean) @ art1 --- [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ art1 --- [WARNING] Using platform encoding (UTF-8 actually) to copy filtered resources, i.e. build is platform dependent! [INFO] skip non existing resourceDirectory /Users/<username>/work/test/art1/src/main/resources [INFO] [INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @ art1 --- [INFO] No sources to compile [INFO] [INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ art1 --- [WARNING] Using platform encoding (UTF-8 actually) to copy filtered resources, i.e. build is platform dependent! [INFO] skip non existing resourceDirectory /Users/<username>/work/test/art1/src/test/resources [INFO] [INFO] --- maven-compiler-plugin:3.1:testCompile (default-testCompile) @ art1 --- [INFO] Changes detected - recompiling the module! [WARNING] File encoding has not been set, using platform encoding UTF-8, i.e. build is platform dependent! [INFO] Compiling 1 source file to /Users/<username>/work/test/art1/target/test-classes [INFO] [INFO] --- maven-surefire-plugin:2.12.4:test (default-test) @ art1 --- [INFO] [INFO] --- gauge-maven-plugin:1.1.0:execute (default) @ art1 --- [Truncated] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException ``` **Observation** The template generates pom file dependency with old gauge-java ``` <dependency> <groupId>com.thoughtworks.gauge</groupId> <artifactId>gauge-java</artifactId> <version>0.3.4</version> <scope>test</scope> </dependency> ``` Answers: username_1: @username_0 can we retest this please? I've just created a new template and its point to 0.6.0 which should work with 0.8+. Status: Issue closed
freeCodeCamp/testable-projects-fcc
587079552
Title: Test environment wont load tests Question: username_0: <!-- freeCodeCamp Testable Projects Issue Template --> I just changed workstation for another computer and now i have a problem with the test environment. #### Issue Description I try to apply <script src="https://cdn.freecodecamp.org/testable-projects-fcc/v1/bundle.js"></script> to my site and it is working but when i try to load the test, i just get the message "load tests" in the background of the field i can see the message #mocha div missing. It worked on my macbook, now im using my older Imac and here it is not working. I checked on my macbook and it seems that it is still working there? google chrome Version 80.0.3987.149 osx yosemite 10.10.5 Imac scripting in Atom ![Screen Shot 2020-03-24 at 17 06 37](https://user-images.githubusercontent.com/59973581/77449141-e6814980-6df1-11ea-901d-8ef1fee57740.png) ![Screen Shot 2020-03-24 at 17 06 37](https://user-images.githubusercontent.com/59973581/77449179-f0a34800-6df1-11ea-95c0-1b3650761015.png) havent even really started my script, so this is it.. <html> <h1>Hello World</h1> <script src="https://cdn.freecodecamp.org/testable-projects-fcc/v1/bundle.js"></script> </html> ``` #### Screenshot <!-- Add a screenshot of your issue --> Answers: username_0: I used firefox instead and here it works... strange.. Status: Issue closed
libsdl-org/sdlwiki
847683808
Title: request to tweak book urls in wiki Question: username_0: https://github.com/libsdl-org/SDL/wiki/Books - Would appreciate editing the title for `Learn C++ By Making Games` book to hide the `target="blank"` - Would appreciate editing the title for `Game Programming in C++: Start to Finish` to also hide the `target="blank"` - For clarity / helpfulness to anyone seeing these entries, can they be tagged somehow as `SDL 1.2` material? Answers: username_1: This is resolved now. username_0: thanks!
Teststandees/hcal_teststand_scripts
107379931
Title: QIE Status Question: username_0: get HF1-2-iBot_StatusReg_PLL320MHzLock # 1 Any value different from the example above indicates bad behavior. these get-commands should be included in the system setup script. Values could be bad if the power is enabled after doing a bkp_reset. they could explain bad behaviors seen last week Tullio ```
salesforce/design-system-react
873927312
Title: How to Setup in CRA 4.x the documentation has no steps for setting up while using the latest CRA version. Answers: username_1: @username_0 and anyone else who may be struggling with this, the following worked for me using the latest version of CRA: 1. Follow all [the steps](https://github.com/salesforce/design-system-react/blob/master/docs/create-react-app-2x.md#step-by-step-instruction) in the CRA 2.x guide, but also install `react-app-rewire-babel-loader` to your dev dependencies 2. Replace the contents of config-overrides.js with the following: ```js const path = require("path"); const fs = require("fs"); const rewireBabelLoader = require("react-app-rewire-babel-loader"); const appDirectory = fs.realpathSync(process.cwd()); const resolveApp = (relativePath) => path.resolve(appDirectory, relativePath); module.exports = function override(config, env) { config = rewireBabelLoader.include( config, resolveApp("node_modules/@salesforce/design-system-react"), ); return config; }; ```
jrwaltz/intro-data-capstone-biodiversity
302539876
Title: Summary Question: username_0: Great work! I probably won’t have much to say here because you pretty much nailed it all the way through. Your Python work was excellent and very readable. Maybe because you copy pasted your code from the codecademy platform, there are multiple occurrences of the same thing. Presentation: I think your graphs look excellent. They don't belong at the last slides with a `Graphs Created` title. Embed your plots on the slides where you are describing the data. I liked the way you described your results including p-values and tables. Your recommendations are resonable. Excellent work overall.
griggheo/ansible-consul-template
125793823
Title: Tag releases Question: username_0: Would you mind tagging a release so that I can use this role in a [Galaxy requirements file](https://docs.ansible.com/ansible/galaxy.html#advanced-control-over-role-requirements-files) without having to use master? Answers: username_1: Tagged release 1.0.0 code name "avocet" (I like birds :) https://github.com/username_1/ansible-consul-template/releases/tag/1.0.0 Status: Issue closed
desktop/desktop
493531504
Title: rebase modal window resizes sporadically Question: username_0: Hi there! This is not a bug, but a minor UX detail. When doing a long-running rebase, the modal window changes width and height depending on the length of the filename being processed. I think it would be easier on the eyes if the modal window always stayed the same size. My gif-fu is not good right now. Here's a YouTube video: https://youtu.be/9uVKxsdG3kE Answers: username_1: Here is some screenshots. ![Screen Shot 2019-09-16 at 3 47 17 PM](https://user-images.githubusercontent.com/14828183/65072875-74551700-d92d-11e9-9936-68515068cd21.png) ![Screen Shot 2019-09-16 at 3 47 47 PM](https://user-images.githubusercontent.com/14828183/65072876-74551700-d92d-11e9-822c-35b046f195e4.png)
betagouv/mon-entreprise
760289277
Title: Intégrer « Place des entreprises » Question: username_0: https://place-des-entreprises.beta.gouv.fr Peut-être pour commencer sur la page [Gérer mon activité](https://mon-entreprise.fr/g%C3%A9rer) ? Voir avoir des liens direct vers les thématiques ? D'autres idées ? cc @MathieuGens Answers: username_0: Place des entreprises se développe bien (produit refinancé sur 3 ans), et je pense qu'il y a une complémentarité entre nos deux sites : * sur [mon-entreprise](https://mon-entreprise.fr/), des simulateurs et outils interactifs pour apporter rapidement l'information la plus précise possible * sur [place-des-entreprises](https://place-des-entreprises.beta.gouv.fr/), la mise en relation avec le bon conseiller Avant de l'intégrer sur nos pages il faut que place-des-entreprises soit disponible au niveau national (je crois que c'est en cours ?). On serait probablement un apporteur de trafic important, en effet il y a actuellement 400 demandes par mois sur place des entreprises, or c'est un service qui intéressera sans doute quelques pourcents de nos ~300k visiteurs mensuels. cc @Flightan @be-mercier username_0: Vu avec l'équipe place-des-entreprises : - le service sera déployé région par région mais cela va prendre du temps (pas d'ETA), les demandes venant de régions non éligibles ne sont pas traitées - le service n'est pas pertinent pour les artistes-auteurs, professions libérales, particuliers employeurs - on peut dans un premier temps intégrer un lien générique vers le place-des-entreprises, puis dans un second temps intégrer directement les blocs thématiques. place-des-entreprises peut fournir une iframe (cf. https://brexit.hautsdefrance.fr/echanger-avec-un-conseiller/) ou si on le demande un json des cartes. Pour démarrer on peut l'ajouter sur la page "Gérer" si la personne remplit son entreprise et qu'elle est dans une région éligible (IdF, hauts de France) et avec un statut compatible (SAS / SARL). Ça fera pas beaucoup de trafic mais ça permet de tester et l'équipe place-des-entreprises pourra suivre le volume et le type de demande effectué. username_0: En plus de la section « Gérer », on peut aussi intégrer des blocs thématiques sous les simulateurs : Bloc « contacter un conseiller à propos d'un recrutement » https://place-des-entreprises.beta.gouv.fr/aide-entreprises/recrutement-formation/demande/recruter#section-breadcrumbs - simulateur cout d'embauche - simulateur aides à l'embauche Bloc « contacter un conseiller pour résoudre un problème de trésorerie, faire face à vos charges » https://place-des-entreprises.beta.gouv.fr/aide-entreprises/entreprise-en-difficulte/demande/tresorerie#section-breadcrumbs - simulateur indépendant - simulateur profession libérale place-des-entreprises peut fournir des intégrations iframe des formulaires pour que l'utilisateur reste sur le site mon-entreprise (on pourrait l'ouvrir dans une modale par exemple) username_0: J'ai fait un point avec @Adeline-Lrn aujourd'hui concernant l'intégration de place-des-entreprises. On n'a pas eu la bande passante pour l'intégrer en 2020, et par ailleurs le fait que le service était limité à quelques régions rendait l'intégration plus compliquée. Mais en 2021 on va avoir plus de moyens, et place-des-entreprises sera déployé au niveau national, avec quasiment toutes les régions pour mars 2021 et un lancement national officiel un peu plus tard dans l'année. Je propose donc que l'on vise fin T1 2021 pour déployer cette intégration. À valider avec Evelyne, mais je pense qu'on est sur un sujet qui a un bon ratio d'impact par rapport au temps de développement, donc ça ne devrait pas poser problème. Par ailleurs point positif, <NAME> a validé le principe d'intégrer place-des-entreprises en mars 2021 sur l'espace connecté urssaf.fr (à la place du bouton oups.gouv qui disparaît). S'aligner sur ce point avec le calendrier et la feuille de route Dicom permet à peu de frais de montrer à quel point on fait partie de l'écosystème Urssaf et qu'on participe à une stratégie de communication cohérente auprès des portails de la branche [insérer un PPT ici]. Pour le travail d'intégration en lui-même : - À quels endroits on ajoute ce module ? Sur quelles catégories de questions ? cf. message ci-dessus pour les réflexions précédentes - Est-ce qu'on veut ré-implémenter les formulaires à partir d'une API place-des-entreprise pour avoir le meilleur niveau d'intégration UI possible ? Ou bien préfère-t-on une intégration iframe qui a un coût de développement et de maintenance plus faible ? - Quelles informations peut-on pré-remplir à partir des données déjà saisies lors du parcours sur mon-entreprise ? - Comment on peut suivre les demandes initiées depuis mon-entreprise et les ajouter à notre page /stats ? Status: Issue closed
lark-parser/lark
617361617
Title: Turning Lark into a modularized mutli-paradigm parsing toolbox Question: username_0: Hi there! Since a long time, I have the dream of an ultimate toolbox for all parsing-related stuff, which fits all my needs and can be used for any parsing-related issue, including experimentation. The main goals of my dream are: 1. Write a grammar once in one kind of EBNF-language 2. Use a parser generated from this grammar in several target languages and technologies (for me, it currently is C, Python and Rust) 3. Work with AST traversals instead of semantic code attached to the grammar 4. Combine different scanners (Pythons `re`, self-contained DFA), but also support of scanner-less parsing 5. Use the same toolbox for experimentation and fiddling with algorithms and ideas Right now, I'm experimenting with the GLL parsing algorithm, but with the need to not completely rewrite all my grammars again, and with the need to use the resulting parser in Python and Rust. Therefore, I'm currently undetermined whether I write another tool on my own or modify an existing tool, like Lark. While taking a look into Larks source code and documentation, I found out that there are [plannings](https://github.com/lark-parser/lark/blob/master/docs/features.md#planned-features-not-implemented-yet) to generate code in other languages and implement an own DFA construction for lexical analysis without using Pythons regex-module for this. For me, the construction of scanner-less parsers is also an important thing which should be supported. This could be done by breaking up the regular expressions and turning them into a part of the grammar, obviously a process of grammar rewriting which might be done internally, including whitespace insertion. All these targets could only be achieve when Lark is broken up into a modular, plugin-based design, where different parser generators, different scanner generators and different code generators are stacked together, and Lark itself only serves as the core to handle and rewrite grammars and use the plugins attached. I've noticed that Lark already has a modular structure, but my concept would be to separate all these modules and make them entirely plug-able and replaceable. This is my first draft of an idea how this could look like: ![lark-modularized](https://user-images.githubusercontent.com/16870072/81804330-37234200-9519-11ea-9d80-6eeb3e31fd6e.png) Please don't feel offended or worried, I don't want to break Lark's current structure or principles with this draft. It is only a proposal of what might be done, and I'm very interested into what your future plans for Lark are. I need a little decision help, whether my concept could become a part of Lark, or if I have to do it on my own. For now, I am not using Lark for my projects, because it does not fit into my requirements. But it could do so when some features would be supported. Python as implementation language for this toolbox is absolutely the right tool. Code generation could be done easily using template engines and a serialized structure of the generated parse-tables for each Parser module. The generated parsers generally should support the features of their implementation language for sake of performance, but should share the same node structuring for resulting tree traversals and semantic evaluation. What are your opinions on this proposal? Answers: username_1: Hi Jan! My goals are very similar to those that you listed, and I would say Lark satisfies all of them, to some degree. Lark can already generate a LALR parser in Julia. Lark also used to have support for scannerless parsing in Earley, which involved rewriting the strings as a list of character terminals. I decided to remove it because dynamic lexing was just as capable and is significantly faster. But I'll consider bringing it back if there's a good reason. If you want to contribute a GLL parser to Lark, I'll be happy to help as much as I can. Same if you want to contribute a Rust parser generator. Regarding lexical analysis, we have made some progress in that area too. The idea was to detect collisions in regexps by intersecting them: https://github.com/MegaIng/interegular That reminds me, @MegaIng, why didn't we move forward with that? Did you ever submit a PR, or were there still some performance issues? I'm not at all offended by the diagram, it seems like you have the right idea, in theory. I'm not sure it would hold up in practice. I would say all the modules in Lark are already pluggable. The standard lexer can serves both LALR(1) and Earley. But each one also has specialized lexers which only fit itself, and I don't see how that could change. To summarize the problem more abstractly: Each parser has its own unique configuration and abilities, and I don't think it's possible to know them in advance and pre-expose the right interface. The obvious solution would be to expose all the information, but that would actually make it less modular. So what I've done so far is to abstract only as much as necessary to support the algorithms that I have, and I'll be happy to keep abstracting whenever necessary, as long as it actually makes the code better. Having said that, it's possible that I've overlooked things, so I'm open to hear another plan of action. And of course there's always room for improvement. Regarding templated parsers, I think it's a good idea, in theory. I don't know how easy it would be to write a lexer+parser+treebuilder like that for each target. Especially if we want it to support the contextual lexer, and all the tree manipulation logic that the grammar allows. At any rate, Lark already knows to create a JSON of everything that it analyzed, so applying to a template should be fairly easy. So bottom line, I like it. I want to see Lark grow, and I think those are good directions for it to go in. So if you want to use Lark, I'm absolutely for it, as long as you submit clean and modular code. (I have to set a standard to keep it maintainable!) Let me know what you think, and if you have more questions for me. username_0: Sure, and I think Lark is great base for this. I'm aware that Lark has an already a huge and in some detail parts very skilled code base (for some algorithms, I first need to look them up and dive into them, and that's a time problem, though!), but generally I would like to contribute. Feel free to close the issue, or we can continue some discussion here. Best regards Jan Status: Issue closed username_0: Hey there, to keep issues low I now decided to close this here for now. Due to other, personal projects, I currently have too less time to participate and provide code to Lark, but I'll stay tuned about further developments in here and try to help when possible. username_1: Hi, it seems like I should have replied here, sorry that it slipped my mind! It's perfectly understandable, I have my own personal projects going. If you ever want to come back and contribute to Lark, I'll be happy to help you do that. P.S. I don't think a scannerless LALR(1) is a good idea. I think they ideal algorithm will try a regexp, and default to scannerless, and in the same sense will try LALR(k), and default to Earley only when needed. In theory, these different layers can be written in different languages (well, regexps already are). However, such a layered parser is hard to write. But Lark is going in that direction, and the suggestions you proposed might push it even further there. So I definitely dig that direction.
w3c/findtext
113055920
Title: order of case fold and normalization operations Question: username_0: http://www.w3.org/International/track/issues/503 [I18N-ISSUE-503] http://www.w3.org/TR/2015/WD-findtext-20151015/#performing-a-search-operation In the section "Search Algorithm" within this section, case fold mapping appears as step 11. Normalization appears as step 13. Since normalization (particularly the compatibility mapping) may change the character sequence, it would be better to apply the case folding after the normalization, particularly if the case fold is language-sensitive.
kubernetes/test-infra
567107643
Title: No status reconciler label Question: username_0: ```yaml { insertId: "1am6xuhg2k1c23t" jsonPayload: { author: "sbueringer" component: "hook" event-GUID: "aa9d1680-527e-11ea-8ba2-0a355f16f694" event-type: "pull_request" file: "prow/plugins/owners-label/owners-label.go:119" func: "k8s.io/test-infra/prow/plugins/owners-label.handle" level: "warning" msg: "Unable to add nonexistent labels: ["area/prow/status-reconciler"]" org: "kubernetes" plugin: "owners-label" pr: 16317 repo: "test-infra" url: "https://github.com/kubernetes/test-infra/pull/16317" } labels: {…} logName: "projects/k8s-prow/logs/hook" receiveTimestamp: "2020-02-18T18:46:01.873818640Z" resource: {…} severity: "ERROR" timestamp: "2020-02-18T18:45:56Z" } ``` /assign @stevekuznetsov /area prow
scottsilverlabs/raspberrystem-ide
93005516
Title: File rename does not update titlebar Question: username_0: When a file is renamed, it does not update the titlebar if used in the following way: * Start with a new or existing file * Click File Open * Click settings on the file being edited * Rename the file and click OK * Click Cancel button (this may be a Close button if #59) * Titlebar will still have old name Note - the file DOES get renamed properly, its just the titlebar is not updated. Reopening the file resolves it manually.<issue_closed> Status: Issue closed
HMS-Core/hms-flutter-plugin
801336405
Title: when request to sign in twice , the app crash [huawei_account] Question: username_0: **Description** i just used the huawei account version 5.0.3+303 and i implemented the **Expected behavior** What you expected to happen? i want to login and if the user not include the email to retry again with email , i will be more accept if there's a method thats makes the email required **Current behavior** What actually happened? when i press to sign in it will navigate me to sign in screen and if i canceled or signed in it will gives a throw with message to make user re-try to login with email but i got this type of exception : **Logs** ``` E/flutter (31393): #8 TapGestureRecognizer.handleTapUp package:flutter/…/gestures/tap.dart:598 E/flutter (31393): #9 BaseTapGestureRecognizer._checkUp package:flutter/…/gestures/tap.dart:287 E/flutter (31393): #10 BaseTapGestureRecognizer.acceptGesture package:flutter/…/gestures/tap.dart:259 E/flutter (31393): #11 GestureArenaManager.sweep package:flutter/…/gestures/arena.dart:157 E/flutter (31393): #12 GestureBinding.handleEvent package:flutter/…/gestures/binding.dart:362 E/flutter (31393): #13 GestureBinding.dispatchEvent package:flutter/…/gestures/binding.dart:338 E/flutter (31393): #14 RendererBinding.dispatchEvent package:flutter/…/rendering/binding.dart:267 E/flutter (31393): #15 GestureBinding._handlePointerEvent package:flutter/…/gestures/binding.dart:295 E/flutter (31393): #16 GestureBinding._flushPointerEventQueue package:flutter/…/gestures/binding.dart:240 E/flutter (31393): #17 GestureBinding._handlePointerDataPacket package:flutter/…/gestures/binding.dart:213 E/flutter (31393): #18 _rootRunUnary (dart:async/zone.dart:1206:13) E/flutter (31393): #19 _CustomZone.runUnary (dart:async/zone.dart:1100:19) E/flutter (31393): #20 _CustomZone.runUnaryGuarded (dart:async/zone.dart:1005:7) E/flutter (31393): #21 _invoke1 (dart:ui/hooks.dart:265:10) ``` **Environment** - Platform: Huawei nova 7i - Kit: huawei_account - Kit Version 5.0.3+303 - OS Version Android 10 - VS code Answers: username_1: Hi @username_0, Thanks for reporting the issue. We have spotted the error and it will be fixed on next release. Also you can follow these steps to create a workaround. 1. Go to **AuthServiceMethodHandler.java** file. 2. Then change the **onActivityResult** method as following: ```java @Override public boolean onActivityResult(int requestCode, int resultCode, Intent data) { final MethodChannel.Result incomingResult = mResult; mResult = null; Task<AuthHuaweiId> authIdTask = HuaweiIdAuthManager.parseAuthResultFromIntent(data); if (requestCode == rCode) { authIdTask.addOnSuccessListener(authId -> { HashMap<String, Object> resultMap = HwIdBuilder.createHwId(authId); Account account = authId.getHuaweiAccount(); if (account != null) { resultMap.put("account", HwIdBuilder.createAccount(account)); } HMSLogger.getInstance(activity.getApplicationContext()).sendSingleEvent("signIn"); if (incomingResult != null) { incomingResult.success(resultMap); } }).addOnFailureListener(e -> { HMSLogger.getInstance(activity.getApplicationContext()).sendSingleEvent("signIn", String.valueOf(((ApiException) authIdTask.getException()).getStatusCode())); if (incomingResult != null) { incomingResult.error(Constant.SIGN_IN_FAILURE, String.valueOf(((ApiException) authIdTask.getException()).getStatusCode()), e.getMessage()); } }); } return true; } ``` Run the application again and let us know how it goes. username_0: thank you Status: Issue closed
pombase/curation
206212007
Title: Linking "within process" molecular functions that are not known to be direct substrates Question: username_0: This ticket is spawned from: https://github.com/pombase/canto/issues/1308#issuecomment-278334643 we need a way to annotate things which are regulating a molecular function indirectly. In the example from here https://github.com/pombase/curation/issues/1278#issuecomment-277270238 - hsp9 where we don't know exactly how it is regulating cdc25 phosphatase activity - hsp9 has not been shown to be a direct regulator of cdc25 so I can't put it on the MF term - We want to link hsp9 to cdc25 as it is part of the pathway - Currently we use BP annotations "regulation of MF" for these. - PomBase have been evaluating if we can get rid of them for https://github.com/geneontology/go-ontology/issues/12859 (and for general consistency and LEGO compliance) - possibly the best way to do this would be to have a relation which can be used with a MF term and specifies that a target is not a direct substrate (causally upstream of *and* within). This does not exist. - [ ] Q what are the existing extensions for MF allowed by GO (Ihave a feeling we self imposed a narrower set) - [ ] once decision made, retrofit existing (mainly should be the ones sitting in the GO logs?) Answers: username_1: Anything that has local_domain = GO:0003674 or any of its is_a descendants, or BFO:0000015 ("process" in the formal-ontology sense that encompasses both MF and BP in GO). That's kind of a pain to get an from the documentation in GitHub (https://github.com/geneontology/annotation_extensions/tree/master/doc), which has a page for each relation to slog through ... username_0: I wonder if for these we can use "negatively regulates" and "positively regulates" This would not work for all molecular functions, but for most of the phosphorylation examples, it is regulatory. The example above would be hsp9 some_molecular_function negatively_regulates cdc25 part_of "negative regulation of G2/M transition" that seems to work? Status: Issue closed
microsoft/Microsoft-UI-UIAutomation
716803950
Title: Consider adding BreakIf and ContinueIf helpers to the C++ Abstraction library Question: username_0: When writing remote operation loop, consumers will frequently end up with a line such as this in the body of the loop: ```c++ operationScope.If(ancestorElement.IsNull(), [&]() { operationScope.Break(); }); ``` It could be helpful to add a helper that lets consumers write: ```c++ operationScope.BreakIf(ancestorElement.IsNull()); ``` This would be a little bit less verbose. The same thing could be said for `Continue`...<issue_closed> Status: Issue closed
srehwald/eat-api
304966458
Title: IPP: „risch“ statt „Frisch“ Question: username_0: https://github.com/username_1/eat-api/blob/gh-pages/ipp-bistro/2018/12.json Answers: username_1: Probably, the dish name was a litte longer than usual. This is a little difficult to fix as the parser might not be able to tell to which column a specific character belongs to in such a case. As this was 3 weeks ago and the menu changed anyway, I assume I can close this (known) issue. Status: Issue closed
vlang/v
484844230
Title: printing a nested struct instance fails Question: username_0: **V version:** 0.1.18 **OS:** Mac OS 10.14.6 **What did you do?** struct Inner { val i32 } struct Outer { inner Inner } p := Outer{Inner{}} println(p.inner) **What did you expect to see?** successful compilation **What did you see instead?** error: expected ')' println ( p tos2((byte*)"{ val: $.inner.val }") ) ; ^ note: to match this '(' println ( p tos2((byte*)"{ val: $.inner.val }") ) ; ^ 1 error generated. V panic: C error. This should never happen. Please create a GitHub issue: Answers: username_1: it seems that the variable was not initialized, this example works: module main struct Inner { val int } struct Outer { inner Inner } fn (o Outer) str() string { return o.inner.val.str() } fn main() { o := Inner{69} p := Outer{o} println(p.str()) // outputs "69" } Status: Issue closed username_2: Compiled with current master and it works. ``` struct Inner { val int } struct Outer { inner Inner } p := Outer{Inner{}} println(p.inner) ``` Only thing was to rename `i32` to `int` because BDFL @medvednikov changed `i32` lately to `int` and `u8` to `byte` - this is for the time being final. The output was: { val: 0 }
ccxt/ccxt
851163330
Title: ByBit new futures throws error in private endpoints Question: username_0: - OS: MacOS - Programming Language version: Python - CCXT version: Latest ``` bybit.fetch_closed_orders('BTC/USD') # works bybit.fetch_closed_orders('BTCUSDM21') # does not work, same with BTCUSDU21 ``` ``` bybit {"ret_code":10001,"ret_msg":"invalid symbol","ext_code":"","ext_info":"","result":null,"time_now":"1617696874.327896","rate_limit_status":599,"rate_limit_reset_ms":1617696874327,"rate_limit":600} ``` Symbols can be used in public endpoints, but not in private endpoints (tested against closed_orders and fetch_my_trades). Thanks for a great package! /Clev
varnishcache/varnish-cache
581829547
Title: We set Accept-Range on pass, is that OK ? Question: username_0: Shouldn't we leave it to the backend to set A-R on pass ? Answers: username_1: bugwash: Right now, we add `Accept-Range` to and 200 response when `http_range_support` is on. This happens after `vcl_deliver`, so there is no VCL control over it. We agree that we want more VCL control and three alternative approaches have been discussed: * add a switch to vcl if `Accept-Range` should be added if not present * set `Accept-Range` before calling `vcl_deliver`, so it can be removed * do not set `Accept-Range` in core code, but rather in `builtin.vcl` We agree to handle this as part of #3246 username_1: I realized that the actual cause of the problem is that we mixed handling of ranges in core code and the range vdp. By moving all relevant parts to that one, whether or not we send `Accept-Ranges` can be controlled by having the `range` vdp active or not. See #3289 username_0: How this should work: 1. vcl_backend_deliver{} should be able to set/unset Accept-Ranges as desired. 2. On pass'ed responses, the backend controls if A-R should be in resp.http going into vcl_deliver{} username_1: How do you address the concern? Status: Issue closed username_0: You can test `obj.uncachable` if it is true, the header came from the backend.
macacajs/app-inspector
235502219
Title: 有的时候可以,有的时候就不行,需要重启设备才可以 Question: username_0: Error: connect ECONNREFUSED 127.0.0.1:8001 at Object.exports._errnoException (util.js:1050:11) at exports._exceptionWithHostPort (util.js:1073:20) at TCPConnectWrap.afterConnect [as oncomplete] (net.js:1093:14) Answers: username_1: 需要插拔usb线,不过不是根本办法 username_2: @username_1 需要插拔 usb 说明 iproxy 的 usb 连接出问题了,可以提供更多信息吗? Status: Issue closed
ryansheehan/terraria
631519436
Title: Docker CI build error Question: username_0: Hi, First of all I'd like to thank you for updating this image so quickly, great work! However it seems like there's something going wrong in the DockerCI builds. I'd help but it seems like I can't access its logs (set to private?) Answers: username_1: It is most likely related to the faulty zip file name error that should be solved by this PR : https://github.com/username_3/terraria/pull/44 username_2: why is the docker container not build automatically as soon as the github gets updates? username_3: Those are from a temporary setup. For a while I was exploring building from source, and looking to contribute the dockerfile back to tshock. For a number of reasons I had to switch back to building from the release zips. I still hope to reactivate that effort of building from source. But I'm going to wait until the pre release phase is over. username_0: @username_3 That's a great idea! Tshock having an official image would be awesome. I think using a development or feature branch might provide a more ideal workflow for working on that though 🙂 Thanks for solving this so quickly and creating this image! Status: Issue closed
aonez/Keka
860432540
Title: Keka not launching on Big Sur from Synthetic.conf remapped directory Question: username_0: I usually keep my non-system programs in a directory located here `/Shared Files/Utilities` Prior to Catalina this was its real path but due to new security measures in Catalina/Big Sur the `/Shared Files` directory is actually located here `/System/Volumes/Data/Shared Files` and is remapped to the root of the volume using `/etc/synthetic.conf` (discussion here): https://derflounder.wordpress.com/2020/01/18/creating-root-level-directories-and-symbolic-links-on-macos-catalina/ Keka stopped working but I found that it works fine when the app bundle is placed in these directories: `/Applications` `~/Applications` `~/Desktop` but not `/Shared Files/*` I also tried manually launching from the "correct path" `/System/Volumes/Data/Shared Files` with no difference. The program seems to hang with no windows or menu bar menus produced. here what appears to be the relevant console output: ``` error 11:07:58.583755-0400 sandboxd Sandbox: Keka(35918) deny(1) file-issue-extension target:/Shared Files/Utilities/Keka.app class:com.apple.app-sandbox.read Violation: deny(1) file-issue-extension target:/Shared Files/Utilities/Keka.app class:com.apple.app-sandbox.read Process: Keka [35918] Path: /System/Volumes/Data/Shared Files/Utilities/Keka.app/Contents/MacOS/Keka Load Address: 0x106fcb000 Identifier: com.aone.keka Version: 4541 (1.2.13) Code Type: x86_64 (Native) Parent Process: launchd [1] Responsible: /System/Volumes/Data/Shared Files/Utilities/Keka.app/Contents/MacOS/Keka User ID: 503 Date/Time: 2021-04-17 11:07:58.573 EDT OS Version: macOS 11.1 (20C69) Report Version: 8 MetaData: {"build":"macOS 11.1 (20C69)","responsible-process-uid":503,"platform_binary":"no","primary-filter":"path","file-flags":0,"container":"\/Users\/yashka\/Library\/Containers\/com.aone.keka\/Data","profile-in-collection":false,"flags":5,"pid":35918,"signing-id":"com.aone.keka","platform-policy":false,"target":"\/Shared Files\/Utilities\/Keka.app","apple-internal":false,"file-mode":511,"errno":1,"vnode-type":"DIRECTORY","extension-class":"com.apple.app-sandbox.read","profile-flags":0,"hardware":"Mac","operation":"file-issue-extension","primary-filter-value":"\/Shared Files\/Utilities\/Keka.app","responsible-process-path":"\/System\/Volumes\/Data\/Shared Files\/Utilities\/Keka.app\/Contents\/MacOS\/Keka","action":"deny","rdev":0,"platform-binary":false,"summary":"deny(1) file-issue-extension target:\/Shared Files\/Utilities\/Keka.app class:com.apple.app-sandbox.read","process-path":"\/Shared Files\/Utilities\/Keka.app\/Contents\/MacOS\/Keka","hardlinked":false,"matched-extension":false,"uid":503,"mount-flags":76583424,"responsible-process-user-uuid":"1708BC0E-644A-4C78-A658-46320EF17422","matched-user-intent-extension":false,"path":"\/Shared Files\/Utilities\/Keka.app","normalized_target":["Shared Files","Utilities","Keka.app"],"team-id":"4FG648TM2A","process":"Keka"} Thread 0 (id: 69905753): 0 libsystem_kernel.dylib 0x00007fff2033c376 __mac_syscall + 10 1 LaunchServices 0x00007fff20884bd0 _LSApplicationCheckIn + 1840 2 HIServices 0x00007fff256d791c _RegisterApplication + 6665 3 HIServices 0x00007fff256d5e28 GetCurrentProcess + 23 4 HIToolbox 0x00007fff286bcabf MenuBarInstance::GetAggregateUIMode(unsigned int*, unsigned int*) + 63 5 HIToolbox 0x00007fff286bca49 MenuBarInstance::IsVisible() + 51 6 AppKit 0x00007fff22c4ba7b _NSInitializeAppContext + 35 7 AppKit 0x00007fff22c4977a -[NSApplication init] + 417 8 AppKit 0x00007fff22c493b9 +[NSApplication sharedApplication] + 120 9 AppKit 0x00007fff22c47c5b NSApplicationMain + 409 10 libdyld.dylib 0x00007fff2038a621 start + 1 11 Keka 0x0000000000000001 Thread 1 (id: 69905774): 0 libsystem_kernel.dylib 0x00007fff2033c53e __workq_kernreturn + 10 1 libsystem_pthread.dylib 0x00007fff2036b467 start_wqthread + 15 Thread 2 (id: 69905778): 0 libsystem_kernel.dylib 0x00007fff2033c53e __workq_kernreturn + 10 1 libsystem_pthread.dylib 0x00007fff2036b467 start_wqthread + 15 Thread 3 (id: 69905779): Binary Images: 0x106fcb000 - 0x10702efff com.aone.keka (1.2.13 - 4541) <73a74552-3c64-3607-9214-9817cbabb20f> /System/Volumes/Data/Shared Files/Utilities/Keka.app/Contents/MacOS/Keka 0x7fff2033a000 - 0x7fff20368fff libsystem_kernel.dylib (7195.60.75) <4bd61365-29af-3234-8002-d989d295fdbb> /usr/lib/system/libsystem_kernel.dylib 0x7fff20369000 - 0x7fff20374fff libsystem_pthread.dylib (454.60.1) <8dd3a0bc-2c92-31e3-bbab-ce923a4342e4> /usr/lib/system/libsystem_pthread.dylib 0x7fff20375000 - 0x7fff203afff7 libdyld.dylib (832.7.1) <2f8a14f5-7cb8-3edd-85ea-7fa960bbc04e> /usr/lib/system/libdyld.dylib 0x7fff20882000 - 0x7fff20ab1e0f com.apple.LaunchServices (1122.11 - 1122.11) <caeec254-68ae-39b5-8452-ec3e1ee8577b> /System/Library/Frameworks/CoreServices.framework/Versions/A/Frameworks/LaunchServices.framework/Versions/A/LaunchServices 0x7fff22c44000 - 0x7fff239a6c6f com.apple.AppKit (6.9 - 2022.20.119) <4cb42914-672d-3af0-a0a5-2209088a3da0> /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit 0x7fff256d3000 - 0x7fff2572efe7 com.apple.HIServices (1.22) <9af2cdd9-8b68-3606-8c9e-1842420acda7> /System/Library/Frameworks/ApplicationServices.framework/Versions/A/Frameworks/HIServices.framework/Versions/A/HIServices 0x7fff286b9000 - 0x7fff289b8ffd com.apple.HIToolbox (2.1.1) <93518490-429f-3e31-8344-15d479c2f4ce> /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox ``` Answers: username_0: I didnt actually have any other sandboxed apps to test, so I just downloaded and ran "fatFileFinder" from here `https://github.com/Ravbug/FatFileFinderCPP/releases/tag/2.2` which appears to be sandboxed - it launches fine from the same folder. username_1: I suppose it will work too for me. In your logs Keka is not able to launch for a sandbox violation. Seems like it maybe has no read permission and the entitlement `com.apple.security.files.user-selected.read-write` might be violating the sandbox. What read permissions does it have in your `Shared files` folder? This is how it looks to me: ``` aone@aONe-Mini ~ % ls -la /System/Volumes/Data/Shared\ Files total 16 drwxr-xr-x 4 aone wheel 128 Apr 19 08:59 . drwxr-xr-x 27 root wheel 864 Apr 19 08:58 .. -rw-r--r--@ 1 aone wheel 6148 Apr 19 08:59 .DS_Store drwxr-xr-x@ 3 aone wheel 96 Apr 15 12:31 Keka.app ``` For what is worth, I've created the `Shared files` folder using Finder (asked admin password) and copied Keka there using Finder too. No Terminal used here. username_1: Maybe you could try to reproduce with [HandBrake](https://handbrake.fr/rotation.php?file=HandBrake-1.3.3.dmg) since it is notarized and most probably hardened, like Keka. That "FatFileFinder" is not notarized so maybe the sandbox checks are not the same. username_1: @username_0 had you the change to try if another notarized app runs there? username_0: Yeah, Handbrake fails in an identical fashion. My permissions in `ls -al` in that directory look identical to yours too. I tried creating a brand new non-remapped directory inside `System/Volumes/Data/` as well, using Finder, and copied Keka and Handbrake there as well, and the results are the same, with the same sandbox error in the console. So it doesn't seem to be related to Synthetic.conf. username_1: I've replicated the issue now. So it seems this is the way to go from Big Sur. You can use `/Users/Shared` instead, can be accessed by any user and it hasn't this sandbox restriction.
typestyle/typestyle
190269238
Title: Share typings between all cssinjs libs Question: username_0: Hi, I am author of [jss](https://github.com/cssinjs/jss/). I like the idea of having typings for styles a lot. I wonder if you are interested to make it available for all cssinjs libs? Answers: username_1: Sorry for the late reply. I've let it sit in my brain and can't commit to it. Typestyle's type definitions are slightly bound to its internal implementation e.g. its color helpers http://typestyle.io/#/colors need `ensureString` https://github.com/typestyle/typestyle/blob/a58b0dbee8c5050a4567881a3e58c8c5a62e777b/src/index.ts#L32-L36 and the `csx/flex` flexbox uses the vendor prefixing as supported by `freestyle` whereas many other frameworks prefer *automatic* vendor prefixing. Thanks for your interest though :rose: Status: Issue closed
ingydotnet/testml-pm6
335924679
Title: TestML fails to install with zef Question: username_0: docker run --entrypoint /bin/sh -it rakudo-star:latest # zef install TestML ===> Searching for: TestML ===> Updated cpan mirror: https://raw.githubusercontent.com/ugexe/Perl6-ecosystems/master/cpan.json ===> Updated p6c mirror: http://ecosystem-api.p6c.org/projects.json ===> Fetching [FAIL]: TestML:ver<0.2.0>:auth<github:ingydotnet> from git://github.com/ingydotnet/testml-pm6/archive/0.2.0.zip.git Aborting due to fetch failure: TestML:ver<0.2.0>:auth<github:ingydotnet> (use --force-fetch to override) in code at /usr/share/perl6/site/sources/7926F4F3ED4C81AA5DA2A54C8AE1E03D03424CCE (Zef::Client) line 227 in method fetch at /usr/share/perl6/site/sources/7926F4F3ED4C81AA5DA2A54C8AE1E03D03424CCE (Zef::Client) line 197 in method fetch at /usr/share/perl6/site/sources/7926F4F3ED4C81AA5DA2A54C8AE1E03D03424CCE (Zef::Client) line 185 in sub MAIN at /usr/share/perl6/site/sources/E4784A2A0FA00D16808817186E95FE74BEF3FE2D (Zef::CLI) line 149 in block <unit> at /usr/share/perl6/site/resources/3065D08F5332CA244672D7F8A05B603F92BB8A7D line 3 in sub MAIN at /usr/share/perl6/site/bin/zef line 2 in block <unit> at /usr/share/perl6/site/bin/zef line 2 Answers: username_0: Maybe it is just an intermittent failure to fetch from Github the way it is specified? FWIW, It installs with this: `zef install https://github.com/ingydotnet/testml-pm6.git` username_1: I ran into this also today while trying to install YAML. It would be best if there were CPAN releases as that seems to be more reliable at this point. username_1: This does appear to be fixed and installable from CPAN now. However, I have reported a new problem with installation on 2019.03 and it is a weird problem.
zyedidia/micro
381992825
Title: Feature request: Mouse support on tty Question: username_0: Hello, I have noticed mouse support works flawlessly on X11 terminals, but not in linux tty with gpm mouse, at least for me. I am using version 1.4.1 of micro. If it's not too much trouble, could you add mouse support for tty? Lots of thanks. Answers: username_1: This please! username_0: If you are in a hurry, [lcxterm](https://gitlab.com/klamonte/lcxterm) might help you. It's a terminal application for the console that translates GPM events into control sequences. username_1: Thanks for the tip. Is this something you have tried? I installed and enabled libgpm support, however micro still does not support the mouse. In fact, I believe the gpm is just for passthrough on remote shells. username_0: Yes, it totally works for me. Make sure that you enabled the GPM service (`systemctl enable gpm.service`, `systemctl start gpm.service`), and that lcxterm is being built with GPM support (`checking for Gpm_Open in -lgpm... yes`).
godotengine/godot
436316760
Title: Index out of size when trying to drag file to scripts panel Question: username_0: **Godot version:** 3.2 2fc2d82 **OS/device including version:** Ubuntu 18.04 **Issue description:** When I try to drag to empty script panel any type of file(gd or tscn) then this error occurs: ``` ERROR: move_child: Index p_pos=-1 out of size (data.children.size() + 1=2) At: scene/main/node.cpp:323. ERROR: set_current_tab: Index p_current=-1 out of size (get_tab_count()=1) At: scene/gui/tab_container.cpp:464. ``` **Steps to reproduce:** https://streamable.com/x7szx Answers: username_1: The video is dead. I can only get ``` scene\gui\tab_container.cpp:552 - Index p_current=0 out of size (get_tab_count()=0) ``` when dragging non-script file on empty script list. username_1: Doesn't happen anymore in 8958e1b Status: Issue closed
liberapay/liberapay.com
602415219
Title: Creators can't pause their patrons' donations to them Question: username_0: I was reminded of this missing feature by <https://mastodon.social/@dansup/104016437070987358>. Answers: username_1: I have some questions for this feature: 1. For how long should donations be paused for? (e.g. a certain duration such as a month or indefinitely until the creator manually re-enables donations) 2. How should the donors be notified if their donations are being paused or resumed? (e.g. notifications, email) 3. What should happen in the case that a patron goes to a project page and tries to donate? Should the donate button be disabled or should the option to donate still be available? Thank you! username_2: My answers are below, For how long should donations be paused for? (e.g. a certain duration such as a month or indefinitely until the creator manually re-enables donations) I personally think it should be until the creator resumes donations again. That was one thing I disliked about Patreon was that I had to keep logging in to pause my upcoming billing cycle. On the other hand, I do see scenarios where creators will pause billing then forget about it completely, and forget to unpause it again, so either way would be fine with me but I prefer pausing until the creator resumes billing again. How should the donors be notified if their donations are being paused or resumed? (e.g. notifications, email) By email, yes. What should happen in the case that a patron goes to a project page and tries to donate? Should the donate button be disabled or should the option to donate still be available? In this case, I think it should still be available to donate because I've had many Patreon supporters still willing to set everything up for when the creator resumes billing again. Perhaps a little text banner on the page could say this creator has temporarily paused billing. You will be charged when they resume billing. username_1: I'm not sure if Patreon does this, but would it make sense to also include the pause feature for a team? Or should I focus on implementing it for individual/shared accounts first? username_0: The Goal page is probably the most appropriate. For teams, you can either disallow pausing, or implement sending a notification to the other members when the team is paused.
nulpoet/mjkey
752743925
Title: 福州哪里有的士发票-福州哪里有的士发票 Question: username_0: 福州哪里有的士发票【徴:ff181一加一⒍⒍⒍】【Q:249⒏一加一357⒌⒋0】学历低加上外貌的原因,很多工作都受限制。最后她找到了一份通讯公司客服的工作,不用见面,只用声音和人打交道。从小乐观的天性,加上后天的积极努力,表妹在单位成了受欢迎的红人。 一些热心的姑妈姨妈总是喜欢牵红线给大 https://github.com/nulpoet/mjkey/issues/785 https://github.com/nulpoet/mjkey/issues/786 https://github.com/nulpoet/mjkey/issues/787
sumpfork/dominiontabs
159853691
Title: Add A3 papersize Question: username_0: Hi, would be possible to add "A3" paper size with no page margin? This would allow to print 16 dividers on each page. Thanks :-) Answers: username_1: The code for this already exists using the stand alone program. The options are: `--papersize A3 --minmargin 0x0` When I run it, I get: - horizontal dividers: 6 rows of 3 cards - vertical dividers: 4 rows of 5 cards (it just fits, almost going to the edge) The work is making this available on the website. The biggest part would be allowing minimum margins to be given as input. An interim step might be adding an A3 option on the page size pull down. username_0: Adding the option to the website would be awesome! :+1: Status: Issue closed username_2: Deployed.
Homebrew/homebrew-core
213580953
Title: `brew upgrade readline` failing with 401 Unauthorized Question: username_0: Please note that these warnings are just used to help the Homebrew maintainers with debugging if you file an issue. If everything you use Homebrew for is working fine: please don't worry and just ignore them. Thanks! Warning: "config" scripts exist outside your system or Homebrew directories. `./configure` scripts often look for *-config scripts to determine if software packages are installed, and what additional flags to use when compiling and linking. Having additional scripts in your path can confuse software installed via Homebrew if the config script overrides a system or Homebrew provided script of the same name. We found the following "config" scripts: /Applications/Wine Devel.app/Contents/Resources/wine/bin/xml2-config /Applications/Wine Devel.app/Contents/Resources/wine/bin/xslt-config Warning: Some installed formula are missing dependencies. You should `brew install` the missing dependencies: brew install gdbm sqlite xvid Run `brew missing` for more details. ``` Status: Issue closed Answers: username_1: Fixed by 77cfba0d7aedeb8f48400c1ff1cfcb362398fae7; thanks for reporting.
daydaychallenge/leetcode-go
615538431
Title: 161. One Edit Distance Question: username_0: #### [161. One Edit Distance](https://leetcode.com/problems/one-edit-distance/) Given two strings ***s\*** and ***t\***, determine if they are both one edit distance apart. **Note:** There are 3 possiblities to satisify one edit distance apart: 1. Insert a character into ***s\*** to get ***t\*** 2. Delete a character from ***s\*** to get ***t\*** 3. Replace a character of ***s\*** to get ***t\*** **Example 1:** ``` Input: s = "ab", t = "acb" Output: true Explanation: We can insert 'c' into s to get t. ``` **Example 2:** ``` Input: s = "cab", t = "ad" Output: false Explanation: We cannot get t from s by only one step. ``` **Example 3:** ``` Input: s = "1203", t = "1213" Output: true Explanation: We can replace '0' with '1' to get t. ```
latex3/latex3
678972310
Title: [doc] xfp: possible typo, "x >? y" should be "x ? y" Question: username_0: from https://tex.stackexchange.com/q/558606 Related source lines: https://github.com/latex3/latex3/blob/8e237df69c46e0625f352e4c9d6b13189008e698/l3packages/xfp/xfp.dtx#L90-L92 From related documentation for floating point expressions in `texdoc interface3`, it seems the ```tex $x\mathop{\mathtt{>?}}y$ ``` in line 91 should be ```tex $x\mathop{\mathtt{?}}y$ ``` which compares if `x` and `y` are not ordered (for example, one of them is NaN). Answers: username_1: Not a bug. These operators are composable. Status: Issue closed username_2: We could actually allow x??y if people find this clearer than x!<=>y. The reason not to allow x?y is the ambiguity with the ternary operator x?y:z. username_1: @username_2 Why does `?` exist in the first place? Isn't the `nan` case covered by `x != x`?
code4craft/webmagic
205817496
Title: 项目依赖Commons-Collections3.2.1存在反序列化漏洞隐患 Question: username_0: 目前webmagic项目中依赖使用的Commons-Collections3.2.1存在严重的反序列化漏洞隐患,建议提高一个小版本到Commons-Collections3.2.2即可 Answers: username_1: 附上官方jira: [https://issues.apache.org/jira/browse/COLLECTIONS-580](https://issues.apache.org/jira/browse/COLLECTIONS-580) WebMagic本身没有使用反序列化API,不过考虑到适用方,还是升级了。 Status: Issue closed
CocoaPods/CocoaPods
381022912
Title: PLCrashReporter-DynamicFramework can't build with Xcode 10 Question: username_0: Hi man, I'm using your PLCrashReporter-DynamicFramework '~> 1.3.0.1' in my iOS app; But it cannot build with Xcode10, the segment registers are 16 bit, Zeroing out the other 48 bits in case client code depends on it. So can you update code from [here](https://github.com/AbletonAppDev/plcrashreporter) and supply a new version asap? Answers: username_1: Please file an issue with the corresponding repo. This is about the CocoaPods library. https://github.com/plausiblelabs/plcrashreporter Status: Issue closed
agusibrahim/Aplikasi-PPOB-Xamarin
440604752
Title: Fix System.FormatException in JsonSerializerInternalReader.EnsureType (Newtonsoft.Json.JsonReader reader, System.Object value, System.Globalization.CultureInfo culture, Newtonsoft.Json.Serialization.JsonContract contract, System.Type targetType) Question: username_0: ### Version 1.0.1(1) ### ### Stacktrace ### ### Reason ### System.FormatException ### Link to App Center ### * [https://appcenter.ms/users/username_0/apps/Retross-Android/crashes/errors/2625443542u](https://appcenter.ms/users/username_0/apps/Retross-Android/crashes/errors/2625443542u)
psf/black
651950220
Title: please address these concerns Question: username_0: I would like @ambv to address the accessibility concerns asked for in this screenshot. I am not making any claims about the specific single quotes issue at hand but my claim is that you were very disrespectful for not even acknowledging @username_1 's statement. This does not align with principles outlined in the Python [code of conduct](https://www.python.org/psf/conduct/). @username_1 I am sorry that your voice was not heard. <img width="802" alt="Screen Shot 2020-07-06 at 7 33 54 PM" src="https://user-images.githubusercontent.com/6046841/86693835-e3385500-bfbf-11ea-84e6-357d8677e5b5.png"> Answers: username_1: @username_0 Hello there! While I appreciate your concern, I believe as long as `-S, --skip-string-normalization` exists then all is fair. The CoC was most definitely not broken as the situation is just a differing of opinions. Also, I'm not sure if you're aware but the use of the word "voice" nowadays is often the prelude to a certain series of events, of which I want no part of 😉 Could you please close this? username_0: Sure, sorry I just wanted to make sure you weren’t left without a solution. I would still personally encourage @ambv to rethink how he handled that situation. Status: Issue closed
andriirogulin/ARSlidingPanel
246283108
Title: How to implement into a tabbarcontroller without hiding or pushing up the tabbar? Question: username_0: When I try to implement I have 2 options... overlap or push the tabbar to stay up to the swipablezone. I would like to have the swipablezone up to the tabbar without hiding it Answers: username_1: Hi @username_0, have you solved this? username_0: Nope, I quit unfortunately username_1: Okay thanks, one more thing, do you know any alternative libraries? I’m currently using this one but it’s unupdated long time ago already. Thanks! G
pandas-dev/pandas
640991515
Title: 1.0 as boolean dtype cause ValueError in pd.read_csv() Question: username_0: ``` Is this intended? I am using Windows 10, Python 3.7.6 and pandas 1.0.4 Answers: username_1: ``` Is this intended? I am using Windows 10, Python 3.7.6 and pandas 1.0.4 username_0: @username_1 I already made a proper issue for this in the pandas repo. Sorry, for opening it in pandas2 and thanks for moving it. Status: Issue closed
epics-base/pvDatabaseCPP
560289714
Title: pvRecord destructor throws when clients are still connected Question: username_0: Any attempt to destruct a pvRecord instance while there are clients (or listeners) still connected throws a bad_weak_ptr exception. When an instance of a pvRecord derived structure is being destructed (e.g. when the last user of a shared_ptr to it goes away), the managed object (pvRecord) destructor is being run. It calls `notifyClients()`, and trying to remove an existing client, the call to `client->detach(shared_from_this());` uses the `shared_from_this()` mechanism, which holds a weak_ptr inside the managed object to be able to find and join an existing shared_ptr from the managed object. In this case - being run as the shared_ptr being destructed - that weak_ptr is not valid anymore, and shared_from_this() thows a bad_weak_ptr exception. ``` terminate called after throwing an instance of 'std::bad_weak_ptr' what(): bad_weak_ptr ==53188== ==53188== Process terminating with default action of signal 6 (SIGABRT) ==53188== at 0x5270081: raise (raise.c:51) ==53188== by 0x525B534: abort (abort.c:79) ==53188== by 0x50DA692: ??? (in /usr/lib/x86_64-linux-gnu/libstdc++.so.6.0.28) ==53188== by 0x50E6035: ??? (in /usr/lib/x86_64-linux-gnu/libstdc++.so.6.0.28) ==53188== by 0x50E5138: ??? (in /usr/lib/x86_64-linux-gnu/libstdc++.so.6.0.28) ==53188== by 0x50E5A63: __gxx_personality_v0 (in /usr/lib/x86_64-linux-gnu/libstdc++.so.6.0.28) ==53188== by 0x522C7B2: ??? (in /lib/x86_64-linux-gnu/libgcc_s.so.1) ==53188== by 0x522D015: _Unwind_Resume (in /lib/x86_64-linux-gnu/libgcc_s.so.1) ==53188== by 0x48F725A: epics::pvDatabase::PVRecord::notifyClients() (pvRecord.cpp:82) ==53188== by 0x48F72E0: epics::pvDatabase::PVRecord::~PVRecord() (pvRecord.cpp:105) ==53188== by 0x11407D: Record::~Record() (pva-one-pps.cpp:49) ==53188== by 0x114099: Record::~Record() (pva-one-pps.cpp:49) ``` Example application attached. Removing the sleep() before the `pvrecord` shared_ptr goes out of scope will show the behavior. [pva-test-loop.tar.gz](https://github.com/epics-base/pvDatabaseCPP/files/4159019/pva-test-loop.tar.gz) Answers: username_1: Without the reset and additional sleep there were weak pointer exceptions. I think this was caused by a client disconnect together with removeRecord. I hate using the additional sleep but I see no way to have the provider disconnect all channels and not return until all channels are removed. username_0: Obviously a misunderstanding. This issue is not about any past version of pvDatabaseCPP. It is about an obvious bug in the destructor of pvRecord. The pvRecord destructor uses `shared_from_this()`, meaning that any pvRecord must be used in the context of a shared_ptr. (Else the destructor will throw a bad_weak_ptr exception when `shared_from_this()` is called.) However, at the moment when the pvRecord destructor is run in the context of a pvRecord shared_ptr, that shared_ptr is already being destructed, so the call to `shared_from_this()` will throw a bad_weak_ptr exception. No matter how, the pvRecord destructor will *always* throw a bad_weak_ptr exception if any clients or listeners are still connected to the pvRecord. username_1: The problem was involved with removing a PVRecord. There are two methods for removing a record: PVDatabase::removeRecord and PVRecord::remove. PVRecord::remove first calls PVDatabase::removeRecord and the calls all attached client to notify them that the record is being removed. This did work OK. PVDatabase has a std::map that has a shared_ptr to each PVRecord. PVDatabase::removeRecord just removes shared_ptr to the PVRecord from the std::map. Normally nothing besides PVDatabase has a shared_ptr to PVRecord. So when it is removed from the map the PVRecord destructor is called. The destructor for PVRecord calls any attached clients to notify that the record is going away. But since it was doing this from the PVRecord destructor the bad weak point exception occurs. A new branch issue53 fixes the problem. What now happens is that PVDatabase::removeRecord optionally calls PVRecord::remove It does NOT call it if PVDatabase::removeRecord is called by PVRecord::remove. If PVRecord::remove is called first is calls PVDatabase::removeRecord asking not to be called back. ALL destructors in pvDatabaseCPP are now trivial, i.e. at most that have a debug statement. username_1: fixed by merge request #54 Status: Issue closed
GoogleContainerTools/skaffold
803857312
Title: Create an application Log event endpoint Question: username_0: 1. Create application log event endpoint. 2. Create an application log event as per the definition ```protobuf message ApplicationLogEvent { string containerName = 1; // container that the log came from string podName = 2; // pod that the log came from string message = 3; // contents of the log } ``` 3. All application logs will be available via this end point.. Answers: username_1: starting on this Status: Issue closed
vueuse/vueuse
952954838
Title: @ Symbol conflicts the Typescript src import Question: username_0: Please am using VUE3 with typescript. And as standard the `tsconfig.json` file has an src directory import denoted as @. ```JSON { "compilerOptions": { ... "paths": { "@/*": [ "src/*" ] }, } } ``` And the import for This vueuse is. `import { usePackage} from '@vueuse/<package>'` But doing so. Would give error. **Module not found: Error: Can't resolve '@vueuse/\<package\>' in 'VUE\project\src'** I believe it's a conflict due to the **@** symbol and the name of the module. Please is there a fix on how to import this library on typescript Vue3? Thanks 🙏 Status: Issue closed Answers: username_1: I don't think it's related. There are hundreds of packages that starts with @ scopes. Guess there might be some misconfigure in your project. username_2: @username_0 You can check out https://github.com/vuestorefront/shopware-pwa/issues/1541 or https://github.com/vuejs/vue-cli/issues/1198 That might help you fix your tsconfig for this current issue.
LoneGazebo/Community-Patch-DLL
255723941
Title: AI not spending faith Question: username_0: this is lisbon. the portuguese don't have a their own religion, so they're not spending on missionaries. don't spend their faith on anything else either though: ![image](https://user-images.githubusercontent.com/1358966/30131758-424119a2-934e-11e7-8daf-6de0c00263bd.png) (they could go for great people as well ...) Answers: username_1: Religion log - are they saving up for something ? G Sent from my iPhone > username_0: ok, the log says 261, 1966, Portugal, Saving up for a Prophet, as we need to convert Non-Puppet Cities. 261, 1966, Portugal, Saving up for a Prophet, as we need to convert Non-Puppet Cities., Faith: 9607 so it seems they do have a holy city, but have been overwhelmed by the indonesians. maybe a special case. pretty sure they could have bought a prophet for far less, but since they don't have a city with the right religion, they're stuck. username_1: Yeah it is tough to tell the AI when to give up on an owned religion. Sent from my iPhone > username_0: took a look at the code but i don't understand it ... when you buy a prophet, which religion does he adhere to? username_1: your 'currentreligion' which is always your founder religion if you've founded one and still own the holy city. Status: Issue closed
ionide/ionide-atom-fsharp
303499549
Title: HTMLElement.rootElement is deprecated. Question: username_0: The contents of `atom-text-editor` elements are no longer encapsulated within a shadow DOM boundary. Please, stop using `rootElement` and access the editor contents directly instead. ``` HTMLElement.rootElement (C:\Users\MadsK\AppData\Local\atom\app-1.24.0\resources\app.asar\src\text-editor-element.js:23:10) n.Tag (C:\Users\MadsK\.atom\packages\ionide-fsharp\lib\fsharp.js:9300:16) Option__Map$Element__HTMLElement_Element__HTMLElement_ (C:\Users\MadsK\.atom\packages\ionide-fsharp\lib\fsharp.js:7058:37) ViewsHelpers__getElementsByClass$Element_Element_ (C:\Users\MadsK\.atom\packages\ionide-fsharp\lib\fsharp.js:9298:9) TooltipHandler__initialize$ (C:\Users\MadsK\.atom\packages\ionide-fsharp\lib\fsharp.js:8877:11) <unknown> (C:\Users\MadsK\.atom\packages\ionide-fsharp\lib\fsharp.js:8803:14) ```
microsoft/Windows-Containers
630223463
Title: WAC-Create new image - Web applications from WebDeploy Question: username_0: To support customers who currently have a web application deployed on an IIS web server, the Create New Image functionality on the containers extension on WAC should allow customers to containerize an application that was exported from IIS with WebDeploy. Answers: username_0: This feature is now in preview. Please add bugs as new issues. Status: Issue closed
weavejester/clojure-toolbox.com
486839698
Title: Missing project Question: username_0: Only to find a 404, with not just that project missing, but also the user under which the project is hosted. Just checked on Clojars and the project is there: https://clojars.org/clj-sql-up Answers: username_1: @username_2 I think this issue can be closed, as the reference to `clj-sql-up` was removed by this commit : https://github.com/username_2/clojure-toolbox.com/commit/666019bbfb2583c201ae9b39841ab96a5c317b60 Status: Issue closed
accordproject/techdocs
697940245
Title: AP ESLint Config Question: username_0: # Feature Request 🛍️ Create an AP ESLint configuration to be shared among our repositories ## Use Case We should have our own ESLint configuration to help keep code across our ecosystem consistent and approachable ## Possible Solution This should run during tests to enforce linting ## Context Use [`@clausehq/eslint-config`](https://www.npmjs.com/package/@clausehq/eslint-config) as a starting point reference Answers: username_1: We may need different linting rules for core part of the stack and web components username_1: Somewhat different issue, in the area of code checks: - license checks on all project username_0: Yes maybe we could have a core stack config and a ui stack config? username_1: I'm not super versed in linter configurations, but one difference is the JavaScript style where the core stack uses CommonJS "requires" v the web components using the ES6 style. Also the standard indentation in the core stack is 4 whitespace rather than 2 in the web components (I wouldn't mind getting rid of that difference though). username_2: Hey, I'm Interested in this :) I'm a newcomer here, I was just able to close this simple issue :) #341 I want to discuss this issue a little bit, I had joined Slack today, can you tell me to whom should I approach and discuss this issue there? username_0: Happy to discuss this more @username_2, I know others can also help: @DianaLease @username_3 username_0: I think we should use the [`@clauseHQ/eslint-config`](https://github.com/clauseHQ/eslint-config) ESLint configuration as a starting point. But I think we should be making our own configuration repository. One for “core” and one for “ui” to start with. So, `@accordproject/eslint-core` and `@accordproject/eslint-ui`. We could make this a monorepo (`accordproject/linting`?), which would make future linting configurations easier to add. @username_1 as a side note, this makes me think we may want to think about having a more structured naming convention for repositories and projects. username_2: and then we can discuss it and move on to further monorepo creation and so on... username_2: Hello @username_0 , I had created a `.eslintrc.json` file [here](https://gist.github.com/username_2/292f886a670af2e88dfebce822fba398) as a sample, So that we can discuss some more rules that we should add in it... For convenience I'm pasting that file below... ``` { "extends": [ "@clausehq/eslint-config", "eslint:recommended", "plugin:import/errors", "plugin:react/recommended", "plugin:jsx-a11y/recommended" ], "rules": { "react/prop-types": 0, "indent": ["error", 2], "linebreak-style":1, "require-jsdoc":0 }, "parser": "babel-eslint", "plugins": [ "react", "import", "jsx-a11y" ], "parserOptions": { "ecmaVersion": 2021, "sourceType": "module", "ecmaFeatures": { "jsx": true } }, "env": { "es6": true, "browser": true, "node": true }, "settings": { "import/resolver": { "node": { "paths": ["src"], "extensions": [".js", ".jsx", ".ts", ".tsx"] } }, "react": { "version": "detect" } } } ``` 1. The import plugin helps ESLint catch commons bugs around imports, exports, and modules in general 2. jsx-a11y catches many bugs around accessibility that can accidentally arise using React, like not having an alt attribute on an img tag. 3. react is mostly common React things, like making sure you import React anywhere you use React. babel-lint allows ESLint to use the same transpiling library, Babel, that Parcel uses under the hood. Without it, ESLint can't understand JSX. 4. eslint-plugin-react now requires you to inform of it what version of React you're using. We're telling it here to look at the package.json to figure it out. 5. Also as you had given me the starting point [@clausehq/eslint-config](https://github.com/clauseHQ/eslint-config) I had also extended that. 6. I had added the indentation rule of 2 whitespaces as suggested by @username_1 , we can add other rule for core-stack in different file once the monorepo is created... @DianaLease & @username_3 , I would like to know your views and opinions on this as you were the contributors in [@clausehq/eslint-config](https://github.com/clauseHQ/eslint-config)... username_3: Thanks @username_2 for your efforts here. @username_0 why don't we just adopt an existing lining ruleset? There are plenty to choose from and we avoid the overhead of maintaining our own. username_0: I'm going off the assumption that @username_1 is correct in that we [need two separate linters](https://github.com/accordproject/techdocs/issues/308#issuecomment-690526639). username_2: Okay @username_0 , I didn't make two files before because I was waiting for your views on this [file](https://gist.github.com/username_2/292f886a670af2e88dfebce822fba398)...will make other file soon :) username_0: @username_2 I think we'll want more input from @username_1 on this before moving forward. username_2: Yeah sure! Waiting for your reply @username_1 😅 username_2: @username_0 any update on this? I had created two sample files... [Core stack](https://gist.github.com/username_2/864a79cea1a078616941f69e11735b31) [Web Components](https://gist.github.com/username_2/292f886a670af2e88dfebce822fba398) Looking forward to having your input on this... username_2: Also the standard indentation in the core stack is 4 whitespace rather than 2 in the web components (I wouldn't mind getting rid of that difference though). As suggested by Jerome, I had added these rules in those files...Also you may get a little information in this [comment](https://github.com/accordproject/techdocs/issues/308#issuecomment-787502875) about various plugins I had added Looking forward to having your suggestions on adding some specific rules which I missed adding and then we can move on creating Accords ESlint Config repo and further npm publishing work... username_3: Can't we take an off-the-shelf linter for each pattern? I still don't understand the motivation for creating one or more custom house styles username_4: hey can i get to work on this issue? username_0: @username_4 I suggest following @username_3's suggestion and implement an existing linter.
sveltejs/sapper
671678879
Title: Add ability to set response attributes on a per-route basis Question: username_0: There's a desire to set certain response attributes on a per-route basis. E.g. the `Cache-Control` header, other headers, response type, etc. A few ideas: * Handle it with middleware * Export some setting in a page or layout or call some function there * Do it in a config file * Expose the router and allow configuring it Answers: username_1: My knee-jerk reaction here would be something along the lines of the second option above, where we have a new module-level export in the component, which is passed the `req`/`res` objects when doing a server rendering of that route. username_2: Being able to affect the request and response before/after render would be a powerful tool. Might be good to provide a single "around" hook to this effect, or maybe just two different hooks: "before" and "after". Though I think "around" would be better. username_3: How about allowing this in `preload`? So either exposing a `this.setHeader`, as in ``` export function preload(page, session) { this.setHeader('Cache-Control', 'max-age=3600'); } ``` or potentially just exposing the entire response, i.e. ``` export function preload(page, session) { this.res.setHeader('Cache-Control', 'max-age=3600'); } ``` What is the use case for working doing something to request/response *after* the request? username_0: `preload` might not be the right place to do it because it runs on the server and client One example @username_2 gave elsewhere was timing how long `preload` took. He wanted to instrument that call and so having a before/after hook would be useful for that. I imagine he might be wanting to do the same thing here of instrumenting how long it takes the entire request to be handled
ant-design/ant-design
446068865
Title: Modal组件和Drawer组件组合使用导致页面滚动条错误 Question: username_0: - [ ] I have searched the [issues](https://github.com/ant-design/ant-design/issues) of this repository and believe that this is not a duplicate. ### Reproduction link [![Edit on CodeSandbox](https://codesandbox.io/static/img/play-codesandbox.svg)](https://codesandbox.io/s/495zk) ### Steps to reproduce 1、点击打开抽屉 2、接着打开Modal框 3、接着关闭Modal框 ### What is expected? 点击打开抽屉,这时底层div无法滚动。打开Modal框接着关闭Modal框,底层div不出现滚动条,不能滚动 ### What is actually happening? 点击打开抽屉,这时底层div无法滚动。打开Modal框接着关闭Modal框,底层div出现滚动条,能滚动。 | Environment | Info | |---|---| | antd | 3.18.2 | | React | react 16.8.6 | | System | windows 10 | | Browser | Google Chrome 版本 74.0.3729.131(正式版本) (32 位) | <!-- generated by ant-design-issue-helper. DO NOT REMOVE --> Answers: username_1: 重复了,看 #17323 ,,,这个先关闭了.... Status: Issue closed username_2: is it fixed? username_1: 见 17323
ossrs/srs
274773740
Title: Docker CI Question: username_0: [https://github.com/username_0/docker-srs](https://github.com/username_0/docker-srs) [https://hub.docker.com/r/pm9551/docker-srs](https://hub.docker.com/r/pm9551/docker-srs) [http://hub.daocloud.io/repos/638f16fa-9220-4c72-94fa-f43aff69c447](http://hub.daocloud.io/repos/638f16fa-9220-4c72-94fa-f43aff69c447) I hope that docker CI can be provided by official.
wireservice/csvkit
232884751
Title: csvcut columns ranges Question: username_0: Are there any plans to add support for column ranges in csvcut? Currently you have to explicitly specify a list of columns ad indicated in the docs. Would be great if an option for a range of columns can be added. -c COLUMNS, --columns COLUMNS A comma separated list of column indices or names to be extracted. Defaults to all columns. Answers: username_1: Have you tried it or did you just rely on the docs? I experienced that you can give a range for csvcut and it will work. E. g.: `csvcut -c 1-13 data.csv` Status: Issue closed username_2: Indeed. I've updated the `--help` message to describe ranges.
AprilRobotics/apriltag
580623614
Title: Missing target Threads::Threads Question: username_0: Hey there, I was just about to use the apriltag library in a cmake project and encountered the following error message: ``` CMake Error at CMakeLists.txt:10 (add_library): Target "apriltag-test" links to target "Threads::Threads" but the target was not found. Perhaps a find_package() call is missing for an IMPORTED target, or an ALIAS target is missing? ``` Shouldn't the `apriltagConfig.cmake` handle finding the Threads library if it's a public dependency? A minimal example can be found [here](https://github.com/username_0/apriltag-cmake-configuration-failure-example). Cheers, Stefan Status: Issue closed Answers: username_1: Thanks for the example! username_2: Note: The change in 35f4ef2 adds a namespace to the exported target: https://github.com/AprilRobotics/apriltag/blob/35f4ef2825745a08ec4690ce8c15f8dc903c2868/CMakeLists.txt#L54-L62 which breaks projects that import that target via `target_link_libraries(${PROJECT_NAME} apriltag)`. To include that namespace, this has to be changed to `target_link_libraries(${PROJECT_NAME} apriltag::apriltag)`. username_3: @username_2 how to install apriltags so that ROS2 wrapper finds it? ``` --- stderr: apriltag_ros CMake Error at CMakeLists.txt:24 (add_library): Target "AprilTagNode" links to target "apriltag::apriltag" but the target was not found. Perhaps a find_package() call is missing for an IMPORTED target, or an ALIAS target is missing? --- Failed <<< apriltag_ros [1.67s, exited with code 1] username_4: I believe target `apriltag::apriltag` does currently not exist. Can you try linking to `apriltag`? We may need to create an ALIAS. username_2: @username_3 What did you try so far? Ideally, `rosdep` resolves all the binary dependencies: `rosdep install --from-paths src -i`. But I think since the namespace was added, there was no new binary release. So you have to clone https://github.com/AprilRobotics/apriltag manually into your workspace and build everything from source. username_3: @username_2 Thank you for your response! So far I tried: 1. cloning both https://github.com/AprilRobotics/apriltag https://github.com/AprilRobotics/apriltag-ros to catkin_ws/src ``` :~/catkin_ws_dashing$ cd apriltag/ @linux:~/catkin_ws_dashing/apriltag$ mkdir build @linux:~/catkin_ws_dashing/apriltag$ cd build @linux:~/catkin_ws_dashing/apriltag/build$ cmake .. @linux:~/catkin_ws_dashing/apriltag/build$ make -j8 @linux:~/catkin_ws_dashing/apriltag/build$ sudo make install ... -- Installing: /usr/local/include/apriltag/tagStandard41h12.h -- Installing: /usr/local/include/apriltag/tagStandard52h13.h -- Installing: /usr/local/lib/libapriltag.so.3.1.0 -- Up-to-date: /usr/local/lib/libapriltag.so.3 -- Up-to-date: /usr/local/lib/libapriltag.so ``` then I would try to build the dashing workspace that has apriltag_msgs apriltag_ros cloned from https://github.com/username_2/apriltag_ros colcon build Starting >>> apriltag Starting >>> apriltag_msgs --- stderr: apriltag CMake Error: The source directory "/home/agx002/catkin_ws_dashing/src/apriltag" does not exist. Specify --help for usage, or press the help button on the CMake GUI. make: *** [cmake_check_build_system] Error 1 --- Failed <<< apriltag [0.47s, exited with code 2] Aborted <<< realsense2_camera_msgs [0.71s] Aborted <<< apriltag_msgs [0.69s] Summary: 0 packages finished [1.22s] 1 package failed: apriltag 1 package had stderr output: apriltag 3 packages not processed ``` as apriltags_msgs requires the apriltag folder in the src I moved it to there ``` but now apriltag_ros won't find ``` Starting >>> apriltag_ros --- stderr: apriltag_ros CMake Error at CMakeLists.txt:24 (add_library): Target "AprilTagNode" links to target "apriltag::apriltag" but the target was not found. Perhaps a find_package() call is missing for an IMPORTED target, or an ALIAS target is missing? --- Failed <<< apriltag_ros [1.59s, exited with code 1] Summary: 5 packages finished [12.9s] 1 package failed: apriltag_ros 1 package had stderr output: apriltag_ros ``` username_3: another error that comes out if to delete all other packages from src but apriltag_ros/ apriltag ``` - stderr: apriltag_ros In file included from /home/agx002/catkin_ws_dashing/src/apriltag_ros/src/AprilTagNode.cpp:1:0: /home/agx002/catkin_ws_dashing/src/apriltag_ros/include/AprilTagNode.hpp:11:10: fatal error: image_transport/camera_subscriber.hpp: No such file or directory #include <image_transport/camera_subscriber.hpp> ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ``` ``` locate camera_subscriber /opt/ros/dashing/include/image_transport/camera_subscriber.h /opt/ros/melodic/include/image_transport/camera_subscriber.h ``` there seem no hpp file username_2: It looks like you mixed up your workspace with different ROS distributions (`melodic` is ROS1, `dashing` is ROS2). There is an independent library called `apriltag` at https://github.com/AprilRobotics/apriltag, which works without ROS. The ROS1 node is at https://github.com/AprilRobotics/apriltag_ros, and the ROS2 node is at https://github.com/username_2/apriltag_ros. Both ROS nodes are called `apriltag_ros` as they live in different groups and are supposed to be used in separate workspaces, e.g. one for ROS1 and one for ROS2. You should not mix both nodes in the same workspace. If you want to build the ROS2 workspace, you need the library from https://github.com/AprilRobotics/apriltag, the message definitions from https://github.com/username_2/apriltag_msgs and the ROS2 node from https://github.com/username_2/apriltag_ros. username_3: @username_2 Thank you for following up! Exactly! ROS1 works fine with default apriltag_ros from https://github.com/AprilRobotics/apriltag_ros also with https://github.com/AprilRobotics/apriltag But ROS2 Dashing seems have some issues. Look: ``` linux:~$ mkdir ws_dashing @linux:~$ cd ws_dashing @linux:~/ws_dashing$ mkdir src @linux:~/ws_dashing$ cd src @linux:~/ws_dashing/src$ @linux:~/ws_dashing/src$ git clone https://github.com/username_2/apriltag_ros @linux:~/ws_dashing/src$ git clone https://github.com/username_2/apriltag_msgs Unpacking objects: 100% (35/35), done. @linux:~/ws_dashing/src$ cd .. @linux:~/ws_dashing$ git clone https://github.com/AprilRobotics/apriltag Cloning into 'apriltag'... @linux:~/ws_dashing$ cd apriltag/ @linux:~/ws_dashing/apriltag$ mkdir build @linux:~/ws_dashing/apriltag$ cd build @linux:~/ws_dashing/apriltag/build$ cmake .. -- The C compiler identification is GNU 7.5.0-- Generating done -- Build files have been written to: /home/void/ws_dashing/apriltag/build @linux:~/ws_dashing/apriltag/build$ sudo make install ``` at this point it should work as the mixing of ROS yet did not occur, right? but the output is as follows: ``` ~/ws_dashing$ colcon build Starting >>> apriltag Starting >>> apriltag_msgs Finished <<< apriltag [14.5s] Finished <<< apriltag_msgs [18.0s] Starting >>> apriltag_ros --- stderr: apriltag_ros In file included from /home/ws_dashing/src/apriltag_ros/src/AprilTagNode.cpp:1:0: /home/ws_dashing/src/apriltag_ros/include/AprilTagNode.hpp:11:10: fatal error: image_transport/camera_subscriber.hpp: No such file or directory #include <image_transport/camera_subscriber.hpp> ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ compilation terminated. make[2]: *** [CMakeFiles/AprilTagNode.dir/src/AprilTagNode.cpp.o] Error 1 make[1]: *** [CMakeFiles/AprilTagNode.dir/all] Error 2 make: *** [all] Error 2 --- Failed <<< apriltag_ros [4.89s, exited with code 2] Summary: 2 packages finished [23.3s] 1 package failed: apriltag_ros 1 package had stderr output: apriltag_ros ``` I tried to look for that specific file but only found ``` /opt/ros/dashing/include/image_transport/camera_subscriber.h ``` username_3: @username_2 resolved with ``` sudo bash -c 'echo "#include \"image_transport.h\"" > /opt/ros/dashing/include/image_transport/image_transport.hpp' sudo bash -c 'echo "#include \"camera_subscriber.h\"" > /opt/ros/dashing/include/image_transport/camera_subscriber.hpp' ``` suggested patch from https://github.com/NVIDIA-AI-IOT/ros2-nvapriltags/issues/7#issuecomment-820750262 username_2: @username_3 I would recommend that you clone and build all packages as part of your ROS2 workspace via colcon. You manually installed the `apriltag` via `make install`, which works because it's a generic CMake package. But if you build and install it this way system-wide, it does not have to be part of your workspace. The issue with the headers that you are facing is resulting from an API change in between ROS2 distributions (`dashing` and `foxy`) See the change at https://github.com/username_2/apriltag_ros/commit/d8d533406995c46d2de30de6401d61c052c6e3ca. On `master` branch is usually compatible with the latest ROS2 release (`foxy` at the moment). If you have to use an older release (e.g. `dashing`), you have to use a different branch, e.g. https://github.com/username_2/apriltag_ros/tree/dashing. username_3: Thank you very much!
ArnoldV/Our.Umbraco.GMaps
495433160
Title: Mapstyle not being set? Question: username_0: Hi there, Awesome package! There seems to be a small bug though regarding mapstyle. I just tried adding JSON for the standard Silver theme from https://mapstyle.withgoogle.com/, but the map style is still the same and MapConfig.Style is empty when being output in the view (the other properties works just fine). Any ideas on how to fix? Here's the mapstyle JSON: `[ { "elementType": "geometry", "stylers": [ { "color": "#f5f5f5" } ] }, { "elementType": "labels.icon", "stylers": [ { "visibility": "off" } ] }, { "elementType": "labels.text.fill", "stylers": [ { "color": "#616161" } ] }, { "elementType": "labels.text.stroke", "stylers": [ { "color": "#f5f5f5" } ] }, { "featureType": "administrative.land_parcel", "elementType": "labels.text.fill", "stylers": [ { "color": "#bdbdbd" } ] }, { "featureType": "poi", "elementType": "geometry", "stylers": [ { "color": "#eeeeee" } ] [Truncated] }, { "featureType": "water", "elementType": "geometry", "stylers": [ { "color": "#c9c9c9" } ] }, { "featureType": "water", "elementType": "labels.text.fill", "stylers": [ { "color": "#9e9e9e" } ] } ]` Answers: username_0: Argh, I just realized that the default settings only apply to items where you havn't already provided an address since setting is being stored per item.. Not a bug then, just a feature I guess :) Status: Issue closed
Iterable/iterable-android-sdk
827437486
Title: Mobile inbox opens as a blank page Question: username_0: I've got the Iterable SDK working nicely otherwise (receiving push notifications and inapp messages), but setting up and opening mobile inbox according to these instructions https://support.iterable.com/hc/en-us/articles/360038744152 ...opens a new activity which shows just a blank white page. Using ``` implementation 'com.iterable:iterableapi:3.2.10' implementation 'com.iterable:iterableapi-ui:3.2.10' ``` Running on Android emulator API 30. Code to open activity: `startActivity(Intent(context, IterableInboxActivity::class.java))` No hints found from logcat. Answers: username_1: Hi @username_0, Did the inapp had `inbox` flag turned on? IterableInboxActivity will only show in-apps which have the checkbox ticked. Otherwise it will be a blank screen. <img width="191" alt="image" src="https://user-images.githubusercontent.com/22951141/111366112-59618100-8650-11eb-88af-0cc6aa06926f.png"> username_0: Yeah, that was it, thanks! Would be a better user experience if the empty state had something in it, maybe some "No messages" text? Also, when I opened the inbox the message list was drawn below status bar. Maybe something I can fix with styling or tweaking AndroidManifest? ![image](https://user-images.githubusercontent.com/1794792/111498558-b8a3bc00-874a-11eb-9bb8-b388d6deddb8.png) Status: Issue closed username_1: @username_0, you can [refer to this doc here](https://support.iterable.com/hc/en-us/articles/360039091471#empty-state) to add title and message to empty inbox screen. Thank you :)
phetsims/rosetta
587139446
Title: Prevent submission of translations with no title Question: username_0: The build server fails if a simulation has no strings, and this case can come up if there is a sim that uses all common strings except for its title. There is an issue logged for this against the build server, see https://github.com/phetsims/perennial/issues/83, but it's been around for over two years without being attended to. I received six emails this week for failed builds due to this problem (all related to the diffusion sim, which is a subset of Gas Properties), and it just made me think that I should prevent it from occurring at all by adding code to rosetta that checks it. Answers: username_0: I ended up adding a check that verifies that at least one sim string has been translated. This has been tested (a little) and deployed. username_0: I've asked @oliver-phet to notify translators that were recently affected by this issue. There isn't much else that I can do to verify it other than make sure that we don't see this type of build failure anymore, but we only get those sporadically and sometimes it can be months between them. I'm going to assume we're good to go here and close it, and will re-open if I see any of the related failures, or anything else that might be related. Status: Issue closed
VSpaceCode/VSpaceCode
641410739
Title: Evaluate integration with vscode-advanced-open-file. Question: username_0: The gif of [this](https://github.com/jit-y/vscode-advanced-open-file) project has a behavior which is very similar to original spacemacs `SPC f f`. You should evaluate this project in order to see if it exposes some actions which may be integrated in our [setting.json](https://github.com/VSpaceCode/VSpaceCode/blob/master/settings.json). Answers: username_1: It works pretty well for me, although there is no `<TAB>` support. the search and enter key works very well. I am imaging the same menu structure could work well for folder opening, and I opened an issue to track that work https://github.com/jit-y/vscode-advanced-open-file/issues/24 Btw, there is a feature request on the vscode to support separate the command to open folder/files. This will be helpful in distinguishing `<spc> p p` and `<spc> f r`, show your support so the feature at least will get to backlog. https://github.com/microsoft/vscode/issues/100421 username_1: Another direction to support file/directory opening would be creating another extension that do something similar to edmagit that uses a buffer than we will have `<TAB>` support. Ultimately, I think the best solution would be somewhere between `vscode-advanced-open-file` and a custom buffer. However, that would require more engineering work. username_1: I took this out of the `v0.5.0` as the extension doesn't have a folder opening feature. This seems heavy for something as simple as opening a file unless the extension also covers other file/folder opening features. There's a lot of edge cases needed to handle like opening local file in the vscode environment. For people who want to use that as your <spc> f f, you can bind it in your overrides. Interesting, VS code remote has file opening prompt that's closer to what we need. username_1: Closing this in favor of #65 Status: Issue closed
christodenny/ordr
482736756
Title: All orders page doesn't show all orders Question: username_0: If there are multiple users, only the last user's items are shown in the All orders page. ![image](https://user-images.githubusercontent.com/9013097/63333814-ea466e00-c2ee-11e9-9176-6fbed986ccba.png) (the previous user had submitted $64.01 worth of items) Answers: username_1: fixed in b75d711fd5bac322ca1fe2c7f837543efd8be5c9 Status: Issue closed
rdn1983/cardiology
358310886
Title: Дневники после КАГ Question: username_0: 1. ![2018-09-08_19-47-06](https://user-images.githubusercontent.com/40735638/45256539-6a0a4f80-b3a0-11e8-8332-4b5b1b8fea4d.png) Answers: username_0: Предлагать "конкретное" время, после КАГ (оно, действительно, через час) username_1: Напишите, пожалуйста, рамки показателей ЧСС, ЧД, АД для трепетания предсердий и синусового ритма. Status: Issue closed username_0: Ранее описано
kubernetes/kubernetes
441788555
Title: Token request: proactively remove unused secret based SA tokens Question: username_0: Once it is possible to opt out of secret based SA tokens, any preexisting secrets should be removed. Action items: 1. Clean up legacy tokens for the controller service accounts after a release /sig auth /sig api-machinery @kubernetes/sig-auth-feature-requests xref: #70679 #71275 #72179 #77599 /priority important-longterm Answers: username_1: /remove-sig api-machinery username_3: We can start considering how we'd enable a service account to indicate it didn't want secret tokens in 1.15, but yes, we would not have the controller-manager switch to using that until at least 1.16 username_4: /remove-lifecycle rotten username_4: /lifecycle frozen username_3: Stopping auto-generating new secret-based tokens and removing existing unused secret-based tokens is being addressed in KEP 2799 (https://github.com/kubernetes/enhancements/pull/2800) /close
tensorflow/models
454975167
Title: Know the tensorflow version from the checkpoint and/or frozen graph(.pb) files Question: username_0: There are a lot of public/opensource models listed here. https://github.com/tensorflow/models/tree/master/research/slim I want to know the version of the tensorflow on which these models were trained on. Answers: username_1: Thank you for your post. We noticed you have not filled out the following field in the issue template. Could you update them if they are relevant in your case, or leave them as N/A? Thanks. What is the top-level directory of the model you are using Have I written custom code OS Platform and Distribution TensorFlow installed from TensorFlow version Bazel version CUDA/cuDNN version GPU model and memory Exact command to reproduce username_1: Hi There, We are checking to see if you still need help on this, as this seems to be an old issue. Please update this issue with the latest information, code snippet to reproduce your issue and error you are seeing. If we don't hear from you in the next 7 days, this issue will be closed automatically. If you don't need help on this issue any more, please consider closing this. Status: Issue closed
geneontology/go-ontology
142346994
Title: TPV sister chromatid biorientation Question: username_0: GO:0031134 sister chromatid biorientation ancestor to GO:0051310 metaphase plate congression should be removed. biorientation and plate congression should be siblings. Biorientation is achecved when all pairs of sister kinetochores are attached to microtubules. This is after metaphase plate congression. Everything else looks OK Answers: username_1: https://github.com/geneontology/go-ontology/issues/12244#issuecomment-173562674 has: mitotic metaphase plate congression . <- po attachment of mitotic spindle microtubules to kinetochore . . <- is_a: NEW: lateral attachment of mitotic spindle microtubules to kinetochore . . <- is_a: sister chromatid biorientation . <- po NEW: microtubule plus-end directed mitotic chromosome migration I thought this is what we agreed on that ticket. It's what I implemented. username_0: my fault I misunderstood. It should be mitotic metaphase plate congression . <- po: NEW: lateral attachment of mitotic spindle microtubules to kinetochore . <- po NEW: microtubule plus-end directed mitotic chromosome migration attachment of mitotic spindle microtubules to kinetochore . . <- is_a: NEW: lateral attachment of mitotic spindle microtubules to kinetochore . . <- is_a: sister chromatid biorientation (this is end-on attachment) so biorientation follows congression, and both have (different parts) which are part of "attachement of spindle microtubules to kinetochore" username_1: Fixed (finally, I hope...) Status: Issue closed username_0: GO:0031134 sister chromatid biorientation ancestor to GO:0051310 metaphase plate congression should be removed. biorientation and plate congression should be siblings. Biorientation is achecved when all pairs of sister kinetochores are attached to microtubules. This is after metaphase plate congression. see http://www.ebi.ac.uk/QuickGO/GTerm?id=GO:0031134#term=ancchart Everything else looks OK username_0: HI David, you are going to swear at me. I realised the next day I had made a b***x here but I thought I'd wait until you got back to resurrect. You were completely correct before. The arrangement for congression should have been mitotic metaphase plate congression . <- po attachment of mitotic spindle microtubules to kinetochore . . <- is_a: NEW: lateral attachment of mitotic spindle microtubules to kinetochore . . <- is_a: sister chromatid biorientation . <- po NEW: microtubule plus-end directed mitotic chromosome migration I was confused because I thought "microtubule attachment" was not part of "localization" the parent of congression. But now I think it probably is (and if it wasn't the localization parent of congression could move down to the localization part). But the attachment is *definitely* part of congression. Sorry about that. Very embarrassed to change it twice. I can hear you muttering.... username_0: It has to be this way: ![confirm](https://cloud.githubusercontent.com/assets/7359272/14272317/f09cff8c-faf9-11e5-8098-651236e331ad.jpg) Congression isn't complete until bipolar attachment is completed. Status: Issue closed username_1: Full changed back (finally!)
TechnionYP5777/SmartCity-Accessibility
217882549
Title: edit the parsing of Queries Question: username_0: since Parse cannot save complex objects, last semester I implemented a parsing to and from String, adding the name field would result in a need to change that parsing #212 Answers: username_0: after last commit, users are dissallowed to name queries as "--This is a String Splitter--" or search for "--This is a String Splitter--" which is...pretty legitamate, but need to enforce this in gui (maybe throw exception from Search as well) username_0: need to make sure gui doesm't allow that inside the string, for the time being an exception is thrown from the class should be enough Status: Issue closed
electron/electron
159711531
Title: Focused elements don't blur when Spotlight or similar windows appear Question: username_0: In non-Electron applications they do. * Electron version: 1.2.2 * Operating system: OS X Status: Issue closed Answers: username_1: This is because unlike native elements, web pages do not have the concept of [key windows](https://developer.apple.com/library/mac/documentation/Cocoa/Conceptual/WinPanel/Concepts/ChangingMainKeyWindow.html), so browsers do not change focused state when the window is not key window anymore but still being a main window. You can reproduce the same behaviors in the web pages opened in Chrome and Safari. We are not able to fix on Electron's side, to fix it we have to make the page keep focused state while do not have a focused element, which is not possible without changes to Chromium and probably web standards. I'm closing this won't fix.
bigcommerce/stencil-cli
345916205
Title: Can't run stencil start due to libsass error Question: username_0: Binary found at /Users/username_0/bigcommerce/cornerstone/node_modules/node-sass/vendor/darwin-x64-51/binding.node Testing binary Binary is fine [email protected] /Users/username_0/bigcommerce/cornerstone/node_modules/node-sass ``` Trying to start stencil and being asked to rebuild node-sass ``` ➜ cornerstone git:(master) ✗ stencil start (node:86424) DeprecationWarning: os.tmpDir() is deprecated. Use os.tmpdir() instead. /Users/username_0/.nvm/versions/node/v7.9.0/lib/node_modules/@bigcommerce/stencil-cli/node_modules/@bigcommerce/node-sass/lib/extensions.js:158 throw new Error([ ^ Error: The `libsass` binding was not found in /Users/username_0/.nvm/versions/node/v7.9.0/lib/node_modules/@bigcommerce/stencil-cli/node_modules/@bigcommerce/node-sass/vendor/darwin-x64-51/binding.node This usually happens because your node version has changed. Run `npm rebuild node-sass` to build the binding for your current node version. at Object.sass.getBinaryPath (/Users/username_0/.nvm/versions/node/v7.9.0/lib/node_modules/@bigcommerce/stencil-cli/node_modules/@bigcommerce/node-sass/lib/extensions.js:158:11) at Object.<anonymous> (/Users/username_0/.nvm/versions/node/v7.9.0/lib/node_modules/@bigcommerce/stencil-cli/node_modules/@bigcommerce/node-sass/lib/index.js:16:36) at Module._compile (module.js:571:32) at Object.Module._extensions..js (module.js:580:10) at Module.load (module.js:488:32) at tryModuleLoad (module.js:447:12) at Function.Module._load (module.js:439:3) at Module.require (module.js:498:17) at require (internal/module.js:20:19) at Object.<anonymous> (/Users/username_0/.nvm/versions/node/v7.9.0/lib/node_modules/@bigcommerce/stencil-cli/node_modules/@bigcommerce/stencil-styles/lib/styles.js:7:14) ``` Answers: username_1: Could you try `npm rebuild @bigcommerce/node-sass`? username_0: ``` ➜ cornerstone git:(master) ✗ npm rebuild @bigcommerce/node-sass -verbose npm info it worked if it ends with ok npm verb cli [ '/Users/username_0/.nvm/versions/node/v7.9.0/bin/node', npm verb cli '/Users/username_0/.nvm/versions/node/v7.9.0/bin/npm', npm verb cli 'rebuild', npm verb cli '@bigcommerce/node-sass', npm verb cli '-verbose' ] npm info using [email protected] npm info using [email protected] npm info readInstalled object npm verb exit [ 0, true ] npm info ok ``` ``` ➜ cornerstone git:(master) ✗ stencil start (node:35394) DeprecationWarning: os.tmpDir() is deprecated. Use os.tmpdir() instead. /Users/username_0/.nvm/versions/node/v7.9.0/lib/node_modules/@bigcommerce/stencil-cli/node_modules/@bigcommerce/node-sass/lib/extensions.js:158 throw new Error([ ^ Error: The `libsass` binding was not found in /Users/username_0/.nvm/versions/node/v7.9.0/lib/node_modules/@bigcommerce/stencil-cli/node_modules/@bigcommerce/node-sass/vendor/darwin-x64-51/binding.node This usually happens because your node version has changed. Run `npm rebuild node-sass` to build the binding for your current node version. at Object.sass.getBinaryPath (/Users/username_0/.nvm/versions/node/v7.9.0/lib/node_modules/@bigcommerce/stencil-cli/node_modules/@bigcommerce/node-sass/lib/extensions.js:158:11) at Object.<anonymous> (/Users/username_0/.nvm/versions/node/v7.9.0/lib/node_modules/@bigcommerce/stencil-cli/node_modules/@bigcommerce/node-sass/lib/index.js:16:36) at Module._compile (module.js:571:32) at Object.Module._extensions..js (module.js:580:10) at Module.load (module.js:488:32) at tryModuleLoad (module.js:447:12) at Function.Module._load (module.js:439:3) at Module.require (module.js:498:17) at require (internal/module.js:20:19) at Object.<anonymous> (/Users/username_0/.nvm/versions/node/v7.9.0/lib/node_modules/@bigcommerce/stencil-cli/node_modules/@bigcommerce/stencil-styles/lib/styles.js:7:14) ``` username_2: This happened to me when I had Node installed globally and NVM to manage versions. For me installing `node 8.x.x` through nvm, then the `stencil-cli`, then stencil start bypasses this issue. username_0: Will try it out... username_1: Good catch @username_0 ! We've updated the required node versions in the official docs, but missed the GitHub README. We'll get this updated. username_0: Screenshot taken just now: ![image](https://user-images.githubusercontent.com/978462/43463989-8be9b6a6-949f-11e8-9933-69e79806d249.png) username_1: Hey @username_0 , yes the documentation in your screenshot is correct. Node version 4.x is deprecated, but 6.x + should be fine. To clarify, stencil CLI will work with node v 8.x, but it isn't a requirement. I think resolving your error was due more to reinstalling stencil CLI and npm dependencies than upgrading to node 8.x, but if you're running into the same error with 7.9.0 after reinstalling CLI/npm dependencies please let me know username_0: Ok, thank you Karen. 8 seems to be running fine for now, but will let you know if I need to go back to 7.9.0 if there are any issues. username_3: I installed everything and tried to start stencil and received this same error. I tried to rebuild it, remove it and reintall it but am still getting the same error. Do I need to uninstall everything and start over or is there an easier way? Can someone clarify the order of installation? username_2: @username_3 What is "likely" happening here is that you have a version of Node installed that's handling `node-sass` and a version of the `stencil-cli` installed that was installed by a different version. The ideal order of things is: - Remove all traces of node on your system. - install NVM (notes about NVM below) - install a version of node, using NVM (I'm currently using `8.16.0`) - now install the stencil-cli globally - now pull down the stencil theme for your store (or clone cornerstone) - type `nvm list` in your console. Make sure you're using `8.16.0` - run npm install - run stencil start NVM stands for Node Version Manager. Think of it like a hidden folder that contains everything about a particular node version. So you'd have a file structure like `./nvm/node_versions/8.1.2` and in that folder there would be everything related to that node version. So when you installed `npm install -g stencil-cli` you'd get the stencil-cli in that folder. If then do `nvm install 8.11.0` you'll be in a separate folder that has no stencil-cli installed. This allows you to manage versions in a more controlled environment and avoids conflicts like you are "most likely" experiencing. *Please note*: I'm not 100% sure that this will fix your problem (although it's worked for basically everyone I've helped through this, say an n of 6) and some of what I explained above is a simple/rough example of how it works, not "actually" how it works. Good luck! username_4: Hey Guys, running into the same problem only I'm running on v10.16.3 `$ stencil start (node:8148) [DEP0022] DeprecationWarning: os.tmpDir() is deprecated. Use os.tmpdir() instead. C:\Users\thoyt\AppData\Roaming\npm\node_modules\@bigcommerce\stencil-cli\node_modules\@bigcommerce\node-sass\lib\extensions.js:158 throw new Error([ ^ Error: The `libsass` binding was not found in C:\Users\thoyt\AppData\Roaming\npm\node_modules\@bigcommerce\stencil-cli\node_modules\@bigcommerce\node-sass\vendor\win32-x64-64\binding.node This usually happens because your node version has changed. Run `npm rebuild node-sass` to build the binding for your current node version. at Object.sass.getBinaryPath (C:\Users\thoyt\AppData\Roaming\npm\node_modules\@bigcommerce\stencil-cli\node_modules\@bigcommerce\node-sass\lib\extensions.js:158:11) at Object.<anonymous> (C:\Users\thoyt\AppData\Roaming\npm\node_modules\@bigcommerce\stencil-cli\node_modules\@bigcommerce\node-sass\lib\index.js:16:36) at Module._compile (internal/modules/cjs/loader.js:778:30) at Object.Module._extensions..js (internal/modules/cjs/loader.js:789:10) at Module.load (internal/modules/cjs/loader.js:653:32) at tryModuleLoad (internal/modules/cjs/loader.js:593:12) at Function.Module._load (internal/modules/cjs/loader.js:585:3) at Module.require (internal/modules/cjs/loader.js:692:17) at require (internal/modules/cjs/helpers.js:25:18) at Object.<anonymous> (C:\Users\thoyt\AppData\Roaming\npm\node_modules\@bigcommerce\stencil-cli\node_modules\@bigcommerce\stencil-styles\lib\styles.js:7:14) at Module._compile (internal/modules/cjs/loader.js:778:30) at Object.Module._extensions..js (internal/modules/cjs/loader.js:789:10) at Module.load (internal/modules/cjs/loader.js:653:32) at tryModuleLoad (internal/modules/cjs/loader.js:593:12) at Function.Module._load (internal/modules/cjs/loader.js:585:3) at Module.require (internal/modules/cjs/loader.js:692:17)' I've tried everything you've instructed. If you could advise. username_4: Using windows npm 10.16.3* username_2: Hi @username_4! I haven't worked on a stencil site in a bit. But at a glance, that looks like the exact issue that most people are encountering when they end up here. My best suggestion, given that you're on widows, is install WSL, then install NVM, then use NVM to install a node version (10+ should be good), then from there you should be fresh. The following line makes me think that your attempts at uninstalling node may have been in vain. ``` C:\Users\thoyt\AppData\Roaming\npm\node_modules@bigcommerce\stencil-cli\node_modules@bigcommerce\node-sass\vendor\win32-x64-64\binding.node ``` Step one is vital, completely uninstall any trace of node on your system. Then install NVM and get node through that path. Then when you install a global package (like `stencil-cli`) you'll be able to find it's files in an `.nvm` directory. Here's a helpful link that doesn't really explain why, but sort of shows where things "should" be installed. https://flaviocopes.com/where-npm-install-packages/ username_5: This is over 6 months old so I am going to close it. I would recommend you please try the latest 3.x version and please open a new issue if you run into any issues. Status: Issue closed
facebook/react-native
107458260
Title: textInput bugs with secureTextEntry value Question: username_0: I found a interesting thing.When I use the TextInput with secureTextEntry property.If <TextInput placeholder='<PASSWORD>' secureTextEntry={true} /> both Android and IOS works fine.But When I use it like this <TextInput placeholder='<PASSWORD>' secureTextEntry='true' /> the Android version just go wrong.But the IOS version works fine. ![image](https://cloud.githubusercontent.com/assets/5960165/9987008/884f95c6-6075-11e5-88de-45c4f3eb35d4.png). Status: Issue closed Answers: username_1: Duplicate #2907
pecigonzalo/pre-commit-shfmt
232178511
Title: A few little issues Question: username_0: Hey! I just came across this repo and noticed a few small issues. * You match `*.sh` - what about `*.bash` and shell scripts without extension? If you run `shfmt` on a directory, it will automatically use only shell files, even if they have no extension, by looking for shebangs. * Instead of `for file in "$@"; do shfmt -l -i 4 -w "$file"; done` just do `shfmt <opts> "$@"` (or a directory, as explained above) * This tool is hard-coding `-i 4`? Answers: username_1: @username_0 Thanks for the feedback * That is correct, although I would say its standard to have `.sh` on most bash/sh/etc files, if `shfmt` detects that then ill give it a try. * About `-i 4` yes its hard coded, I can put a VAR over there to allow for config if `pre-commit` allows me to. Ill give this a look soon, otherwise feel free to submit a PR username_2: This seems to be a good one-liner you can use to find shell scripts in your repo: ```shell find . -type f \( -name "*" ! -path "./.[a-zA-Z]*/*" ! -wholename './.[a-zA-Z]*' \) -exec file {} \; | \ grep -Pi 'shell script' | \ cut -d : -f 1` ``` Status: Issue closed username_1: Updated the hook to find all shell files from the ones provided and added indent as a parameter. Let me know if this helps :D
NautiluX/slide
1091215236
Title: SEG Fault after opening first image Question: username_0: First time user. - Raspberry Pi 4B running Raspbian Buster. - Qt 5.15.2 - libexif 0.6.22-3 - Not using -c JSON file - The folder gdrive has only two .jpg files that can be displayed in other viewers - The images were rclone sync'd but I stopped rclone in case there was a sharing issue during sync After starting slide, the first image displays and then I receive the following seg fault. ``` pi@picframe:~ $ slide -p ~/gdrive updating image: /home/pi/gdrive/20210629_075427.jpg qt.qpa.xcb: QXcbConnection: XCB error: 148 (Unknown), sequence: 192, resource id: 0, major code: 140 (Unknown), minor code: 20 Segmentation fault ``` Answers: username_0: FIXED - updated to latest slide_pi_0.9.13 release Although I'm still getting one qt warning upon first image display, but no segv. Maybe a revision diff with qt dynamic linking or a Xserver difference. I am ssh'd into RPI4 via VNC, although it also happens via MobaXterm. The warning can be suppressed by setting an env var, but it also suppresses other qt.qpa warnings/errors ` export QT_LOGGING_RULES='qt.qpa.*=false'` username_0: This warning does not occur when connected directly to RPI4 Status: Issue closed
qiwihui/pocket_readings
242243742
Title: How Django REST Framework Changed My Life Question: username_0: How Django REST Framework Changed My Life In my 12 years of being a web and mobile developer, I have never been so excited about one particular framework or library as I am about Django REST Framework (DRF). Tags: via Pocket http://ngenworks.com/technology/how-django-rest-framework-changed-my-life/ Sep 21, 2015 at 16:40PM
SeleniumHQ/selenium
335816750
Title: Python client has to check for a still active keep-alive connection before sending the command Question: username_0: E BadStatusLine: '' ``` Answers: username_0: @username_3 can you please have a look? It would be good to also have such a test in the Selenium test suite. username_1: As a workaround, for the moment, I force the webdriver not to use keep_alive connection. ie : selenium/webdriver/firefox/webdriver.py line 165 ` executor = FirefoxRemoteConnection(remote_server_addr=self.service.service_url, keep_alive=False)` username_0: I know that @username_3 rewrote that part of the Python bindings to make use of urllib3, but the patch doesn't seem to have landed yet. David, can you please give us an update? username_2: PR is currently under review #6103 username_3: as there is a PR no need to keep a bug too Status: Issue closed
ststeiger/PdfSharpCore
643737473
Title: The right way to MigraDocCore AddImage(...) Question: username_0: Hello~ Is it a bug?Is there something wrong with me? ``` //The file path exists. string imgFile = "/Users/jess/myun/Upload/logonew.png"; //The following line of code throws a null reference exception. var a = MigraDocCore.DocumentObjectModel.MigraDoc.DocumentObjectModel.Shapes.ImageSource.FromFile(imgFile); //The following line of code throws a null reference exception. var b = MigraDocCore.DocumentObjectModel.MigraDoc.DocumentObjectModel.Shapes.ImageSource.FromStream("logonew.png", () => System.IO.File.OpenRead(imgFile)); MigraDocCore.DocumentObjectModel.Shapes.Image image = section.AddImage(a); ``` ## My running environment is: ``` .NET Core SDK (reflecting any global.json): Version: 3.1.301 Commit: <PASSWORD> Runtime Environment: OS Name: Mac OS X OS Version: 10.13 OS Platform: Darwin RID: osx.10.13-x64 Base Path: /usr/local/share/dotnet/sdk/3.1.301/ Host (useful for support): Version: 3.1.5 Commit: <PASSWORD> ``` Answers: username_1: Hi username_0 This is how I do it (I store images as embedded resources): ``` protected ImageSource.IImageSource GetEmbeddedImage(string Name, string Path) { Assembly assembly = Assembly.GetExecutingAssembly(); if (ImageSource.ImageSourceImpl == null) ImageSource.ImageSourceImpl = new ImageSharpImageSource<Rgba32>(); return ImageSource.FromStream(Name, () => assembly.GetManifestResourceStream(Path)); } ``` Try to insert this code: ``` if (ImageSource.ImageSourceImpl == null) ImageSource.ImageSourceImpl = new ImageSharpImageSource<Rgba32>(); ``` before you call "FromFile" or "FromStream". Hope this helps. username_0: Hi @username_1 Thank you very much for your effective help. I successfully inserted the picture into the PDF file according to your method. ``` if (ImageSource.ImageSourceImpl == null) ImageSource.ImageSourceImpl = new ImageSharpImageSource<Rgba32>(); Image image = section.AddImage(ImageSource.FromFile(imgFile)); ``` thank you very much! Status: Issue closed
VyacheslavRadyuk/sf-senla-courses
836868491
Title: 2.2 input validation improve Question: username_0: https://github.com/VyacheslavRadyuk/sf-senla-courses/blob/d65d2994f0fd9abc61416946f5a9080513971660/lecture_2.2/TouristManager.cls#L26 что не так: A) смотри, если здесь **ages** бует null - вылетит Exception. for (Integer age : **ages** ) { B) такой подход уже неверный: for (Integer age : ages ) { if (age == null) { return new List<Tourist__c>(); } } потому что: Представь что в запрос придет список age где все значения будут null, и в запросе будет ... where Age__c IN (null, null, null). База нормально это воспримет и просто вернет пустой список Подход неверный потому что если один из списка будет null, a остальные будут цифры - твой меод вернет пустой списко, что неправильно. Проверкой на null всех членов обычно не занимаются, но если очень надо, то можно сделать так: преобразовать на входе метода List в Set и потом проверить, буден короче цикл что поправить: 1) достаточно проверить входящий список на null и на пустоту (пункт А). 2) исправь пункт В Status: Issue closed Answers: username_0: -
JamesPD85/FlexMileage
282595135
Title: TypeError Question: username_0: c:\Python36\Scripts\Sublime\Mileage\FlexMileage>python tkgui.py Traceback (most recent call last): File "tkgui.py", line 149, in <module> app = FlexMileageApp() File "tkgui.py", line 30, in __init__ frame.grid(row=0,column=0,sticky="nsew") TypeError: wm_grid() got an unexpected keyword argument 'row' Code: ``` for F in (HomePage,History,EnterMileage,LoginPage,Costs): frame = F(container,self) self.frames[F] = frame # unexpected keyword argument 'row' frame.grid(row=0,column=0,sticky="nsew") # error^ ``` Answers: username_0: Turns out class Costs was inherited from tk.Tk, when it should have been tk.Frame. Status: Issue closed
MoboPlayer/TestMyMoboPlayer
50515057
Title: 多次开关闭播放界面,内存占用会越来越大,直到系统结束掉程序 Question: username_0: 你好,请教个问题: 当在ViewController中点击视频cell,会弹出MyVideoController。点击MyVideoController的OK按钮,播放界面会dismiss。当多次开关闭播放界面,内存占用会越来越大,到250M左右时,系统就会结束掉程序。 这种情况应该怎么处理?谢谢了。 Answers: username_0: 测试了下,现在新版的MoboPlayer多次开关闭播放界面,内存还是会增长,以几兆的速度增长,比以前的内存增长速度少很多。
Markyparky56/WasmNoise
527607956
Title: Set individual frequencies Question: username_0: Is it an option to add individual frequency settings for each of the noise generator input dimensions? This can be of use in the strip, square and cube functions. This way there's more control over the step size of the bulk generation. _I've used WasmMoise to feed noise to an animation, but when the noise has a small enough frequency it also changes too fast on the z-axis (time). I patched it by adjusting the frequency of the z-axis because I didn't find another option (except getting the noise per frame) instead of multiple frames at once._ Answers: username_1: While per-axis frequencies aren't a feature of FastNoise, neither are the bulk functions. I quite like the idea and doubt it would be much effort to set up. Work on WasmNoise is basically stalled at the moment but I'm hoping to be able to pick it back up in the New Year, so I'll see about adding it then! Awesome to hear that people are using my project, are you able to share what kind of stuff you're using it for?
playgameservices/play-games-plugin-for-unity
55130965
Title: Authentication Return calls before initialization done Question: username_0: If i do init game services first and straight away continue with Social.Active.Authenticate, I'll directly receive a return false as I can still choose which account i want to login. This happens only after i had login GooglePlayServices before and I reinstall the game. The problem I found is the AuthState haven't changed to silent pending when I'm calling Authenticate functions. To solve the problem i create an init state for the AuthState and reject any callback when Player is in This State. Answers: username_1: Could you elaborate a little bit more? How did you accomplished this in your code? username_2: I am having similar issue. I have a login button inside my game and it works fine. Only problem is when I signOut all accounts from Settings->Accounts section. I run my game and inside I click on LoginButton, it asks me to enter account details and shows me I signed in successfully but it returns 'false' as Social.localUser.Authenticate call back. The next time if I click its fine, only happens once at start. username_3: I also had the same issue with 0.9.12 and 0.9.15 (didn't try with older versions). After some testing I found out that the authentication process fires incorrectly before initialization is done **only** if saved games are enabled: ```c# PlayGamesClientConfiguration config = new PlayGamesClientConfiguration.Builder() .EnableSavedGames() .Build(); PlayGamesPlatform.InitializeInstance(config); PlayGamesPlatform.DebugLogEnabled = true; PlayGamesPlatform.Activate(); ``` Noteworthy is that the correct authentication process starts normally, after the initialization is done, but the callback is already fired by the process that started too early. So basically authentication transition is handled twice (HandleAuthTransition() method in NativeClient.cs). And as said before, this happens only after launching the game first time after one has authenticated before and then re-installed the game. username_4: @username_3 thanks for the details - we'll make sure to address this soon. username_5: I added a note to issue #443. Changing the callback to an actual method rather than a bool success in has resolved my issue with authentication. I have tested this with saved games and without. username_4: This is resolved in version 0.9.20 of the plugin. Please re-open if there are still questions. Status: Issue closed username_6: I really need this fix, but can't update since we're due to release in a few days. Is there any way I can get a patch to fix this? I'm on version 0.9.11 of the plugin.
flutter-webrtc/flutter-webrtc
907406573
Title: How to close debugging info? Question: username_0: When debugging,the webrtc lib print too many logs,How to close it? I/org.webrtc.Logging(26300): EglRenderer: Duration: 4004 ms. Frames received: 0. Dropped: 0. Rendered: 0. Render fps: .0. Average render time: NA. Average swapBuffer time: NA. Status: Issue closed Answers: username_1: same issue username_2: same here. Even after dispose the local stream
bazelbuild/rules_docker
730499341
Title: [puller] Feature request: support registry-mirrors global configuration Question: username_0: Basically I replaced: ```python container_pull( name = "redis_latest", registry = "index.docker.io", repository = "library/redis", tag = "latest" ) ``` with ```python container_pull( name = "redis_latest", registry = "mirror.gcr.io", repository = "library/redis", tag = "latest" ) ``` I found it a little awkward that I can't configure `registry-mirrors` globally by global bazelrc. In non-bazel builds that use native docker client they just added this global config to their docker daemon: ```json { "registry-mirrors": ["https://mirror.gcr.io"] } ``` I know we love everything to be defined in our committed source code but in this case - I think there's a great value in allowing global definition. BTW: I see that `docker_toolchain_configure` has `client_config` property - but I'm pretty sure it's not used in case of the puller (am I wrong?). Maybe the best case would be to read that property directly from the client_config. Answers: username_1: Hi @username_0, Generally you're right that injecting files whose contents aren't part of the source tree is an anti-pattern. However, if the docker client supports this, it makes sense for rules_docker to support this as well. Note that client_config is a standard docker client configuration JSON and I would make rules_docker accept arbitrary fields from this file unless docker recognizes it as well. Put another way, if docker client doesn't expect `registry-mirrors` in the client config JSON, I wouldn't update rules_docker to expect it in that file. OTOH, if the client config JSON is where docker client reads the registry mirrors from, I'm open to having rules_docker support it as well. I'm open to PR(s) that implements this. It's possible part of the implementation needs to be in go-containerregistry. username_2: #1789 would solve our use-case. @username_0 would it solve yours too? username_3: Just FYI we have had a lot of discussion around the approach to solve this in Bazel Build Slack https://bazelbuild.slack.com/archives/CA3NW13MH/p1618411736057600 https://bazelbuild.slack.com/archives/CA3NW13MH/p1618501747076700 The general consensus is that the Container eco system (docker, podman, etc...) support too many versions of `registry-mirror` setting. And we are currently unsure between several choices: A. Should we support `registry-mirrors` or `http_proxy` or `.netrc` or bazel's own `--experimental-downloader-config` any combination of those configs while pulling container images. B. If we were to support `registry-mirrors` config, should that abstraction live in `rules_docker` or it's dependency, `google/go-containerregistry`. @username_4, maintainer of go-containerregistry, seems to be in favor of having that supported in that repo and willing to review a PR for it. username_0: @username_2 sorry for late response - we eventually solved it by introducing an indirection to `container_pull`: ```python load( "@io_bazel_rules_docker//container:container.bzl", container_pull_real = "container_pull" ) load("@bazel_tools//tools/build_defs/repo:utils.bzl", "maybe") _MIRROR_URL="..." def container_pull(name, registry, **kwargs): # rewriting docker-hub registry to our proxy in artifactory since soon they will limit anonymous requests actual_registry = _MIRROR_URL if registry == "index.docker.io" else registry kwargs['registry'] = actual_registry maybe( name = name, repo_rule = container_pull_real, **kwargs ) ``` username_4: @username_0, that's awesome. I think for now we can recommend that as a short-term fix and in the long term correctly address it in rules_docker/go-containerregistry when we find a good cross-system solution.
modflowpy/flopy
991613775
Title: Can not access head file (.hds) using Flopy Question: username_0: `head = gwf.output.head().get_data()` The above line on the FlopyQucik.py shows following error: head = gwf.output.head().get_data() Traceback (most recent call last): File "C:\Users\.../3093209259.py", line 1, in <module> head = gwf.output.head().get_data() File "H:\Anaconda3\envs\Flopy\lib\site-packages\flopy\mf6\mfmodel.py", line 218, in __getattr__ raise AttributeError(item) AttributeError: output` Answers: username_1: What version of flopy are you using? The `.output` property is a relatively new addition to flopy's mf6 functionality. username_0: Hello jlarsen, I have been using the Flopy 3.3.4. username_2: @username_0 have you tried the development version of flopy? If so and you still are having this problem can you confirm that 1) you are using MODFLOW 6 (this functionality only works with MODFLOW 6 models) and 2) that the size of your head file is not 0 bytes. username_2: @username_0 is this still an issue? I will close this issue if I don't hear otherwise. Status: Issue closed
mhutchie/vscode-git-graph
1064169017
Title: Open multiple files from a commit Question: username_0: **Describe the feature that you'd like** From the Git Graph graph, it is very easy to find a commit, open it and find the files that were worked on. But to open multiple of them (for editing or diff-view) takes several clicks back and forth from the Git Graph graph since the files always open on top. It would be nice to be able to mark several files from a commit to then open them all at once. Alternatively, being able to "scroll wheel click" them (like links in a web browser) and have them open in a new tab in the background, without losing focus of the Git Graph graph, would also work. **Additional context (optional)**
ionic-team/ionic-docs
397318438
Title: Float Elements docs can be misleading Question: username_0: **I'm submitting a ...** [x] bug report [ ] feature request **URL:** https://beta.ionicframework.com/docs/layout/css-utilities#element-placement **Current behavior:** The markup does not actually leverage `float-*` attributes since it uses a grid. **Expected behavior:** A relevant markup, without a grid. **Steps to reproduce:** - Copy sample markup, - Change with responsive float attributes that should reverse the float positions, for instance: ```html <ion-grid> <ion-row> <ion-col float-left float-md-right> <div> <h3>float-left (should be float-right above md breakpoint)</h3> Lorem ipsum dolor sit amet, consectetur adipiscing elit. Sed ac vehicula lorem. </div> </ion-col> <ion-col float-right float-md-left> <div> <h3>float-right (should be float-left above md breakpoint)</h3> Lorem ipsum dolor sit amet, consectetur adipiscing elit. Sed ac vehicula lorem. </div> </ion-col> </ion-row> </ion-grid> ``` - Notice that it doesn't work as advertised, - Debug and see that Flexbox is handling the actual layout. **Other information:** Can be fixed with the following markup and CSS that both mimic the grid structure: **Template** ```html <div class="ion-grid"> <div class="ion-row"> <div class="ion-col ion-col-33" float-left float-md-right> <div> <h3>float-left (should be float-right above md breakpoint)</h3> Lorem ipsum dolor sit amet, consectetur adipiscing elit. Sed ac vehicula lorem. </div> </div> <div class="ion-col ion-col-33" float-right float-md-left> <div> <h3>float-right (should be float-left above md breakpoint)</h3> Lorem ipsum dolor sit amet, consectetur adipiscing elit. Sed ac vehicula lorem. </div> </div> </div> </div> [Truncated] } .ion-col { padding: 5px; } .ion-col-33 { width: 33%; } .ion-grid { padding: 0 0 20px 0; } .ion-col div { background-color: #f5f7fa; border: 1px solid #e4e8f0; padding: 5px; } ```
MinaProtocol/mina
763124529
Title: update charts for libp2p metrics (after libp2p prometheus refactor) Question: username_0: Once libp2p is exposing new metrics and is properly integrated with the -libp2p-metrics-port command line flag, we should support this flag in the majority of deployments and ensure that prometheus/grafana has the proper annotations to read from libp2p's prometheus endpoints in addition to the daemon's.
spacemeshos/SMIPS
745843665
Title: Updates Part II: protocol updates Question: username_0: Requirements * Protect network integrity and security. Don't allow the network to fail, or security assumptions to be violated, after an update. * Minimize trust assumptions required to have confidence that an update is non-malicious * Don't take agency away from node operators or make minority forks unreasonably burdensome * Minimize node downtime around an update * Give clarity and transparency into which nodes are running which version of the protocol * Subject to these constraints, make it as fast and easy as possible to update the protocol, and minimize implementation complexity Design * Node operator downloads and manually installs a node update that contains a protocol update, or the node software auto-downloads it (see #32). The update contains a hardcoded layer number `t` for activation of protocol changes. As laid out in #32, the node adds a "protocol version signature" to its blocks and/or ATXs to indicate a.) that it's received the update and b.) commits to triggering the protocol update at layer `t`. * In addition to the protocol update trigger layer, `t`, an update includes a cutoff layer height `s` < `t`, by which sufficient support must be indicated or else the update won’t be carried out. There should be at least one epoch between `s` and `t`. * After downloading and activating a client update, count the number of clients that have upgraded/“voted for” the change for one epoch (by spacetime committed/block weight). If the support level is not high enough by layer `s`, do not trigger the protocol upgrade. * If a node detects that an update has been approved but the node software has not been upgraded, it will begin printing a countdown and warnings of an impending shutdown with instructions. Some number of layers before the update is due to trigger, the node will panic and exit and print instructions again. The node operator may choose to run the node with a `--ignore-updates` flag that will explicitly continue running the old protocol, they may restart with the `--auto-updates` flag enabled, or they may manually download and install an update. The printed message should contain a link to a document with a more complete explanation of how and why it works this way. * The next release after a “triggered” upgrade can remove the vote counting logic and hardcode the protocol update as of layer `t`—this is equivalent to a checkpoint. A subsequent upgrade is likely to use the same logic to trigger a subsequent update at a later layer. Tasks * Finalize node software and protocol version signature format. Consider the differences among node implementation version, P2P protocol version, and core Spacemesh protocol version. See #32. * Develop node logic that counts update/version “signature” votes and triggers a protocol update, or not, on this basis * Finalize update params (minimum spacetime/block weight "voting" threshold, minimum time to trigger update) * Finalize copy, order, and timing of new warning messages, and update longer doc explaining this * Add tests for voting mechanism * Discuss what to do in case of a failed update
jrowberg/i2cdevlib
156009200
Title: I2Cdev & ADXL345 compilation error Question: username_0: I am trying to compile a code using I2Cdev.h and ADXL345.h and it gives me the following error: I2Cdev.h:46:1: error: unknown type name ‘class’ class I2Cdev { ^ I2Cdev.h:46:14: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘{’ token class I2Cdev { ^ In file included from IMU.c:12:0: ADXL345.h:163:1: error: unknown type name ‘class’ class ADXL345 { ^ ADXL345.h:163:15: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘{’ token class ADXL345 { I haven't modified the code. Has anyone had this problem before? Thank you for your time! Answers: username_1: First some mandatory questions.. * What is your target platform? * What happens when you try to compile the ADXL345_raw example? * Are you using the current i2cdev master branch? Judging from the error text I would suggest you to look at what happens just before `#include <I2Cdev.h>` in your sketch. Try to temporarily move the failing #include to the top or the bottom of the #include block. If compiler complains about something else the problem is outside of I2Cdev.h. username_0: My target platform is Raspberry Pi 2B, actually I compiled ADXL345 example without problems. Thank you for the help! username_0: The code I am using is in [https://github.com/username_0/IMU_gy85](url) username_1: I think the main problem is the macro: `#define ADXL345 0x53` in sensors.h username_1: Are you using g++ when compiling your code? ADXL345.h and I2Cdev.h are C++ files. username_0: In fact I was compiling using gcc. I just used the i2c-dev.h and adapted your libraries to mine. Thank you for your help! Status: Issue closed
VirtusLab/strapi-plugin-content-manager-extension-hierarchical
704334140
Title: Bug in the extension Question: username_0: Can you please check the extension for version 3.1.5???Or help me fix the issue??thank you in advance !!! Answers: username_1: @username_0 can you please provide more details? username_2: Hello, @username_1 ! I'm creating new strapi 3.2.4 project ```yarn create strapi-app my-project --quickstart``` and run it with ```yarn develop``` it works fine as expected: I'm able to create and to edit content types. After installing extension by copying three folders (admin/extensions/hooks) from the root of this repo to the root of the strapi 3.2.4 project, I'm able to run application, but with problems: Problem #1. I'm not able to see existing content types any more. No errors - no content type. Problem #2. When I try to create new content type I get error after clicking save button: POST to http://localhost:1337/content-type-builder/content-types ends with 400 code and response: {"error":{"contentType":["contentType field cannot have keys not specified in the object shape"]}} username_3: I unfortunately have the same issue on 3.2.5 TypeError: undefined is not an object (evaluating 'g.split') username_4: I am getting the same issues on 3.6.6, would be thankful if someone could check this! username_5: I have a "contentType field has unspecified keys: draftAndPublish, pluginOptions"
Kinovarobotics/kortex
1045269284
Title: Request to read external torques Question: username_0: ## Summary Request to read external torque at every joints. ## Use case I would like to use the external force to estimate contact between the robot arm and the environment. This requires knowing detailed mass distribution of the robot arm. The equation to compute external torque can be found here: https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=8059840, described in section III. ## Alternative This external torque can be measured torque subtracting other terms in the dynamics equation. If I know the distribution of mass in the robot arm links, I may be able to compute this by myself. ## Possible issues No, this only provides an extra value to read. ## Additional context I am working on to reproduce the result from this paper https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=9561761. They work with a Kuka robot arm, that seems to have this feature. Answers: username_1: Hi @username_0, I think what you are looking for is all documented in the Gen3 [User Guide](https://artifactory.kinovaapps.com/ui/api/v1/download?repoKey=generic-documentation-public&path=Documentation%2FGen3%2FTechnical%20documentation%2FUser%20Guide%2FUser%20Guide%20Gen3%20-%20R06.pdf). In the *Guidance for advanced users* section (found at the end), you can find Mass, Center of Mass, Inertia matrix as well as transformation matrix (and other information you might find usefull) for every joint : ![image](https://user-images.githubusercontent.com/83971099/140431758-aad6c7fa-ef4f-4105-a06e-c09326d5c03f.png) Regards, Felix username_2: Hi @username_0, similar request was discussed in issue #52. Cheers username_0: @username_2 Thank you for your reply! I wonder if you rely on the parameters provided in their documents. Did you find the inertial or mass value inaccurate? I tried the robot arm without any load and the measured torques are different from gravitational torques at static state. May I ask if you did any system identification on the mass matric or gravitational term?
microsoft/TypeScript-Website
626844178
Title: Docs: Move "Migrating from JavaScript" to a new toplevel box? Question: username_0: "Migrating from JavaScript" sounds like an impusername_1nt enough document (and different from the other entries in the tutorials section) to go into its own toplevel box, with links to the sections. Answers: username_1: Yep, that's a goal! username_1: This is in now, there's now a "JavaScript" section which is focused on JS users using TS tooling Status: Issue closed
andymass/vim-matchup
401535562
Title: Conditional compilation #ifdef and splitting if() then statement over two lines Question: username_0: I was trying out the features **(a.1) jump between matching words** (command `%`) and **(a.2) jump to open and close words** (commands `[%` and `]%`) to jump, for example, from one `if` to the next `elseif`. However, I work with programs that contain `#ifdef` for conditional compilation. Also, sometimes `if() then` are split over two lines by means of the character `&`. In these cases, the commands `%`, `[%` or `]%` seem not to be working. Here is an example: ``` program ifElseIfElseProg implicit none integer :: a = 100 if( a == 10 ) then a = a + 1 #ifdef three_d + 3 #endif else if( a == 20 ) & then print*, "Value of a is 20" else if( a == 30 ) then print*, "Value of a is 30" else print*, "None of the values is matching" end if print*, "exact value of a is ", a end program ifElseIfElseProg ``` Answers: username_1: Can you tell me what language this is, and if you use any filetype plugins for it? username_0: The language is FORTRAN. My .vimrc reads: ``` call plug#begin('~/.vim/plugged') Plug 'username_1/vim-matchup' call plug#end() " Uncomment the following to have Vim jump to the last position when reopening a file if has("autocmd") au BufReadPost * if line("'\"") > 0 && line("'\"") <= line("$") \| exe "normal! g'\"" | endif endif ``` Status: Issue closed username_1: Thanks. Note that I can't handle multi-line matches at the moment so "then" isn't highlighted. Also I can't support every possible line continuation.
BoevaLab/FREEC
748195521
Title: Segmentation fault Question: username_0: I am trying to run Control-FREEC on a set of samples (WES) and I get segmentation faults in some of the samples while others complete fine. ``` ..Annotation of CNVs for X 1 0.00875069 2 0.935849 3 0.0070744 4 0.0478207 5 0.000315901 6 3.9343e-05 15 2.07169e-06 zsh: segmentation fault freec -conf config.txt ``` Is this something expected? Thanks! Answers: username_1: Dear user, no, it is not expected.. But still happens sometime. To help you I will need your config file and the complete output into the command line. Thank you Valentina username_0: Dear Valentina, I've attached here both the config and the console output files. Best, Jose [out.txt](https://github.com/BoevaLab/FREEC/files/5581956/out.txt) [config.txt](https://github.com/BoevaLab/FREEC/files/5581957/config.txt) username_0: Hi, have you been able to look at this issue? username_1: Dear <NAME>, From your output, it is difficult to say what exactly caused the error. Can you tell me which files have been created? Also, can you try to run this simplified config, please, and let me know whether it created an error? [general] BedGraphOutput = TRUE chrFiles = /Users/username_0/Projects/Scitron/Results/Chromosomes chrLenFile = /Users/username_0/Projects/Scitron/Results/Homo_sapiens_assembly38.len #forceGCcontentNormalization = 1 maxThreads = 8 #minimalSubclonePresence = 30 #ploidy = 2,3 ploidy=2 sex = XX readCountThreshold = 50 breakPointThreshold = 1.2 breakPointType = 4 window = 0 [control] inputFormat = pileup mateFile = B190986_RG1297_32_gDNA.pileup mateOrientation = FR [sample] inputFormat = pileup mateFile = 1297_34_cfDNA.pileup mateOrientation = FR [BAF] SNPfile = /Users/username_0/Projects/Scitron/Results/dbsnp_138.hg38.vcf.gz [target] captureRegions = /Users/username_0/Projects/Scitron/intervals/intersect_collapsed.bed username_0: Dear Valentina, Thanks for your reply. So, I played a bit and I got to learn that it was the `minimalSubclonePresence = 30` parameter the one causing troubles even though 30 is the recommended value for WES data. I think for now I will not use that option since it is not really needed for the post-analysis. Best, Jose username_1: Good to know. The subclonal search is still in Beta. Hope to debug it when have time. Sent from my iPhone >
ppiastucki/recast4j
765687009
Title: Unable to run the .jar in macOS Question: username_0: Hi, I tried to run the demo but it is not working on macOS 10.15.7 ``` java -jar recast-demo-1.2.2-SNAPSHOT.jar 20:32:25 Exception in thread "main" java.lang.ExceptionInInitializerError at org.lwjgl.glfw.GLFW.glfwCreateWindow(GLFW.java:1831) at org.recast4j.demo.RecastDemo.start(RecastDemo.java:220) at org.recast4j.demo.RecastDemo.main(RecastDemo.java:593) Caused by: java.lang.IllegalStateException: GLFW windows may only be created on the main thread and that thread must be the first thread in the process. Please run the JVM with -XstartOnFirstThread. For offscreen rendering, make sure another window toolkit (e.g. AWT or JavaFX) is initialized before GLFW. at org.lwjgl.glfw.EventLoop$OffScreen.<clinit>(EventLoop.java:39) ... 3 more ``` ^ The above seems to be solved adding `-XstartOnFirstThread` And then: user@machine:/Users/user/Desktop/recast4j-parent-1.2.5/recast-demo/target: java -XstartOnFirstThread -jar recast-demo-1.2.2-SNAPSHOT.jar 20:32:34 20:36:06.321 [main] DEBUG org.recast4j.demo.RecastDemo - Intel Inc. 20:36:06.323 [main] DEBUG org.recast4j.demo.RecastDemo - 2.1 INTEL-14.7.8 20:36:06.323 [main] DEBUG org.recast4j.demo.RecastDemo - Intel(R) Iris(TM) Plus Graphics 655 20:36:06.323 [main] DEBUG org.recast4j.demo.RecastDemo - 1.20 Exception in thread "main" java.lang.IllegalStateException at org.recast4j.demo.ui.NuklearGL.<init>(NuklearGL.java:184) at org.recast4j.demo.ui.NuklearUI.<init>(NuklearUI.java:89) at org.recast4j.demo.RecastDemo.start(RecastDemo.java:331) at org.recast4j.demo.RecastDemo.main(RecastDemo.java:593) ``` Which is throwing at: `if (glGetShaderi(vert_shdr, GL_COMPILE_STATUS) != GL_TRUE) {` Do you know why? PS: It never runs on a windows Virtualbox Windows 10 machine which shows as the graphic device: ``` 23:32:14.486 [main] DEBUG org.recast4j.demo.RecastDemo - VMware, Inc. 23:32:14.486 [main] DEBUG org.recast4j.demo.RecastDemo - 2.1 Mesa mesa-17.3.9 23:32:14.486 [main] DEBUG org.recast4j.demo.RecastDemo - SVGA3D; build: RELEASE; 23:32:14.486 [main] DEBUG org.recast4j.demo.RecastDemo - 1.20 ``` And then the same error for `NuklearGL.java:184` Answers: username_1: Hi, The following entry in the logs contains the version of the GLSL: ``` 23:32:14.486 [main] DEBUG org.recast4j.demo.RecastDemo - 1.20 ``` The shaders used in RecastDemo require GLSL 1.5 or higher. The code tries to ensure the proper version of the GLSL by settings the following: ``` glfwWindowHint(GLFW_CONTEXT_VERSION_MAJOR, 3); glfwWindowHint(GLFW_CONTEXT_VERSION_MINOR, 2); ``` but apparently it is not respected on your device. Status: Issue closed username_1: Closing inactive issue
QuickLogic-Corp/quicklogic-fpga-toolchain
669195775
Title: Dockerfile contains link to non-existent installer Question: username_0: The v1.1.0_Beta directory is no longer available, so this line is broken. Remove _Beta to fix it. RUN wget https://github.com/QuickLogic-Corp/quicklogic-fpga-toolchain/releases/download/v1.1.0_Beta/Symbiflow_v1.1.0.gz.run Answers: username_0: Dockerfile has been updated and merged. Status: Issue closed
NativeScript/nativescript-cordova-support
358980746
Title: Installation instructions problem Question: username_0: Hi there, Is this plugin still maintained? I just tried running the install command as provided by the documentation but it doesn't work. Using nativescript 4.2 on Windows. Cheers! Answers: username_1: @username_0 Thank you for reporting this inaccuracy and sorry for not responding earlier. The plugin has to be installed from a local path where the repository is cloned. I've fixed the instructions in the README. The plugin is not officially supported but nevertheless if there's something wrong with it we will do our best to fix it and if demand increases it may at some point become officially supported.
aonez/Keka
204367161
Title: p7zip error 2 on big files over 4GB Question: username_0: As commented on #15 by @username_1. Answers: username_0: I've managed to replicate the issue, creating zip files with the bundled mac archiver. Those created with Keka work. username_0: Seems like macOS system is creating corrupt files if they are bigger than 4GB. Already reported a bug to Apple. username_0: Though its not really a p7zip fail but a corruption in macOS (and earlier, tested Mac OS X 10.9), maybe <NAME> can enlighten us: https://sourceforge.net/p/sevenzip/bugs/2038 username_1: Why is the native Archiver still able to unzip it when it is a corrupt file? username_0: Also unzip (via Terminal) can extract its contents, not without trowing some warnings of data corruption. WinZip is unable of opening the file, and WinRAR extracts it also throwing some corruption messages. username_2: I just got this error with a *.7z I just created with 1.0.7 (maximum compression). The archive itself is about 18 MB. Got the error when attempting to unpack with 1.0.7 and 1.0.8. However, was able unpack via p7zip 16.02 directly ("7z x *.7z") on CentOS. username_0: @username_2 this seems to be another issue, since this one just affects files bigger than 4GB. Is it possible to you to share the file? username_3: FWIW, I encountered a similar issue (or something similar) on El Capitan too. But it happened with an 11.75 GB file zipped by Keka. The compressed directory on my Mac was named "December 29, 1 (by Keka)". Below are the results of using the built-in/default Unzip utility from Terminal: ``` $ unzip -t "December 29, 1 (by Keka)" Archive: December 29, 1 (by Keka).zip warning [December 29, 1 (by Keka).zip]: 7458748132 extra bytes at beginning or within zipfile (attempting to process anyway) error [December 29, 1 (by Keka).zip]: start of central directory not found; zipfile corrupt. (please check that you have transferred or created the zipfile in the appropriate BINARY mode and that you have compiled UnZip properly) ``` Some research uncovered that El Capitan uses an outdated Unzip utility by Info-Zip, i.e., version 5.52, which doesn't support Zip64 files, a format created automatically when an archive begins to exceed a 4 GB boundary. (See [this answer on StackExchange](https://superuser.com/a/249689/466273).) Upgrading to Info-Zip's Unzip version 6.00 via Homebrew and trying again showed that all was actually well with the archive: ``` $ unzip6 -t "December 29, 1 (by Keka)" Archive: December 29, 1 (by Keka).zip testing: December 29, 1/ OK [...] No errors detected in compressed data of December 29, 1 (by Keka).zip. ``` username_0: No timeline, but eventually sure #139 username_4: Today I updated to 1.1.5 and this error happened to me with a 30,85 GB (according to Finder) large UDRO DMG source file I cannot share with you. Rolled back to 1.1.4 and it works flawlessly again. This has never occured to me before and I regularly use 7zip on files larger than 4 GB. Just letting you know! username_5: I can shed some light on this error. From my Mac environment after enabling logging (and searching console.app for 'keka') I found this error: `ERROR: stat error for /<omitted>/target/generated-sources/annotations/. (Permission denied)` Looks like the directory 'annotations' was missing the owner:execute bit (644). username_0: Thanks a lot @username_5! Clearly this can be solved by improving the error messages displayed #116. username_6: // , I get the same error on a 104mb file. Why is this closed??
EA31337/EA-Tester
642611346
Title: mt_modify: Support for multiple -f params Question: username_0: When multiple `-f` params are specified for `mt_modify.py`, modify multiple files at once. E.g. ``` mt_modify.py -f EURUSD1_0.fxt -f EURUSD1_1.fxt -t fxt-header -m spread=10 ``` Alternatively we can have files separated by comma.
robotpy/robotpy-cscore
443126721
Title: gettting nothing while running cam in process Question: username_0: while trying to run a usb cam in a process and grabbing its frames the value of the image is 0 ` #camera is a usb camera object multiprocessing.Process(target=self.camera_process, args=(cam,camera_server)) def camera_process(self, camera,cs): cv_sink = cscore.CvSink("test") cv_sink.setSource(camera) while True: _, image = cv_sink.grabFrame(image) ` Answers: username_1: No, this won't work. You're going to have to send the images to the subprocess yourself (e.g. via a pipe). username_2: @username_0 what are you trying to accomplish? If you're trying to run this code on a roborio, see the documentation at https://robotpy.readthedocs.io/en/stable/vision/roborio.html username_0: no im trying to run this on my computer in and to process each camera in parallel username_2: If you create your camera objects inside the `camera_process` function, it should work. username_2: Closed due to age. Status: Issue closed
andymeneely/squib
113424799
Title: Release v0.8 Question: username_0: - [ ] CHANGELOG is written for all new changes - [ ] README is updated - [ ] Samples are updated - [ ] `rake doc` - [ ] Check `sample_regression_spec.rb` regression tests are all enabled (i.e. `overwrite_sample` is commented out) - [ ] Bump version.rb - [ ] Do a full rake locally - [ ] `rake sanity`, and check visually - [ ] Travis is passing on dev branch - [ ] Merge master branch - [ ] Create GitHub release tag - [ ] `gem push pkg/squib-x.y.z.gem` - [ ] Github milestone closed - [ ] Push `rake doc` to website - [ ] Bump version.rb to the next alpha - [ ] Publish on BoardGameGeek thread<issue_closed> Status: Issue closed
thuguerre/WebSiteQuickAddingForTodoist
514960810
Title: init a release checklist Question: username_0: create a documentation file in which are listed all actions to perform before releasing a new version of the project: - manual tests to perform (if not possible to automate them) once deployed - verify message translation - verify ascendant compatibility - ... Answers: username_0: PR #68 Status: Issue closed
chapmanb/cloudbiolinux
527426402
Title: Broad FTP login error is causing bcbio-nextgen data upgrade to fail Question: username_0: See related issue filed over at bcbio-nextgen: https://github.com/bcbio/bcbio-nextgen/issues/3021 Looks like the gsapubftp login might need to be updated? Here's the relevant code: https://github.com/username_1/cloudbiolinux/search?q=gsapubftp&unscoped_q=gsapubftp Answers: username_0: See related GATK thread: https://gatkforums.broadinstitute.org/gatk/discussion/1215/how-can-i-access-the-gsa-public-ftp-server username_1: Mike; This is the error message that the Broad FTP server reports when it's overloaded. I don't believe anything has changed, it's just that it doesn't handle high loads very well. We don't have a good fix for this short of hosting ourselves so the best recommendation is to retry and hopefully it'll be less stressed and work cleanly. Hope this helps. Status: Issue closed username_0: Thanks Brad, I thought that might be what was going on. I wasn't sure though because I could log into the FTP server with an external client and list the files; I was only seeing errors with the Python script. username_0: A couple of days later, I'm still seeing this error: ``` Running GGD recipe: hg38 1000g_snps 20160105 --2019-11-26 08:44:02-- ftp://gsapubftp-anonymous:*password*@ftp.broadinstitute.org/bundle/hg38//1000G_p$ ase1.snps.high_confidence.hg38.vcf.gz => ‘variation/1000G_phase1.snps.high_confidence.vcf.gz’ Resolving ftp.broadinstitute.org (ftp.broadinstitute.org)... 192.168.127.12 Connecting to ftp.broadinstitute.org (ftp.broadinstitute.org)|192.168.127.12|:21... connected. Logging in as gsapubftp-anonymous ... Login incorrect. ``` username_2: Hi Mike, Sorry about that, there isn't much we can do about it unfortunately. It should eventually work when the servers stop being overwhelmed. username_0: Following up, on a fresh bcbio-nextgen 1.1.9 install, I'm still seeing this: ``` --2019-12-13 09:40:53-- ftp://gsapubftp-anonymous:*password*@ftp.broadinstitute.org/bundle/hg38//Mills_and_1000G_ gold_standard.indels.hg38.vcf.gz.tbi => ‘variation/Mills_and_1000G_gold_standard.indels.vcf.gz.tbi’ Resolving ftp.broadinstitute.org (ftp.broadinstitute.org)... 192.168.127.12 Connecting to ftp.broadinstitute.org (ftp.broadinstitute.org)|192.168.127.12|:21... connected. Logging in as gsapubftp-anonymous ... Login incorrect. Upgrading bcbio Upgrading bcbio-nextgen data files List of genomes to get (from the config file at '{'genomes': [{'dbkey': 'hg38', 'name': 'Human (hg38) full', 'inde xes': ['seq', 'twobit', 'bwa', 'hisat2'], 'annotations': ['transcripts', 'RADAR', 'rmsk', 'salmon-decoys', 'fusion -blacklist', 'ccds', 'capture_regions', 'coverage', 'prioritize', 'dbsnp', 'hapmap_snps', '1000g_omni_snps', 'ACMG 56_genes', '1000g_snps', 'mills_indels', '1000g_indels', 'clinvar', 'qsignature', 'genesplicer', 'effects_transcri pts', 'varpon', 'vcfanno', 'viral'], 'validation': ['giab-NA12878', 'giab-NA24385', 'giab-NA24631', 'platinum-geno me-NA12878', 'giab-NA12878-remap', 'giab-NA12878-crossmap', 'dream-syn4-crossmap', 'dream-syn3-crossmap', 'giab-NA 12878-NA24385-somatic', 'giab-NA24143', 'giab-NA24149']}, {'dbkey': 'mm10', 'name': 'Mouse (mm10)', 'indexes': ['s eq', 'twobit'], 'annotations': ['transcripts', 'rmsk', 'problem_regions', 'dbsnp', 'vcfanno']}, {'dbkey': 'rn6', ' name': 'Rat (rn6)', 'indexes': ['seq', 'twobit'], 'annotations': ['transcripts']}, {'dbkey': 'canFam3', 'name': 'D og (canFam3)', 'indexes': ['twobit'], 'annotations': ['transcripts', 'dbsnp']}, {'dbkey': 'BDGP6', 'name': 'D mela ngogaster (BDGP6)', 'indexes': ['seq'], 'annotations': ['transcripts']}], 'genome_indexes': ['bowtie2', 'rtg', 'st ar'], 'install_liftover': False, 'install_uniref': False}'): Human (hg38) full, Mouse (mm10), Rat (rn6), Dog (canF am3), D melangogaster (BDGP6) Running GGD recipe: hg38 seq 1000g-20150219_1 Moving on to next genome prep method after trying ggd GGD recipe not available for hg38 bowtie2 Downloading genome from s3: hg38 bowtie2 Moving on to next genome prep method after trying s3 No pre-computed indices for hg38 bowtie2 Preparing genome hg38 with index bowtie2 Moving on to next genome prep method after trying ggd GGD recipe not available for hg38 rtg Downloading genome from s3: hg38 rtg Moving on to next genome prep method after trying s3 No pre-computed indices for hg38 rtg Preparing genome hg38 with index rtg Moving on to next genome prep method after trying ggd GGD recipe not available for hg38 star Downloading genome from s3: hg38 star Moving on to next genome prep method after trying s3 No pre-computed indices for hg38 star Preparing genome hg38 with index star hg38 detected, building a simple reference with no alts, decoys or HLA from /data00/n/app/bcbio/1.1.9/install/geno mes/Hsapiens/hg38/seq/hg38.fa to /data00/n/app/bcbio/1.1.9/install/genomes/Hsapiens/hg38/seq/hg38-simple.fa. Preparing STAR index from /data00/n/app/bcbio/1.1.9/install/genomes/Hsapiens/hg38/seq/hg38-simple.fa. Removing /data00/n/app/bcbio/1.1.9/install/genomes/Hsapiens/hg38/seq/hg38-simple.fa. Running GGD recipe: hg38 transcripts 2018-10-10_92 Running GGD recipe: hg38 RADAR v2-20180202 Running GGD recipe: hg38 rmsk 20180319 Running GGD recipe: hg38 salmon-decoys 94 Running GGD recipe: hg38 fusion-blacklist 2 Running GGD recipe: hg38 ccds r20 Running GGD recipe: hg38 capture_regions 20161202 Running GGD recipe: hg38 coverage 2018-10-16 Running GGD recipe: hg38 prioritize 20181227 Running GGD recipe: hg38 dbsnp 151-20180418 Running GGD recipe: hg38 hapmap_snps 20160105 Running GGD recipe: hg38 1000g_omni_snps 20160105 [Truncated] _prep_genomes(env, genomes, genome_indexes, ready_approaches, data_filedir) File "/mnt/resource/tmp.1YauZ8ZyEP/bcbio/tmpbcbio-install/cloudbiolinux/cloudbio/biodata/genomes.py", line 480, in _prep_genomes retrieve_fn(env, manager, gid, idx) File "/mnt/resource/tmp.1YauZ8ZyEP/bcbio/tmpbcbio-install/cloudbiolinux/cloudbio/biodata/genomes.py", line 872, in _install_with_ggd ggd.install_recipe(os.getcwd(), env.system_install, recipe_file, gid) File "/mnt/resource/tmp.1YauZ8ZyEP/bcbio/tmpbcbio-install/cloudbiolinux/cloudbio/biodata/ggd.py", line 30, in in stall_recipe recipe["recipe"]["full"]["recipe_type"], system_install) File "/mnt/resource/tmp.1YauZ8ZyEP/bcbio/tmpbcbio-install/cloudbiolinux/cloudbio/biodata/ggd.py", line 62, in _r un_recipe subprocess.check_output(["bash", run_file]) File "/n/app/bcbio/1.1.9/install/anaconda/lib/python3.6/subprocess.py", line 336, in check_output **kwargs).stdout File "/n/app/bcbio/1.1.9/install/anaconda/lib/python3.6/subprocess.py", line 418, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bash', '/data00/n/app/bcbio/1.1.9/install/genomes/Hsapiens/hg38/txtmp/gg d-run.sh']' returned non-zero exit status 6. ``` username_2: Thanks, it will be random that this happens unfortunately.
bigdargon/hostsVN
650450015
Title: Chặn tracker trên thoibao.com Question: username_0: [//]: # (***Vui lòng nhập thông tin phía dưới dòng có dấu "[//]:" các dòng này sẽ bị ẩn khi đăng bài. Xin cám ơn!) [//]: # (Nếu bạn muốn thêm tên miền mới, vui lòng đưa tên miền vào thẻ bên dưới.) ***Tên miền***: ``` infolinks.com ``` Status: Issue closed Answers: username_1: @username_0 Tên miền đã có trong bộ lọc rồi nha bạn!
Ouranosinc/Magpie
685067901
Title: [Feature] Alternative listing of services Question: username_0: As of latest versions, `GET /services` returns services nested under their respective `service-type`, which are then again provided under each element's `service_type` field. There is therefore redundancy of information, but even more, this format makes it hard to parse since one always needs to loop service-type keys to then exact sub-services keys, to finally obtain a flat list of "service" objects. A query parameter `list=true|false` should be added to render the same information as flat list of service objects. Original format must still be returned for backward compatibility (an avoid massive rewrites), but the new flag could help skipping a few steps by requesting user that wants it as such.<issue_closed> Status: Issue closed