repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
OliverHellwig/sanskrit
543037397
Title: What license is the code and data released under? Question: username_0: The documentation/repo does not say what others are authorized to do with this code or data. Answers: username_0: Ah, I found the CC license for the DCS, but still haven't found one for the code. I believe the CC license requires linking to the full text as well, though? username_1: I have to add a license statement to the code, but it would be similar to CC. username_0: Note that CC advises against using their licenses for software/code: https://creativecommons.org/faq/#can-i-apply-a-creative-commons-license-to-software
electron/electron
239773328
Title: None Question: username_0: @p3x-robot make sure you run this code only after the `app.on('ready')` event is triggered, not before. Status: Issue closed Answers: username_1: It appears that @username_0 has resolved this issue for @p3x-robot, so in the meantime i'm going to go ahead and close this.
dotnet/docfx
267009206
Title: Microsoft.DocAsCode.Dfm (or any) is not depending on System.Composition.AttributedModel Question: username_0: **DocFX Version Used**: Any **Template used**: Any **Steps to Reproduce**: [Same steps with tutorial on documentation](https://dotnet.github.io/docfx/tutorial/howto_customize_docfx_flavored_markdown.html) **Expected Behavior**: The plugin should builds. **Actual Behavior**: `ExportAttribute` is missing. More info: ```c# [Export(typeof(IDfmEngineCustomizer))] ``` I got stuck in step 7, the `ExportAttribute` is missing. After hours of searching I found it was a class in another [NuGet](https://www.nuget.org/packages/System.Composition.AttributedModel/) named `System.Composition.AttributedModel` version `1.0.31` (It still won't compile if latest stable version `1.1.0` is used). And then I take a quick look on the dfm depedencies and found this screenshot confirms my hypothesis: ![screenshot 71](https://user-images.githubusercontent.com/20214420/31797132-f20fb256-b556-11e7-9d4f-71d0923cbd5c.png) Answers: username_1: Thank you for raising this! Fixed in the document. Status: Issue closed
liyaSileshi/Spaceman
494811000
Title: code review Question: username_0: - Nice use of docstrings - You could improve commenting throughout your code within the functions to describe what is happening more clearly - Good job making your code modular and organizing everything into functions - Clean up the TODOs after you are finished with them, many of your TODOs would make nice comments if you simply delete the “TODO” part - Good job using semantic variable naming - Nice spaceman =D
Miniand/brdg.me-issues
69993195
Title: Cathedral - improve representation of direction Question: username_0: _From @username_0 on April 16, 2015 2:44_ @baconheist is shit with spatial puzzles, make it more noob friendly :stuck_out_tongue: Suggestion is to show an arrow. _Copied from original issue: Miniand/brdg.me#68_
strapi/strapi
569776913
Title: Filter combined with search is not working in admin Question: username_0: **Describe the bug** Filter combined with search is not working in admin **Steps to reproduce the behavior** 1. Go to a content-type overview 2. Apply a filter (e.g. Id is xxx) 3. Now type something in the search bar 4. The filter is still applied (visually) but not working anymore **Expected behavior** The search is queried along with the configured filters **Screenshots** Without search: ![image](https://user-images.githubusercontent.com/39594925/75146598-bb74de80-56fb-11ea-9b54-7336e25dccd6.png) With search: ![image](https://user-images.githubusercontent.com/39594925/75146669-da737080-56fb-11ea-8543-72488d6a9357.png) **System** - Node.js version: v10.17.0 - NPM version: 6.11.3 - Strapi version: 18.7-beta - Database: MongoDB - Operating system: MacOS / Ubuntu 18.04 Answers: username_1: I believe the entire search/filtering system here in the admin is getting an overhaul username_0: @username_1 ah okay. It also is not working with the internal API (e.g. `application::deal.deal?_q=portugal&_limit=20&_start=0&is_published_deal=true`). `is_published_deal=true` is not applied there anymore. username_1: I don't think `_q` was designed to be used with filters, it's original purpose was a bit of a "type-a-head" type thing for the CTM when searching relations. It kinda acts as a fuzzy search across the full-text indexes. username_1: @lauriejim not a discussion as @username_0 already working on PR Status: Issue closed username_2: This was merged a while back now. Closing
onnx/onnx
612958285
Title: A "name" field for SparseTensorProto? Question: username_0: This issue was raised from https://gitter.im/onnx/Lobby. Indeed, a "name" is needed for SparseTensorProto if it's used to store the model weights for example. Thoughts? @username_1 Answers: username_1: A SparseTensorProto has a “TensorProto values” which has a name. Won’t it be enough to use the name in the “values” field? Do we need a separate name? username_0: Fair enough. Let's add some comment in place to clarify this "naming" please. Status: Issue closed username_2: This is documented now: https://github.com/onnx/onnx/blob/master/onnx/onnx.proto#L608
MicrosoftDocs/azure-docs
404662682
Title: Not able to run app locally Question: username_0: I have tried to run application but I am getting error. System.Fabric.FabricConnectionDeniedException: 'Not authorized to connect' Inner Exception : COMException: Exception from HRESULT: 0x80071C43 --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3aa58af5-02ea-e425-1fc5-9f588d394016 * Version Independent ID: 86440381-783a-6e49-bacb-59b1c1581d81 * Content: [Create a .NET app on Service Fabric in Azure](https://docs.microsoft.com/en-in/azure/service-fabric/service-fabric-quickstart-dotnet#prerequisites) * Content Source: [articles/service-fabric/service-fabric-quickstart-dotnet.md](https://github.com/Microsoft/azure-docs/blob/master/articles/service-fabric/service-fabric-quickstart-dotnet.md) * Service: **service-fabric** * GitHub Login: @MikkelHegn * Microsoft Alias: **mikhegn** Answers: username_1: Thanks for the feedback! We are currently investigating and will update you shortly. username_1: @username_0 Any update? username_0: Thanks. Its working now. I guess I missed something on my end. I probably did not restarted machine after I install service fabric SDK. Or I did not created a local cluster or I did not started it. But now its working fine. Regards Nagesh Status: Issue closed username_2: I also tried to run the application on local environment and i install the all required configuration on machine. But i am getting COMException: Exception from HRESULT: 0x80071CC0 error. Also i restarted the window after installing all dependency. Please provide the solution. ![image](https://user-images.githubusercontent.com/83421728/116599822-33fdad80-a946-11eb-8971-9dc4b595938e.png)
chef/supermarket
122788773
Title: Increase visibility of adoption process Question: username_0: #1031 added the available for [adoption page](https://supermarket.chef.io/available_for_adoption), but you have to go off into the weeds to find it. (Under the temporary announcements drop down, or go off to 'Cookbooks' and scroll down to the bottom). I've looked and I don't think these things happen currently: * email followers when a cookbook is put up for adoption * include a cookbook available for adoption notification in the cookbook activity on the home page * include adopted cookbook events in the available_for_adoption page for 3-6 months. * put dates on the available_for_adoption page that show when a cookbook was made adoptable both in date and time since date. Answers: username_1: @username_2 Feel free to ping me about this one when you are ready to work on it :) username_1: Things I think we can implement: 1) Emailing followers when cookbook put up for adoption 2) When cookbooks are put up for adoption, show this in the dashboard of followed cookbooks activity 3) Change adoptable cookbooks list to be chronological 4) Put date a cookbook was put up for adoption on the adoptable cookbooks page username_2: My current plan is to break this out into three different parts and take the following steps: 1. Emailing followers when a cookbook goes up for adoption ✅ 2. Adding `up_for_adoption_since` to the Cookbook model and replacing the `up_for_adoption` boolean currently in the model, handling these migrations. 3. Refactoring out the `update` controller action into `update` and `adopt` actions in Cookbook. This will make all logic surrounding updating and putting up for adopting clearer and will work well with the cookbook/tools and the existing polymorphic path infrastructure. username_3: ![i love this plan](https://cloud.githubusercontent.com/assets/517302/15436807/a3e95abc-1e91-11e6-8938-c2f0505435ad.gif) I love this plan. I'm excited to be a part of it. Status: Issue closed
Flowpack/Flowpack.SimpleSearch
133798025
Title: Indexing nodes with custom defined roles Question: username_0: After updating to the newest version i got a problem with indexing nodes with a custom defined role. In version 1.3 it works well. Version: 1.3.2 Neos: 2.1.3 Flowpack.SimpleSearch: 1.3.2 I configured a custom role called SpanishEditor, which has only access to edit content in the spanish dimension. You find the policy file attached: [Policy.txt](https://github.com/Flowpack/Flowpack.SimpleSearch/files/131317/Policy.txt) If any user tries to pusblish or save some changes, the following exception occurs: Exception: Access denied for method Method: TYPO3\TYPO3CR\Domain\Service\Context::validateWorkspace() Exception file: [exception.txt](https://github.com/Flowpack/Flowpack.SimpleSearch/files/131278/exception.txt) Did i miss something in the role configuration or is it a bug in the package? Would be nice to know. Thanks for a hint! Answers: username_0: Ok, its related to: https://github.com/Flowpack/Flowpack.SimpleSearch.ContentRepositoryAdaptor/issues/13 Status: Issue closed
coursier/coursier
594981505
Title: Can't use Microsoft Azure DevOps Maven Repository as Resolver Question: username_0: I don't want to do cross-posting but looks like there's a fix need from Coursier as well, https://github.com/sbt/sbt/issues/5492 Answers: username_1: Here are steps and problem from https://github.com/sbt/sbt/issues/5492 ## steps Add a Microsoft [Azure DevOps Maven](https://docs.microsoft.com/en-us/azure/devops/artifacts/get-started-maven?view=azure-devops) repository as resolver: ```sbt lazy val scala211 = "2.11.12" ThisBuild / organization := "com.foo" ThisBuild / scalaVersion := scala211 ThisBuild / useCoursier := true name := "baz" crossScalaVersions := Seq(scala211) libraryDependencies ++= Seq( "foo" %% "bar" % 0.1 % "provided" ) resolvers += "nabu-releases" at "https://pkgs.dev.azure.com/foo/Foo/_packaging/foo-releases/maven/v1" credentials += Credentials(Path.userHome / ".sbt" / ".credentials") ``` sbt version ``` [info] This is sbt 1.3.9 ``` ## problem Coursier throws: ``` [warn] Note: Unresolved dependencies path: [error] stack trace is suppressed; run last update for the full output [error] (update) sbt.librarymanagement.ResolveException: Error downloading com.foo:bar_2.11:1.0.0 [error] Not found [error] Not found [error] not found: /Users/username_0/.ivy2/local/com.foo/bar_2.11/1.0.0/ivys/ivy.xml [error] not found: https://repo1.maven.org/maven2/com/foo/bar_2.11/1.0.0/bar_2.11-1.0.0.pom [error] checksum format error: /Users/username_0/Library/Caches/Coursier/v1/https/pkgs.dev.azure.com/Foo-backend/Foo/_packaging/Foo-releases/maven/v1/com/foo/bar_2.11/1.0.0/bar_2.11-1.0.0.pom.sha1 ``` ## notes ``` sbt:foo> show credentials [info] * FileCredentials("/Users/username_0/.sbt/sonatype_credential") [info] * FileCredentials("/Users/username_0/.sbt/.credentials") ``` ``` cat /Users/username_0/.sbt/.credentials realm=foo host=pkgs.dev.azure.com user=foo password=xxx ``` The checksum actually has a HTML from Azure: [Truncated] If I turn off Coursier```ThisBuild / useCoursier := false``` I have the following error from ivy: ``` sbt:foo> test [error] Unable to find credentials for [https://pkgsprodcus1.pkgs.visualstudio.com/ @ pkgs.dev.azure.com]. [error] Is one of these realms misspelled for host [pkgs.dev.azure.com]: [error] * Foo-releases [error] Unable to find credentials for [https://pkgsprodcus1.pkgs.visualstudio.com/ @ pkgs.dev.azure.com]. [error] Is one of these realms misspelled for host [pkgs.dev.azure.com]: [error] * Foo-releases [error] Unable to find credentials for [https://pkgsprodcus1.pkgs.visualstudio.com/ @ pkgs.dev.azure.com]. [error] Is one of these realms misspelled for host [pkgs.dev.azure.com]: [error] * Foo-releases [error] Unable to find credentials for [https://pkgsprodcus1.pkgs.visualstudio.com/ @ pkgs.dev.azure.com]. [error] Is one of these realms misspelled for host [pkgs.dev.azure.com]: [error] * Foo-releases [error] Unable to find credentials for [https://pkgsprodcus1.pkgs.visualstudio.com/ @ pkgs.dev.azure.com]. [error] Is one of these realms misspelled for host [pkgs.dev.azure.com]: [error] * Foo-releases ``` username_2: This seems to be related to #1421 and the associated PR #1948 The fix in #1948 would **not** be sufficient to fix this, but here's an interesting nugget: ```text curl -v https://pkgs.dev.azure.com/nabu-backend/Nabu-Bots/_packaging/nabu-releases/maven/v1/com/modak/spark-profiler_2.11/1.0.6-2.4.0/spark-profiler_2.11-1.0.6-2.4.0.pom.sha1 2>&1 |grep -i www-authenticate < www-authenticate: Bearer authorization_uri=https://login.windows.net/fdcf6fad-c3f0-4c66-8ca1-1e3aaac65150 < www-authenticate: Basic realm="https://pkgsprodcus1.pkgs.visualstudio.com/" < www-authenticate: TFS-Federated ``` AIUI, the current code in https://github.com/coursier/coursier/blob/8431fef09561afb97551a844a56ede0481a930f2/modules/cache/jvm/src/main/scala/coursier/cache/CacheUrl.scala#L437 would only see the last line ("TFS-Federated") according to https://docs.oracle.com/javase/8/docs/api/java/net/URLConnection.html#getHeaderField-java.lang.String-, and would therefore not be able to use it at all. The correction would be to use the `getHeaderFields` API and iterate through all instances of the www-authenticate header until we find a valid Basic auth challenge. And then perform a coursier/sbt-coursier/sbt release train for general availability. username_3: https://github.com/sbt/sbt/issues/5492#issuecomment-850692474 username_3: This configuration works in sbt ``` csrConfiguration := { val conf = csrConfiguration.value val resolvers = csrResolvers.value val adoRepoHost = "pkgs.dev.azure.com" val adoCredentialsOpt = credentials.value.collectFirst { case creds: DirectCredentials if creds.host == adoRepoHost => creds } val newConfOpt = adoCredentialsOpt.map { adoCredentials => val auths = resolvers .collect { case repo: MavenRepository if repo.root.startsWith(s"https://$adoRepoHost/") => repo.name -> Authentication(adoCredentials.userName, adoCredentials.passwd) } auths.foldLeft(conf) { case (conf, (repoId, auth)) => conf.addRepositoryAuthentication(repoId, auth) } } newConfOpt.getOrElse(conf) } ``` username_4: That seems to allow `sbt update` and `compile` but importing the project into intellij or vscode+metals gives the checksum error still username_3: @username_4 I updated my comment, you need to specify the updateClassifiers configuration as well. ``` updateClassifiers / csrConfiguration := csrConfiguration.value .withClassifiers(Vector("sources")) .withHasClassifiers(true), ```
anuraghazra/github-readme-stats
1160549766
Title: [Re-architecture]: Rendering Of Cards Question: username_0: ## The Problem The current architecture of rendering the card follows a more traditional approach where we concatenate raw svg strings to build a valid SVG markup. - Build svg template - Manipulate string - Interpolate dynamic props - Render card #### Problem 1 The problem with this approach is that it's not very scalable specially when we consider the fact we have to handle precise pixel values and by default SVGs don't support flexible layouts, text wrapping and other vital layout engine features which are needed to build a complex card layout. This approach has implemented since the start of github-readme-stats which worked fine early on, but as the cards grow in complexity it's becoming more and more harder to manage the layouts. For example we had implement all sorts of hacky workarounds to use [FlexBox](https://github.com/username_0/github-readme-stats/blob/6b6162588109c6f4fc262fe53bd46c288814c690/src/common/utils.js#L148), [TextWrapper](https://github.com/username_0/github-readme-stats/blob/6b6162588109c6f4fc262fe53bd46c288814c690/src/common/utils.js#L223), [MeasureText](https://github.com/username_0/github-readme-stats/blob/6b6162588109c6f4fc262fe53bd46c288814c690/src/common/utils.js#L304) and other features which are already present in any modern layout engine. #### Problem 2 Second point is that manipulating strings like we do now poses a huge burden for us to properly interpolate all the pixel values which are needed and it's a manual process. There is no %, rem, em, or flexible units to work with in SVGs. Every single element on the SVG has to be carefully crafted to maintain their size/width/height/position. It's not fun to build cards. ## Solution The solution which I'm proposing to implement is by taking advantage of [foreignObject](https://developer.mozilla.org/en-US/docs/Web/SVG/Element/foreignObject#example) in SVG. Advantages of foreginObject - Easy to build cards, we have all the power of flexible layouts, CSS & browser's layout engine. - Browser support is [good](https://developer.mozilla.org/en-US/docs/Web/SVG/Element/foreignObject#browser_compatibility) - Trivial to build templates for cards ### Low hanging There are few things which we have to check first, one of them is if foreignObject is supported in github mobile or not, That's an issue because we don't know what kind of engine the github mobile app using. Other thing is checking various browsers for compatibility and consistency of the styling and layout. Even though browser support is good, we still need to do some testing to make sure it's consistent across browsers. ## Implementation plan Over the next few weeks I'll be start implementing this feature and deploy it in a vercel preview link to test things out. ## Feedback I'll be needing feedback from the community and the core members (@username_1) to implement this & the feasibility of this. Let me know what you folks are thinking about this approach we can discuss more on the pros and cons. Answers: username_1: @username_0 I like your proposal wasn't aware of [foreignObject](https://developer.mozilla.org/en-US/docs/Web/SVG/Element/foreignObject#example) object, but it looks like the right tool for the job! I think it might clean up the codebase! And if not, it will save us hours trying to interpolate pixel values 🥲. Feel free to tag me if you need my help.
tkuri/papers
645957999
Title: Learning to Super Resolve Intensity Images from Events Question: username_0: ## 論文概要 画素毎に輝度の変化量をバイナリで出力するイベントカメラは高速・低消費電力が強みだが解像度が低いのが課題であった。そこでイベントストリームから高解像度・HDRの画像を直接生成するためのEnd2Endのネットワークを提案。パイプライン型の従来手法よりも高精度に生成可能。 イベントストリームからEnd2Endでの超解像生成としてはFirst Effort。時間的な不連続さを緩和するためにRNNを採用している。 ## 感想 結果がわかやすく、First Effortとしては手法が洗練されている。イベントストリームという特殊な入力信号だが、アーキテクチャ自体はシンプルなのも良い。CVPR2020オーラル。
appertly/cleopatra
156643178
Title: Incorrect usage of `async` tests Question: username_0: First I would like to say thanks for using HackUnit! I'm pretty happy with where HackUnit is currently as well as where it could go, but I only have my own use cases covered so far. I also love the concept of your package. I've attempted a few times to make a similar one and have never been happy with the result. The main point of this issue is that you are using the `async` keyword for all of your tests, even though none of them actually make use of Hack's async functionality. Basically to make `async` tests useful, the test method must `await` for something, which is really only useful for real IO (to disk or network). I'm not sure if there's a performance penalty for making functions async without actually using the async feature, but it would confuse me as a contributor to your project. Answers: username_1: Since there's really no problem with putting the async modifier on every method, I just did so for consistency sake. I have several projects using HackUnit and a few of them *do* have awaitable methods that get called. To alleviate your concern about confusing contributors, do you think it'd be useful to include a note explaining this in a CONTRIBUTING.md file?
Azure/azure-sdk-for-ruby
229752713
Title: Waiting new messages from service bus queue continuously Question: username_0: I m getting the queue message by following code: `message = sbs.receive_queue_message("test-queue", { :peek_lock => false })` It listens about 60 seconds then get time out. I'd like to do it in listener to process the messages without break the listener. It means after get the new message `receive_queue_message` could be renew itself and wait for new message without time out. Answers: username_1: @username_0 we're not actively working on Service Bus code from azure gem, feel free to send us a PR to address this issue, we'd be happy to review it. username_2: @username_0 You can reopen the issue if you are planning to submit the PR. Closing the issue Status: Issue closed
othreecodes/Zapier-Twitter-Jobs
362534687
Title: RT @inioge: Specialization isn't encouraged in Nigeria's tech ecosystem, employers are looking for an all in one developer who will do a jo… Question: username_0: RT @inioge: Specialization isn't encouraged in Nigeria's tech ecosystem, employers are looking for an all in one developer who will do a jo… <a href="http://twitter.com/download/android" rel="nofollow">Twitter for Android</a> https://twitter.com/153501220/status/1043071646184210432 Fri Sep 21 09:37:14 +0000 2018
TheAlgorithms/MATLAB-Octave
584650829
Title: Add computer vision algorithm in matlab Question: username_0: I have done blob detection in Matlab, so should I add this algorithm in the repository. Answers: username_1: Go ahead. However, please be patient with the review process. Thanks! username_1: I am closing this becoz #26 Feel free to open it later for any other issues. Status: Issue closed
pavankarthiktanikanti/FoodOrderingAppBackend
569462361
Title: Category Controller Get Category by Id - “/category/{category_id}” Question: username_0: - It should be a GET request. - This endpoint must request the following value from the customer as a path variable: - Category UUID - String - If the category id field is empty, throw “CategoryNotFoundException” with the message code (CNF-001) and message (**Category id field should not be empty**) and their corresponding HTTP status. - If there are no categories available by the id provided, throw “CategoryNotFoundException” with the message code (CNF-002) and message (**No category by this id**) and their corresponding HTTP status. - If the category id entered by the customer matches any category in the database, it should retrieve that category with all items within that category and then display the response in a JSON format with the corresponding HTTP status. Also, the name searched should not be case sensitive. - Here is a sample JSON response. { "id": "2ddf59f6-ecd0-11e8-8eb2-f2801f1b9fd1", "category_name": "Sweet Dish", "item_list": [ { "id": "24614e76-a238-11e8-9077-720006ceb890", "item_name": "vanilla icecream", "price": 230, "item_type": "VEG" }, { "id": "24615498-a238-11e8-9077-720006ceb890", "item_name": "chocolate icecream", "price": 240, "item_type": "VEG" }, { "id": "2ddf4524-ecd0-11e8-8eb2-f2801f1b9fd1", "item_name": "butter paneer", "price": 190, "item_type": "VEG" }, { "id": "2ddf4768-ecd0-11e8-8eb2-f2801f1b9fd1", "item_name": "chillie chowmine", "price": 210, "item_type": "NON_VEG" } ] }<issue_closed> Status: Issue closed
resque/resque-scheduler
103001155
Title: unclear and very confusing docs Question: username_0: Resque.enqueue_in(5.days, SendFollowUpEmail, :user_id => current_user.id) With that, `enqueue_in(5.days` is understood to mean a DELAYED job. User continues to read docs, he finds "Also supported is `Resque.enqueue_at` which takes a timestamp to queue the job". !!!!!????????? This is a total bullshit man. If you're going to distinguish between delayed and scheduled, make so. Don't confuse user. I have read your doc 5 times and I still don't understand clearly how to: 1. Invoke a cron-like job 2. Schedule a delayed one-time job. Also, the language should be changed. For non-native, it's a catastrophe to read all that. First of all, make a clear separation between cronjob and delayed job. Don't name it a "scheduled" job, it's very confusing, name it either "cronjob" or "recurring" or "repeating". A delayed job is what it is, one-time, no need to change that. Thanks. Answers: username_1: +1 this is indeed somewhat confusing. Looked at the code to find out that `enqueue_in`/`enqueue_at` are the same with the exception that the former is "Identical to enqueue_at but takes number_of_seconds_from_now instead of a timestamp.". This difference doesn't really come across in the README. username_2: @username_0 This is the first time I've read your full comment. I'm hoping that since it's been over eight months that you're not as angry anymore, or that perhaps you're not even using resque-scheduler anymore. In any case, you can probably figure out that this project is in need of contributors. *Contributors*. Ranting in an issue might make *you* feel better, but it doesn't do much to improve the situation for everyone else. Patches welcome, as always. Stay safe out there. 💖 🎷 🐻
LartTyler/MHWDB-Docs
316610364
Title: Defense property in Weapons Question: username_0: Hi, Some weapons have defense bonuses (I.E. Barroth Blaster 3), but they are not being listed. Are you planning on adding them as well? Thanks, and keep up the good work! Answers: username_1: I am now! I'd honestly overlooked them, but it should definitely be in there. I'll get it added to the backlog. Thanks! Status: Issue closed username_1: This has been added in [v1.9.1](https://github.com/username_1/MHWDB-Docs/releases/tag/1.9.1).
elBukkit/MagicPlugin
167418268
Title: Upgrade spells Question: username_0: I can't found what is the command to upgrade a enchant, can you please tell me what is it? thanks :) (I'm using the potter pack) Answers: username_1: As in you want to level up a spell, add more spells to a wand, or rank up a wand? username_0: level up a spell, thanks for your help :) username_1: Leveled up spells are unique spells on their own. So you simply add them to a wand, they will replace any lower-levelled versions already on the wand. They have special "key names", so like "/wand add apparate|2" will add apparate level 2. The character there is a pipe, normally shift+backspace. username_0: Thanks :) Status: Issue closed
TempoX-Ltda/Gestao-Linha-UV
629496920
Title: Crop da imagem na ROI detectada pelo TemplateMatching Question: username_0: Boa tarde a todos. O @MAd0G está me auxiliando a partir de hoje com a função de pós-processamento na classe de TemplateMatching para posterior integração com o módulo de HistogramCalculation Answers: username_0: Caso alguém mais esteja disponível para auxiliar na integração, fico a disposição para repassar mais detalhes. username_0: - [ ] Função de crop da área do retângulo desenvolvida - [ ] Função de crop da área do retângulo testada
keycloak/keycloak
1166224462
Title: Planning issues for 17.0.1 Question: username_0: ### Describe the bug SA ### Version AS ### Expected behavior _No response_ ### Actual behavior _No response_ ### How to Reproduce? _No response_ ### Anything else? _No response_ Answers: username_1: Would love to also see these inside: https://github.com/keycloak/keycloak-web/pull/288 / https://github.com/keycloak/keycloak-web/pulls/dguhr -> merge the PRs for the openshift starter to work again https://github.com/keycloak/keycloak/pull/10347 -> makes troubleshooting way easier when no write rights on dirs https://github.com/keycloak/keycloak/pull/10665 -> fix start-dev in windows. username_2: Can we also include: * https://github.com/keycloak/keycloak/issues/10411 * https://github.com/keycloak/keycloak/issues/10579 * https://github.com/keycloak/keycloak/issues/10400 * https://github.com/keycloak/keycloak/issues/10338 username_3: @username_0 would be possible to include https://github.com/keycloak/keycloak/issues/10678 and the RESTEasy upgrade https://github.com/keycloak/keycloak/issues/9599 ? username_1: @username_2 @username_0 should we do a round of checking the performance regressions we observed before 17.0.1 then? username_5: -- <NAME> Principal Software Engineer Red Hat <https://www.redhat.com/> ***@***.*** IM: username_5 @RedHat <https://twitter.com/redhat> Red Hat <https://www.linkedin.com/company/red-hat> Red Hat <https://www.facebook.com/RedHatInc> <https://www.redhat.com/> username_3: @username_1 FYI PR https://github.com/keycloak/keycloak/pull/10665 was closed by the author. username_1: @username_3 was superseeded by https://github.com/keycloak/keycloak/pull/10805 - now with a commit message that includes the issue #. so its merged username_1: Update to 2.7.5 is done via #10823, can be ticked off. Status: Issue closed
cerkut/wpAudioPlugin
427841574
Title: 01-Basics/myWidth.m working ok? Question: username_0: People could not hear the width effect on myWidth.m, but the code looks correct? https://github.com/username_0/wpAudioPlugin/blob/master/01-Basics/myWidth.m Remedy: increase the max width value from 4 to 10? Status: Issue closed Answers: username_0: Works ok on Ableton Live 10 with head phones
dotnet/codeformatter
119145925
Title: Support DNX Question: username_0: This is a sort of bigger feature request version of issue #172. In the brave new world, DNX-based projects and solutions are going to be more and more common. I haven't yet figured out a way of telling `codeformatter` to format all the code within a DNX solution or DNX project, as you can't specify individual source files (entirely reasonably, given that it's doing more than formatting). Answers: username_1: This is definitely in the works now. The intent is to add support for project.json based projects. A couple of other pieces need to fall into place before we get full parity with MSBuild based project. But this will be coming. Sent using OWA for iPhone username_2: The workspace implementation should be good enough (hopefully) https://github.com/dotnet/cli/tree/master/src/Microsoft.DotNet.ProjectModel.Workspaces. We need to start producing packages for these so we can consume them here. username_3: Now that VS2017 is released, can we get a version of code formatter that will work on .NET Core based csproj projects (#256), and on a machine with only VS2017 installed, i.e. where .NET 4.5.2 isn't installed (#255)? username_1: @username_3 it's on my radar to do. Still a lot of VS 2017 ship work going on right now. Once that calms down though going to be taking a look at this. username_3: I'll keep my fingers crossed that things calm down soon then ;-) username_4: .net core formatter would be great...
caktus/margarita
191760581
Title: Improve the Let's Encrypt process Question: username_0: It's about time we start doing this in a way that isn't so fragile. - pip install certbot (the new name for the letsencrypt automated cert script) - clean up the old git checkout of letsencrypt - remove the old renewal cron job - add a new cron job that executes `certbot-auto renew`, instead of having to match all of the subdomains and webroots used in the initial invocation - use whichever nginx reload/restart command that won't silently fail Status: Issue closed Answers: username_1: Thanks @username_0. This work is now included in release 2.2.0.
electron-userland/electron-builder
184204424
Title: Icons missing for uninstall and install warning Question: username_0: <!-- Which version of electron-builder are you using? --> * **Version**: 7.12.2 <!-- What target are you building for? --> * **Target**: Squirrel The installer, shortcut and application while running all have the correct icon. When I'm going to add/remove programs the icon is missing. ![uninstall icon missing](https://cloud.githubusercontent.com/assets/951290/19557349/94d90016-96c5-11e6-8da2-120e899f52f6.PNG) When installing you get a dialog and that dialog is also missing the icon. ![install icon missing](https://cloud.githubusercontent.com/assets/951290/19557366/a8f919f0-96c5-11e6-9dad-715dd49a6151.PNG) We use a gulp script to create the installer: ```js builder.build({ targets: Platform.WINDOWS.createTarget('Squirrel', Arch.ia32);, publish: true, githubToken: 'xxx', devMetadata: { build: { win: { loadingGif: 'resources/windows/installer_animated_' + lang + '.gif', msi: false, icon: 'resources/windows/icon.ico' } }, directories: { output: 'dist/' + lang } } }) ``` Package.json ```js { "repository": "https://github.com/gynzy/xxx-releases", "licence": "UNLICENSED", "private": "private", "build": { "appId": "com.xxx.xxx", "category": "public.app-category.productivity", "extraResources": "xxx-resources" }, "directories": { "buildResources": "resources", "app": "build" }, "devDependencies": { "asar": "0.12.3", "chai": "3.5.0", "electron": "1.4.3", "electron-builder": "7.12.2", "electron-mocha": "3.1.1", "fs-jetpack": "0.9.2", "gulp": "3.9.1", "gulp-batch": "1.0.5", "gulp-less": "3.1.0", [Truncated] "gulp-replace": "0.5.4", "gulp-watch": "4.3.10", "request": "2.75.0", "run-sequence": "1.2.2", "semver": "5.3.0", "yargs": "6.0.0" }, "scripts": { "build": "gulp build", "install-app-modules": "cd app && npm install && cd ..", "create-installer": "gulp create-installer", "preinstall": "npm run install-app-modules", "start": "gulp start", "pretest": "gulp build --env=test", "test": "electron-mocha build --renderer" } } ``` Another question: how do we change the help link and add update information link? (see first screenshot) We want them to be different. Answers: username_1: Hmm... Please try to set https://docs.npmjs.com/files/package.json#homepage in your app package.json Status: Issue closed username_0: We need the squirrel auto-update and have it working. We've released our app a month a go and a lot of users have installed it. So we cannot change to NSIS because that'll break our working auto-update mechanism. I guess changing the homepage url works, but how do we get two different urls for help link and update information link, some applications even have a 3rd support link. I'd really like to specify all 3 to different urls. username_1: @username_0 Filed as separate issue #839 Please note — I cannot promise time frame. username_0: Thank you very much, I can contribute if you'd like. But we don't have to rush 😃 . username_2: Did you ever figure out a fix for the missing icon in the add/remove programs screen (see the OP's first image). I'm using an NSIS installer with electron-builder v11.2.1. and the installer itself has the correct icon (unlike the OP's second image), but the add/remove programs screen isn't showing the icon. username_0: Yes I fixed it by using iconUrl but we used Squirrel. username_2: Thanks for the response. The squirrel is working just fine for me as well, but I'd like to be able to use NSIS installer. This seems to be the only remaining issue. Should I create a separate issue for it, given that it is not relevant to the squirrel target? username_1: @username_2 Yes, please create separate issue. username_3: He created the issue here: https://github.com/electron-userland/electron-builder/issues/1108 I'm linking it so that future readers don't have to go searching for it.
MicrosoftDocs/windows-driver-docs
515194878
Title: "downloads the package from WMIS" Question: username_0: I am troubleshooting this process. When you refer to the information in my subject, could you include more detail on that - where exactly is WMIS - URL, ports? event logs of all our devices have messages in the logs saying the devices can't access so I'm assuming we need a Proxy/firewall config change to allow it, but can't find that required information. error message being - "A connection to the Windows Metadata and Internet Services (WMIS) could not be established." --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9d5f648a-f925-7bca-7bad-b97fded54baf * Version Independent ID: e6a51140-c16d-8096-7888-2536b23a3ebc * Content: [Installing Device Metadata Packages from WMIS - Windows drivers](https://docs.microsoft.com/en-us/windows-hardware/drivers/install/installing-device-metadata-packages-from-wmis#feedback) * Content Source: [windows-driver-docs-pr/install/installing-device-metadata-packages-from-wmis.md](https://github.com/MicrosoftDocs/windows-driver-docs/blob/staging/windows-driver-docs-pr/install/installing-device-metadata-packages-from-wmis.md) * Product: **windows-hardware** * Technology: **windows-devices** * GitHub Login: @username_2 * Microsoft Alias: **username_2** Answers: username_1: @username_0. As per my knowledge, WMIS online sever contains OEM certified device drivers for particular drivers username_2: Thanks for the feedback and apologies for the delay. It's possible that group policy might be set to prevent the WMIS acquisition. I'm afraid I don't have any expertise with it though, so if this is still an issue, I'd recommend posting in the following forums, where other driver developers and Microsoft engineers can be found. The community there is quite responsive. * [Windows Hardware WDK and Driver Development](https://social.msdn.microsoft.com/Forums/windowsdesktop/en-US/home?forum=wdk) * [OSR's ntdev List](http://www.osronline.com/showlists.cfm?list=ntdev) Again, thanks for reaching out. Best of luck with this issue. Status: Issue closed
geonetwork/schema-plugins
115886769
Title: RNDT: multiple iPA Question: username_0: Current implementation only allows to handle a single iPA code, so the catalog is bound to a single Organization. Add the ability to handle multiple iPA/Organizations. Answers: username_0: Closed in 2.10.x with 2789e2e Status: Issue closed
Heigvd/Wegas
789961077
Title: PMG _ Onglet ressources, coûts des heures non travaillées / pourrait travailler Question: username_0: Il y a un truc que je ne comprends pas dans la légende. C'est quoi ce coûts des heures non travaillées ? Il me semble qu'on pourrait simplement le retirer ou bien ? Autre question: le "pourrait travailler" n'est pas lié au mode Agile uniquement ? <img width="876" alt="Capture d’écran 2021-01-20 à 14 01 44" src="https://user-images.githubusercontent.com/1556490/105178368-34f19880-5b28-11eb-885b-d9b5eb65e5dc.png"> Answers: username_1: Pourrait travailler, c'est si tu mets ASAP ![image](https://user-images.githubusercontent.com/6583528/105179260-64ed6b80-5b29-11eb-8d55-270a72129dc5.png) username_1: "coûts des heures non-travaillées", c'est un effet de bord de la mise à jour de l'onglet projet. ça devrait être marqué "heures non-travaillées". On va corriger ce texte. La légende est là car quand une ressource facture des heures non travaillées, on le montre dans son planning : ![image](https://user-images.githubusercontent.com/6583528/105179528-bbf34080-5b29-11eb-8d7b-58b8eaaa5302.png) username_0: OK ! username_0: OK ! T'as le droit de dire que je suis insupportable, mais tu crois qu'on pourrait l'afficher seulement si on a un ASAP activé ? Si pas possible, pas de soucis username_1: Ça doit pas être bien compliqué. Je regarde plus tard si je m'ennuie Status: Issue closed
thoughtbot/administrate
149430804
Title: Overriding the index page for a specific model. Question: username_0: I'm trying to get the index listing for a specific model to show an extra link where the "Edit" and "Destroy" buttons are located. The model name being `Foo`, I've created a view within `app/views/admin/foos/_collection.erb`. It works great on the root `Foo` listing page. However, an instance of `Foo` can have many of `Bar`. When I select a `Foo`, the detail page shows a list of `Bar` object there, and unfortunately the same collection template is used. Is that a bug, or am I missing something? Answers: username_1: It's because the `render collection partial` looks for the partial via a relative path instead of fully qualified. How did you solve this problem @username_0? I'm debating generalizing that functionality to all models username_2: there's a solution for this issue in #522, which has been working for me... username_0: @username_1 no, and we ended up completely rewriting the view. We're thinking about abandoning Administrate for multiple reasons now, one of them being this, the other one being the fact that Administrate is super-confused when asked to sort on "virtual columns". There should be a flag to tell Administrate about those. username_3: A potential solution for this would be to pass an additional parameter to the `_collection` partial. I'll be happy to look into a PR providing this (or a different solution). Status: Issue closed
godotengine/godot
290264024
Title: Something wrong with Depth/Parallax Mapping in Spatial Material Question: username_0: <!-- Please search existing issues for potential duplicates before filing yours: https://github.com/godotengine/godot/issues?q=is%3Aissue --> **Godot version:** Godot 3.0 RC 2 **OS/device including version:** Ubuntu 16.04 64-bit **Issue description:** The default parallax mapping change bump angle when you rotate scene view. I changed this line of code: `vec3 view_dir = normalize(normalize(-VERTEX)*mat3(TANGENT,-BINORMAL,NORMAL));` to this one (removed minus) `vec3 view_dir = normalize(normalize(-VERTEX)*mat3(TANGENT,BINORMAL,NORMAL))` and it works but I don't know it is real fix. **Steps to reproduce:** Open the project and slowly rotate scene view above the planes **Minimal reproduction project:** [Test5.zip](https://github.com/godotengine/godot/files/1649698/Test5.zip) Answers: username_1: I had the same problem, I thought it was a problem with my parallax maps at first. Good to know there's a workaround for now. username_2: Looks like parallax also affected by this #16009. I did some tests and I see that different inbuilt geometry (planes, quads) shows different parallax behavior. For example material with fix in test scene from this issue will work for inbuilt Plane, but will not work for Quad and for imported geometry. username_3: Unfortunately, this fix does not work with all meshes. I wanted to submit a pull request (see [this commit](https://github.com/username_3/godot/commit/fd54b81adcf3e68512ea6a03406f8713dd1bb7dd)) also incorporating changes from https://github.com/godotengine/godot/issues/15794, but while it looks good on a cube exported from Blender, it looks broken on the test mesh from the Material Tester demo. **Good depth mapping:** ![depth_mapping_good](https://user-images.githubusercontent.com/180032/38200120-6b17e21e-3693-11e8-8840-411c8f67a821.png) **Broken depth mapping:** ![depth_mapping_bad](https://user-images.githubusercontent.com/180032/38200119-6af78c26-3693-11e8-96ee-b1da840c59eb.png) username_2: Tangents are different everywhere. They affect how normal maps and parallax works. I've tried to unify tangents in my fix #16096 (not merged) to match those from Material Tester and to work correctly with OpenGL style normal maps, like in Material Tester. (But maybe something was changed since then?) username_4: I am not really sure about this because, as mentioned before, it really depends on the application, Maybe options to flip binormal/tangent should be added.. username_5: After doing a lot of experimenting with parallax maps in Godot, I also could only get them to work using the same method as @username_0 , by removing the minus from the binormal code. As noted by @username_2 in #16009 parallax mapping will not work on Godot's native primitives. Also, care must be taken to use an actual depth map with white as the furthest depth, as Substance and other engines/renderers use a height/displacement map which has white as the closest depth. I have attached a project file with 2 parallax materials using the aforementioned steps. Since I posted this on twitter and my website, I have heard no feedback that this approach doesn't work on various hardware configurations. Hopefully this easy fix (removing the negative sign) can be added before 3.1 is released. [parallax_heads.zip](https://github.com/godotengine/godot/files/2448878/parallax_heads.zip) ![parallax_heads_screenshot1](https://user-images.githubusercontent.com/7741669/46515217-d1cf8f80-c8a5-11e8-884c-7ccc557c1b7a.jpg) username_4: I am not quite sure what to do with this one. Technically, it can fail completely depending on the geometry and how it was created, given tangent space is not warranted to be universal. I will add an option to flip x and y in tangent space to make sure it can work with everything, but it needs to be done manually. Status: Issue closed
luechtdiode/KuTu
285318121
Title: Für die Erfassung optimierte Reihenfolge in der Riegen-Auswahlliste Question: username_0: Die Reihenfolge der Riegen in der Auswahl-Liste ist gleich sortiert, wie jene, die für den Druck der Riegenblätter aufbereitet wird. Bei der Druckversion macht die Reihenfolge Sinn, weil dort die Riegenblätter pro Gerät-Posten verteilt werden: `Durchgang -> Gerät -> Runde`. Bei der Erfassung der Riegenresultate aber werden pro Runde die Riegenblätter mit den Resultaten der gerade geturnten Geräte eingesammelt: `Durchgang -> Runde -> Gerät`. Mit dieser Optimierung kann die Liste bei der Erfassung der Resultate möglichst lückenlos von oben nach unten erfasst werden.<issue_closed> Status: Issue closed
Azure/azure-sdk-for-js
434005037
Title: [Service Bus] Fix tests' README instructions Question: username_0: The tests' README instructions for setting up for use with CLEAN_NAMESPACE option enabled does not seem complete. Receiving error that permissions are not existing for deleting entities even though the role assignment is 'owner' Answers: username_1: Can you verify if the app has `Owner` role in the `Role assignments` section under `Access control (IAM)` tab in your namespace in the portal ? Status: Issue closed username_1: [ Discussed offline with @username_0 ] - Issue - `does not have authorization to perform action ` - Cause - value of `RESOURCE_GROUP` environment variable is incorrect. - Setting it to the actual corresponding RESOURCE_GROUP fixed the issue. Closing the issue!
openaddresses/openaddresses
603559927
Title: Statewide Wyoming Update! Question: username_0: Current Source: http://results.openaddresses.io/sources/us/wy/statewide @username_1 do you recall where this came from? Not important, just curious. Thanks! New: https://services3.arcgis.com/r0iJ85SKZ4zAzz3P/ArcGIS/rest/services/Wyoming_Parcels_for_2019/FeatureServer/0 Source: https://www.arcgis.com/apps/webappviewer/index.html?id=4bb9a66f7287402b8f650aa9f21d3fa5 **https://twitter.com/vonparcel sent to me ;) Answers: username_1: See the note here: https://github.com/openaddresses/openaddresses/blob/master/sources/us/wy/statewide.json username_2: @username_1 @username_0 not sure what to do here, am I adding the MapServer that @username_0 found, or should we recache and host per the note.
pwweb/artomator
467683181
Title: Standalone controller errors Question: username_0: When creating standalone controllers without creating a model (i.e. for the standard laravel user model), the following issues are observed: - File name is incorrect (using User instead of UserController) - Class name is incorrect (using User instead of UserController) - Line 55 of stub is not replaced, leaving variable {{schema_data}} - Line 98 of stub is not replaced, leaving variable {{schema_data}} Answers: username_1: Is this when you use artomator:all or artomator:controller? username_0: When using artomator:controller username_1: Going to need a little more information on this: what's the exact command that you ran to get the issue? Also, when prompted did you specify that a model should be created? Looking at this if you run: `artomator:controller Standalone/TestController` then it works as expected. I think I'll adda a check in the function to see if the `Controller` words has been provided on the end of the name and then add if not. username_0: Must have been that the call was without the Controller. I'd agree to check for Controller and if missing adding it. username_1: Yes, if the command is called without controller on the end then it won’t name it right. In the “all” command it’s added to the call for you. However the Schema replacements still doesn’t make sense. If you select “no” to the model generation, the command should pick the stub file without the schema. Puzzling. I’ve added the code to the dev branch for testing, but there is a breaking change at the moment, so double check the readme before using it. Status: Issue closed username_1: This has been corrected in the latest version v0.2.0-beta.
bastikr/python-ubx
361942912
Title: Usage examples and documentation Question: username_0: This library seems perfect, but I need some help understanding how to use it. It would be helpful to have some usage examples or other documentation. E.g., example of connecting to UBX over serial, and reading data from a message. Answers: username_1: @username_3 First off, thank you for making this library! I'm working on adapting it for Micropython and it's going to save me a lot of time! I've read through the code and still can't figure out for the life of me what it's expecting as input. I've tried feeding it a byte at a time, a stream of bytes, a string.. etc.. can't seem to get anywhere with it. If you could provide even the simplest example of usage with serial, it would be very much appreciated. Is there any way I can can contribute back to your project? username_2: Hi @username_3, any chance you could add a usage example? The work you've done on the descriptions is very good, and I'd like to know better how to use that work. username_3: I switched jobs and now neither have an u-blox device nor any files I can use for testing. I think from reading the code it should be something like this (untested): ```python f = open(path_or_socket, "rb") reader = ubx.Reader(f.read) while True: try: rawmessage = reader.read_rawmessage() except ubx.ChecksumError: # Handle checksum error continue except EOFError: # Normal exit break try: message = parser.parse(rawmessage) except KeyError: # Handle unknown message continue except ubx.PayloadError: # Handle payload error continue # Do something with the message ``` Maybe you could provide some test files? Then I can add some tests and provide an example. All in all I'm not very happy about how this library is designed but it worked fine for me and might still be useful for others.
nnguyetnam-group/common
116244531
Title: Try to raise an issue Question: username_0: Hi anh @username_1 , anh có thấy issue này ko anh ? Answers: username_1: Có Anh reply trực tiếp qua email nè username_1: Định có thể assign và gắn label cho issue này thử xem coi có quyền ko đi. Thanks, Nam username_0: test label and assign feature of git :) username_1: Nice… Status: Issue closed username_1: i think it is OK to close it username_1: Test send picture…. username_1: ![img_3887](https://cloud.githubusercontent.com/assets/13391376/11109336/077a12ac-8922-11e5-8ac1-f83951cdfe4f.JPG) we must send via web reply...
coding-coworking-club/python-2019-fall
528709569
Title: [Homework #] problem_title Question: username_0: <!-- 我是註解符號,不會顯示在內文中,送出前請先預覽確認! --> ## 提交連結 <!-- https://judge.ccclub.io/status/6d2469c10e893bba6d0633901c7ebc6a --> ## 程式碼 <!-- 請在下方填入你的程式碼並注意縮排(若有提交資訊可留空)。 請勿上傳螢幕截圖! --> ```python #input N+1 rows title = input() book_lst = [] while title != '0': book_lst.append(title) title = input() #create a function to encode books def encode(lst): #simply the original list lst_R = [] for item in lst: if item not in lst_R: lst_R.append(item) else: continue #empty output list output = [] for i in range(len(lst_R)): code1 = "{0:0=3d}".format(i+1) counter = 0 for j in range(len(lst)): if lst[j] == lst_R[i]: counter += 1 code2 = "{0:0=3d}".format(counter) code3 = "{0:0=4d}".format(j+1) code = code1 + code2 + code3 output.append(code) else: continue #make codes into a dic output_dic = {} start = 0 for i in range(len(lst_R)): N = 0 for j in range(len(lst)): if lst[j] == lst_R[i]: N += 1 else: continue output_dic[lst_R[i]] = output[start:start+N] start += N return lst_R, output_dic #print result for book in encode(book_lst)[0]: code = " ".join(encode(book_lst)[1][book]) print (book, code) ``` ## 錯誤訊息 <!-- 若有錯誤訊息請在下方提供(可截圖),若無則請留空。 --> Time Limit Exceeded ## 問題描述 <!-- 請在下方簡單敘述一下你的問題,並盡量附上你的解題思路。 --> 已考慮字典無順序性,輸出結果也和範例output一致,但出現"Time Limit Exceeded"的error,請問助教是哪個地方造成程式運行的過程中花太多時間。 Answers: username_0: @alexypdu 請助教幫忙解答,謝謝! username_1: 你好,我們發現測資有誤,修正過後你的程式沒錯。 再麻煩重新提交你的程式,應該就沒問題了。 不好意思讓你困惑這麼久 >< Status: Issue closed
drush-ops/drush
43629565
Title: Allow contrib_destination to be managed per project (backport 7.x patch to 6.x). Question: username_0: Please backport this great patch, 087e51eabba013e28dde4be11aaf01e6ace0d8d3 from https://github.com/drush-ops/drush/pull/582, to 6.x for mainstream Drush users. From my tests, the patch applies cleanly and works as expected. Here's a pull request: https://github.com/drush-ops/drush/pull/835. Status: Issue closed Answers: username_1: This was fixed.
featherweight-design/component-library
807809903
Title: `Icon` component tests Question: username_0: ## Description Write Cypress tests for the `Icon` component ## Acceptance Criteria * [ ] Writes tests for all usage scenarios * [ ] Refactors component to use `copyContent` * [ ] Updates `pull_request` workflow for new tests
irods/irods
57029870
Title: [unixODBC][Driver Manager] error with psqlodbs.so Question: username_0: I was attempting to run `./plugins/database/packaging/setup_irods_database.sh` on a built-in-place instance of 4-0-stable and the process fails on step 3 of 5: ```bash Step 3 of 5: Configuring database security... Configuring security... Skipped. Existing Postgres configured by DBA. Testing database communications... Configuration problem: Communications with the database failed. ERROR: cllConnect: SQLConnect failed: -1 ERROR: cllConnect: SQLConnect failed:odbcEntry=postgres,user=irods,pass=XXXXX ERROR: cllConnect: SQLSTATE: 01000 ERROR: cllConnect: Native Error Code: 0 ERROR: cllConnect: [unixODBC][Driver Manager]Can't open lib '/usr/pgsql-9.3/lib/psqlodbc.so /usr/lib64/psqlodbc.so' : file not found userName=irods password=<PASSWORD> cllConnect failed with error -1. Found 0 processes: There are no iRODS servers running. Abort. ``` The error shows the odbc files are present and I can see that they are there, and ```./find_odbc_postgres.sh``` returns results; ```bash $ ./find_odbc_postgres.sh /usr/pgsql-9.3/lib/psqlodbc.so /usr/lib64/psqlodbc.so ``` Is this behavior cause by the fact that there are two psqlodbc.so files found? Thanks, Answers: username_1: Yes, this is due only to the fact that there are two .so files found. If you can rename or remove one of them, the other will get picked up and installation should get a bit further. username_0: Thanks, that was it! and as you predicted, it failed right after that. I'll keep working on it, thanks for the help! Status: Issue closed
fellesverkstedet/workshop-wiki
731279311
Title: Try turning extraction in back of shopbot room into vacuum cleaner as well Question: username_0: We have a splitter for the tubes that run from the extra extraction in the shopbot room. Graham and I have been discussing of using the splitter so you could use the extraction as a vacuum cleaner as well for the shopbots, instead of finding and moving vacuum cleaners somewhere in the workshop.
broadinstitute/cromwell
194652570
Title: Link Cromwell README to other websites and github files Question: username_0: @username_0 commented on [Thu Dec 08 2016](https://github.com/broadinstitute/dsde-docs/issues/1515) - [ ] Link to [Dev Docs](https://software.broadinstitute.org/wdl/devzone), latest version and release notes, blog, contact us (Slack and Forum), WDL Spec, [WDL website](https://software.broadinstitute.org/wdl/), [GATK website](https://software.broadinstitute.org/gatk/) - [ ] Link and describe the following files: [Authors](https://github.com/broadinstitute/cromwell/blob/develop/AUTHORS), [Changelog](https://github.com/broadinstitute/cromwell/blob/develop/CHANGELOG.md), [Apache License](https://github.com/broadinstitute/cromwell/blob/develop/LICENSE-ASL-2.0), [Broad License](https://github.com/broadinstitute/cromwell/blob/develop/LICENSE.txt), [Migration](https://github.com/broadinstitute/cromwell/blob/develop/MIGRATION.md), [Making a Backend](https://github.com/broadinstitute/cromwell/blob/develop/MakingABackend.MD), [Notice](https://github.com/broadinstitute/cromwell/blob/develop/NOTICE), [Security](https://github.com/broadinstitute/cromwell/blob/develop/SecurityRecommendations.md).<issue_closed> Status: Issue closed
artemnovichkov/swift-project-template
245568951
Title: Carthage is an undocumented dependency Question: username_0: Running cookiecutter failed for me: ``` Initialized empty Git repository in ... /<KEY>: line 6: carthage: command not found ERROR: Stopping generation because post_gen_project hook script didn't exit successfully Hook script failed (exit status: 127) ``` Resolved by running `brew install carthage`. Answers: username_1: Hi, @username_0! cookiecutter runs post_gen_project hook which contains Carthage installation in my template , because we use it in iOS projects by defaults. If you want to fix it for further using, just remove `carthage update --no-use-binaries --platform iOS` from `hooks/post_gen_project.sh` script. username_0: Hey @username_1 -- I only meant that it should be documented. Would you accept a PR? Status: Issue closed
opencypher/openCypher
211074062
Title: Allow MERGE on graph patterns Question: username_0: Currently the `MERGE` clause only accepts a single pattern. This means you can express a path, but not anything more complex than that. You can break your pattern down into multiple `MERGE` clauses, but that gives you different semantics, since the scope of what to _match-or-create_ applies per `MERGE` clause. In some use cases you actually have an entire subgraph where you want to either find a matching instance of that pattern, or create a whole new instance of it. Allowing this also aligns the allowed pattern syntax between all the pattern based clauses: (`OPTIONAL`) `MATCH`, `CREATE`, and `MERGE`. Answers: username_0: Example: This pattern works in both `MATCH` and `CREATE` ``` CREATE (left:Two)<-[:REL1]-(root:One)-[:REL2]->(right:Two), (:Three)<-[:REL1]-(left)-[:REL1]->(:Three), (:Three)<-[:REL2]-(right)-[:REL2]->(:Three) RETURN root ``` It would be nice if it also worked in `MERGE`, with _match-or-create_ semantics: ``` MERGE (left:Two)<-[:REL1]-(root:One)-[:REL2]->(right:Two), (:Three)<-[:REL1]-(left)-[:REL1]->(:Three), (:Three)<-[:REL2]-(right)-[:REL2]->(:Three) RETURN root ```
wireservice/csvkit
162989041
Title: non ascii chars in field values causing csvcut to create new fields Question: username_0: Hi There, I have been using csvcut tool. My data is in csv with combination of ascii and non ascii characters. The issue I am facing is when there are some non ascii chars in the filed value it is causing new fields and new lines. Could you please help me to handle non ascii chars. I have attached the input and output files. The command I have used is " csvcut -c Col4,preview,Col2,searchable_body,Col3 testNonASCII > result.csv" I need to carry these non-ascii chars to the output file with out changes. Answers: username_1: Have you tried with the latest version from GitHub? pip install -e git+git://github.com/username_3/csvkit.git@master#egg=csvkit username_0: I got the error "Command python setup.py egg_info failed with error code 1 in /tmp/pip-build-adaptive/dbfread" when I tried "pip install -e git+git://github.com/username_3/csvkit.git@master#egg=csvkit". Attached full log. [pip.txt](https://github.com/wireservice/csvkit/files/340210/pip.txt) username_1: What version of Python are you running? python --version username_0: 2.6.6 username_1: The latest version of csvkit only supports 2.7 and higher. username_0: Is there a way I can have python 2.7 and use it for csvkit on same machine where python 2.6.6 is default? username_1: If you install Python 3, it should not clash with Python 2. Then, you can use [virtualenv](https://virtualenv.pypa.io/en/stable/) to use csvkit in Python 3. username_0: I have tried csvkit latest with Python 2.7.12 on virtualenv, but the result is same as before. username_1: I can't reproduce. Try: pip install --upgrade -e git+git://github.com/username_3/csvkit.git@master#egg=csvkit Also, I get different output when using the newest version of csvkit, so I think the issue is fixed. username_0: (myPython2.7)[me@mysys ~]$ python -V Python 2.7.12 (myPython2.7)[me@mysys ~]$ rm result.csv (myPython2.7)[me@mysys ~]$ csvcut -c Col4,preview,Col2,searchable_body,Col3 testNonASCII.csv > result.csv Output: (myPython2.7)[me@mysys ~]$ cat result.csv Col4,preview,Col2,searchable_body,Col3 cool4,,,3 accounting at everyone.�, �,,,, hi in info mbi... mbi.instructure.com more of the week ,0,cool3,ALLOWED, Input: (myPython2.7)[me@mysys ~]$ cat testNonASCII.csv Col4,searchable_body,result,sms_message_type,preview,subscriber_reputation,Col1,Col2,Col3 cool4,3 accounting at everyone.�� hi in info mbi... mbi.instructure.com more of the week ,ALLOWED,none,"Hi everyone.��CR_2x1$1x4_CR Week 3 of accounting in the MBI...CR_2x1$1x4_CR CR_2x1$1x4_CR More info at mbi.instructure.comCR_2x1$1x4_CR ",0,cool1,cool2,cool3 Could you please point me where I am doing wrong here. I have ran "pip install --upgrade -e git+git://github.com/username_3/csvkit.git@master#egg=csvkit" before this test. The output is creating 3 rows for the input of 1 record. username_1: I get one row: ``` Col4,preview,Col2,searchable_body,Col3 cool4,"Hi everyone.

CR_2x1$1x4_CR Week 3 of accounting in the MBI...CR_2x1$1x4_CR CR_2x1$1x4_CR More info at mbi.instructure.comCR_2x1$1x4_CR ",cool2,3 accounting at everyone.

 hi in info mbi... mbi.instructure.com more of the week ,cool3 ``` username_1: Seems to be an issue in Python 2, but not Python 3. username_1: @username_3 @username_2 Any idea of UTF-8 issues in Python 2? username_2: Possibly related to: https://github.com/wireservice/csvkit/issues/596 and https://github.com/wireservice/agate/issues/624 but it's a different kind of error. I'll try to look into these this week. username_0: Thank you very much for the support. Just wanted to update you with my findings over the issue (thought of useful to fix the issue). Found that "e280 a8e2 80a8" hex string causing new record when it goes through csvcut. I need to retain the special chars as it is in the input inside the field after csvcut output. Any idea of such UTF-8 char-set may create new lines in csvcut? so that I can replace with constant before csvcut and replace back the constants with original char-set. username_1: The issue is with `\u2028` (LINE SEPARATOR) in Python 2, which now causes: Your file is not "utf-8" encoded. Please specify the correct encoding with the -e flag. Use the -v flag to see the complete error. Status: Issue closed username_3: Fixed in csvkit `1.0.0`
SunstriderEmu/BugTracker
622393530
Title: Concerted Efforts - not affected by Faction Balance Question: username_0: **Describe the bug** Bring 1 Alterac Valley Mark of Honor, 1 Arathi Basin Mark of Honor, 1 Warsong Gulch Mark of Honor and 1 Eye of the Storm Mark of Honor to an Alliance Brigadier General in any Alliance Capital City, or Horde General in any Horde Capital City will grant you 314honor(1) - which is not affected by faction balance - additional 50% ally / 25% horde (1) https://wowwiki.fandom.com/wiki/Quest:Concerted_Efforts - column to the right, says 314honor **To Reproduce** bring 4 marks, redeem them - get 314 honor without faction balance bonus Status: Issue closed Answers: username_1: Duplicate of #3375
ropensci/skimr
934694914
Title: Skimming when all values are NA Question: username_0: ``` I think the base columns are okay (n_missing, complte_rate) but probably we should not do the other statistics. @username_1 thoughts? Answers: username_0: I guess it could be that we push the count to 0 so it works like the _NA_NUMERIC_ case. username_1: I think the issue is primarily how we handle NA's in some of the summary stats that we include: count and hist. We could probably add some simple updates to check if all the data is NA, and if so, have them return NA_character_ too. How does that sound?
e-mission/e-mission-data-collection
119887248
Title: Fine location tracking does not start properly on geofence exit Question: username_0: Sometimes, when we start fine location tracking on a geofence exit, fine location tracking does not actually appear to start. Here is an example: On the test iphone3: - 7:40: Application launched with LaunchOptionsLocationKey = NO - 7:40: tripDiaryStateMachine = (null), relaunchLocationManager = 0, recreating the state machine - 7:40: initializing with WAITING_FOR_TRIP_START - 7:45: EXITED_GEOFENCE - 7:45: started fine location tracking - 7:45: TRIP_STARTED - 7:45: WAITING_FOR_TRIP_START -> ONGOING_TRIP - 7:45: pushAndClearData - 7:45: no data - 7:45: deleting geofence - 7:45: TRIP_STARTED - 7:45: pushAndClearData - 7:45: no data - 7:45: EXITED_GEOFENCE in silent push - 7:45: pushAndClearData - 7:45: no data - 7:45: TRIP_RESTARTED - 7:45: pushAndClearData - 7:45: no data - NO points!! - 7:58: LaunchOptionsLocationKey = YES - 7:58: initializing to ONGOING_TRIP - 7:58: started fine location tracking - points received continously Is the key perhaps that the app was already running when the geofence was exited? Or is it just that the trip was largely over before the geofence triggered? Looking at the points from the other iphones and from the android phones, it looks like the trip ended at around 7:52, so that does not appear to be the reason. According to the android example, Let's look at other examples. Answers: username_0: It is not 1. above because the same thing happened today morning. - 8:00: launched with LaunchOptionsLocationKey = NO - 8:00: Received remote push, about to check whether a trip has ended - 8:00: T_RECEIVED_SILENT_PUSH in WAITING_FOR_TRIP_START - 8:24: T_VISIT_ENDED - 8:24: pushAndClearData - 8:24: EXITED_GEOFENCE - 8:24: started fine location tracking - 8:24: pushAndClearData - 8:24: location point - 8:25: location point ... username_0: Let's go back to see if there are any differences between last night and today morning. Last night looks pretty similar: - 7:00: launched with LaunchOptionsLocationKey = NO - 7:00: Received remote push, about to check whether a trip has ended - 7:00: T_RECEIVED_SILENT_PUSH in STATE_WAITING_FOR_TRIP_START ... However, when the geofence exit is triggered at 7:45, and we call pushAndClearData, we return with fetch result = no data. And today morning, when we invoked pushAndClearData, we returned new data. Hm. I wonder if that is the reason... 8:26: successfully pushed 332 entries to the server 8:26: Returning with fetch result = new data (repeat 3 times) 8:26: location point 8:27: location point 8:28: location point ... username_0: Another difference is the visit notification today morning as opposed to yesterday. Maybe there are no visit notifications triggered right after restart? Right now, I am going to change this to return new data every time. And then actually make multiple trips tomorrow and test it out. Cleaning up and simplifying this code to only call it once will happen later?! Status: Issue closed username_0: this has been working properly for a while
relay-tools/relay-query-lookup-renderer
602391050
Title: Does not behave same as QueryRenderer - Does multiple fetch requests Question: username_0: @username_1 The react-relay query render will performance once fetch, regardless of how many times it appears in the react tree. For example ``` import {QueryRenderer} from 'react-relay'; const something = [1,2,3] {something.map( () => <QueryRenderer {...props} /> ) ``` Will cause a single request. ``` import {QueryLookupRenderer} from 'relay-query-lookup-renderer'; const something = [1,2,3] {something.map( () => <QueryLookupRenderer {...props} /> ) ``` Will fire 3 requests, if it can not find the data in the store. Only difference between these two screenshots was using switch `QueryRenderer` for `QueryLookupRenderer` ![image](https://user-images.githubusercontent.com/9244507/79629983-c26f1900-81a1-11ea-9cfd-89f191fdc2fa.png) ![image](https://user-images.githubusercontent.com/9244507/79629992-d87cd980-81a1-11ea-87e8-9843a87d0107.png) Answers: username_1: @username_0 this was forked from Relay's QueryRenderer before they introduced a query cache. I don't plan to make any updates to this package. It will be deprecated in favor of the `store-or-network` fetch policy currently in Relay experimental. Status: Issue closed username_0: Thanks for the info Rob.
rancher/rancher
278178117
Title: OpenStack Controller deletes the ssh-key when deleting a node Question: username_0: **Rancher versions:** rancher/server: 1.6.11 rancher/agent: Irrelevant **Docker version: (`docker version`,`docker info` preferred)** docker version Client: Version: 1.12.6 API version: 1.24 Go version: go1.6.4 Git commit: 78d1802 Built: Tue Jan 10 20:20:01 2017 OS/Arch: linux/amd64 Server: Version: 1.12.6 API version: 1.24 Go version: go1.6.4 Git commit: 78d1802 Built: Tue Jan 10 20:20:01 2017 OS/Arch: linux/amd64 **Operating system and kernel: (`cat /etc/os-release`, `uname -r` preferred)** 3.10.0-327.36.1.el7.x86_64 **Type/provider of hosts: (VirtualBox/Bare-metal/AWS/GCE/DO)** OpenStack **Setup details: (single node rancher vs. HA rancher, internal DB vs. external DB)** Just a Rancher Server **Environment Template: (Cattle/Kubernetes/Swarm/Mesos)** Cattle **Steps to Reproduce:** 1) Create a new node using the OpenStack driver 2) Delete the node **Results:** The node is deleted but I can't reuse the same ssh key anymore because has been deleted after the DELETE request to OpenStack. I suggest not to send the key name in the DELETE request for the node deletion Answers: username_1: docker-machine creates a unique SSH key for each host and deletes it when the host is removed. You shouldn't be reusing them. Status: Issue closed
marmelab/react-admin
603201472
Title: Expand not responding in custom query Datagrid Question: username_0: **What you were expecting:** When using a `Datagrid` to display a custom query result (following the example in the documentation), I expect that I can use the `expand` feature. **What happened instead:** The expand icons is displayed, but nothing happens upon clicking it. **Steps to reproduce:** - Create a `Datagrid` outside of a `List` or any other component that injects the data - Provide your own data (`ids` and `data`) to the component - Add a `expand` component to the `Datagrid` - Try to expand a row **Related code:** The simple example code wasn't loading for me in codesandbox, so I had to reproduce it in the tutorial sandbox. Sorry for that. * Preferably, a CodeSandbox forked from https://codesandbox.io/s/github/marmelab/react-admin/tree/master/examples/simple * A link to a GitHub repo with the minimal codebase to reproduce the issue ``` insert short code snippets here ``` **Other information:** <!-- List any other information that is relevant to your issue. Stack traces, related issues, suggestions on how to fix, Stack Overflow links, forum links, etc. For visual or layout problems, please include images or animated gifs.--> **Environment** * React-admin version: * Last version that did not exhibit the issue (if applicable): * React version: * Browser: * Stack trace (in case of a JS error): Answers: username_1: That's because the expand feature stores its state per resource in redux. Just add a `resource` prop on the `Datagrid` (it's usually injected by the `List` component) and it will work. See https://codesandbox.io/s/eloquent-jennings-2n57h?file=/src/users.js Status: Issue closed username_2: which example in the documentation were you referring to? username_0: Thanks for the answer! @username_2 The example I referred to is in the `DataGrid` documentation, the very last tip before the `SimpleList` component section.
trailofbits/ebpfpub
765960534
Title: 酒泉妹子真实找上门服务z Question: username_0: 酒泉哪有特殊服务的洗浴【十(微)7813╧72524漂亮】是影视寒冬年,与此同时,针对网络内容的政策逐渐出台,台网剧审查趋向统一标准,两者的界限越来越模糊。整体剧集市场在呼吁现实题材、更加接地气的内容、弘扬主旋律和主流价值观的作品。在大环境下,响应市场号召的北京时代光影在这一年,产出了电视剧《我怕来不及》和两部网剧《当你恋爱时》《将军家的小娘子》。时代光影董事长王锦不惧影视寒冬的“威胁”,发挥台网联动的优势,拓宽题材类型,在原有小人物创作的基础上,大胆创新,将视角定位在“以小见大”的社会话题,以及融入多种感情的家国情怀。聚焦小人物,反映真实情感《我怕来不及》正在央视八套黄金强档热播,这是一部聚焦大时代背景下的小人物奋斗史的剧集。改革背景下,工人李春生一面实施绿色矿山计划,一面照顾白家老小。李春生越挫越勇的奋斗精神感染了大批观众,《我怕来不及》连续两周收视夺冠,引发讨论。高收视和高关注的背后,也有不少网友在网上开轰。“李春生是圣父”、“不可能有这样的人”……针对这些非议,总制片人王锦特别强调了《我怕来不及》的情感浓度,将“我怕来不及”诠释为“子欲养而亲不待”,传达出正能量的价值观。他从小人物李春生的视角出发,赞扬了李春生牺牲自我、成全大家的勇气和坚忍不拔的品质,强调守住亲情就是守住了家。从年成立以来,时代光影一直选择这些朴实善良的小人物为主角,并将他们身上的隐忍品质和真善美放大,例如前些年的《满仓进城》《俺娘田小草》《我的小姨》,还有正在热播的《我怕来不及》,这些剧集从不同角度阐述了小人物身上的大能量。在一次采访中,被问到为何选择这些“普通人”,王锦表示,一方面是觉得自己就是个小人物,能够抓到小人物的温度,体会得到小人物的情感,比如李春生为了爱与责任,抚养岳母,照顾毫无血缘关系的侄女晶晶,无条件帮助白家度过难关;另一方面,李春生这个人物确实来自于自己的亲身经历,“我爸就属于这种人,就是李春生式的人物。小时候,舅舅过世的早,舅妈改嫁,剩下两个孩子,父亲就把孩子接到家里养。后来叔叔和大伯家也经历了相似的遭遇,父亲除了照顾我之外还一直坚持照顾另外三个家庭的孩子。他付出这么多,甚至于还不被理解。”亲情的回归和担当是王锦在展现小人物和底层人民时尤为关注的部分,王锦觉得父亲是个很伟大的人,所以他自己对李春生式的人物很有认同感。这种精神更不应该被谴责,而应该被弘扬。正如王锦所说:《我怕来不及》并不是要教导每一个人去做李春生,而是想要表达他没有错,不应该被抨击,甚至期望有人能认可这样的平民英雄。我们大多数人的生活都是平淡的,但这并不意味不真实存在。截止目前,《我怕来不及》即将收官,就全国测量仪央卫晚间电视剧、猫眼数据显示来看,《我怕来不及》的收视率一直处于递增状态,我们可以看到,《我怕来不及》的观众一半是通过电视台收看,另一半则是通过网络获取视频。很多网友也通过平台弹幕发表对剧集的看法,话题性增强。从这一角度看,原本的电视受众开始往网络迁移,视频网站也需要更多元的内容来满足,台网剧间的壁垒慢慢消除,电视剧集的投放渠道更加丰富,这更有利于电视剧的影响力,促进制作公司的发展。拓宽类型,促进多元发展究其原因,视频网站受众越来越多,年龄层越来越广,必然要求内容更多元,大众化的内容也有了空间。这也是现在很多电视剧或台网联动的剧同样能在网络上引爆的原因。面对影视寒冬的影响,在综合分析行业优势利弊的基础上,年,时代光影优化了企业布局,向网剧市场进军。目前,时代光影便依托自身的优势在与视频平台打交道,与优酷合作了《将军家的小娘子》这类集爱情、甜宠、轻虐元素为一体的古装剧,以及轻体量的都市迷你偶像剧《当你恋爱时》。无论形式还是题材,都更具有网感和互动感。王锦坦言,其实所有制作公司现在都是在寒冬中爬行,大家都是痛并快乐着,对于内容制作公司,现阶段要拼的地方太多。纵观目前的内容制作市场,优质的内容供应商往往具备许多共性:绝对敏锐的政策嗅觉,优良的制作和宣发。王锦表示:“可能在政策上觉悟比较高是我们唯一小小的优势。我们知道什么东西能做,什么东西不能做,底线在哪里。此外,还要尊重创作规律,不能拔苗助长。”市场一直在变化,观众、制作手法、故事、政策都在变化,制作公司唯一能做的就是不断跟上变化。这几年剧集市场频频波动,无论是政策变化还是受众喜好,抑或平台玩法都对内容制作方提出了更高的要求。但万变不离其宗。对受众保持深刻洞察、有核心制作团队、对项目有足够的把控能力、对市场和政策有敏锐感知,永远是一家内容公司在竞争中脱颖而出的核心竞争力。声明:中华娱乐网刊载此文出于传递更多信息之目的,并非意味着赞同其观点或证实其描述。版权归作者所有,更多同类文章敬请浏览:综合资讯亲也孟檀铰https://github.com/trailofbits/ebpfpub/issues/109?S4YA4 <br />https://github.com/trailofbits/ebpfpub/issues/1168?73zN0 <br />https://github.com/trailofbits/ebpfpub/issues/713?vfkzl <br />https://github.com/trailofbits/ebpfpub/issues/1019?hraxu <br />https://github.com/trailofbits/ebpfpub/issues/1324?rsqgv <br />https://github.com/trailofbits/ebpfpub/issues/1631?57460 <br />https://github.com/trailofbits/ebpfpub/issues/251?86593 <br />ucxtjyjyjqlhurtcmfnzpgpindbtwcapttj
eclipse/xacc
1099910788
Title: Errors when running GradientStrategy in shots mode Question: username_0: Simple repro: in `TEST(GradientStrategiesTester, checkDeuteronVQE)` (`GradientStrategiesTester.cpp`) set `{{"shots", 1024}}` when initializing `qpp`. This will hit `xacc::error` added in https://github.com/eclipse/xacc/commit/156d56aa8c0a033d5329f0be07ef9e3e62b81c14 Root cause: the `observed` circuits (constructed by the gradient strategy impl) contain a base circuit for identity term with no `Measure`. Hence, no measurement shots data is added. Need to investigate if this is a bug in the gradient strategy (e.g., should filter identity term) or we could simply remove `xacc::error` in the above commit. e.g., some accelerators may throw if sending circuits with no measure. Answers: username_0: This is indeed a problem with gradient strategy. Sending base circuit w/o measurement (identity term) will cause problems for `aer` and potentially `ibm` as well. e.g., https://github.com/eclipse/xacc/blob/master/quantum/plugins/ibm/aer/accelerator/aer_accelerator.cpp#L196: this will crash since the `counts` field is not present. In VQE, we explicitly handle this (removing non-observing circuits): https://github.com/eclipse/xacc/blob/master/quantum/plugins/algorithms/vqe/vqe.cpp#L135 but not in the gradient strategies. Hence, the proper fix is to address this bug in the gradient strategy. Status: Issue closed
python/pythondotorg
819377269
Title: Create "box" on python.org/jobs for sponsor recognition Question: username_0: <!-- This is the repository and issue tracker for https://www.python.org website. If you're looking to file an issue with CPython itself, please go to https://bugs.python.org Issues related to Python's documentation (https://docs.python.org) can also be filed in https://bugs.python.org, by selecting the "Documentation" component. --> **Is your feature request related to a problem? Please describe.** One of the benefits for our new sponsorship program includes logo recognition on https://www.python.org/jobs/. **Describe the solution you'd like** It would be great if we can add a new box that I can edit going forward from /admin in the right hand margin of the homepage. See snap shot of my mocked up drawing: <img width="1109" alt="Screen Shot 2021-03-01 at 4 43 40 PM" src="https://user-images.githubusercontent.com/2590029/109569347-ec75b500-7aad-11eb-9f3e-a87f349383a1.png"><issue_closed> Status: Issue closed
spring-cloud/spring-cloud-dataflow
485750925
Title: allow definition of environment variables for a task when launched on cloudfoundry Question: username_0: **Problem description:** Following scenario: - I run SCDF on cloudfoundry. - I define an app. (a Spring Boot Boot app requiring java 11) - I define a task from the given app. - I launch the task - CF uses the `java_buildpack` to create the app container with java 8 (java 8 is the default) - CF starts the app - the app fails to start, because the wrong java version is used (8 instead of 11) **Solution description:** There should be a way to pass environment variables. To tell the java_buildpack to use the correct java version, one has to define the following environment variable (e.g. in the `manifest.yml`: `JBP_CONFIG_OPEN_JDK_JRE: '{jre: { version: 11.+ }}'` A possible solution could be additional params like this: `deployer.cloudfoundry.env.JBP_CONFIG_OPEN_JDK_JRE='{jre: { version: 11.+ }}'` `deployer.cloudfoundry.env.XXXX=zzzzzzz` `deployer.cloudfoundry.env.YYYY=wwwww` It is important to understand that defining only one environment variable might not be enough. How ever, these environment variables must then be passt to cloudfoundry while deploying the app. **Additional context:** I also created an issue which describes that currently the buildpack parameter is not interpreted correctly: https://github.com/spring-cloud/spring-cloud-dataflow/issues/3466 Answers: username_1: Resolved as part of https://github.com/spring-cloud/spring-cloud-deployer-cloudfoundry/pull/308 Status: Issue closed username_0: @username_1 do you provide SNAPSHOT builds of SCDF? username_2: @username_0 Given this is fixed in Spring Cloud Deployer CF `2.1.0.BUILD-SNAPSHOT`, it is available in [Spring snapshot repo](https://repo.spring.io/libs-snapshot-local/org/springframework/cloud/spring-cloud-deployer-cloudfoundry/2.1.0.BUILD-SNAPSHOT/ ) and since the latest SCDF master has `2.1.0.M1` CF deployer, you can update the SCDF to use `2.1.0.BUILD-SNAPSHOT` version of CF deployer and test. Let us know how it goes. Thanks! username_0: hmm, but I already run `2.2.0.RELEASE` - should I downgrade the version? username_2: No, you are referring to the SCDF `2.2.0.RELEASE` and what meant above is the CF deployer version. If you are building SCDF locally, then update your parent POM to use `2.1.0.BUILD-SNAPSHOT` of `spring-cloud-deployer-cloudfoundry` and try this fix. username_0: @username_2 unfortunate it seems not to be enough to just define: ``` <dependency> <groupId>org.springframework.cloud</groupId> <artifactId>spring-cloud-deployer-cloudfoundry</artifactId> <version>2.1.0.BUILD-SNAPSHOT</version> </dependency> ``` because a class seems to be missing now: `java.lang.NoClassDefFoundError: org/springframework/cloud/deployer/spi/cloudfoundry/CloudFoundry2630AndLaterTaskLauncher` username_2: What version of SCDF do you use? You need to be on `2.2.2.BUILD-SNAPSHOT` - the latest `master` that is. username_0: @username_2 I still did not manage to get the dependencies to work without getting: ``` Caused by: java.lang.NoClassDefFoundError: org/springframework/cloud/deployer/spi/cloudfoundry/CloudFoundry2630AndLaterTaskLauncher at java.base/java.lang.Class.getDeclaredMethods0(Native Method) at java.base/java.lang.Class.privateGetDeclaredMethods(Class.java:3167) at java.base/java.lang.Class.getDeclaredMethods(Class.java:2310) at org.springframework.util.ReflectionUtils.getDeclaredMethods(ReflectionUtils.java:489) ``` This is the source I use for testing: https://gist.github.com/username_0/b7998965c31c4dd850999613c5db8460 username_0: @username_1 if understand this all correctly, then this means it will not be possible to create task on CF that run with java11 and need parameters - as normal parameters will end up as environment variables and these are fixed on the created CF app/container. I think this quite a limitation of the current implementation and should be clearly documented. username_3: @username_0 - I'm not sure what you mean by `normal parameters` here. Typically task/job parameters are passed as command line arguments, specified as part of the `task launch` command. These can be different for every launch and will not be affected by this. SCDF sets certain app properties as environment variables and, by default, these are converted to JSON and set as the environment variable `SPRING_APPLICATION_JSON`. This is a standard Spring Boot property source, that will be parsed as JSON to bind all the properties. There may be some edge cases, such as property names containing property placeholders, in which this might make a difference but since these property values are already defined in the SAJ environment variable, they are static, will be applied to every task container and cannot be changed via the SCDF API without destroying the task. If you experience any specific problems related to this, please let us know. username_0: @username_3 indeed, I did mix up some stuff here - sorry for the noise!
godotengine/godot
332998474
Title: Proposal: Consider using bgfx as a rendering backend? Question: username_0: I started using Godot lately, and I find it ingenious. I am using a Mac, and while I don't intend to release something Mac-only in the end, it is the best development platform for me, as well as many others. Following the news regarding the deprecation of OpenGL for Apple platforms I saw this issue on GitHub, https://github.com/godotengine/godot/issues/19368 and I personally feel that the issues raised by the Godot team are all valid. It also seems that the team has already planned Vulkan support, and Mac / iOS will be supported using MoltenVK. However, before this happens, I would like to throw another option to the table, which is using bgfx as a rendering backend. https://github.com/username_4/bgfx BGFX has DirectX, OpenGL, and Metal support out of the box and will soon add Vulkan support (https://github.com/username_4/bgfx/issues/274). MoltenVK for Mac also seems like a viable option, but it has some limitations atm (https://github.com/KhronosGroup/MoltenVK/blob/master/Docs/MoltenVK_Runtime_UserGuide.md#limitations) which I have to admit I don't know how much they affect Godot (if at all). Since BGFX's API allows supporting multiple rendering backends without changing its core API, perhaps using this instead of investing time only in Vulkan will be better for Godot in the long run? I totally respect Godot team's decisions regarding Vulkan, and I trust that if Godot chooses this path it will work as good as now - if not better. I just threw the BGFX idea in order to have some feedback on whether BGFX was something that was already considered - and what is the team's opinion on choosing BGFX instead of investing time only on Vulkan support (also having to trust that MoltenVK will always work as advertised). Answers: username_1: See the reason why SDL was not accepted at. https://github.com/godotengine/godot/pull/16470 username_0: Can you please elaborate on how the SDL issue is relevant? SDL was considered as an abstraction layer on user input systems and getting rendering contexts ready to use. Bgfx, however, allows full and API-agnostic rendering. Do you mean that maintenance of an additional abstraction layer in Godot would be painful? username_1: See: https://github.com/godotengine/godot/pull/16470#issuecomment-364070439 Summary of points: \<Reduz> Imagine a situation where we use SDL2 and drop the existing backends. What happens every time we need a platform specific feature not available on SDL2? (Something it happens often) . We have the following scenarios: \<Reduz> Added to that, the problem is that we may need to add something that SDL does not support and, while for us it's something specific with a simple use, while adding this function to SDL may involve creating a large API with all the functions that are needed for abstraction. \<Reduz> I'm sorry, no matter how I try to think of ways we could use SDL, it's always more disadvantages than advantages.. Now replace SDL with BGFX. There are two parts of abstraction. A) low level abstraction B) construction of the rendering pipeline for mobile (gles2) vs pc (vulkan) Using bfgx only helps for A), B) needs to rebuilt. username_2: possible dumb question incoming: but, when vulkan rendering stuff gets complete, it would negate this issue right? i imagine time spent on vulkan will far supersede bgfx, or other rendering systems username_0: @username_2 correct. @username_1 the way I understand your comments, it's an issue of third party support (what if Godot needs something that bgfx does not offer in the future), and having full control over the implemented features. I get it, but I also hope that Vulkan works as advertised as well. username_3: This option was discussed and discarded in the past (as you can imagine), but since this was not discussed in an issue, I will take the time to explain why BGFX is not an option. 1) There are two main platforms that need to be supported. Modern Desktop PC (OpenGL ES 3, and later Vulkan) and mobile including medium and low end (OpenGL ES 2). These APIs are not programmed the same way, so a "wrapper" to simplify the work is impossible. GLES3/Vulkan uses UBOs, VAOs, TBOs. shaders with integers and plenty of features. GLES2 is very basic and supports none of that, so different approaches need to be used to write a back end. Modern hardware also uses certain rendering techniques (HDR/clustered/single pass shading), while low end uses other techniques (LDR/multipass). As such, BGFX does not save the work of having to write two backends for different hardware. 2) Having an extra layer of complexity in the middle makes things more complex, and makes debugging more difficult. Contributors would need to learn BGFX to write rendering code instead of a more standard API like Vulkan or OpenGL, out of which there is plenty of documentation and examples. As net worth, middle layer is a negative point here, not a positive one. 3) Khronos has the [Vulkan Portability Initiative](https://www.khronos.org/vulkan/portability-initiative), where they aim to run it over Metal, DirectX, etc. This is pretty much the same as BGFX, thus decreasing even further the value of BGFX. 4) Using Vulkan allows us to take advantage of extensions and new features on bleeding edge hardware much faster than using BGFX. 5) Our contributors are part of the Khronos Advisory Board, so by using Vulkan we can give valuable feedback of our experience to hardware manufacturers, making sure they hear us when woring on future versions of the specification. Hope it's clearer now! There is more than plenty of reasons to not use BGFX. Status: Issue closed username_0: @username_3 this was exactly the kind of answer I was hoping for, thanks. username_4: If you just want to use single renderer in bgfx, and you don't care about cross-platform and feature parity between renders it's as fast to add features to single renderer with bgfx too. username_3: @username_4 I never said anything negative about BGFX itself, only in the context of Godot. The point of the discussion was mainly that, given two render pipelines need to be written anyway, using BGFX is not an advantage or solution regarding to that. You answers completely missed it. For the other answers, I was not questioning whether BGFX was high or low level, or whether it can be debugged, or anything else. The point was simply that _it is_ added complexity (which is undeniable) for an use case that is redundant and provides no advantages. Apologies, but I think you are being overly defensive when the whole argument was about Godot, not BGFX. username_4: This is what I'm talking about. You're saying that bgfx would add complexity into your code base, but it's actually quite oposite, by using bgfx it removes complexity from usual code base. Once someone replaces their own renderer with bgfx, they usually have redundant code on higher level, they don't have to do state tracking anymore since bgfx deals with that, they don't have to do multiple passes over scene because bgfx allows you to submit out of order and orders draw calls for you, etc. But bgfx is complex on it's own. In the past you made statements that you want to control whole stack, and you don't want to introduce risk by adding 3rd party open source software. I find this is more honest response, since there is nothing that I can add or remove from bgfx that would make you reconsider your strategy about renderer. So no need to justify your decision about renderer by saying bgfx does or doesn't do X, Y, Z, if your decision was unrelated to bgfx as you stated before. username_3: @username_4 It definitely _is_ complexity. You seem to be assuming that just because of adding BGFX support, everything will be fine and no one will ever have to learn how the underlying APIs work. This is, unfortunately not the case and how things work in real life. This may work for APIs or libraries where you truly no longer need to care how internal implementations of things work as long as they do what they have to. In my experience if we were to use BGFX, contributors would still need to also know the underlying APIs as well as BGFX (OpenGL or Vulkan). How else are you supposed to fine tune performance, or understand why something fails when it does? Just thinking BGFX will fit like a glove and just work is naive. So, as I already made it pretty clear that there is no advantage to it, due to us needing two rendering pipelines, and that contributors will still need to understand the underlying rendering APIs. I hope you understand that to me it's just extra unnecessary complexity that is best kept away. username_5: That's exactly what the above points amount to. So we're glad that BGFX exists and that it's a great solution for many applications, but as of today it's not something that we're interested in for Godot. Thanks for respecting our decision. username_0: I gained a lot of interesting information here, not only by the Godot members, but also from @username_4 . Just wanting to clarify the reasoning behind my question: Personally, when developing, I want options. In the case of a game engine, I want it to be as flexible as it can be - this is why I would choose bgfx - because I want to be able to not be unaffected by changes and politics of the graphics ecosystem (like Apple not supporting OpenGL anymore, etc). The reasoning behind my question was if Godot shared the same thoughts. In the case of Godot, the team wants full control over _one_ rendering stack, and flexibility comes second. This will probably allow them to implement more features and be completely unaffected by 3rd party frameworks in the long run. Which is also a very good approach. -- From this discussion, the thing I am keeping is that @username_4 was right to point out that using bgfx actually saves a ton of code (something I have seen in many testimonies around), but @username_5 and @username_3 were also right to point out that if the flexibility of supporting more than one rendering backends is not a primary goal, then relying on a 3rd party middle layer just for having only one rendering backend may hinder the process later down the road. username_6: The problem is you need to support many APIs if you want to ship a cross-platform product of acceptable quality. If multiple backends isn't a primary goal then by consequence cross-platform support isn't a goal too. It's not pretty, it's not what people want to hear, but it's the reality we live in today. In order to ship a quality cross-platform product you can only to accept that and do what it takes to conform to that reality. And yes, I'm ready for the thumbdowns... let them come! username_3: @username_0, @username_6 You are both wrong. We can still rely on more rendering back-ends via lower level wrappers like MoltenVK or GFX-RS, so we already get best of both worlds. username_7: My 5 cents: maintaining your own renderer is more actual work than it sounds. And it will drain your resources on fixing things that don't necessary add end user value. Image one day you need to ship a game on PS4/Switch/etc, what do you do? Spend month porting Godot to a random platform? bgfx is already ported IIRC. And then another day you get strange rendering bug reports on random Android phone that you never know existed, what you gonna do? etc. To get real feeling of what it takes to actually ship something GL based nowadays, try to look at UE4: https://github.com/EpicGames/UnrealEngine/tree/release/Engine/Source/Runtime/OpenGLDrv/Private - it is full of hacks for random devices. Value of rendering libs is not that they are "cross-platform", but rather that they are proven to work on platforms. IMHO Value of Godot is not in "%insert_your_gapi_of_choice% support", but rather in features in the editor. username_6: Exactly. Bgfx is being used in production by renowned studios to deploy cross-platform products. So it's reasonably battle-tested and I don't see any compelling reason not to take advantage of that, not to mention the API is very well designed too. Angle isn't a good option if you care about performance. username_3: @username_7 , @username_6, @username_8 : Guys, thanks a lot for your enthusiasm, but I am the one doing the rendering work in Godot, not you. I've been working on 3D rendering for 25 years so, if I am telling you that things as they are now are optimal and BGFX will just stand in the way to being productive I hope you believe me. As always with Godot, you are free to make your own renderer with BGFX, show me that it's flawless, has better better performance, works better than mine (of course while supporting the full Godot feature set) and uses less resources and code to prove me wrong.. as well as commiting full time over the next years to maintain it. If you want a revolution, begin with it yourself. If not, keep pestering all you want and it will be ignored. username_8: I only remember ever mentioning bgfx once or maybe twice in the past. All I can find is this comment back in [2017](https://godot.eska.me/irc-logs/devel/2017-07-21.log) asking _"What about bgfx?"_. I don't understand why you would include me. I guess it's because I up-voted a comment I consider brings an interesting point to the table. Regarding the others, I don't understand why such a negative response. I don't see anything wrong in their comments. They are not pestering in any way... username_9: What a classic example of elitism and douchebaggery. I will stay miles away from Godot and any projects you maintain, thanks. username_5: This issue has long outlived its usefulness. And insulting community members is not tolerated in this community, so you are indeed more than encouraged to stay clear from this project.
atom/settings-view
214799360
Title: The install button is shown for installed packages in safe mode Question: username_0: ### Prerequisites * [X] Put an X between the brackets on this line if you have done all of the following: * Reproduced the problem in Safe Mode: http://flight-manual.atom.io/hacking-atom/sections/debugging/#using-safe-mode * Followed all applicable steps in the debugging guide: http://flight-manual.atom.io/hacking-atom/sections/debugging/ * Checked the FAQs on the message board for common solutions: https://discuss.atom.io/c/faq * Checked that your issue isn't already filed: https://github.com/issues?utf8=✓&q=is%3Aissue+user%3Aatom * Checked that there is not already an Atom package that provides the described functionality: https://atom.io/packages ### Description In safe mode the install button is shown for installed packages. This behavior changed recently. **Image from Atom 1.16.0-beta0** ![image](https://cloud.githubusercontent.com/assets/1058982/24012041/e1ca0750-0a7c-11e7-9daa-5344fbfa9bb1.png) **Image from Atom 1.13.1** ![image](https://cloud.githubusercontent.com/assets/1058982/24012278/b4bfba24-0a7d-11e7-83da-8955d0399998.png) ### Steps to Reproduce 1. Open Atom in safe mode from `cmd.exe` using `atom --safe` 2. Open the `settings-view` 3. Navigate to the Packages tab 4. Check the Package Card for the Community packages **Expected behavior:** Disable and install buttons to show up like before. Or not install button because the packages are already installed. **Actual behavior:** Install button is visible. **Reproduces how often:** 100% ### Versions 1.16.0-beta0 x64 Windows 10 ### Additional Information /cc: @username_1 We discussed this in Slack as a potential jQuery removal issue but it changed before that. Answers: username_1: I am unable to reproduce this on my machine. Maybe it's an issue just on Windows? Could have it been caused by https://github.com/atom/settings-view/pull/894? username_0: https://github.com/atom/settings-view/pull/894 is the most suspicious change between the releases where this changed. For me running `atom.packages.getAvailablePackageNames()` in the developer tools console only returns core packages in safe mode and all packages outside of safe mode. username_2: I was able to reproduce this on Ubuntu 1.22.0-dev-1880e14 running on --safe and -d --safe I'll attempt to find the cause and make a pr if I find a solution username_3: Yep, the array returned by `atom.packages.getAvailablePackageNames()` [does not return community packages in safe mode](https://github.com/atom/atom/blob/6c6c82b328ac0b0a91e2b39006ceaf683a565bb5/src/package-manager.js#L61). However, this package [uses `apm`](https://github.com/atom/settings-view/blob/8729f2b114aca74b4e80b1b356d75717c99e65da/lib/package-manager.coffee#L82) to get the list of all packages on the computer, so community packages appear in the cards regardless of any modes. Because the package is thought to exist but not declared as active, it is given the `install` button.
VictoriaMetrics/VictoriaMetrics
1005055665
Title: Regexp grouping does not work in v1.66.0 Question: username_0: Describe the bug Regexp grouping does not work. Works fine in v.1.65.0. This seems to be a connected PR [1610](https://github.com/VictoriaMetrics/VictoriaMetrics/pull/1610). To Reproduce Queries like `{__name__=~"(foo|bar)"}` does not work in v1.66.0. It produces zero results. Other regexp syntax however does like `{__name__=~"f.+"}` Expected behavior Regexp grouping works :) Answers: username_1: This also may be related to https://github.com/VictoriaMetrics/VictoriaMetrics/issues/1601 username_1: I cannot reproduce the bug in v1.66.0 and v1.66.1 - the query `{__name__=~"up|scrape_duration_seconds"}` works as expected - it returns time series with `up` and `scrape_duration_seconds` names. @username_0 , could you provide more details for your case? E.g. the exact time series names and the exact query which doesn't return the expected results. username_0: `sum(ascent_over_time({__name__=~"deals_count_(Market1|Market2)"}[1i]))` I have not tried v1.66.1, it was not the latest version when I pulled the container, I should mention I'm running standalone version. Same query works if I replace `(Market1|Market2)` with `(.+)` (but then I cannot filter on specifics, there are more metrics named similarly). I did test some parts of it with just bare metric names (no functions) and it didn't work. username_0: Seems like the problem occurs when matching other labels in the query. This query returns 0 results `sum(ascent_over_time({__name__=~"deals_count_(Market1|Market2)", environment=~"prod"}[1i]))` This however does `sum(ascent_over_time({__name__=~"deals_count_(Market1|Market2)"}[1i]))` Also each metric works when addressing them directly (or with `(.+)`) `sum(ascent_over_time({__name__=~"deals_count_Market1", environment=~"prod"}[1i]))` username_0: This query works: `sum by (market,environment) (ascent_over_time(label_replace({__name__=~"deals_count_(Market1|Market2)"}, 'market', '$1', '__name__', 'deals_count_(.+)')[1i]))` and produces the results in the image below ![image](https://user-images.githubusercontent.com/5592150/134509361-b2d8ceb9-7079-4cda-9c25-5884f60a5e2a.png) This query does not work either: `sum by (market,environment) (ascent_over_time(label_replace({__name__=~"deals_count_(Market1|Market2)", environment=~'.+'}, 'market', '$1', '__name__', 'deals_count_(.+)')[1i]))` Leaving environment label blank (`environment=''`) makes it work thou. username_1: Confirmed the issue. It should be fixed in the commit 718eca33ab62695248831f26a13efdc3d5d3d870 . @username_0 , could you build VictoriaMetrics from this commit according to [these docs](https://docs.victoriametrics.com/#how-to-build-from-sources) and verify whether the issue is gone? username_1: @username_0 , could you verify whether the issue is fixed in [v1.66.2](https://github.com/VictoriaMetrics/VictoriaMetrics/releases)? username_0: Yes it is working as expected now :) Status: Issue closed username_1: Thanks for the confirmation! Then closing the issue as fixed.
ovn-org/ovn-kubernetes
460587513
Title: ovnkube-db-64ddbfcd94-xmtsv 1/2 CrashLoopBackOff Question: username_0: the container sb-ovsdb inside ovnkube-db- pod has an issue : ``` Unable to connect to the server: tls: either ServerName or InsecureSkipVerify must be specified in the tls.Config ``` the pod restarts. here is the log: ``` kubectl logs ovnkube-db-64ddbfcd94-xmtsv -c sb-ovsdb -n ovn-kubernetes ================== ovnkube.sh --- version: 3 ================ ==================== command: sb-ovsdb =================== hostname: dm1002 =================== daemonset version 3 =================== Image built from ovn-kubernetes ref: refs/heads/master commit: <PASSWORD> =============== run sb_ovsdb ========== MASTER ONLY ovn_log_sb=-vconsole:info -vfile:info ovn_sb_log_file=/var/log/openvswitch/ovsdb-server-sb-2019-06-25-18:50:38.log info: Waiting for pid_ready ovnsb_db.pid to come up, waiting 1s ... 2019-06-25T18:50:38Z|00001|vlog|INFO|opened log file /var/log/openvswitch/ovsdb-server-sb-2019-06-25-18:50:38.log 2019-06-25T18:50:38Z|00002|ovsdb_server|INFO|ovsdb-server (Open vSwitch) 2.11.1 pid_ready ovnsb_db.pid came up in 1 5 sec tries =============== sb-ovsdb ========== RUNNING Unable to connect to the server: tls: either ServerName or InsecureSkipVerify must be specified in the tls.Config Failed to create endpoint with host for ovnkube-db service ``` Answers: username_0: adding ``` --insecure-skip-tls-verify kubectl create --insecure-skip-tls-verify -f ovnkube-db.yaml ``` enabled the container to run Status: Issue closed username_0: after a white the ovnkube-db- pod have same issue again username_0: the container sb-ovsdb inside ovnkube-db- pod has an issue : ``` Unable to connect to the server: tls: either ServerName or InsecureSkipVerify must be specified in the tls.Config ``` the pod restarts. here is the log: ``` kubectl logs ovnkube-db-64ddbfcd94-xmtsv -c sb-ovsdb -n ovn-kubernetes ================== ovnkube.sh --- version: 3 ================ ==================== command: sb-ovsdb =================== hostname: dm1002 =================== daemonset version 3 =================== Image built from ovn-kubernetes ref: refs/heads/master commit: 7b781d84aa7e0795dac2b966637cae6f2e2b29f7 =============== run sb_ovsdb ========== MASTER ONLY ovn_log_sb=-vconsole:info -vfile:info ovn_sb_log_file=/var/log/openvswitch/ovsdb-server-sb-2019-06-25-18:50:38.log info: Waiting for pid_ready ovnsb_db.pid to come up, waiting 1s ... 2019-06-25T18:50:38Z|00001|vlog|INFO|opened log file /var/log/openvswitch/ovsdb-server-sb-2019-06-25-18:50:38.log 2019-06-25T18:50:38Z|00002|ovsdb_server|INFO|ovsdb-server (Open vSwitch) 2.11.1 pid_ready ovnsb_db.pid came up in 1 5 sec tries =============== sb-ovsdb ========== RUNNING Unable to connect to the server: tls: either ServerName or InsecureSkipVerify must be specified in the tls.Config Failed to create endpoint with host for ovnkube-db service ```
jupyterhub/mybinder.org-deploy
297922915
Title: Document the cloud services we use in mybinder.org Question: username_0: In running mybinder.org, we use many cloud services. We should document what these are, why we use them, and how we use them! Current list is: - [ ] Google Kubernetes Engine - [ ] Google Container Registry - [ ] StackDriver logging - [ ] Google Network Load Balancer This will also let us be explicit and clear about the cloud-specific things we are doing. None of these services are unique to Google, and we should continue to stay that way! Answers: username_1: @username_0 If you can do this easier as handwritten versus typed, I would be happy to craft the text to go with it on Monday. username_0: @username_1 Thank you for the generous offer! I generally prefer text however, so have made a PR #488. @minrk / @betatim I think either of you know more about the current state of stackdriver than I do, so I've left a stub in the PR around stackdriver. Hopefully one of you can help fill it up afterwards! Status: Issue closed
decred/decrediton
766592170
Title: Unclickable menu items on seed copy/entry Question: username_0: `About Decrediton`, `Settings` and `Logs` are displayed during the wallet creation pages - see top right in the screenshots. These were not displayed in 1.5. `About Decrediton` is clickable, but the other two are not. I suggest these should be hidden, as they were previously. ![seed-entered](https://user-images.githubusercontent.com/6762864/102098937-a8b0cf00-3e1f-11eb-8089-b79ada883626.png) ![wallet-seed](https://user-images.githubusercontent.com/6762864/102098940-a8b0cf00-3e1f-11eb-96a2-addf1d821a6b.png) Answers: username_1: on it Status: Issue closed
rust-lang/compiler-team
580244745
Title: Cranelift backend for rustc Question: username_0: # Meeting proposal info - **Title:** Cranelift backend for rustc - **Type:** technical # Summary The current LLVM backend for rustc is very good at optimizing functions. However it is not the fastest, even with optimizations disabled. When compiling in release mode this is not a big problem, but during development this can be annoying. [Cranelift](https://github.com/bytecodealliance/wasmtime/tree/master/cranelift) has the potential to improve compilation time, as it is optimized for compilation time as opposed to being optimized for good optimizations like LLVM. Over the course of the past ~1.5 year I have been working on a Cranelift based codegen backend for rustc ([rustc_codegen_cranelift](https://github.com/username_0/rustc_codegen_cranelift) or cg_clif for short). It is currently complete enough to compile many programs. While there are cases where LLVM is faster, Cranelift is already faster than LLVM in [many cases](https://github.com/username_0/rustc_codegen_cranelift/issues/878#issuecomment-597871730). Rustc_codegen_cranelift is currently at a point where it may make sense to talk about bringing support in tree. This meeting proposal is about talking if we do want this and to talk about the future of rustc_codegen_cranelift in general. # About this issue This issue corresponds to a meeting proposal for the compiler team [steering meeting][]. It corresponds to a possible topic of discussion. You can read more about the [steering meeting procedure here][proc]. # Comment policy These issues are meant to be used as an "announcements channel" regarding the proposal, and not as a place to discuss the technical details. Feel free to subscribe to updates. We'll post comments when reviewing the proposal in meetings or making a scheduling decision. In the meantime, if you have questions or ideas, ping the proposers on [Zulip] (or elsewhere). [Zulip]: https://github.com/rust-lang/compiler-team/blob/master/content/about/chat-platform.md [steering meeting]: https://github.com/rust-lang/compiler-team/blob/master/content/about/steering-meeting.md [proc]: https://forge.rust-lang.org/compiler/steering-meeting/submit.html [this template]: https://github.com/rust-lang/compiler-team/blob/master/content/proposals/_index.md [hackmd]: https://hackmd.io/ [gist]: https://gist.github.com/ Answers: username_0: More information can be found at https://hackmd.io/@username_0/HJL5ryFS8. username_1: This is also [a problem for Miri](https://github.com/rust-lang/miri/issues/932). Would be great to find a unified solution. :) Status: Issue closed
mystor/git-revise
523158850
Title: Removed lines are not treated like index lines as revise -i suggests Question: username_0: Calling `git revise -i HEAD~3` and removing a line to move it to the index does not work. The help text in the `git-revise-todo` buffer says: `# If a line is removed, it will be treated like an 'index' line.` but removing a line (even the last) shows an error: `invalid value: Unexpected commits missing from TODO list`. Maybe it should also be mentioned that `index` lines must be last. Answers: username_1: Thanks for the report. That's a mistake in the documentation from an earlier version which did treat missing commits as 'index' lines. We should definitely update the documentation to be correct here. https://github.com/username_1/git-revise/blob/162ac90cba010a0f90f2290bf267d130466417d6/gitrevise/todo.py#L198-L212 username_0: Ah, not a new feature :) See #56 username_1: This has been fixed since #56 Status: Issue closed
facebook/react-native
539033250
Title: __blobCollectorProvider issues with Blob.slice() Question: username_0: If you have a Blob blob and call blob.slice() a couple of times, the newly created blobs inherit the same blobId but with a different size and offset. The new __blobCollectorProvider will deallocate the blobId as soon as even one sliced blob gets collected by JS. Happens for both android and iOS. <!-- Please provide a clear and concise description of what the bug is. Include screenshots if needed. Please test using the latest React Native release to make sure your issue has not already been fixed: http://facebook.github.io/react-native/docs/upgrading.html --> React Native version: <!-- Run `react-native info` in your terminal and copy the results here. --> System: OS: macOS 10.15.2 CPU: (16) x64 Intel(R) Core(TM) i9-9980HK CPU @ 2.40GHz Memory: 2.28 GB / 32.00 GB Shell: 5.7.1 - /bin/zsh Binaries: Node: 13.3.0 - /usr/local/bin/node npm: 6.13.1 - /usr/local/bin/npm SDKs: iOS SDK: Platforms: iOS 13.2, DriverKit 19.0, macOS 10.15, tvOS 13.2, watchOS 6.1 Android SDK: API Levels: 28, 29 Build Tools: 28.0.3, 29.0.2 System Images: android-29 | Google Play Intel x86 Atom IDEs: Android Studio: 3.5 AI-191.8026.42.35.6010548 Xcode: 11.3/11C29 - /usr/bin/xcodebuild npmPackages: react: ^16.11.0 => 16.12.0 react-native: ^0.61.5 => 0.61.5 ## Steps To Reproduce 1. 2. <!-- Issues without reproduction steps or code are likely to stall. --> Describe what you expected to happen: Snack, code example, screenshot, or link to a repository: <!-- Please provide a Snack (https://snack.expo.io/), a link to a repository on GitHub, or provide a minimal code example that reproduces the problem. You may provide a screenshot of the application if you think it is relevant to your bug report. Here are some tips for providing a minimal example: https://stackoverflow.com/help/mcve. -- -->
vuetifyjs/vuetify
499109215
Title: [Bug Report] Time picker does not with full-width props Question: username_0: ### Environment **Vuetify Version:** 2.0.19 **Last working version:** 2.0.18 **Vue Version:** 2.6.10 **Browsers:** Chrome 77.0.3865.90 **OS:** Windows 10 ### Steps to reproduce Add the full-width props to a time picker ### Expected Behavior The clock should be visible ### Actual Behavior The clock is not visible ### Reproduction Link <a href="https://codepen.io/username_0/pen/KKPYXXP" target="_blank">https://codepen.io/username_0/pen/KKPYXXP</a> <!-- generated by vuetify-issue-helper. DO NOT REMOVE --> Answers: username_0: I saw this issue raised also [there](https://github.com/username_1/vuetify-datetime-picker/issues/42) username_1: Reset the style `.v-picker__body`: ```css .v-picker__body { margin: initial; } ``` username_0: Another solution is to apply this to the picker: `style="width:100%"` It's a bit dirty but it does the trick. Status: Issue closed
zzyztyy/pyIGRF
567072875
Title: Incorrect calculations for 2020.0 and onward Question: username_0: Calculations with the latest IGRF 13 model are incorrect. They are very different with IGRF reference implementation. For example: lat 60.0, lon 40.0, alt 0.0, date 2020.13. The results are almost 0.8 degree different. And there is a huge leap between 2019.98 and 2020.0, which is impossible. Answers: username_1: The index to use in igrf13coeffs interpolation for year 2020 is wrong. Need to fix it in the file "loadCoeffs.py", i.e. change the line below "pointer for last coefficient in pen-ultimate set of MF coefficients..." to "ll = 3255" username_0: @username_1 tried you solution, works well for me, but there are no unittests in this project. Who knows what else could be wrong? Anyway, would you make a PR? This line looks strange: https://github.com/username_2/pyIGRF/blob/35e9b3deb7b264940fa713256863e72013169f42/pyIGRF/loadCoeffs.py#L50 Can not find such a thing in reference C implementaion. username_2: This project is directly recompiled from Fortran code. The magic number like "ll=3255" is everywhere. The var "ll" represent the positions of first g10. So adding 199 to "ll" means using the next row i.e. 2020. It is the solution to fix this error. Status: Issue closed
pauljherrera/Oanda-Forex-Trader
278219515
Title: Cancelar limit orders anteriores a una fecha especificada Question: username_0: Cliente: Add function to close limit orders, selected by date. Obtener todas las limit orders que no se han ejecutado y cerrar todas aquellas que se hayan abierto antes de cierta fecha. Interfaz: def cancel_limit_orders(self, date):
mirsaeedi/dotnet-apim
724615027
Title: ARM template parameters in config Question: username_0: Hi, Is it possible to use ARM parameters in the config file. For our backend we make use of Azure Functions with function keys. Function keys are secret and are not allowed to be committed in our repository (e.g. config files). That's why we prefer to use parameters in our ARM deployment. Is this already existing functionality? If so, can you please explain how that works? If not, is there an alternative way to change parameters / variables in config files over different environments? Thanks in advance.
researchstudio-sat/webofneeds
413358567
Title: Archived Need view breaks when opening Question: username_0: ![grafik](https://user-images.githubusercontent.com/32457902/53239201-b40b9700-369b-11e9-9a4e-c9c2f8efaff2.png) ![grafik](https://user-images.githubusercontent.com/32457902/53239184-a6561180-369b-11e9-932a-5cc9fa47f780.png) ![grafik](https://user-images.githubusercontent.com/32457902/53239224-c4237680-369b-11e9-9243-aab160cfdfa7.png) Occured on live (22.02.19) Answers: username_1: pretty sure pr #2762 would fix this username_1: nvm the issue is not fixed within that pr... Status: Issue closed
JuliaLang/julia
198237774
Title: deprecate rand! etc. in favor of x .= rand.()? Question: username_0: Since we can now do `x .= rand.()` to evaluate `rand()` for each element of `x`, do we still need `rand!` etcetera? See also #12277. Answers: username_0: (Though `rand!()` does have an optimized SIMD implementation IIRC.) username_1: Can't the specific broadcast function that `x .= rand.()` lowers into be overloaded to use the optimized version? username_0: Yes, it can. Of course, it won't work if you fuse the `rand.()` with additional operations. username_1: Yes, but that should be fine I think. You couldn't used the optimized one if you wrote the fused version manually anyway, right? username_0: Right. username_2: Can one replace use cases like `rand!(similar(X))` with `similar(X) .= rand.()`? username_0: @username_2, that would work. I agree it's more verbose, but I don't think that construct is very common; I can only find two packages that use it in test scripts, and in both cases they could have used `rand(size(X)...)` as well. username_0: On the other hand, `x .= rand.()` is not a drop-in replacement for `rand!(x)` if `eltype(x)` != `Float64`, so I think probably we should keep `rand!`. (The argument for replacing `fill!` and `scale!` remains, however.) Status: Issue closed username_3: Perhaps `fill!` and `scale!` deserve dedicated issues? Best! username_3: (Also relevant to `randexp!` and `randn!`.)
mobxjs/mobx-state-tree
559646432
Title: Type safety when setting a reference with actions Question: username_0: **_Question_** * [*] I've checked documentation and searched for existing issues * [*] I tried the [spectrum channel](https://spectrum.chat/?t=dad48299-3dfc-4e10-b6da-9af1e39498a3) Here's my model ```javascript export const AccountModel = types.model({ id: types.identifier, username: types.string }) export const AccountsModel = types.model('Accounts', { list: types.array(AccountModel), selectedAccount: types.maybeNull(types.reference(AccountModel)) }) ``` Here's my actions: ```javascript .actions(self => { return { selectAccount(newAccount: IAccountModel) { // We select the new account self.selectedAccount = newAccount.id // TypeScript Error } }) ``` You can see where typescript goes wrong on the "TypeScript Error" line. I searched a lot through the documentation and the spectrum chat but I the only way to do this without a typescript warning is by typing `self.selectedAccount = newAccount.id as any` . But it's obviously not very good practice. Any way to do this in a cleaner way ? Answers: username_1: Note that you can do just `self.selectedAccount = newAccount` if is a model instance, MST will pick the `id` automatically. If that doesn't help, please provide a small reproduction in a code sandbox so that others can inspect it :) username_2: @username_1 I also have an issue with this and made a minimal sandbox. Would be nice if you could take a look on it and tell me what I do wrong: https://codesandbox.io/s/mst-set-reference-ts-6kul3 username_1: @username_2 references are references to things that are stored already somewhere else in the tree. Since that doesn't seem to be the case in your example, I guess you should be using just `currentEntry: types.maybeNull(Entry)`, and not a reference at all. If you have further questions, please open a new issue as this doesn't seem to relate to the current thread. Status: Issue closed username_1: Also closing this as the original question has been answered. username_2: I was actually referencing things in the tree, I just overlooked the fact that I need to use an identifier type. Got it fixed, thanks!
krzosa/OddJobsServiceFrontend
619580933
Title: "Rejestracja" > entering the registration form Question: username_0: 1. run the application 2. click on `Rejestracja` ** in fields that expect only letters, you can enter numbers and Vice versa, should it be so?** ![image](https://user-images.githubusercontent.com/62210045/82132864-c57f1880-97e4-11ea-8ab6-6aafcd409aa4.png)
nozavroni/collections
198378416
Title: Create _ function Question: username_0: Create an alias for "invoke" and call it "_". This will give PHP lisp-like syntax. ```php $lib = [ 'add' => function($a) { return function($b) use ($a) { return $a + $b; }; }, 'mult' => function($a) { return function ($B) use ($a) { return $a * $b; }; } }; function _($call, ...$args) use ($lib) { $func = $lib[$call]; return $func(...$args); } _(_('add', 10), 10); // 20 _(_('mult', 10), 10); // 100 ```
plotly/dash
743587529
Title: I cant see the same result Question: username_0: Hi, Nowadays im focused on dash, and i wanna create a side by side graph but when I tried the code from [this tutorial](https://dash-building-blocks.readthedocs.io/en/latest/examples/multigraph.html), I cant see the same result as expected. My graphs are one under the other. Is this tutorial is not up to date or what is the problem? Can you recommend me to another way to do clean side by side graphs like this or with the dropdowns? Thanks in advance! Answers: username_0: I solved the problem by using dbc.Row() function. It is so useful.
dtolnay/syn
282800190
Title: Parse dyn trait Question: username_0: https://github.com/rust-lang/rust/blob/master/src/test/run-pass/dyn-trait.rs ```rust let x: &(dyn 'static + Display) = &BYTE; let y: Box<dyn Display + 'static> = Box::new(BYTE); ``` Answers: username_0: Implemented in e45b59fe322544ee145615e8e1bfa6a0fe284d94. Status: Issue closed
robb0wen/synthwave-vscode
444344505
Title: can't edit and no glow Question: username_0: sorry forgot to do that Answers: username_1: Where is a description? username_0: sorry forgot to do that username_0: it has a description now username_2: If you edit the CSS you have to re-run the 'enable CSS and JS plugin' from your command palette before the changes will be visible. However, if you aren't seeing the glow, then it could be that the plugin isn't working. Have you followed the installation guide for the CSS and JS plugin? there are permissions errors that some users have had that you might also be experiencing username_0: I did follow the installation guide but the glow is still not working. Also I am re-running the 'enable CSS and JS plugin' but it doesn't change anything even if i remove a couple of sections it still doesn't change. username_0: i just found it that even if i remove every file in the theme folder in the .vscode extension folder that is changing, as long as the them folder (either with or without files) is in the extention folder i can use the theme but its all still the same. username_0: is there maybe a way to just combine the theme and the glow cuz i found someone who made some changes to the code so you wont have to run the glow seperetly maybe try that. https://gist.github.com/iamkilo/afbbadc92b504deba093ba96174452cd this is the link to that guy's post it would be a huge help if you can do that and it would safe a lot of problems username_0: @username_2 could you consider doing this in the next update
defold/editor2-issues
217274172
Title: Frame selection in outline Question: username_0: Ability to frame the current selection in outliner Answers: username_1: This should already be possible, see "Scene -> Camera -> Frame Selection". The shortcut keybinding does not currently work on macOS though https://github.com/defold/editor2-issues/issues/425. username_0: Depends, it doesn't work on mac as you say so I can't test it. When I use the menu way it only frames the scene not the outliner. But in a recent update the outline in the selection behaves better anyway, when I reported this the item you selected in the scene didn't get focus in the outliner. So it was hard to find it in the hierarchy. Maybe that is enough. username_1: We merged some big changes to how manipulators (and selection) work yesterday so that may be what you are seeing. That the keybinding doesn't work is of course poor, will see if I can fix it a bit sooner. username_0: The editor frames stuff as it should, so this can be closed Status: Issue closed
boto/boto3
126147011
Title: ApiGat Question: username_0: Hello, I have trouble with Api Gateway to create a new Model I have the follow function ```Python def create_model(api_id, name, description, schema): try: client.create_model( restApiId=api_id, name=name, description=description, schema=schema, contentType='application/json' ) return True except Exception: return False ``` I don't know how to pass the schema var to the function, my schema.json is ```json { "$schema": "http://json-schema.org/draft-04/schema#", "title": "login", "type": "object", "properties": { "error_code": { "type": "integer" }, "error_message": { "type": "string" } } } ``` I open the json file, the doc says schema is a string so I do: ```python ... json_conifg='schema.json' json_data=open(json_conifg).read() model = json.loads(json_data) if create_model(api_id, response['model'], 'Test', str(model)): print "The Model was created successfully" else: print "Error creating the Model " + response['model'] ``` But I receive the error: ```json botocore.exceptions.ClientError: An error occurred (BadRequestException) when calling the CreateModel operation: Invalid model specified: Validation Result: warnings : [], errors : Invalid model schema specified ``` How can I parse or convert my schema.json to string? Answers: username_0: I solve it chaging ```python model=json.loads(json.dumps(json_data)) if create_model(api_id, response['model'], 'Test', model): print "The Model was created successfully" else: print "Error creating the Model " + response['model'] ``` Status: Issue closed username_0: model=json.loads(json.dumps(json_data)) if create_model(api_id, response['model'], 'Test', model): print "The Model was created successfully" else: print "Error creating the Model " + response['model']
olofk/edalize
1099181245
Title: ModelSim makefile is problematic on Windows Question: username_0: The ModelSim makefile contains a check to detect if vco is run as 32 or 64 bits. This check explicitly calls for shell which seems like it tries to invoke sh, even on windows. Suggestion would be to remove or somehow rewrite this check. Users can already today set MTI_VCO_MODE to force 32- or 64-bit behaviour. Perhaps this is good enough if it's documented somewhere
macbre/analyze-css
793251295
Title: Freezing on a syntax Question: username_0: Hi @username_1, I've noticed that a specific syntax in a CSS file can freeze analyze-css, which in return freezes phantomas. How to reproduce: 1/ Copy-paste the following code in a file: ```css .someclass:not([ng-show="gcdmShowSystemNotAvailableMessage()"]){display:none} ``` 2/ Run analyze-css on this file 3/ It freezes. Here is the log: ``` DEBUG=* bin/analyze-css.js --file bugged-syntax.css analyze-css:bin analyze-css v0.12.12 +0ms analyze-css:bin opts: {"file":"error-file.css","noOffenders":false} +4ms analyze-css:runner Loading local CSS file: /Users/gaelmetais/Documents/GitHub/phantomas/error-file.css +0ms analyze-css:preprocessors Initializing... +0ms analyze-css opts: {"noOffenders":false,"preprocessor":false,"file":"/Users/gaelmetais/Documents/GitHub/phantomas/error-file.css"} +0ms analyze-css:rules Rules to be loaded: base64, bodySelectors, childSelectors, colors, comments, complex, duplicated, emptyRules, expressions, ieFixes, import, important, length, mediaQueries, minified, multiClassesSelectors, parsingErrors, prefixes, propertyResets, qualified, specificity, stats +0ms analyze-css:rules "base64" loaded: Reports on base64-encoded images +36ms analyze-css:rules "bodySelectors" loaded: Reports redundant body selectors +3ms analyze-css:rules "childSelectors" loaded: Reports redundant child nodes selectors +14ms analyze-css:rules "colors" loaded: Reports number of unique colors used in CSS +116ms analyze-css:rules "comments" loaded: Reports too long CSS comments +2ms analyze-css:rules "complex" loaded: Reports too complex CSS selectors +5ms analyze-css:rules "duplicated" loaded: Reports duplicated CSS selectors and properties +4ms analyze-css:rules "emptyRules" loaded: Total number of empty CSS rules +2ms analyze-css:rules "expressions" loaded: Reports CSS expressions +6ms analyze-css:rules "ieFixes" loaded: Reports fixes for old versions of Internet Explorer (IE9 and below) +5ms analyze-css:rules "import" loaded: Number of @import rules +4ms analyze-css:rules "important" loaded: Number of properties with value forced by !important +2ms analyze-css:rules "length" loaded: Length of CSS file +2ms analyze-css:rules "mediaQueries" loaded: Reports media queries +4ms analyze-css:rules "minified" loaded: Reports not minified CSS +3ms analyze-css:rules "multiClassesSelectors" loaded: Reports selectors with multiple classes +4ms analyze-css:rules "parsingErrors" loaded: CSS parsing errors +26ms analyze-css:prefixes Using data generated on 2020-09-08 using autoprefixer-core v6.0.1 +0ms analyze-css:rules "prefixes" loaded: Reports outdated vendor prefixes +6ms analyze-css:rules "propertyResets" loaded: Reports accidental property resets +4ms analyze-css:rules "qualified" loaded: Reports qualified selectors +2ms analyze-css:rules "specificity" loaded: Reports rules specificity +37ms analyze-css:rules "stats" loaded: Emit CSS stats +2ms analyze-css:parser Going to parse 0.08 kB of CSS +0ms analyze-css:parser CSS parsed +7ms analyze-css {"type":"rule","selectors":[".someclass:not([ng-show=\"gcdmShowSystemNotAvailableMessage()\"])"],"declarations":[{"type":"declaration","property":"display","value":"none","position":{"start":{"line":1,"column":65},"end":{"line":1,"column":77}}}],"position":{"start":{"line":1,"column":1},"end":{"line":1,"column":78}}} +304ms [freeze] ``` Note: The [W3C CSS validator](https://jigsaw.w3.org/css-validator/validator) says the syntax is correct. Status: Issue closed Answers: username_1: @username_0 - finally dealt with it. Will publish v2.0 soon 🚀
alphacep/vosk-api
545090479
Title: Task :lint FAILED Question: username_0: Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.0/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 5s 1 actionable task: 1 executed Answers: username_0: 1. It seems that this `vosk-api` repo needs modifications for either `Android build` or `Python module build`. Especially I need to modify the `CMakeLists.txt` files for both to run correctly. 2. Could you also specify which directory exactly we should clone the kaldi source to? It seems for `Android build`, we should create a folder `build/kaldi_${KALDI_SUFFIX}` under `android` folder in this repo, and then we clone the kaldi source under `android/build/kaldi_${KALDI_SUFFIX}` username_1: You shouldn't clone kaldi yourself, gradle does it automatically inside build-kaldi script. username_0: Task :generateJsonModelDebug debug|armeabi-v7a :-- Check for working C compiler: /home/liuyi/Android/Sdk/ndk/20.1.5948944/toolchains/llvm/prebuilt/linux-x86_64/bin/clang debug|armeabi-v7a :-- Check for working C compiler: /home/liuyi/Android/Sdk/ndk/20.1.5948944/toolchains/llvm/prebuilt/linux-x86_64/bin/clang -- works debug|armeabi-v7a :-- Detecting C compiler ABI info debug|armeabi-v7a :-- Detecting C compiler ABI info - done debug|armeabi-v7a :-- Detecting C compile features debug|armeabi-v7a :-- Detecting C compile features - done debug|armeabi-v7a :-- Check for working CXX compiler: /home/liuyi/Android/Sdk/ndk/20.1.5948944/toolchains/llvm/prebuilt/linux-x86_64/bin/clang++ debug|armeabi-v7a :-- Check for working CXX compiler: /home/liuyi/Android/Sdk/ndk/20.1.5948944/toolchains/llvm/prebuilt/linux-x86_64/bin/clang++ -- works debug|armeabi-v7a :-- Detecting CXX compiler ABI info debug|armeabi-v7a :-- Detecting CXX compiler ABI info - done debug|armeabi-v7a :-- Detecting CXX compile features debug|armeabi-v7a :-- Detecting CXX compile features - done debug|armeabi-v7a :-- Configuring done debug|armeabi-v7a :-- Generating done debug|armeabi-v7a :-- Build files have been written to: /home/liuyi/Documents/VoiceAssist/vosk-api/android/.cxx/cmake/debug/armeabi-v7a debug|arm64-v8a :-- Check for working C compiler: /home/liuyi/Android/Sdk/ndk/20.1.5948944/toolchains/llvm/prebuilt/linux-x86_64/bin/clang debug|arm64-v8a :-- Check for working C compiler: /home/liuyi/Android/Sdk/ndk/20.1.5948944/toolchains/llvm/prebuilt/linux-x86_64/bin/clang -- works debug|arm64-v8a :-- Detecting C compiler ABI info debug|arm64-v8a :-- Detecting C compiler ABI info - done debug|arm64-v8a :-- Detecting C compile features debug|arm64-v8a :-- Detecting C compile features - done debug|arm64-v8a :-- Check for working CXX compiler: /home/liuyi/Android/Sdk/ndk/20.1.5948944/toolchains/llvm/prebuilt/linux-x86_64/bin/clang++ debug|arm64-v8a :-- Check for working CXX compiler: /home/liuyi/Android/Sdk/ndk/20.1.5948944/toolchains/llvm/prebuilt/linux-x86_64/bin/clang++ -- works debug|arm64-v8a :-- Detecting CXX compiler ABI info debug|arm64-v8a :-- Detecting CXX compiler ABI info - done debug|arm64-v8a :-- Detecting CXX compile features debug|arm64-v8a :-- Detecting CXX compile features - done debug|arm64-v8a :-- Configuring done debug|arm64-v8a :-- Generating done debug|arm64-v8a :-- Build files have been written to: /home/liuyi/Documents/VoiceAssist/vosk-api/android/.cxx/cmake/debug/arm64-v8a debug|x86_64 :-- Check for working C compiler: /home/liuyi/Android/Sdk/ndk/20.1.5948944/toolchains/llvm/prebuilt/linux-x86_64/bin/clang debug|x86_64 :-- Check for working C compiler: /home/liuyi/Android/Sdk/ndk/20.1.5948944/toolchains/llvm/prebuilt/linux-x86_64/bin/clang -- works debug|x86_64 :-- Detecting C compiler ABI info debug|x86_64 :-- Detecting C compiler ABI info - done debug|x86_64 :-- Detecting C compile features debug|x86_64 :-- Detecting C compile features - done debug|x86_64 :-- Check for working CXX compiler: /home/liuyi/Android/Sdk/ndk/20.1.5948944/toolchains/llvm/prebuilt/linux-x86_64/bin/clang++ debug|x86_64 :-- Check for working CXX compiler: /home/liuyi/Android/Sdk/ndk/20.1.5948944/toolchains/llvm/prebuilt/linux-x86_64/bin/clang++ -- works debug|x86_64 :-- Detecting CXX compiler ABI info debug|x86_64 :-- Detecting CXX compiler ABI info - done debug|x86_64 :-- Detecting CXX compile features debug|x86_64 :-- Detecting CXX compile features - done debug|x86_64 :-- Configuring done debug|x86_64 :-- Generating done debug|x86_64 :-- Build files have been written to: /home/liuyi/Documents/VoiceAssist/vosk-api/android/.cxx/cmake/debug/x86_64 Task :externalNativeBuildDebug FAILED Build kaldi_jni_armeabi-v7a ninja: error: '../../../../build/kaldi_arm_32/local/lib/libfst.a', needed by '../../../../build/intermediates/cmake/debug/obj/armeabi-v7a/libkaldi_jni.so', missing and no known rule to make it FAILURE: Build failed with an exception. What went wrong: Execution failed for task ':externalNativeBuildDebug'. Build command failed. Error while executing process /home/liuyi/Android/Sdk/cmake/3.6.4111459/bin/cmake with arguments {--build /home/liuyi/Documents/VoiceAssist/vosk-api/android/.cxx/cmake/debug/armeabi-v7a --target kaldi_jni} ninja: error: '../../../../build/kaldi_arm_32/local/lib/libfst.a', needed by '../../../../build/intermediates/cmake/debug/obj/armeabi-v7a/libkaldi_jni.so', missing and no known rule to make it Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.0/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 8s 16 actionable tasks: 16 executed username_1: Update and try again. username_0: 1 actionable task: 1 executed (base) liuyi@liuyi-HP:android$ ls username_1: You don't have ndk installed in your sdk. Status: Issue closed
Emraldis/Destiny2LFGDiscordBot
663439866
Title: Add command to start an activity Question: username_0: Add a command to start an activity. !LFG -start n This will ping *matching* players who have marked themselves as available for an activity. matching to be used in the future with a list of eligible activities. n indicates the number of players the user is looking for. Should be compatible with the use of --list in the future Spam Prevention: A user should only be able to ping once every 5 minutes. A user should only be able to be pinged 1 time per minute.
apache/camel-quarkus
718255376
Title: Add a filter to discard AnnotationTypeConverterLoader Question: username_0: Citing from the comment in `org.apache.camel.quarkus.core.deployment.CamelProcessor.typeConverterRegistry(CamelRecorder, RecorderContext, ApplicationArchivesBuildItem, List<CamelTypeConverterLoaderBuildItem>)`: we could add a filter to discard AnnotationTypeConverterLoader but maybe we should introduce a marker interface like StaticTypeConverterLoader for loaders that do not require to perform any discovery at runtime. cc @lburgazzoli
mvysny/photocloud-frame-slideshow
299130173
Title: DLNA support stopped working Question: username_0: Since one of the recent releases, browsing or showing DLNA feeds stopped working. The exception reported is the following: org.xmlpull.v1.XmlPullParserException: undefined prefix: sec (position:START_TAG <{null}sec:dcmInfo@>2:167 in java.io.StringReader ...) Answers: username_0: ![screenshot_20180221-221619](https://user-images.githubusercontent.com/34660840/36505845-101f4c56-1755-11e8-95eb-99679cd0abac.png) username_1: I'm sorry to hear that. However, it seems that the problem is not in PhotoCloud - it seems that your DLNA server is producing an unparsable XML file. Could it be that a firmware has been upgraded on your DLNA device recently? I have tested out my Synology diskstation DLNA integration and it works as before. Can you please try other DLNA apps whether they work or whether they are broken as well? username_0: Hi Martin - Thanks for your reply. I am running minidlna 1.1.2 on a raspberry pi. There has been no version change in quite some time. Older versions of Photocloud used to work with this server. I just tried a few other DLNA clients now and they could connect to my server and browse fine: I tried MediaHouse-Pro and BubbleUPnP on my Android phone and also tried my LG PF1500 projector. They show no sign of problems. If you have APKs for the 1.10.x series I can try if they still work. Cheers, Wilfried username_0: Actually, sorry, ignore that. I just tried various older photocloud versions going back until 1.10.3 and they show the same problem, so it must still somehow be related to my server. Maybe I have some content on it now that I didn't have in the past. I will keep investigating Cheers On February 22, 2018 22:44:34 <NAME> <<EMAIL>> wrote: > Hi Martin - > > Thanks for your reply. I am running minidlna 1.1.2 on a raspberry pi. There > has been no version change in quite some time. Older versions of Photocloud > used to work with this server. > > I just tried a few other DLNA clients now and they could connect to my > server and browse fine: > > I tried MediaHouse-Pro and BubbleUPnP on my Android phone and also tried my > LG PF1500 projector. They show no sign of problems. > > If you have APKs for the 1.10.x series I can try if they still work. > > Cheers, Wilfried > > username_1: Thanks for the investigation. I also have a testing installation of linux-based DLNA server on my ubuntu server, I'll check the version and whether it is still compatible. I have also checked whether I haven't upgraded the Android client DLNA library lately, but nope - it has been cling 2.1.1 from the very beginning. It might be that some funky files might cause funky DLNA XML file which cling can't parse. username_0: Yes, thanks Martin, ok. My workaround in the meantime is to use the ssh/sftp feed type. For my purposes that works just as well, maybe even better than DLNA. If you find no problem with your Ubuntu server, then please feel free to close the ticket. Thanks a lot for your time and have a nice weekend Wilfried Status: Issue closed username_1: I have set up minidlna 1.2.0+dfsg-2 on my Ubuntu 17.10 and I have failed to reproduce the issue - it works properly. That being said, it definitely may be a case that there is a bug in minidlna triggered by a particular file type, or a set of files. If you can spare the time, you can investigate which folder exactly causes this issue to happen; you can then send me the file in question and I can file a bug report for minidlna.
jcmturner/gokrb5
405683554
Title: webasm Question: username_0: Is it possible to use gokrb5 with web asm (https://github.com/golang/go/wiki/WebAssembly) ? Answers: username_1: @username_0 I don't know much about web assembly. What's your use case? have you tried it to see what happens? Status: Issue closed
Azure/azure-iot-sdk-node
586819858
Title: NotConnectedError: connect ECONNREFUSED 172.18.0.2:8883 Question: username_0: <!-- Hi there! thank you for discovering and submitting an issue! # Need Support? - **Have a feature request for SDKs?** Please post it on [User Voice](https://feedback.azure.com/forums/321918-azure-iot) to help us prioritize - **Have a technical question?** Ask on [Stack Overflow with tag "azure-iot-hub"](https://stackoverflow.com/questions/tagged/azure-iot-hub) - **Need Support?** Every customer with an active Azure subscription has access to [support](https://docs.microsoft.com/en-us/azure/azure-supportability/how-to-create-azure-support-request) with guaranteed response time. Consider submitting a ticket and get assistance from Microsoft support team - **Found a bug?** Please help us fix it by thoroughly documenting it and filing an issue on GitHub (See below). Please first tell us a little bit about the environment you're running: The commands in the comments can be run directly in a command prompt. --> # Context - **OS and version used:** Ubuntu 18.04.4 LTS, Jetson nano aarch64 aarch64 aarch64 GNU/Linux - **Node.js version:** v10.19.0 - **npm version:** 6.13.4 - **list of installed packages:** ├──[email protected] ├── [email protected] # Description of the issue: Module client not able to communicate with azure iot hub # Steps followed 1. Installed vscode extensions (Azure account, Azure iot hub, Azure iotedge) 2. Create a iot edge device in Azure iot hub 3. update `/etc/iotedge/config.yaml` file with device connection string. 4. `$ sudo systemctl restart iotedge` 5. `CTRl + SHIFT + P` , select `New IotT Edge solution` 6. Select nodejs samplemodule 7. Right click on the `deployment.template.json` and select `Build and push IoT Edge Solution`. 8. In vscode add the iothub connection string 9. Right click on `/config/deployment.arm64v8.json` and select `Create Deployment For Single device` 10. Select the device. # Code sample exhibiting the issue: app.js file ``` 'use strict'; var Transport = require('azure-iot-device-mqtt').Mqtt; var Client = require('azure-iot-device').ModuleClient; var Message = require('azure-iot-device').Message; console.log('started.....') Client.fromEnvironment(Transport, function (err, client) { if (err) { throw err; } else { client.on('error', function (err) { throw err; }); // connect to the Edge instance client.open(function (err) { if (err) { [Truncated] COPY package*.json ./ RUN npm install --production COPY app.js ./ CMD ["node", "app.js"] ``` # Console log of the issue: ``` /app/app.js:20 throw err; ^ NotConnectedError: connect ECONNREFUSED 172.18.0.2:8883 ``` Answers: username_1: It's not clear this is an issue with the SDK, but seems like things aren't working for you in Edge. I'd open up an issue in https://github.com/Azure/iotedge and try to get help there. Also, you've probably already seen this but just in case here is a tutorial on IoT Edge with Node.js for Linux: https://docs.microsoft.com/en-us/azure/iot-edge/tutorial-node-module Make sure to follow all the steps :) Status: Issue closed username_2: No activity. username_3: @username_0, @username_1, thank you for your contribution to our open-sourced project! Please help us improve by filling out this 2-minute [customer satisfaction survey](https://forms.office.com/Pages/ResponsePage.aspx?id=v4j5cvGGr0<KEY>)
MhLiao/DB
814450770
Title: Training for OpenCV TextDetectionModel_DB Question: username_0: Hi, I'm trying to train a new custom DB model and possibly convert it to ONNX format to use it with OpenCV. I'm already using pre-trained weights which are given [here](https://drive.google.com/drive/folders/1qzNCHfUJOS0NEUOIKn69eCtxdlNPpWbq). I've also managed to train and get good results with eval.py and demo.py. Then I tried to convert that model file to ONNX using _torch.onnx.export._ Here is how I did it after loading the model and weights; `dummy_input = torch.randn(1, 3, None, None, requires_grad=True) torch.onnx.export(model, dummy_input, "db.onnx")` This results in an error `RuntimeError: hasSpecialCase INTERNAL ASSERT FAILED at /pytorch/torch/csrc/jit/passes/alias_analysis.cpp:300, please report a bug to PyTorch. We don't have an op for aten::to but it isn't a special case. (analyzeImpl at /pytorch/torch/csrc/jit/passes/alias_analysis.cpp:300)` To fix this in [model.py(line 39)](https://github.com/MhLiao/DB/blob/6f208690cbb1a84e259f45fff465fad4ae5a6751/structure/model.py#L39) I have commented out paralleling of model. `... self.model = BasicModel(args) # for loading models # self.model = parallelize(self.model, distributed, local_rank) self.criterion = SegDetectorLossBuilder(...` After that it exports an onnx file without an error. When I try to import that model to opencv and do inference, I get these error; `opencv/modules/core/src/matrix.cpp:806: error: (-215:Assertion failed) m.dims <= 2 in function 'Mat'` I'm not sure if the problem is occuring when converting to onnx. Problem might be opencv related. Are there any guidelines to train/use our own DB models with opencv? Answers: username_1: Hi <NAME>! Coulde you show CODE for convert pytorch to ONNX. Many thanks you username_0: I used pytorch's function torch.onnx.export which you can find details [here](https://pytorch.org/tutorials/advanced/super_resolution_with_onnxruntime.html). To be more specific I have modified the demo.py to create the model and load my weights into it. [Added a few lines here](https://github.com/MhLiao/DB/blob/4e341f7c07005fe5cccac8f4042c03f0ef4424cb/demo.py#L125) `def inference(self, image_path, visualize=False): self.init_torch_tensor() model = self.init_model() self.resume(model, self.model_path) dummy_input = torch.randn(1, 3, 736, 992, requires_grad=True) all_matircs = {} model.eval() torch.onnx.export(model, dummy_input, "db.onnx") batch = dict() batch['filename'] = [image_path] img, original_shape = self.load_image(image_path) batch['shape'] = [original_shape] with torch.no_grad(): batch['image'] = img print(img.shape) pred = model.forward(batch, training=False) output = self.structure.representer.represent(batch, pred, is_output_polygon=self.args['polygon']) if not os.path.isdir(self.args['result_dir']): os.mkdir(self.args['result_dir']) self.format_output(batch, output) if visualize and self.structure.visualizer: vis_image = self.structure.visualizer.demo_visualize(image_path, output) cv2.imwrite(os.path.join(self.args['result_dir'], image_path.split('/')[-1].split('.')[0]+'.jpg'), vis_image)` username_1: Thank you your quick reply. Maybe we have to train with non paralelnize mode. I trained with parallel mode and then load with non parallel to inference the results is bad username_0: I doubt that data paralleling effects inference results. It would be great if someone who achieved training and transporting that model to openCV could help. username_1: When i try using your code to convert torch to onnx i got error like below: ![Capture](https://user-images.githubusercontent.com/54969565/115492979-12892b80-a28d-11eb-91c9-3d6e3bee1f8b.JPG) But when i using https://github.com/Media-Smart/volksdep it is ok. Do you know why? username_2: @MhLiao I also encounter this problem, can you help us? username_2: @username_0 @username_1 username_2: Have you solve this problem? username_0: Unfortunately no I didn't and currently working on something else. But I'm still keen to learn if this is possible. username_2: @username_0 Have change DCN module? When I follow your method, get following error: `RuntimeError: ONNX export failed: Couldn't export Python operator ModulatedDeformConvFunction`
mtchavez/circleci
118193027
Title: Missing API -> POST: /user/ssh-key Question: username_0: Adds a CircleCI key to your Github User account. Answers: username_1: Doesn't appear to be a valid API endpoint. Returns a `404` closing until more is known on whether this is supported still. Status: Issue closed username_0: Yes. I tried this too. It doesn't seems to be working. They have given this in API doc. Will check again. Thanks.
spring-projects/spring-hateoas
266173420
Title: Adding new HttpMessageConverters does not work Question: username_0: I'm developing a HAL-based REST API with spring-boot and spring-hateoas. I have a bean in my configuration class, which adds a new `ByteArrayHttpMessageConverter`, because my Controller returns a `ByteArrayResource`. When i debug the application and end up in the `AbstractMessageConverterMethodProcessor` class, i can't see my `ByteArrayHttpMessageConverter` in the `messageConverters` list. My config file looks like this: ``` @Configuration public class MyConfig{ private static final String SPRING_HATEOAS_OBJECT_MAPPER = "_halObjectMapper"; @Autowired @Qualifier(SPRING_HATEOAS_OBJECT_MAPPER) private ObjectMapper springHateoasObjectMapper; @Autowired private Jackson2ObjectMapperBuilder springBootObjectMapperBuilder; @Bean(name = "objectMapper") @Primary ObjectMapper objectMapper() { this.springBootObjectMapperBuilder.configure(this.springHateoasObjectMapper); return springHateoasObjectMapper; } @Bean public HttpMessageConverters customConverters() { ByteArrayHttpMessageConverter arrayHttpMessageConverter = new ByteArrayHttpMessageConverter(); arrayHttpMessageConverter.setSupportedMediaTypes(Arrays.asList(MediaType.APPLICATION_OCTET_STREAM)); return new HttpMessageConverters(arrayHttpMessageConverter); } @Bean public Jackson2ObjectMapperBuilderCustomizer objectMapperBuilder() { return builder -> builder.configure(springHateoasObjectMapper); } } ´´´ Answers: username_1: Duplicates #118, #352, and #149. username_1: Right now, Spring HATEOAS has a static-based configuration that has a BIG side effect of making dynamic updates like this tricky. The linked PR is _a_ solution, but possibly not the best one. We're looking at tackling this in the new 1.0 branch in a more fluid, Spring fashion. username_0: @username_1 Which linked PR are you talking about? Thank you, for the answer. username_1: Sorry, left out #618 username_0: I couldn't solve my problem with the PR. I found a stackoverflow answer, what solved the problem. This should be better handled from spring-hateoas.... My code: ``` @EnableHypermediaSupport(type = HypermediaType.HAL) @Configuration public class ServerConfig extends WebMvcConfigurationSupport { @Autowired private RequestMappingHandlerAdapter adapter; @PostConstruct public void initStuff() { List<HttpMessageConverter<?>> messageConverters = adapter.getMessageConverters(); ByteArrayHttpMessageConverter arrayHttpMessageConverter = new ByteArrayHttpMessageConverter(); ResourceHttpMessageConverter resourceHttpMessageConverter = new ResourceHttpMessageConverter(); messageConverters.add(0, arrayHttpMessageConverter); messageConverters.add(0, resourceHttpMessageConverter); } } ``` username_1: The issue with using`@EnableHypermediaSupport` is you disable all the things supplied by Spring Boot and take it into yourself. So that isn't an optimal solution for most people. It's an issue @olivergierke and I are looking into for a better developer experience. username_2: Is there any update on this? I'm currently having to use strings for my endpoints as I'm using value objects for all my ID path variables. Just wondering if there is a plan? username_1: Superseded by https://github.com/spring-projects/spring-hateoas/issues/833 Status: Issue closed
pytorch/pytorch
445029072
Title: Change function that truncate to scalar_t on each atomicAdd Question: username_0: ## 🚀 Feature <!-- A clear and concise description of the feature proposal --> at https://github.com/pytorch/pytorch/pull/19630/files#diff-5092da792c30694ee4adf0d0ae2a37c6R206 it is truncating to scalar_t on each atomicAdd call. It would be good to change that. See comments by @ngimel https://github.com/pytorch/pytorch/pull/19630#discussion_r281426912 ## Motivation <!-- Please outline the motivation for the proposal. Is your feature request related to a problem? e.g., I'm always frustrated when [...]. If this is related to another GitHub issue, please link here too --> Performance improving ## Pitch <!-- A clear and concise description of what you want to happen. --> ## Alternatives <!-- A clear and concise description of any alternative solutions or features you've considered, if any. --> ## Additional context <!-- Add any other context or screenshots about the feature request here. --> Answers: username_0: @ngimel please feel free to add any more details about your thoughts on ths issue. Thank you so much! username_1: Closing due to age. @username_0, would you please update and reopen this with additional details if this issue still exists? Status: Issue closed username_0: @username_1 IIRC I opened this issue to take notes from a PR review for a follow up task. Not sure if it is still relevant. Thanks!
adonisjs/core
640169717
Title: Change inlineSvg tag to use a config'ed path instead of public path Question: username_0: Currently the inlineSvg tag looks for a svg files relative to the public directory. Instead of hard-coding this to use the public directory, it would be a huge improvement to make this path a configurable value, like `Config.get('app.views.svgPath')`. Give it a default value of the public path so that there are no breaking changes, yet it makes the feature more flexible. ## Why this feature is required (specific use-cases will be appreciated)? Two primary reasons that I have come across: 1. Keeping SVGs out of the public directory that are not mean't to be publicly accessible to anyone. Currently if an SVG has to be somewhere in the public directory then it's URL is accessible to everyone. I have a situation where I have a website with some users that are logged in with special privileges and other users that are not logged in and do not have special privileges. It's important for certain SVG files (that are proprietary art files of users) to only be served up and available to my logged in users. But since the `inlineSvg` tag requires SVGs to be in the public directory then those SVGs are publicly available. They are available for anyone to download with the right URL and that just cannot work for both privacy and copyright reasons. Being able to change the path to something not public would be desirable for this reason. 2. Another use case is simply being able to specify shorter svg paths. If you have your SVGs located in `/public/images/svgs/fontawesome/solid/*.svg` then when you have to do `inlineSvg("images/svg/fontawesome/solid/facebook")`. Would be amazing if you could simply change the config path globally to `"public/images/svg/fontawesome/solid"` and then just type in svgs like `inlineSvg("facebook")`. ## Have you tried any other work arounds? The workarounds I have tried: hard coding the SVGs into the templates, which doesn't really work for a lot of my SVGs that are dynamically determined on page load. Loading the SVG code in the page via AJAX on page load. This works. Not ideal. I have not tried it but you could also create a different custom tag to inline svgs with a custom path. I cannot find the documentation on how to do this exactly. ## Are you willing to work on it with little guidance? Absolutely, I would think this would not be much of an issue. I have been looking through the v5 code but cannot make sense of some of it. Like I see the public path being fetched from `Helper.js` but its not clear where this `Helper.js` is at. If its in some other module or what. Answers: username_1: @username_0 There is no really documentation for the custom tag creation. The [main Adonis docs for custom tags](https://adonisjs.com/docs/4.1/views#_tags_2) refers to the [Edge custom tags section](https://edge.adonisjs.com/docs/tags#_writing_your_own_tags) which in turn refers to the missing video (@thetutlage if possible could you please check if this can be fixed). Nevertheless you could explore the source code for the custom [inlineSvg tag](https://adonisjs.com/docs/4.1/views#_inlinesvg) which is available in the `your-project-root/node_modules/%40adonisjs/framework/src/View/Tags/InlineSvg.js`. Especially look up the following method where the path you need to change is defined: ``` javascript _getAbsPath (filePath) { return path.isAbsolute(filePath) ? filePath : path.join(this._publicPath, filePath) } ``` You could make your own version of it, say `inlineSvgFreePath` for which [the docs section](https://adonisjs.com/docs/4.1/views#_tags_2) I believe would be enough. username_0: Looks like a lot of steps to essentially duplicate the `inlineSvg` tag. Would it be more worth my time to just update the core code with the ability to specify a configurable path? Surely I'm not the first person who has desired this feature. username_1: No. This would come seriously inconvenient as you move your application to stage and production environment. The code in the `node_modules` would be overwrtten every time you reinstall your NPM dependencies under new environment. As well your updates would be lost as you do not commit `node_modules` to version control system (and you should not). username_0: No no no. I mean make a pull request to the AdonisJS v5 core as a new feature. So that anyone could use this feature in the future if/when it was merged into the v5 framework. username_1: @username_0 I see.
SwiftGen/SwiftGen
313793733
Title: Storyboards: Support multiple target modules. Question: username_0: Currently the storyboards check what the current module is, to avoid importing it, either via the `module` parameter, or the environment variable. The user should be able to provide multiple modules they may want to ignore. In #405 I said this would be a breaking change, but it isn't really. If the user passes `--param module="stuff"` once, the context value will be a string. If the user passes multiple values, it'll be a string array. It would only be up to the template to handle this difference. Stencil supports the `in` expression in `if` tags, which supports both strings and arrays: https://github.com/kylef/Stencil/blob/6b02fccf84a0adb561dc1f51eb39c5f1b03cd5f5/Sources/Expression.swift#L101-L120 The problem with this approach is that the string comparison uses `.contains`, which we can't use for exact matching. To really solve this we also need to add a new `array` filter (to StencilSwiftKit) to create an array from a value: - If an existing array, just return it - If a string (or int, or ...), wrap it in an array. Note: This would also solve https://github.com/SwiftGen/StencilSwiftKit/issues/44. Once we have this `array` filter, we can simply update the tempaltes to use the `in` expression mentioned above to check for targets. Answers: username_0: Closing this as #435 fixes the underlying issue of "inherit module from target". Status: Issue closed
github/roadmap
665304795
Title: Actions: Create and share workflows templates in an organization Question: username_0: <!-- Remember to... 1. Assign the issue to the appropriate product manager (probably yourself) 2. Apply a label for: release phase, feature area, product SKU, and deployment model, per the README 3. Apply the *public roadmap* label when ready for review to be published to the public repository or the *internal roadmap* label to keep this internal 4. Add to the current year's project board, in the column for when it will ship next 5. Establish links with other internal issues either in the other issue or in the comments (to keep the description customer-friendly) 6. Head over to github/customer-feedback to comment about this roadmap issue on any relevant feedback issues --> **Summary** <!--(What is it?)--> Today, when developers use the GitHub interface to create a new workflow they see a number of "starter workflows" to choose from: ![Screen Shot 2020-04-24 at 10 03 43 AM](https://user-images.githubusercontent.com/1865328/80227171-e8b60e00-8612-11ea-9083-74bd9aa08a51.png) These workflows are available to all repositories and are defined in the [actions/starter-workflows](https://github.com/actions/starter-workflows) repository. **Intended Outcome** <!--(Why are we building it?)--> With this feature, users will be able to define their own starter workflows. These will be defined at the organization level, and will be made available to every repository within the organization. **How will it work?** <!--(Short description of workflow/functionality)--> When developers create a new workflow, they will see these custom workflows and can choose whether to use them or not. Answers: username_1: https://github.blog/changelog/2020-06-23-github-actions-workflow-templates/
arkayenro/arkinventory
166481112
Title: All Options are locked Question: username_0: When I go to change a bar option, it says the bar is locked, the sort option is locked and the assigned category is locked. I have tried reinstalling it Answers: username_1: go into the config and then to controls setup the layout options in there for each location so they are not using "* Default" username_2: I've tried that myself and have had 0 success, then tried deleting everything and making new with no success, uninstalled and reinstalled with no success and made new names for a profile and new sets in settings and set those instead of default in the controls tap with no success... username_2: Ah finally... completely exited the game after that final one adjusting something else I don't recall and they work now username_0: The only option I see in there id the Default and Guild Bank Default username_3: We need to make a new layout but I can't find a place to do so. username_4: to create a new layout go into Config -> Settings -> Design (tab) and add a new one, then go back into Config -> Controls -> Bag (tab) and change the layout, I also created a new category set (from settings) and set that in Controls Bag as well After then reload UI and it starts working again username_0: That got it!! Thanks All username_0: OOPS, spoke to soon, the Assignable Categories is still messed up so I can not change an items category username_0: Am I the only one having this issue? username_1: i found an issue where the settings may not always update, will be fixed in .10 and you shouldnt have to reload to get them to kick in username_5: I am having this problem as well. I cant edit any bags, all are saying locked. username_6: If you have problems with sorting your bags, bank, void storage and so on: 1. Go to Config >> Controls. 2. On every tab under the Blueprint name change the dropdown menus Style, Category Set, and Layout to the names you'd been using before patch 7.0. 3. The dropdown menus Style, Category Set, and Layout MUST NOT contain the word Default. If it does, assigning categories to items will not work. username_7: Ahh Perfect. The .10 update fixed my Profilesettings :-) Thanks for the good work username_1: .11 should fix the categories username_0: ok, that fixed it so I can edit the items and everything else. Where do I change the size though? username_1: size of? should be under one of the style options username_8: After the patch, and 2 updates from Curse, reloading, creating new category, and doing everything thus far suggested, it still does not work. I am attaching a screenshot of what I see, and have yet to discover what to do about it. It occurred to me that maybe there was some kind of conflict.. Any ideas would be appreciated. ![image](https://cloud.githubusercontent.com/assets/20568983/17011027/7a2b6424-4ed7-11e6-9d6e-04362e361c41.png) ![image](https://cloud.githubusercontent.com/assets/20568983/17011035/91e72e4a-4ed7-11e6-9239-e5bcef2a3957.png) username_6: Imagine your old profile name is ANEW and try this: https://www.youtube.com/watch?v=f_XFlsAYBho username_0: Unless I am having a blond moment, I cannot find anything that says style username_6: Here http://imgur.com/x4j7xGj And give me a screenshot like this http://imgur.com/ocGYGEs Status: Issue closed username_1: fixed in .15 editable default username_8: The latest version seems to have fixed the problem. So Far, it seems all is well. Glad to have it back. It is so superior to Bagnon, and OneBag. username_9: Is this some kind of joke? You are not updated your Wiki and you are going with your old bullshit "go into the config and then to controls", CONTROLS are not presented anymore in the latest version. This shit happened because of your note "remove profile" and now everything is FUCKED UP and bags are looking like shit, no item level, no borders, wasn't able to set filters, JUST PURE NOTHING!
ThibaultLatrille/ControverSciences
123456594
Title: La barre d'outils qui est tout en haut ! Question: username_0: Quand on rédige une synthèse ça peut être long en terme de déroulé de scrollbar, du coup au bout d'un moment, on n'a plus la barre d'outils "gras, italique, lien http, citation, toutçatoutça". Ce serait génial si elle suivait l'écran sur le coté à la verticale, (un peu comme les pubs "collantes" sur certains sites). Mais ça doit être taré à coder (je me dit qu'en disant ça, les programmateurs seront touchés dans leurs ego et qu'ils vont se lancer le défi haha) ou alors plus simple, probablement que si elle était en bas et pas en haut ça permettrait d'y avoir accès en écrivant. Sur la page https://controversciences.org/ Par : <NAME> Navigateur : chrome modern linux webkit<issue_closed> Status: Issue closed
acceptbitcoincash/acceptbitcoincash
299088885
Title: Add 'Contemporary Training' to the 'Education' category Question: username_0: Requesting to add 'Contemporary Training' to the 'Education' category. Details follow: ```yml - name: Contemporary Training url: http://contemporarytraining.co.za/ img: twitter: CTrainingZA facebook: https://www.facebook.com/ContemporaryTrainingZA email_address: <EMAIL> region: af country: ZA city: Johannesburg & Pretoria bch: Yes btc: No othercrypto: Yes doc: http://contemporarytraining.co.za/cima-blog/ ``` Resources for adding this merchant: [Link to Contemporary Training](http://contemporarytraining.co.za/) If needed, try the twitter handles profile image: [twitter.com/CTrainingZA](https://twitter.com/CTrainingZA) If needed, try the facebook handles profile image: [fb.com/https://www.facebook.com/ContemporaryTrainingZA](https://fb.com/https://www.facebook.com/ContemporaryTrainingZA) - [ ] Verify site is legitimate and safe to list. - [ ] Correct data in form if any is innacurate. If everything looks okay, Add it to the site: - [ ] Assign to yourself when you begin work. - [ ] Download and resize the image and put it into the proper img folder. - [ ] Add listing alphabetically to proper .yml file. - [ ] Commit changes mentioning this issue number with 'closes #[ISSUE NUMBER HERE]'.<issue_closed> Status: Issue closed
Azure/autorest.az
602904153
Title: RP: SERVICEHEALTH Question: username_0: Enable CodeGen for service SERVICEHEALTH [1] Please check with Azure Cli team first in case the service will be excluded [2] PM spec/status can be found at https://dev.azure.com/azure-mgmt-ex/Onboarding/_boards/board/t/Onboarding%20Team/Features<issue_closed> Status: Issue closed
AshleyLab/risk_scores
501004892
Title: Determining time of follow-up visit in UKBB data Question: username_0: This document is very useful (see 5 'Repeat assessment data in datasets' section): http://biobank.ndph.ox.ac.uk/~bbdatan/Repeat_assessment_doc_v1.0.pdf The document states how to assess repeat visits etc Answers: username_0: This document is very useful (see 5 'Repeat assessment data in datasets' section): http://biobank.ndph.ox.ac.uk/~bbdatan/Repeat_assessment_doc_v1.0.pdf The document states how to assess repeat visits etc username_0: Question: Are only self-reported codes available at baseline?