repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
mtjo/MacStroke | 957709184 | Title: 开机自启无效
Question:
username_0: 在偏好设置里勾选了自动启动,但是重启之后没有自动启动,并且设置面板上的勾选框变成未勾选了
macos 10.15.7
Answers:
username_1: 我系统 11.5.1 ,刚刚测试是正常的,如果设置仍然不生效可以按以下步骤操作: 系统偏好设置->用户与群组->选择当前用户->登陆项->在列表中删除macstroke然后重新设置(或者删除后重新添加macstroke)
username_0: 公司的机器, 用户与群组 灰色的不能点击,是不是和这个有关
username_1: 是的,设置开机启动要管理员权限才可以
username_0: 可是我安装的mos、clashx这些就可以开机自启诶
username_0: 我晚上用家里的电脑试试
username_1: 你可以删除软件重新安装试试,我从10.13一直升级到现在都没出现过这个问题,也没有人反馈类似的问题
username_0: 找到原因了,我是下载的zip,然后解压之后就直接点开运行了,应该拖进应用程序目录,这样就可以了
软件很好用,给你点赞
Status: Issue closed
|
iulo/iulo.github.io | 354278724 | Title: cli工具编写
Question:
username_0: [commander](https://github.com/tj/commander.js) 用来实现基本命令
[Inquirer](https://github.com/SBoudrias/Inquirer.js/) 实现cli中的问答
Answers:
username_0: ## 发布相关工具
[standard-version](https://github.com/conventional-changelog/standard-version) |
nhagem/AcornNet | 142506855 | Title: Need to verify data files discarded
Question:
username_0: Need to make sure it really is unit 1 dying 100% of the time. Verify by spot checking.
Answers:
username_0: Something is definitely janky. Spot checking suggests that the script is correctly picking up sessions with >=90% "FF" in connection, BUT, there are also false positives - example, "Unit3,10-Oct-2014,06:53:14.425" is pulled as bad, but there are no "FF" connections in it at all.
Holdover from running at >=0%? Does the file re-write?
username_0: File doesn't re-write, so every time you run the script, you have to delete the file or it just adds to it. Otherwise it's doing alright. Missing Unit 6 bad sessions due to a different issue with clean.py, not clean2.py. This script should be fine once that issue is fixed; re-check at that point.
username_0: Fixed!
Status: Issue closed
|
nanovms/ops | 428952509 | Title: ability to list package contents
Question:
username_0: you can simply untar it and tree the dir but it'd be nice to have this rolled up into a single command, some things to think about though:
* some interpreters like ruby/python have a lot of crap so a simple listing might not be best - maybe use paging? idk
* we have plans to hash everything inside the pkg at some point for integrity && verification - maybe that belongs in output as well
Answers:
username_1: `ops package show` and `ops package show | less` would cover both of the cases you mention (paging handled by less or some other Unix command).
One thing I notice is that `package` is becoming a theme so moving it to a sub command would be cool if that sort of thing is acceptable: `ops package list <pkg>`, `ops package show <pkg>`, `ops package get <pkg>`.
username_0: i'm in favor of this although @username_2 is going to be doing work in this area in the coming days
username_2: @username_0 should we close these with #308 checked in ?
Status: Issue closed
|
DCurrent/openbor | 843935295 | Title: Blockodds strange behaviour
Question:
username_0: ```
// Run random chance against blockodds. If it
// passes, AI will block.
if ((rand32()&ent->modeldata.blockodds) == 1)
{
return 1;
}
```
## Debugging
But its not how it works. Because no matter what value I use, there is just two options:
- 1 - Block EVERYTHING, including **during a combo** (the entity receives the first hit and, even if their pain animation isn't over yet, it will automatically block the second hit, breaking from the pain animation)
- Any other value - didn't care to block at all
### Reproduce
Just set the _blockoods 1_ in the character header.
If you set "blockoods" to any number that is not "1", the entity doesn't block at all.
If you set "blockodds 1" and use "nopassiveblock 1", the engine doesn't block at all.
### Expected behavior
The character must try to block in a percentage based on the value of blockodds. And if the first attack has already hit and the pain animation has not yet ended, it cannot block the next attacks.
### Screenshots
### Version
Please provide the SPECIFIC version the issue first appeared. This is very important, it is nearly impossible for us to pour through the entire code base to find singular issues without a starting point.
- Windows
- Build 6330, 6315, 6392. But it was always like that if I remember well, so in version 4432 too.
Answers:
username_1: I'll look into this. I think it's actually two separate issues.
- The formula for odds is set up wrong, that's probably a simple fix.
- No matter what odds are, the engine should not have the native ability to block if the entity is in any state other than idle. Most likely there's a missing check flag check.
username_0: @username_1 thanks. I remember some people reporting that some entities can block even if they are on the ground - right from the fall animation into the block, skipping the whole raise proccess.
username_2: Hey guys! Thanks to your reports and comments I was able to find where the problem is.
Here's the fixes:
Blockodds rate fix
````
if ((rand32()&ent->modeldata.blockodds) == 0)
{
return 1;
}
````
Block state fix, check if the "blocking instance" is gone or not before all other tasks
````
int check_blocking_eligible(entity *ent, entity *other, s_collision_attack *attack)
{
if (!ent->blocking)
{
return 0;
}
````
In the videos below you can see some tests:
Blockodds rate fix
https://www.youtube.com/watch?v=6cgTytt4iOM
Block state fix
https://www.youtube.com/watch?v=IHQbTMw3yRs
I'm doing more tests to see if everything is ok in a test build. |
nathanvda/cocoon | 169000067 | Title: Associations only saving the last one added
Question:
username_0: Hi! I see you've got a slur of issues already but I was hoping you might have some insight on what I (probably) am doing wrong in using it.
I have an app in which objects have a lot of named, self-referential (and normal) associations. For example, the Character model has a `best_friends` association (pointing to at least 1 Character), a `siblings` association (pointing to at least 1 Character), a `hometowns` association (pointing to at least 1 Location), etc. In the `Character#edit` page, I allow users to create these associations using Cocoon, and observe the following:
* Saving associations one at a time works as expected, even if creating multiple associations as long as each one only has 1 (e.g. creating one `best_friends` and one `siblings` works, but creating two `siblings` relations only saves the latter one).
* Removing a relation and adding a new one in-line (e.g. deleting any `best_friends` associations and adding a new one) works, but of course only adds the last one.
Just to give you a sense of context, here's what the UI looks like adding relations:

(Ugly I know, but just trying to get it working right now! :) )
Here's the relevant code:
CharactersController
```
...
def edit
@content = Character.find(params[:id])
end
...
```
relevant edit.html.erb (in the Content class, which Character inherits from)
```
<% content.class.attribute_categories.each do |category, data| %>
<div id="<%= category %>" class="row">
<% data[:attributes].each do |attribute| %>
<div class="col s10 m8 l4">
<% value = content.send(attribute) %>
<% if value.is_a?(ActiveRecord::Associations::CollectionProxy) %>
<% through_class = content.class.reflect_on_association(attribute).options[:through].to_s %>
<%= render 'content/form/relation_input', f: f, attribute: attribute, relation: through_class %>
<% else %>
<%= render 'content/form/text_input', f: f, attribute: attribute %>
<% end %>
</div>
<% end %>
</div>
<% end %>
```
content/form/_relation_input.html.erb (rendered in the case of associations, like this question):
```
<div>
<%= f.label attribute, attribute.humanize.capitalize %>
</div>
<div id="<%= relation %>">
<%= f.fields_for relation do |builder| %>
<%= render 'content/form/groupship_fields', f: builder, attribute: attribute.singularize, parent: f.object %>
<% end %>
<div class="links">
<% color = f.object.send(attribute).build.class.color %>
<%= link_to_add_association "add #{attribute.to_s.singularize}", f, relation,
class: "btn #{color}",
partial: 'content/form/groupship_fields',
render_options: { locals: {
attribute: attribute.singularize,
parent: f.object
}} %>
[Truncated]
<% end %>
</div>
```
And, of course, back to the CharacterController#update that handles processing:
```
def update
@content = Character.find(params[:id])
if @content.update_attributes(content_params)
successful_response(@content, t(:update_success, model_name: humanized_model_name))
else
failed_response('edit', :unprocessable_entity)
end
end
```
If there's any insight you can give into what I might be doing wrong, it'd be much appreciated. I assume I'm just messing up something small to make it only process the last of each association sent.
Please let me know if there's any other information you need to take a look. The relevant code (though not simplified) is all in https://github.com/indentlabs/Indent/tree/master/app/views/content if you want to look at the source directly.
Answers:
username_1: First thing: check what is actually posted to the controller? Sometimes due to an error in your html (e.g. using id's instead of classes) it will only send one of a repeating group. But from what I could see, that _seems_ ok (still check to be safe).
I saw in your code (thanks for sharing it), that you use a handy shortcut to define your relations: `relate :siblings, with: :siblingships` and in the concern you ony add the `accepts_nested_attributes_for` for the connecting relation. Imho you also have to add it for the actual relation you are editing. A bit weird that it partially works, so not sure if that is the case.
Also, you might have a problem: all your "related"-classes are in the folder `content_groupers` but you do not prefix the class with `ContentGroupers` so not sure if the rails autoloading will actually find the class correctly (but maybe you solved that otherwise --I didn't go looking to far).
username_0: Thanks for the insight. Upon inspecting what is actually posted to the controller, you may be right: there's only one ID being passed through: when adding two `mothers` (in addition to one existing one), the relevant posted params look like:
```
"motherships_attributes"=>
{"0"=>{"id"=>"1", "_destroy"=>"false"},
"1470283435380"=>{"mother_id"=>"2", "_destroy"=>"false"},
"1470283438992"=>{"mother_id"=>"", "_destroy"=>"false"}},
```
And after wiping them all (successfully, in one submit), trying to add 3 at the same time:
```
"motherships_attributes"=>
{"1470285441878"=>{"mother_id"=>"3", "_destroy"=>"false"},
"1470285445040"=>{"mother_id"=>"", "_destroy"=>"false"},
"1470285447940"=>{"mother_id"=>"", "_destroy"=>"false"}},
```
Pretty sure those empty `mother_id` params are the culprit here. Do you see anything in the HTML that might be the reason?
Also FWIW on the `relate` shortcut, I tried adding in the actual relation as well, but it didn't seem to make a difference here. I'll take a look around whether I need it in the long run though, thanks for the heads up.
And yep! Adding `config.autoload_paths += Dir[Rails.root.join('app', 'models', '{*/}')]` to `config/application.rb` _should_ be recursively autoloading everything in the models directory, at least.
Thanks again for your help. I'll keep digging at the HTML, but anything you can offer would be much appreciated. And thanks for Cocoon!
username_1: It is a little over my head, just gave it a quick cursory read, but please note that all views are server-side-rendered, so every time you press `link_to_add_association` it just inserts a pre-rendered partial. If you need to set a specific value upon insertion, you will have to use js and the callbacks.
Also, not sure but you seem to make it harder on yourself. To be clear:
* you do not have to specify an id for new elements (cocoon does that: if the id does not exist in the database, rails assumes it is a new element)
* you do not have to specify a parent-id for a nested element, since the hash is nested, it is clear to what rails must add it
(I will read it in more detail later, if needed)
Status: Issue closed
username_0: Just an FYI, the complications seemed to stem from the autocomplete field I was using (rails-jquery-autocomplete) that required me to specify the ID. I swapped it out for a native rails `f.select` and everything works perfectly now!
Thank you for your all your help! :) |
diffplug/spotless | 1071686099 | Title: Updating spotless version - cannot add a configuration as a configuration with that name already exists.
Question:
username_0: If you are submitting a **bug**, please include the following:
- [X] summary of problem
- [X] gradle or maven version
- [X] spotless version
- [X] operating system and version
- [X] copy-paste your full Spotless configuration block(s), and a link to a public git repo that reproduces the problem if possible
- [X] copy-paste the full content of any console errors emitted by `gradlew spotless[Apply/Check] --stacktrace`
I am trying to update spotless from 5.x to 6.x and get the below exception. The project uses `org.gradle.parallel=true` and throws this exception on a machine with 16 cores, presumably causing more parallelism than other machines may cause, so may be hard to repro locally. But hopefully based on the stack trace we can find a location to verify. Perhaps this should use ConcurrentHashMap / computeIfAbsent?
https://github.com/diffplug/spotless/blob/main/plugin-gradle/src/main/java/com/diffplug/gradle/spotless/GradleProvisioner.java#L54
Executing the task individually succeeds, it is when executing on the whole project it fails probably due to sync issue
`Gradle`: 7.3.1
`spotless-gradle-plugin`: 6.0.2
OS: Amazon Linux 2
Repro: https://github.com/username_0/opentelemetry-java-instrumentation/tree/gradle-7.3.1
Run `./gradlew spotlessApply` (may not repro on all machines). Many spotless tasks run and succeed but the error happens and fails the build
Spotless config:
```
You need to add a repository containing the '[com.google.googlejavaformat:google-java-format:1.12.0]' artifact in 'testing/agent-for-testing/build.gradle'.
E.g.: 'repositories { mavenCentral() }'
org.gradle.api.InvalidUserDataException: Cannot add a configuration with name 'spotless865452459' as a configuration with that name already exists.
at org.gradle.api.internal.DefaultNamedDomainObjectCollection.assertCanAdd(DefaultNamedDomainObjectCollection.java:213)
at org.gradle.api.internal.AbstractNamedDomainObjectContainer.create(AbstractNamedDomainObjectContainer.java:77)
at org.gradle.api.internal.AbstractValidatingNamedDomainObjectContainer.create(AbstractValidatingNamedDomainObjectContainer.java:47)
at org.gradle.api.internal.AbstractNamedDomainObjectContainer.create(AbstractNamedDomainObjectContainer.java:56)
at com.diffplug.gradle.spotless.GradleProvisioner.lambda$forProject$1(GradleProvisioner.java:73)
at com.diffplug.gradle.spotless.GradleProvisioner$DedupingProvisioner.provisionWithTransitives(GradleProvisioner.java:61)
at com.diffplug.spotless.JarState.provisionWithTransitives(JarState.java:68)
at com.diffplug.spotless.JarState.from(JarState.java:57)
at com.diffplug.spotless.JarState.from(JarState.java:52)
at com.diffplug.spotless.java.GoogleJavaFormatStep$State.<init>(GoogleJavaFormatStep.java:139)
at com.diffplug.spotless.java.GoogleJavaFormatStep.lambda$create$1(GoogleJavaFormatStep.java:92)
at com.diffplug.spotless.FormatterStepImpl.calculateState(FormatterStepImpl.java:56)
at com.diffplug.spotless.LazyForwardingEquality.state(LazyForwardingEquality.java:56)
at com.diffplug.spotless.LazyForwardingEquality.toBytes(LazyForwardingEquality.java:85)
at com.diffplug.spotless.LazyForwardingEquality.hashCode(LazyForwardingEquality.java:102)
```
Answers:
username_1: Thanks very much for the example, it reproduces fine on my Macbook Air. Importantly, I'm seeing two failures. The first exception is this:
```
E.g.: 'repositories { mavenCentral() }'
org.gradle.api.artifacts.ResolveException: Could not resolve all dependencies for configuration ':testing:agent-for-testing:spotless865452459'.
at org.gradle.api.internal.artifacts.ivyservice.ErrorHandlingConfigurationResolver.wrapException(ErrorHandlingConfigurationResolver.java:105)
at org.gradle.api.internal.artifacts.ivyservice.ErrorHandlingConfigurationResolver.resolveGraph(ErrorHandlingConfigurationResolver.java:76)
...
Caused by: java.util.ConcurrentModificationException
at java.base/java.util.TreeMap$PrivateEntryIterator.nextEntry(TreeMap.java:1208)
at java.base/java.util.TreeMap$KeyIterator.next(TreeMap.java:1262)
at org.gradle.api.internal.collections.FilteredCollection$FilteringIterator.findNext(FilteredCollection.java:121)
at org.gradle.api.internal.collections.FilteredCollection$FilteringIterator.next(FilteredCollection.java:140)
at org.gradle.api.internal.DefaultDomainObjectCollection$IteratorImpl.next(DefaultDomainObjectCollection.java:475)
at org.gradle.api.internal.artifacts.ivyservice.moduleconverter.DefaultRootComponentMetadataBuilder.getRootComponentMetadata(DefaultRootComponentMetadataBuilder.java:95)
at org.gradle.api.internal.artifacts.ivyservice.moduleconverter.DefaultRootComponentMetadataBuilder.lambda$buildRootComponentMetadata$0(DefaultRootComponentMetadataBuilder.java:86)
at org.gradle.api.internal.project.DefaultProjectStateRegistry$ProjectStateImpl.fromMutableState(DefaultProjectStateRegistry.java:393)
at org.gradle.api.internal.artifacts.ivyservice.moduleconverter.DefaultRootComponentMetadataBuilder.buildRootComponentMetadata(DefaultRootComponentMetadataBuilder.java:84)
at org.gradle.api.internal.artifacts.ivyservice.moduleconverter.DefaultRootComponentMetadataBuilder.toRootComponentMetaData(DefaultRootComponentMetadataBuilder.java:70)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration.toRootComponentMetaData(DefaultConfiguration.java:1197)
at org.gradle.api.internal.artifacts.ivyservice.resolveengine.DefaultArtifactDependencyResolver$DefaultResolveContextToComponentResolver.resolve(DefaultArtifactDependencyResolver.java:227)
at org.gradle.api.internal.artifacts.ivyservice.resolveengine.graph.builder.DependencyGraphBuilder.resolve(DependencyGraphBuilder.java:139)
at org.gradle.api.internal.artifacts.ivyservice.resolveengine.DefaultArtifactDependencyResolver.resolve(DefaultArtifactDependencyResolver.java:145)
at org.gradle.api.internal.artifacts.ivyservice.DefaultConfigurationResolver.resolveGraph(DefaultConfigurationResolver.java:186)
at org.gradle.api.internal.artifacts.ivyservice.ShortCircuitEmptyConfigurationResolver.resolveGraph(ShortCircuitEmptyConfigurationResolver.java:85)
at org.gradle.api.internal.artifacts.ivyservice.ErrorHandlingConfigurationResolver.resolveGraph(ErrorHandlingConfigurationResolver.java:74)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration$1.call(DefaultConfiguration.java:644)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration$1.call(DefaultConfiguration.java:635)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204)
at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration.resolveGraphIfRequired(DefaultConfiguration.java:635)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration.lambda$resolveExclusively$4(DefaultConfiguration.java:615)
...
```
And *then* I get the exception you linked above, "a configuration with that name already exists". I think it's the first `ConcurrentModificationException` which is the root problem, I'm digging in on that now...
Status: Issue closed
username_1: Fixed in `plugin-gradle/6.0.3`, please confirm.
username_1: If you are submitting a **bug**, please include the following:
- [X] summary of problem
- [X] gradle or maven version
- [X] spotless version
- [X] operating system and version
- [X] copy-paste your full Spotless configuration block(s), and a link to a public git repo that reproduces the problem if possible
- [X] copy-paste the full content of any console errors emitted by `gradlew spotless[Apply/Check] --stacktrace`
I am trying to update spotless from 5.x to 6.x and get the below exception. The project uses `org.gradle.parallel=true` and throws this exception on a machine with 16 cores, presumably causing more parallelism than other machines may cause, so may be hard to repro locally. But hopefully based on the stack trace we can find a location to verify. Perhaps this should use ConcurrentHashMap / computeIfAbsent?
https://github.com/diffplug/spotless/blob/main/plugin-gradle/src/main/java/com/diffplug/gradle/spotless/GradleProvisioner.java#L54
Executing the task individually succeeds, it is when executing on the whole project it fails probably due to sync issue
`Gradle`: 7.3.1
`spotless-gradle-plugin`: 6.0.2
OS: Amazon Linux 2 (AWS m5.4xlarge)
Repro: https://github.com/username_0/opentelemetry-java-instrumentation/tree/spotless-repro
Run `./gradlew spotlessApply` (may not repro on all machines). Many spotless tasks run and succeed but the error happens and fails the build
Spotless config:
```
You need to add a repository containing the '[com.google.googlejavaformat:google-java-format:1.12.0]' artifact in 'testing/agent-for-testing/build.gradle'.
E.g.: 'repositories { mavenCentral() }'
org.gradle.api.InvalidUserDataException: Cannot add a configuration with name 'spotless865452459' as a configuration with that name already exists.
at org.gradle.api.internal.DefaultNamedDomainObjectCollection.assertCanAdd(DefaultNamedDomainObjectCollection.java:213)
at org.gradle.api.internal.AbstractNamedDomainObjectContainer.create(AbstractNamedDomainObjectContainer.java:77)
at org.gradle.api.internal.AbstractValidatingNamedDomainObjectContainer.create(AbstractValidatingNamedDomainObjectContainer.java:47)
at org.gradle.api.internal.AbstractNamedDomainObjectContainer.create(AbstractNamedDomainObjectContainer.java:56)
at com.diffplug.gradle.spotless.GradleProvisioner.lambda$forProject$1(GradleProvisioner.java:73)
at com.diffplug.gradle.spotless.GradleProvisioner$DedupingProvisioner.provisionWithTransitives(GradleProvisioner.java:61)
at com.diffplug.spotless.JarState.provisionWithTransitives(JarState.java:68)
at com.diffplug.spotless.JarState.from(JarState.java:57)
at com.diffplug.spotless.JarState.from(JarState.java:52)
at com.diffplug.spotless.java.GoogleJavaFormatStep$State.<init>(GoogleJavaFormatStep.java:139)
at com.diffplug.spotless.java.GoogleJavaFormatStep.lambda$create$1(GoogleJavaFormatStep.java:92)
at com.diffplug.spotless.FormatterStepImpl.calculateState(FormatterStepImpl.java:56)
at com.diffplug.spotless.LazyForwardingEquality.state(LazyForwardingEquality.java:56)
at com.diffplug.spotless.LazyForwardingEquality.toBytes(LazyForwardingEquality.java:85)
at com.diffplug.spotless.LazyForwardingEquality.hashCode(LazyForwardingEquality.java:102)
```
username_2: I think the fix for this issue is causing #1024
username_1: @username_2 definitely suspicious looking, but I don't think so, see explanation at #1024.
username_0: Hi @username_1 - thanks for the quick fix. I am trying 6.0.4, and while the error I had isn't showing anymore I get a deprecation warning (which currently we have set to fail the build but for now can workaround by disabling it)
```
Resolution of the configuration :conventions:spotless865452459 was attempted from a context different than the project context. Have a look at the documentation to understand why this is a problem and how it can be resolved. This behaviour has been deprecated and is scheduled to be removed in Gradle 8.0. See https://docs.gradle.org/7.3.1/userguide/viewing_debugging_dependencies.html#sub:resolving-unsafe-configuration-resolution-errors for more details.
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration.resolveToStateOrLater(DefaultConfiguration.java:595)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration.access$1900(DefaultConfiguration.java:152)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration$SelectedArtifactsProvider.getValue(DefaultConfiguration.java:1344)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration$SelectedArtifactsProvider.getValue(DefaultConfiguration.java:1334)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration$ConfigurationFileCollection.getSelectedArtifacts(DefaultConfiguration.java:1406)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration$ConfigurationFileCollection.visitContents(DefaultConfiguration.java:1393)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration.visitContents(DefaultConfiguration.java:498)
at org.gradle.api.internal.file.AbstractFileCollection.getFiles(AbstractFileCollection.java:130)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration_Decorated.getFiles(Unknown Source)
at org.gradle.api.internal.artifacts.configurations.DefaultConfiguration.resolve(DefaultConfiguration.java:488)
at com.diffplug.gradle.spotless.GradleProvisioner.lambda$forProject$1(GradleProvisioner.java:83)
at com.diffplug.gradle.spotless.GradleProvisioner$DedupingProvisioner.provisionWithTransitives(GradleProvisioner.java:61)
at com.diffplug.spotless.JarState.provisionWithTransitives(JarState.java:68)
at com.diffplug.spotless.JarState.from(JarState.java:57)
at com.diffplug.spotless.JarState.from(JarState.java:52)
at com.diffplug.spotless.java.GoogleJavaFormatStep$State.<init>(GoogleJavaFormatStep.java:139)
at com.diffplug.spotless.java.GoogleJavaFormatStep.lambda$create$1(GoogleJavaFormatStep.java:92)
at com.diffplug.spotless.FormatterStepImpl.calculateState(FormatterStepImpl.java:56)
at com.diffplug.spotless.LazyForwardingEquality.state(LazyForwardingEquality.java:56)
at com.diffplug.spotless.LazyForwardingEquality.toBytes(LazyForwardingEquality.java:85)
at com.diffplug.spotless.LazyForwardingEquality.hashCode(LazyForwardingEquality.java:102)
at java.base/java.util.ArrayList.hashCodeRange(ArrayList.java:595)
at java.base/java.util.ArrayList.hashCode(ArrayList.java:582)
at com.diffplug.gradle.spotless.FormatExtension.setupTask(FormatExtension.java:768)
at com.diffplug.gradle.spotless.JavaExtension.setupTask(JavaExtension.java:226)
at com.diffplug.gradle.spotless.SpotlessExtensionImpl.lambda$createFormatTasks$6(SpotlessExtensionImpl.java:100)
at org.gradle.api.internal.DefaultMutationGuard$2.execute(DefaultMutationGuard.java:44)
at org.gradle.api.internal.DefaultMutationGuard$2.execute(DefaultMutationGuard.java:44)
at org.gradle.configuration.internal.DefaultUserCodeApplicationContext$CurrentApplication$1.execute(DefaultUserCodeApplicationContext.java:123)
at org.gradle.api.internal.DefaultCollectionCallbackActionDecorator$BuildOperationEmittingAction$1.run(DefaultCollectionCallbackActionDecorator.java:110)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68)
at org.gradle.api.internal.DefaultCollectionCallbackActionDecorator$BuildOperationEmittingAction.execute(DefaultCollectionCallbackActionDecorator.java:107)
at org.gradle.internal.ImmutableActionSet$SetWithManyActions.execute(ImmutableActionSet.java:329)
at org.gradle.api.internal.DefaultDomainObjectCollection.doAdd(DefaultDomainObjectCollection.java:260)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection.doAdd(DefaultNamedDomainObjectCollection.java:113)
at org.gradle.api.internal.DefaultDomainObjectCollection.add(DefaultDomainObjectCollection.java:254)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.tryCreate(DefaultNamedDomainObjectCollection.java:944)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.access$1401(DefaultTaskContainer.java:654)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider$1.run(DefaultTaskContainer.java:680)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68)
at org.gradle.api.internal.tasks.DefaultTaskContainer$TaskCreatingProvider.tryCreate(DefaultTaskContainer.java:676)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection$AbstractDomainObjectCreatingProvider.calculateOwnValue(DefaultNamedDomainObjectCollection.java:929)
[Truncated]
at com.diffplug.gradle.spotless.SpotlessPlugin.lambda$configureCleanTask$2(SpotlessPlugin.java:59)
at org.gradle.configuration.internal.DefaultUserCodeApplicationContext$CurrentApplication$1.execute(DefaultUserCodeApplicationContext.java:123)
at org.gradle.api.internal.DefaultCollectionCallbackActionDecorator$BuildOperationEmittingAction$1.run(DefaultCollectionCallbackActionDecorator.java:110)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29)
at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47)
at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68)
at org.gradle.api.internal.DefaultCollectionCallbackActionDecorator$BuildOperationEmittingAction.execute(DefaultCollectionCallbackActionDecorator.java:107)
at org.gradle.api.internal.DefaultMutationGuard$2.execute(DefaultMutationGuard.java:44)
at org.gradle.api.internal.DefaultMutationGuard$2.execute(DefaultMutationGuard.java:44)
at org.gradle.api.internal.collections.CollectionFilter$1.execute(CollectionFilter.java:59)
at org.gradle.internal.ImmutableActionSet$SetWithManyActions.execute(ImmutableActionSet.java:329)
at org.gradle.api.internal.DefaultDomainObjectCollection.doAdd(DefaultDomainObjectCollection.java:260)
at org.gradle.api.internal.DefaultNamedDomainObjectCollection.doAdd(DefaultNamedDomainObjectCollection.java:113)
at org.gradle.api.internal.DefaultDomainObjectCollection.add(DefaultDomainObjectCollection.java:254)
```
username_1: For the sake of keeping things organized, I'm gonna close this out and follow-up the warning at #1028
Status: Issue closed
|
bianyun12/project_20 | 759296575 | Title: 安全漏洞提醒
Question:
username_0: 漏洞类型:邮箱 SMTP 信息泄露
漏洞等级:高
漏洞地址:https://github.com/bianyun12/project_20/blob/4865cc45c6468ec195f25fce98029975d383c2fa/project_day20/run.py
漏洞危害:任何人可以通过 SMTP 账号密码收发邮件,进而通过邮箱重置其他平台密码
解决方案:重置 SMTP 密码并检查邮箱是否有敏感信息泄露(请勿只修改代码,历史版本库依旧可见)
本次扫描结果由 [ 码小六 ] https://github.com/4x99/code6 提供(欢迎 star) |
steamclock/bluejay | 284321399 | Title: Add L2CAP channels support
Question:
username_0: I would like to start discussion how these API should look like.
It may be something like
```
extension Bluejay {
func openL2CAPChannel(_ psm: CBL2CAPPSM, completion: Result<CBL2CAPChannel, Error>)
}
``` |
docToolchain/docToolchain | 836476874 | Title: help needed: where is the CSS defined which is used in documents, created by doctoolchain?
Question:
username_0: **Is your feature request related to a problem? Please describe.**
I like the html files generated by doctoolchain.
That's why I wanted to use the https://github.com/docToolchain/arc42-template-project as template for a microsite for documentation, which uses the doctoolchain technology
But I don't like the CSS styles in the files in this website template, especially the toc. The toc style has `position: fixed`, no scrollbars and can't be used for big toc. I can't find a solution how to solve the issue.
On the other side I would like to just using the style which is generated by doctoolchain by default. But I can find the css only in the created html documents. I can't find, where and how this CSS parts are defined. Maybe this is hard coded in doctoolchain, why not, would be fine.
My idea was, that there could be some CSS files which I just could copy into the website template, and then maybe adapt a bit.
**Describe the solution you'd like**
I would like to find in the documentation where and how the CSS style is defined. And maybe also how to apply this style to pages that (like the website template) are created with doctoolchain.
Answers:
username_1: thanx for your report.
it turns out that the arc42-template-project was defined as a long lasting project but soon will be replaced with the new docToolchain approach:
```
curl -Lo dtcw doctoolchain.github.io/dtcw
chmod +x dtcw
./dtcw installTemplate
./dtcw generateSite
````
(there is also a powershell dtcw.ps1 script)
as a result, you will have a repository with _only_ the adoc files plus a config and the doctoolchain-wrapper dtcw.
to be honest, I also don't like the default styles.
So, there is a
./dtcw cooyThemes
command which copies all templates to you projects. You will find the styles in /src/site/content/assets (I hope I remembered this correctly).
Change the files you need to change and delete all other files. docToolchain will overwrite the internal files with your changed ones.
The goal is to then create an external theme from these changes files which can be configured.
dies this help?
username_0: I downloaded dtcw.ps1, changed access
it looks like, the correct downloads are
```
./dtcw downloadTemplate
./dtcw generateSite
```
I used an empty folder and it worked. I got a ready site. Very simple if you know how to do it!
Some remarks:
- It looks like the CSS is the same as used by https://github.com/docToolchain/arc42-template-project
- The site contains some content (for example HomePage content, some content from the About) which is not defined in the src folder. But it is was clear, how.
- Then I followed your instruction `./dtcw copyThemes` , and now I get an idea how this works. Now I get something similar to the https://github.com/docToolchain/arc42-template-project
- when building the site using generateSite, sometimes I get
"There is insufficient memory for the Java Runtime Environment to continue."
I don't get this error with other Java applications.
- It's one thing to be able to quickly create a site from a template. It's another thing to be able to maintain that site. Furthermore, it's not clear to me yet how simple or complicated that can get. So far I see problems with finding better CSS and also with creating the navigation in an easier way. For example for a wiki pages.
I think, I will investigate, and I will also have a look on Antora. Antora describes a separation of content and UI, and navigation can be controlled with asciidoc lists. This is what I currently miss in the project template. How well do doctoolchain and Antora fit together?
username_1: Yes, Antora might be a good alternative. I don't use it because I think there is value in a landing page and blog on the site, which I get with jBake out of the box.
And Antora is "only" a static site generator. You get some more features with docToolchain.
But you can use it both also together. Setup an Antora repository, copy the docToolchain wrapper to it and you can use docToolchain for tasks like exportExcel.
username_0: For my database-documentation-generator I decided to use Antora: https://datahandwerk.github.io/
- Antora default UI can be used nearly "out of the box" and the result is fine
- Antora navigation concept is fine to generate navigation depending on database structure
On the other hand - and compared to docToolchain
- It is more elaborate to understand and set up Antora than docToolchain
- xref are very powerful, but they are a one-way street and only work within Antora
- the sources of the documentation must be in a certain structure
- It is not possible to use any content outside the Antora structure as sources. Or only with great difficulty or effort.
- That's why "code-as-docs" is rather not possible, since code is usually not distributed in folders in an Antora-appropriate form
I wrote a blog post in German: http://datahandwerk.aisberg.de/2021-04-20-docs-code-mit-asciidoc-und-antora/
username_1: thanx for this valuable feedback and your blog post! |
srobo/tasks | 1116538020 | Title: Organise prizes for the competition
Question:
username_0: The prizes for the various award categories need organisation.
### Original
[comp/prizes/main](https://github.com/srobo/recurring-tasks/blob/master/comp/prizes/main.yaml)
### Dependencies
* #857 Acquire trophies
* #858 Acquire edible prizes
* #860 Distribute certificates
* #867 Organise the SR2022 post-finals sequence
* #868 Have a rehearsal of the post-finals sequence |
bastiW/event-ticket-plus-shortcode | 537115755 | Title: Update to work with latest release of event ticket plus.
Question:
username_0: Hi there!
Have you any plan to update the code to work with the latest release of event ticket plus?
Answers:
username_1: No! As I don’t maintain this project anymore. If you are willing to contribute. I can make you admin for this project. |
data-carpentry-for-agriculture/trial-lesson | 570187308 | Title: Reordering sequence
Question:
username_0: - [ ] Brittani and Aolin do a reordering email draft or slack message
- [ ] Group review the reorder theory
- [ ] divide and conquer the implementation of reordering
Answers:
username_0: Brittani laid out the order in Slack, here's the composited reshuffling for group consideration and commentary: https://docs.google.com/document/d/1wNjoRIHxDDo987jRUpKiI-ohcTeEAtynJkG5PzKNqAk/edit#
username_1: Reorg up, closing.
Status: Issue closed
|
sindresorhus/is-html | 267551673 | Title: json string containing html returns true
Answers:
username_1: Well, that's still valid HTML. If you change the extension from `.json` to `.html`, the HTML part of it will render. HTML can be almost anything, so not really possible to do a better check than this except for checking anything it can't be or use machine learning. I'm happy to accept PRs to improve the detection though.
Status: Issue closed
|
contradictioned/areweideyet | 113194227 | Title: VS Code
Question:
username_0: VS Code totally has syntax highlighting for Rust out of the box.
Since it is based on the same core as Atom, I'd likely classify it in the same category as Atom, so text editors.
Answers:
username_1: Do you know if there is more that is supported by VS Code?
Like, can the atom plugins be used there?
Because if it only highlights .rs files, I'd tend not to include it (yet).
And for fairness reasons maybe put it into a list like "syntax highlightning supported" together with geany...
username_2: There is now plugin for Rust- https://marketplace.visualstudio.com/items/saviorisdead.RustyCode. It completes code using Racer and format code using rustfmt.
username_1: It already works fine.. I think, that's enough to include it ;)
Status: Issue closed
|
kotct/dot | 254458961 | Title: CI: Symbol's function definition is void: assert
Question:
username_0: In CI, we're getting errors now with `assert` not being defined. We should definitely have that defined, and I think it's probably something up with byte-compilation or how we're loading `cl` to do assertions.
For reference, here is the [first build where this showed up, Travis build #139](https://travis-ci.org/kotct/dot/builds/268922848).<issue_closed>
Status: Issue closed |
PaddlePaddle/Paddle | 360313922 | Title: checkpoint载入后第一步训练前大量时间空置
Question:
username_0: ```python
def _train_by_any_executor(self, event_handler, exe, num_epochs, reader):
if self.checkpoint_cfg:
epochs = [
epoch_id for epoch_id in range(num_epochs)
if epoch_id >= self.checkpoint_cfg.epoch_id
]
else:
epochs = [epoch_id for epoch_id in range(num_epochs)]
for epoch_id in epochs:
event_handler(BeginEpochEvent(epoch_id))
for step_id, data in enumerate(reader()):
if self.__stop:
if self.checkpoint_cfg:
self._clean_checkpoint()
return
if self.checkpoint_cfg and self.checkpoint_cfg.load_serial \
and self.checkpoint_cfg.step_id >= step_id and self.checkpoint_cfg.epoch_id == epoch_id:
continue
begin_event = BeginStepEvent(epoch_id, step_id)
event_handler(begin_event)
if begin_event.fetch_metrics:
metrics = exe.run(feed=data,
fetch_list=[
var.name
for var in self.train_func_outputs
])
else:
metrics = exe.run(feed=data, fetch_list=[])
if self.checkpoint_cfg:
self._save_checkpoint(epoch_id, step_id)
event_handler(EndStepEvent(epoch_id, step_id, metrics))
event_handler(EndEpochEvent(epoch_id))
if self.checkpoint_cfg:
self._clean_checkpoint()
```
` if self.checkpoint_cfg and self.checkpoint_cfg.load_serial \
and self.checkpoint_cfg.step_id >= step_id and self.checkpoint_cfg.epoch_id == epoch_id:
continue`
这一段非常影响性能。在读取样本较慢、训练总数较大的时候(比如1秒读取一个batch数据的时候,有2万个batch数据)。一个存档点step_id在5000,加载了存档点岂不是要当空读数据5000秒之后才真正开始训练?
在有shuffle的时候重复一个epoch里一些数据重复用于训练问题不是很大吧?
我还以为怎么回事,一下午怎么event_handler一点输出也没有。
建议给个参数用来遗弃存档点的step_id
Answers:
username_1: 感谢您的建议, 不过请问你使用的是Paddle v2版本还是 Paddle Fluid版本呢?
username_0: Fluid
username_2: 感谢您的建议, 我会跟进优化的事情。 |
portainer/portainer | 1127751491 | Title: Search box in the stacks templates and make editor larger
Question:
username_0: It would be so much easier to use if this template editor had a search feature. Currently the browser search doesn't search the text box, so it would it's own if we want to be able to search through larger templates
Also, I find the editor a too small even on a large monitor. Could we move the variables to the top and make them collapsible so that the editor can fill more of the browser window?
And finally the two tables for name/value pairs in the variables section could be just one table with column titles (name and value).
 |
coronalabs/corona | 990418045 | Title: App not building as before (only changed the build version): java.lang.OutOfMemoryError or java heap space error
Question:
username_0: **Attach your build.settings**
settings =
{
window =
{
defaultMode = "normal",
defaultViewWidth = 720,
defaultViewHeight = 720,
resizable = true,
enableCloseButton = true,
enableMinimizeButton = true,
suspendWhenMinimized = true,
titleText = {
-- The "default" text will be used if the system is using a language and/or
-- country code not defined below. This serves as a fallback mechanism.
default = "Aterramentos (Omniscience 42)",
-- This text is used on English language systems in the United States.
-- Note that the country code must be separated by a dash (-).
["en-us"] = "Aterramentos (Omniscience 42)",
-- This text is used on English language systems in the United Kingdom.
-- Note that the country code must be separated by a dash (-).
["en-gb"] = "Aterramentos (Omniscience 42)",
-- This text is used for all other English language systems.
["en"] = "Aterramentos (Omniscience 42)",
-- This text is used for all French language systems.
["fr"] = "Aterramentos (Omniscience 42)",
-- This text is used for all Spanish language systems.
["es"] = "Aterramentos (Omniscience 42)",
},
},
orientation =
{
default ="portrait",
content = "portrait",
supported =
{
"portrait",
},
},
android =
{
usesExpansionFile = false,
useGoogleServicesJson = false,
largeHeap = true,
usesPermissions =
{
"android.permission.VIBRATE",
"android.permission.INTERNET",
"com.android.vending.CHECK_LICENSE",
"android.permission.WRITE_EXTERNAL_STORAGE",
-- Permission to retrieve current location from the GPS.
"android.permission.ACCESS_FINE_LOCATION",
-- Permission to retrieve current location from WiFi or cellular service.
"android.permission.ACCESS_COARSE_LOCATION",
},
},
plugins =
{
["CoronaProvider.native.popup.quickLook"] =
[Truncated]
{
publisherId = "tech.scotth",
marketplaceId="1316qz",
},
["plugin.notifications.v2"] =
{
publisherId = "com.coronalabs"
},
--["plugin.pasteboard"] =
--{
-- publisherId = "com.coronalabs",
--},
--["plugin.voiceToText"] =
--{
-- publisherId="tech.scotth",
-- marketplaceId = "<EMAIL>",
--},
},
}
Status: Issue closed
Answers:
username_0: My application demands a lot of processes to happen at the same time (its not a game).
Builds: 3649 and 3654 where tested initially.
It worked properly on build 2021.3642.
Building on: Windows 7-10
Building for: Android
Regarding this:
Full build log
Enable full build log:
On Windows:
reg ADD "HKEY_CURRENT_USER\Software\Ansca Corona\Corona Simulator\Preferences" /f /v debugBuildProcess /d 5
Didnt manage to do that. im not very good with those registry things ://
Had this error when i had a ".bin" file with more than 40mb size inside the project:
Execution failed for task ':App:signReleaseBundle'.
A failure occurred while executing com.android,build.gradle.internal.tasks.FinalizeBundleTask$BundleToolRunnable
java.lang.OutOfMemoryError (no error message)
Had this error when i had a ".mp4" file with more than 40mb size inside the project:
Execution failed for task ':App:signReleaseBundle'.
A failure occurred while executing com.android,build.gradle.internal.tasks.FinalizeBundleTask$BundleToolRunnable
Java heap space
Attach your build.settings
settings =
{
window =
{
defaultMode = "normal",
defaultViewWidth = 720,
defaultViewHeight = 720,
resizable = true,
enableCloseButton = true,
enableMinimizeButton = true,
suspendWhenMinimized = true,
titleText =
{
-- The "default" text will be used if the system is using a language and/or
-- country code not defined below. This serves as a fallback mechanism.
default = "Aterramentos (Omniscience 42)",
-- This text is used on English language systems in the United States.
-- Note that the country code must be separated by a dash (-).
["en-us"] = "Aterramentos (Omniscience 42)",
-- This text is used on English language systems in the United Kingdom.
-- Note that the country code must be separated by a dash (-).
["en-gb"] = "Aterramentos (Omniscience 42)",
-- This text is used for all other English language systems.
["en"] = "Aterramentos (Omniscience 42)",
-- This text is used for all French language systems.
["fr"] = "Aterramentos (Omniscience 42)",
-- This text is used for all Spanish language systems.
["es"] = "Aterramentos (Omniscience 42)",
},
},
[Truncated]
},
},
plugins =
{
["CoronaProvider.native.popup.quickLook"] =
{
publisherId = "com.coronalabs",
supportedPlatforms = { iphone = true },
},
["plugin.zip"] =
{
publisherId = "com.coronalabs"
},
["plugin.notifications.v2"] =
{
publisherId = "com.coronalabs"
},
}
} |
FlorentF9/DeepTemporalClustering | 663340696 | Title: variable time step
Question:
username_0: Hi in the documentation for the DTC object, found in DeepTemporalClustering.py, it is indicated that the timesteps param can be variable. However when I instantiate as follows:
# Instantiate model
dtc = DTC(n_clusters=3,
input_dim=X_train.shape[-1],
timesteps=None,
n_filters=50,
kernel_size=10,
strides=1,
pool_size=None,
n_units=[50, 1],
alpha=1,
dist_metric='eucl',
cluster_init='kmeans',
heatmap=False)
I get an error. There is an assert which brings up a typeError. Should I be using 0 instead of None?
Answers:
username_1: I assume the error is thrown by the pooling layer. You can use variable series length by setting `timesteps=None` (because the architecture is fully convolutional and recurrent), however you can't set `pool_size=None`.
username_0: They don't work when I assign timesteps=None
Could you show an example of how you would initiate the DTC object for a variable-length dataset
username_1: You are right, I have to take a look at this when I have time. I keep this issue open for now.
username_1: I took a quick look. Actually I was stuck with the variable timestep because of two things:
* Decoder: it needs to know the dimension of its input and I found no way around this. If you put a `None` there, then how is it supposed to know the length of the sequence to reconstruct?
* TSClusteringLayer: to initialize the cluster center weights, the dimension is needed and it depends on the timesteps. I don't see how we can define cluster centers without fixing their length with this architecture.
In a nutshell, I think the timesteps cannot be variable in DTC. But if anyone has an idea about it, please tell! |
davide-casiraghi/ci-global-calendar | 535436266 | Title: Field disappearing on save action when some required fields are missing in EVENT CREATE view
Question:
username_0: **Describe the bug**
Some fields are disappearing on save action when some required fields are missing in EVENT CREATE view.
This fields are:
- Teachers
- Venue
- Organizers
**To Reproduce**
Steps to reproduce the behavior:
1. Login
2. Click on Manager > New event
3. Fill all the fields missing for instance the description
4. Save
5. See fields disappearing
**Expected behavior**
The field value stays<issue_closed>
Status: Issue closed |
HarishTeens/rpsgames | 529337629 | Title: Chat UI not updated
Question:
username_0: When the "joiner" joins the game, the "creator" chat is not updated and its opponent name is not displayed in the chatbox. In other words: only the "joiner" can see its opponent name in the chatbox.
https://github.com/username_1/rpsgames/pull/7/files?file-filters%5B%5D=.js#r344139625
May i fix it ? If it's ok i'll also use this occasion to do a bit of code refactoring and variables renaming too.
Answers:
username_1: Hey @username_0 , sorry for the late reply , yes you can fix it. Go ahead |
pivotal/skenario | 620972024 | Title: Run simulation in-memory
Question:
username_0: Currently stock movements are written to an Sqlite database. And simulation results are returned from queries on the database. This is nice and flexible, but requires a lot of IO. And disk storage.
To make the simulation run much faster, we should store all the relevant data in memory. We don't necessarily need to store every single stock movement either, if we know what metrics we want to collect up front. (If the simulation is deterministic end-to-end then we can always rerun to get more detailed metrics).
Answers:
username_0: From 4.5 seconds to 200 millseconds!
```
2020/06/08 15:11:10 "POST http://localhost:3000/run HTTP/1.1" from [::1]:43940 - 200 834337B in 237.825889ms
2020/06/08 15:11:18 "POST http://localhost:3000/run HTTP/1.1" from [::1]:43940 - 200 834113B in 4.533779147s
```
username_1: I think we could make it even faster if we get rid of an database completely.
Our response to client should include TallyLine, ResponseTimes, RequestsPerSecond.
Here is a suggested solution how to calculate them.
TallyLine:
type TallyLine struct {
OccursAt int64
StockName string
KindStocked string
Tally int64
}
There is a completed movements list - completed.
Sort completed by occursAt field
Create a map:
Key -> StockName
Value -> Tally (the number of entities in stock)
For completed {
If KindStocked in (request, desired, replica){
Update map -> map.put(StockName, +/- 1)
Add to result new TallyLine
}
}
ResponseTimes:
type ResponseTime struct {
ArrivedAt int64
CompletedAt int64
ResponseTime int64
}
Create a map reqToMovements:
Key -> requestEntity
Value -> list(movements)
Create a result map result
arriveAt
completedAt
For req : reqToMovements.keySet() {
For mov : reqToMovements.get(req){
arriveAt = min(arriveAt, mov.occusAt)
completedAt = max(completedAt, mov.occusAt)
}
result.put(req, ResponseTime{arriveAt, completedAt , completedAt-arriveAt})
}
RequestsPerSecond:
type RPS struct {
Second int64
Requests int64
}
Create a map timeToCount :
Key -> occursAt
Value -> numberOfRequest
For mov : completed{
If mov.kind == “arrive_at_routing” {
timeToCount.put(mov.occursAt, +1)
}
}
Status: Issue closed
|
kubernetes-sigs/kind | 531012089 | Title: deployment keeps not ready
Question:
username_0: Debug Mode: false
Server:
Containers: 1
Running: 1
Paused: 0
Stopped: 0
Images: 50
Server Version: 19.03.5
Storage Driver: overlay2
Backing Filesystem: extfs
Supports d_type: true
Native Overlay Diff: true
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: bridge host ipvlan macvlan null overlay
Log: awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Init Binary: docker-init
containerd version: b34a5c8af56e510852c35414db4c1f4fa6172339
runc version: 3e425f80a8c931f88e6d94a8c831b9d5aa481657
init version: fec3683
Security Options:
seccomp
Profile: default
Kernel Version: 4.9.184-linuxkit
Operating System: Docker Desktop
OSType: linux
Architecture: x86_64
CPUs: 2
Total Memory: 1.952GiB
Name: docker-desktop
ID: QGQZ:K3HP:GQMO:NCF4:QTSV:MLRL:CGLI:RRKQ:PQPJ:77M4:NT2W:IOPA
Docker Root Dir: /var/lib/docker
Debug Mode: true
File Descriptors: 35
Goroutines: 47
System Time: 2019-12-02T09:40:48.1635519Z
EventsListeners: 1
Registry: https://index.docker.io/v1/
Labels:
Experimental: false
Insecure Registries:
127.0.0.0/8
Registry Mirrors:
https://hub-mirror.c.163.com/
Live Restore Enabled: false
Product License: Community Engine
- OS (e.g. from `/etc/os-release`):
windows 10 with docker desktop

Status: Issue closed
Answers:
username_0: runs normal after I delete all before kube configs |
kubernetes/minikube | 796738188 | Title: minikube is not starting
Question:
username_0: <!--- Please include the "minikube start" command you used in your reproduction steps --->
**Steps to reproduce the issue:**
C:\Program Files\Kubernetes\Minikube>minikube start --driver=ssh
* minikube v1.17.1 on Microsoft Windows 10 Pro 10.0.18363 Build 18363
* Using the ssh driver based on user configuration
* Starting control plane node minikube in cluster minikube
! StartHost failed, but will try again: config: please provide an IP address
* Failed to start ssh bare metal machine. Running "minikube delete" may fix it: config: please provide an IP address
X Exiting due to GUEST_PROVISION: Failed to start host: config: please provide an IP address
<!--- TIP: Add the "--alsologtostderr" flag to the command-line for more logs --->
**Full output of failed command:**
**Full output of `minikube start` command used, if not already included:**
**Optional: Full output of `minikube logs` command:**
<details>
</details>
Answers:
username_1: Same as #10266, can we provide some better driver documentation or make it more clear somehow ?
https://minikube.sigs.k8s.io/docs/drivers/ssh/
When you supply your own VM, you _will_ need to provide the connection parameters for it (at a minimum, the IP)
Currently there are also some issues with the default SSH keys, so you need to provide those explicitly: https://github.com/kubernetes/minikube/issues/10289#issuecomment-769212722
username_0: thank you so much for helping me.
Status: Issue closed
|
zhanglei1949/Game-21 | 66618701 | Title: comment for your codes
Question:
username_0: 1.Your code style is very good with a explanation. You use a good way to write some functions to this program. The name of your function is understanding. Import some functions.
2.You make a check for your input, but sometimes some bugs will occur. If you enter more than one letter, it will give you more than one questions.
3.The game can only run one time, you can't make it run twice or more.
Answers:
username_1: Ok,I have improved it. |
firasdib/Regex101 | 166501539 | Title: Why not use local font in regex101?
Question:
username_0: Hi,
There is no doubt that regex101 is the best regular expression test site I have encountered.
But most people got a problem that we can not visit fonts.googleapis.com.
Thus, we were hardly to open regex101 to use it.
So, I hope regex101 can use local font and css, so that we can easily visit and use regex101.
Thank you very much.
Answers:
username_1: To describe the issue in slightly more detail:
When google fonts is blocked for some reason, the initialization process for regex101 hangs waiting for the fonts to load forever.
The problem is that the initialization process only ever waits for the first font to be loaded, and instead of failing with a timeout it hangs forever (so fallback fonts are never used or considered). While regex101 works fine with different fonts, if the first one is not available, you will never finish loading.
Local CSS is perhaps not the right way to fix this, and perhaps a global timeout on initialization might be useful, or at the very least consider fallback fonts.
username_2: Would it be a good idea to switch to [jQuery CDN](https://code.jquery.com/)? [This post](http://edjiang.com/post/97299595332/how-googles-cdn-prevents-your-site-from-loading) suggests it as a solution.
username_3: @username_2 not so much, i can remember the times when jQuery site was hacked. So google or local is much more preferred over another CDN
username_4: How about a check if the $ is defined or not, after attempting to load from Googe APIs? if it can not be loaded from the server, try to use the local version. I have the same problem in my designs from time to time, and this is the solution that I use.
P.S. @username_0 I knew Internet blocking and censorship was a thing in China, but your government also blocks access to Googe APIs and jQuery? What do they think with themselves? :|
username_5: v2 does not have jquery as a dependency, and does not make use of any CDN, so this will not be a problem in the future. For now we should just try to work around it.
Status: Issue closed
|
handsontable/docs | 405603669 | Title: Add more examples to the documentation
Question:
username_0: @wojciechczerniak commented on [Tue Oct 16 2018](https://github.com/handsontable/angular-handsontable/issues/126)
### Description
We should improve our documentation by adding most requested examples:
- [ ] Custom ID, Class, Style and other attributes
- [x] Binding settings with an object
- [x] Binding settings with an attributes
- [ ] Binding callbacks/hooks #122 #123
- [x] Custom Context Menu
- [x] Custom Editor
- [x] Custom Renderer
- [ ] Custom Validator (https://github.com/handsontable/angular-handsontable/issues/126#issuecomment-437011533)
- [x] Language change
- [ ] Integration with state manager
- [x] Referencing the Handsontable instance
- [ ] Interacting with Handsontable in your app (ie. checkbox, button)
Bonus:
- [ ] Examples are broken in Fiefox & IE #136
---
@Calidus commented on [Thu Nov 08 2018](https://github.com/handsontable/angular-handsontable/issues/126#issuecomment-437011533)
Validation could also use some examples.<issue_closed>
Status: Issue closed |
nss-day-cohort-22/movie-history-kindhearted-toads | 277175607 | Title: User Login
Question:
username_0: __Given__ that a user *is not* logged into the app
__When__ they load the app
__Then__the user is offered an affordance to login/register an account
__Given__ that a user *is* logged into the app
__When__ they load the app
__Then__ the user is presented with the home screen |
shlomif/fc-solve | 191136828 | Title: depth_dbm_fc_solver / etc. should validate the input board as containing all the cards and exactly once
Question:
username_0: Success!
--------
Foundations: H-6 C-9 D-5 S-7
Freecells: 8H 9S
: KS
: KH QS JH
: TC 9D
:
:
:
:
:
==
Column 2 -> Column 7
--------
Foundations: H-T C-3 D-5 S-7
Freecells: 4C 5C
: KS
: KH QS JH
: TC 9D
:
:
:
:
:
==
Column 2 -> Column 7
--------
Foundations: H-T C-3 D-5 S-7
Freecells: 4C 5C
: KS
: KC QD
: QH JC
: TD
: 9S
: 9D 8C 7D
: 7C
: 6C
==
Column 7 -> Column 5
--------
Foundations: H-4 C-Q D-J S-0
Freecells: 2S 3S
: KS
: KC
: KH
: QD
:
:
:
:
==
Column 3 -> Column 1
--------
Foundations: H-/ C-c D-- S-s
Freecells: 2C 6C
: KD
: JS
: TS
: TD
: TC
: 9D
: 8D
: 5C
==
END
handle_and_destroy_instance_solution end
```
or worse - it exits with an obscure exception . It should <b>validate</b> the board for extra/missing cards and not start the solving process if the board does not validate. With a regression test!
Answers:
username_0: Fixed in commit ad0157c6360b02a93fe48b1fd5af2baa8c871b15 . Closing
Status: Issue closed
|
react-component/color-picker | 274952980 | Title: Page scrolls to top when opening color picker
Question:
username_0: Captured a gif of the issue from the [samples page](http://react-component.github.io/color-picker/examples/simple.html)

Answers:
username_1: i found this bug caused by the ColorPickerPanel.focus() the source code is in

In a further step ,it is caused by the rc-trigger's container,because the position value of top and left is always zero
username_2: Is there any solution found?
username_3: Please deactivate the auto-scroll option
username_4: I played around with this for a while, and wasn't able to create a scenario where this functionality was actually needed. Opened a PR to remove it.
username_1: @username_4 When an color-picker is on the edge of a web page,click it open the color-panel,focus can let the color-panel scoll to your view. if do not, you may see part of color-panel
username_4: @username_1 I understand the idea, but the picker automatically displays above or below, depending on the location. Can you show me an example where the focus/scroll is required?
Here are the top and bottom examples from the PR I opened:
When the input is at the top of the screen:

When the input is at the bottom of the screen:

username_2: @username_3 How I do that any suggestion?
username_3: @username_2 No, I'm just waiting for them to remove the auto-scroll functionality when clicking.
username_4: @username_5 Any chance of merging #58 ?
Status: Issue closed
username_5: bump v0.2.5 |
pyg-team/pytorch_geometric | 1118868434 | Title: Caught AttributeError in DataLoader worker process 0.
Question:
username_0: ### 🐛 Describe the bug
when I run the code:
val_dataloader = torch_geometric.loader.DataLoader(
val_dataset,
batch_size=hparams.batch_size,
num_workers=hparams.num_workers)
for idx,batch in enumerate(val_dataloader):
print(idx,batch)
I got the following error, I don't know how to fix this issue:
Traceback (most recent call last):
File "/home/aiyicen/00_Script/ares_release/ares/train.py", line 103, in <module>
main()
File "/home/aiyicen/00_Script/ares_release/ares/train.py", line 85, in main
for idx,batch in enumerate(val_dataloader):
File "/home/aiyicen/anaconda3/envs/ares/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 345, in __next__
data = self._next_data()
File "/home/aiyicen/anaconda3/envs/ares/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 856, in _next_data
return self._process_data(data)
File "/home/aiyicen/anaconda3/envs/ares/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 881, in _process_data
data.reraise()
File "/home/aiyicen/anaconda3/envs/ares/lib/python3.8/site-packages/torch/_utils.py", line 395, in reraise
raise self.exc_type(msg)
AttributeError: Caught AttributeError in DataLoader worker process 0.
Original Traceback (most recent call last):
File "/home/aiyicen/anaconda3/envs/ares/lib/python3.8/site-packages/torch/utils/data/_utils/worker.py", line 178, in _worker_loop
data = fetcher.fetch(index)
File "/home/aiyicen/anaconda3/envs/ares/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 47, in fetch
return self.collate_fn(data)
File "/home/aiyicen/anaconda3/envs/ares/lib/python3.8/site-packages/torch_geometric/loader/dataloader.py", line 18, in __call__
return Batch.from_data_list(batch, self.follow_batch,
File "/home/aiyicen/anaconda3/envs/ares/lib/python3.8/site-packages/torch_geometric/data/batch.py", line 68, in from_data_list
batch, slice_dict, inc_dict = collate(
File "/home/aiyicen/anaconda3/envs/ares/lib/python3.8/site-packages/torch_geometric/data/collate.py", line 32, in collate
out = cls(_base_cls=data_list[0].__class__) # Dynamic inheritance.
File "/home/aiyicen/anaconda3/envs/ares/lib/python3.8/site-packages/torch_geometric/data/batch.py", line 40, in __call__
return super(DynamicInheritance, new_cls).__call__(*args, **kwargs)
File "/home/aiyicen/anaconda3/envs/ares/lib/python3.8/site-packages/e3nn-0.1.0-py3.8-linux-x86_64.egg/e3nn/point/data_helpers.py", line 186, in __init__
edge_index, edge_attr = neighbor_list_and_relative_vec(
File "/home/aiyicen/anaconda3/envs/ares/lib/python3.8/site-packages/e3nn-0.1.0-py3.8-linux-x86_64.egg/e3nn/point/data_helpers.py", line 28, in neighbor_list_and_relative_vec
N, _ = pos.shape
AttributeError: 'NoneType' object has no attribute 'shape'
### Environment
torch 1.5.0+cu101
torch-cluster 1.5.7
torch-geometric 2.0.4
torch-scatter 2.0.5
torch-sparse 0.6.7
torch-spline-conv 1.2.0
torchmetrics 0.7.0
torchvision 0.6.0+cu101
cudatoolkit 10.1.243
OS: Ubuntu 20.04.3 LTS
GeForce RTX 3090
I installed pytorch and PYG by using the wheels and pip tool:
pip install torch-1.5.0+cu101-cp38-cp38-linux_x86_64.whl
pip install torch_cluster-1.5.7-cp38-cp38-linux_x86_64.whl
pip install torch_scatter-2.0.5-cp38-cp38-linux_x86_64.whl
pip install torch_sparse-0.6.7-cp38-cp38-linux_x86_64.whl
pip install torch_spline_conv-1.2.0-cp38-cp38-linux_x86_64.whl
pip install torchvision-0.6.0+cu101-cp38-cp38-linux_x86_64.whl
pip install torch_geometric
Answers:
username_1: Do you have a minimal example to reproduce this error? Which example are you running? It looks like somewhere in the code, `data.pos` is referenced, but this attribute does not (yet) exist.
username_0: I loaded dataset from lmdb data by atom3d.datasets.load_dataset, then i created dataloader by torch_geometric.data.DataLoader. My code is as follow:
import atom3d.datasets as da
val_dataset = da.load_dataset(hparams.val_dataset, hparams.filetype,
transform=transform)
val_dataloader = torch_geometric.data.DataLoader(
val_dataset,
batch_size=hparams.batch_size,
num_workers=hparams.num_workers)
When i run:
for i,data in enumerate(val_dataloader.dataset):
print(i,data)
I got this result and the pos is not empty:
0 DataNeighbors(x=[552, 3], edge_index=[2, 27600], edge_attr=[27600, 3], pos=[552, 3], Rs_in=[1], label=[1], id='('1q9a_bps_res4_newfrags', 'S_000681_minimize_008')', file_path='/scratch/users/psuriana/ares_psuriana/data/val')
1 DataNeighbors(x=[552, 3], edge_index=[2, 27600], edge_attr=[27600, 3], pos=[552, 3], Rs_in=[1], label=[1], id='('1q9a_bps_res4_newfrags', 'S_001016_minimize_001')', file_path='/scratch/users/psuriana/ares_psuriana/data/val')
2 DataNeighbors(x=[552, 3], edge_index=[2, 27600], edge_attr=[27600, 3], pos=[552, 3], Rs_in=[1], label=[1], id='('1q9a_bps_res4_newfrags', 'S_001067_minimize_002')', file_path='/scratch/users/psuriana/ares_psuriana/data/val')
3 DataNeighbors(x=[552, 3], edge_index=[2, 27600], edge_attr=[27600, 3], pos=[552, 3], Rs_in=[1], label=[1],
......
18 DataNeighbors(x=[345, 3], edge_index=[2, 17250], edge_attr=[17250, 3], pos=[345, 3], Rs_in=[1], label=[1], id='('1kka_bps_res4_newfrags', 'S_001963_minimize_003')', file_path='/scratch/users/psuriana/ares_psuriana/data/val')
19 DataNeighbors(x=[345, 3], edge_index=[2, 17250], edge_attr=[17250, 3], pos=[345, 3], Rs_in=[1], label=[1], id='('1kka_bps_res4_newfrags', 'S_003526_minimize_003')', file_path='/scratch/users/psuriana/ares_psuriana/data/val')
But when i try to print the val_dataloader, i got the error:
for idx,batch in enumerate(val_dataloader):
print(idx,batch)
username_1: Sorry, I have problems reproducing this. Which dataset are you using and what are the values of each attribute in `hparams`?
username_0: I run the train.py and use the ares_release/data/lmdbs dataset in this source https://zenodo.org/record/5088971#.Yf3RQepByuU. I input the lmdbs/train and lmdbs/val dataset and use the default hparams. The code I've run is as follows:
import argparse as ap
import logging
import os
import pathlib
import sys
import atom3d.datasets as da
import dotenv as de
import pytorch_lightning as pl
import pytorch_lightning.loggers as log
import torch_geometric
import wandb
from torch.utils.data import DataLoader
import sys
sys.path.append(r'/home/aiyicen/00_Script/ares_release/ares/')
import data as d
import model as m
root_dir = pathlib.Path(__file__).parent.parent.absolute()
de.load_dotenv(os.path.join(root_dir, '.env'))
logger = logging.getLogger("lightning")
wandb.init(project="ares")
def main():
parser = ap.ArgumentParser()
# add PROGRAM level args
parser.add_argument('train_dataset', type=str, default='/home/aiyicen/00_Script/ares_release/data/lmdbs/train')
# parser.add_argument('--train_dataset', type=str, default='/home/aiyicen/00_Script/ares_release/data/pdbs/S_000028_476.pdb')
parser.add_argument('--val_dataset', type=str,default='/home/aiyicen/00_Script/ares_release/data/lmdbs/val')
# parser.add_argument('--val_dataset', type=str,default='/home/aiyicen/00_Script/ares_release/data/pdbs/S_000041_026.pdb')
parser.add_argument('-f', '--filetype', type=str, default='lmdb',
choices=['lmdb', 'pdb', 'silent'])
parser.add_argument('--batch_size', type=int, default=1)
parser.add_argument('--label_dir', type=str, default=None)
parser.add_argument('--num_workers', type=int, default=20)
# add model specific args
parser = m.ARESModel.add_model_specific_args(parser)
# add trainer args
parser = pl.Trainer.add_argparse_args(parser)
hparams = parser.parse_args()
dict_args = vars(hparams)
transform = d.create_transform(True, hparams.label_dir, hparams.filetype)
# DATA PREP
logger.info(f"Dataset of type {hparams.filetype}")
logger.info(f"Creating dataloaders...")
train_dataset = da.load_dataset(hparams.train_dataset, hparams.filetype,
transform=transform)
train_dataloader = torch_geometric.loader.DataLoader(
train_dataset,
batch_size=hparams.batch_size,
num_workers=hparams.num_workers,
shuffle=True)
val_dataset = da.load_dataset(hparams.val_dataset, hparams.filetype,
transform=transform)
val_dataloader = torch_geometric.data.DataLoader(
val_dataset,
batch_size=hparams.batch_size,
num_workers=hparams.num_workers)
for i,data in enumerate(val_dataloader.dataset):
print(i,data)
if __name__ == "__main__":
logging.basicConfig(stream=sys.stdout,
format='%(asctime)s %(levelname)s %(process)d: ' +
'%(message)s',
level=logging.INFO)
main()
username_1: Thanks @username_0. Can you say something about how your transform looks like? |
guzzle/psr7 | 125477083 | Title: Relative URI > Can't parse_url if colon is present
Question:
username_0: When using **relative** URIs they are sent to the Uri constructor and `parse_url` is called to extract information from them.
If the path component contains a **colon** in it, and if there are no **query parameters** and there is no **fragment** component then `parse_url` will fail to recognize the path component and return `false`.
```php
// Psr\Http\Message\UriInterface\Uri
public function __construct($uri = '')
{
if ($uri != null) {
$parts = parse_url($uri);
....
}
...
public static function resolve(UriInterface $base, $rel)
{
...
if (!($rel instanceof UriInterface)) {
$rel = new self($rel);
}
....
}
```
Answers:
username_1: Can you provide a concrete example?
Please note thought that one of the things I want to avoid is having to implement a custom URI parser to account for things like this. It would be slow and would be a maintenance burden. If this ends up being a limitation of PHP's parse_url, then I think we would just chalk it up as a "known issue".
username_0: I've experimented some more and found even more constraints to the condition. Now I'm pretty sure this should be a known issue.
I'll update the initial description with the further constraints and an example.
username_0: Should I close the issue or will you close it with a particular status/label?
username_1: Hm... I wonder if there's something we can do here, at least for the case of combining URIs... For example, maybe we can have a special case for combining with a relative URI that starts with '/'?
username_0: Simple fix would be (in the context of initializing relative paths) to add a question mark if one is not already present.
Status: Issue closed
username_2: Is there a PHP bug report for this? https://3v4l.org/QMNeo definitely seems like a PHP bug.
username_0: @username_2 not one that I found but `parse_url` might very well have been intended to parse complete URLs (containing scheme, authentication, domain, tld, port, path, query, fragment, etc.).
Using `parse_url` for relative paths could be regarded as improper usage. Which is why I didn't pursue it as a PHP bug per se.
username_2: But it parses relative urls correctly for the most part like `//example.org` and `/path`. So either it parses only absolute URIs and needs to return false for anything else. Or it must parse relative URIs correctly as well. |
explosion/spaCy | 854184466 | Title: can create_kb just build once when train the entity linking model
Question:
username_0: <!-- Describe your issue here. Please keep in mind that the GitHub issue tracker is mostly intended for reports related to the spaCy code base and source, and for bugs and enhancements. If you're looking for help with your code, consider posting a question here:
- GitHub Discussions: https://github.com/explosion/spaCy/discussions
- Stack Overflow: http://stackoverflow.com/questions/tagged/spacy
-->
## Your Environment
<!-- Include details of your environment. If you're using spaCy 1.7+, you can also type `python -m spacy info --markdown` and copy-paste the result here.-->
- Operating System:
- Python Version Used:
- spaCy Version Used:
- Environment Information:
Answers:
username_1: Hi! As a maintenance note - this type of question is better suited for the discussion forum, so I'll move it there. The original issue will be closed, but will link through to the new, open discussion thread.
username_1: @username_0 : can you elaborate a little on your question?
username_0: 
In spacy v2 we can create kb and save it.
But in spacy v3 we must put it with entity_linker train together as a function and can't save KB only
it will be called every time when i train el model and cost much time to repeat the same thing!
Status: Issue closed
|
xJon/The-1.7.10-Pack | 158609476 | Title: Empty Graves
Question:
username_0: 10++ times now we have lost our hole inventory, with no apparent reason.
get killed by another player, go and knock down grave - only getting head in return. this does seem to be random, but only when players killed eachother (morph or not, in diffrent dimensions etc.)
last bit of my client log of said problem.
`[B#347] 2016/06/06 08:25:24 [INFO] [08:25:24] [Client thread/INFO] [TabbyChat]: [CHAT] username_0 was slain by Nejrael
[B#347] 2016/06/06 08:25:25 [INFO] [08:25:25] [pool-5-thread-102/WARN]: Couldn't look up profile properties for com.mojang.authlib.GameProfile@36e1c73a[id=431a8638-5eef-4b5a-a5b2-50e5cdca9498,name=username_0,properties={textures=[com.mojang.authlib.properties.Property@5da1b244, com.mojang.authlib.properties.Property@7b590ab7, com.mojang.authlib.properties.Property@52cd046d, com.mojang.authlib.properties.Property@3dae5133, com.mojang.authlib.properties.Property@12fefd18, com.mojang.authlib.properties.Property@5b85d2f1, com.mojang.authlib.properties.Property@3e95aa7a, com.mojang.authlib.properties.Property@70190036, com.mojang.authlib.properties.Property@5927fe8a, com.mojang.authlib.properties.Property@6888cc61, com.mojang.authlib.properties.Property@4ca563f5, com.mojang.authlib.properties.Property@78c034c5, com.mojang.authlib.properties.Property@5523483e, com.mojang.authlib.properties.Property@460e7ed3, com.mojang.authlib.properties.Property@76484ae, com.mojang.authlib.properties.Property@43571754, com.mojang.authlib.properties.Property@bfcf5c7, com.mojang.authlib.properties.Property@56cc2d21, com.mojang.authlib.properties.Property@3ff8e191, com.mojang.authlib.properties.Property@44047ff, com.mojang.authlib.properties.Property@72083002, com.mojang.authlib.properties.Property@480946c8, com.mojang.authlib.properties.Property@b9b5885, com.mojang.authlib.properties.Property@53231517, com.mojang.authlib.properties.Property@40f53b89, com.mojang.authlib.properties.Property@18f0f3e5, com.mojang.authlib.properties.Property@a676212, com.mojang.authlib.properties.Property@3c30cd35, com.mojang.authlib.properties.Property@7b6f7f3f, com.mojang.authlib.properties.Property@61042d17, com.mojang.authlib.properties.Property@3397190f, com.mojang.authlib.properties.Property@95816d8, com.mojang.authlib.properties.Property@2a5e8269, com.mojang.authlib.properties.Property@4e6bc278, com.mojang.authlib.properties.Property@2067ec76, com.mojang.authlib.properties.Property@3e11ddd5, com.mojang.authlib.properties.Property@33385851, com.mojang.authlib.properties.Property@633a4d07]},legacy=false]
[B#347] 2016/06/06 08:25:25 [INFO] com.mojang.authlib.exceptions.AuthenticationException: The client has sent too many requests within a certain amount of time
[B#347] 2016/06/06 08:25:25 [INFO] at com.mojang.authlib.yggdrasil.YggdrasilAuthenticationService.makeRequest(YggdrasilAuthenticationService.java:65) ~[YggdrasilAuthenticationService.class:?]
[B#347] 2016/06/06 08:25:25 [INFO] at com.mojang.authlib.yggdrasil.YggdrasilMinecraftSessionService.fillProfileProperties(YggdrasilMinecraftSessionService.java:135) [YggdrasilMinecraftSessionService.class:?]
[B#347] 2016/06/06 08:25:25 [INFO] at net.minecraft.client.resources.SkinManager$3.run(SourceFile:106) [brm.class:?]
[B#347] 2016/06/06 08:25:25 [INFO] at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) [?:1.8.0_73]
[B#347] 2016/06/06 08:25:25 [INFO] at java.util.concurrent.FutureTask.run(Unknown Source) [?:1.8.0_73]
[B#347] 2016/06/06 08:25:25 [INFO] at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) [?:1.8.0_73]
[B#347] 2016/06/06 08:25:25 [INFO] at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) [?:1.8.0_73]
[B#347] 2016/06/06 08:25:25 [INFO] at java.lang.Thread.run(Unknown Source) [?:1.8.0_73]
[B#347] 2016/06/06 08:25:26 [INFO] [08:25:26] [Client thread/INFO] [TabbyChat]: [CHAT] You're protected from attack by other players as long as your inventory is empty.
[B#347] 2016/06/06 08:25:26 [INFO] [08:25:26] [Client thread/INFO] [TabbyChat]: [CHAT] Welcome back to Town.
[B#347] 2016/06/06 08:25:26 [INFO] [08:25:26] [pool-5-thread-102/WARN]: Couldn't look up profile properties for com.mojang.authlib.GameProfile@3a51f7a6[id=431a8638-5eef-4b5a-a5b2-50e5cdca9498,name=username_0,properties={},legacy=false]
[B#347] 2016/06/06 08:25:26 [INFO] com.mojang.authlib.exceptions.AuthenticationException: The client has sent too many requests within a certain amount of time
[B#347] 2016/06/06 08:25:26 [INFO] at com.mojang.authlib.yggdrasil.YggdrasilAuthenticationService.makeRequest(YggdrasilAuthenticationService.java:65) ~[YggdrasilAuthenticationService.class:?]
[B#347] 2016/06/06 08:25:26 [INFO] at com.mojang.authlib.yggdrasil.YggdrasilMinecraftSessionService.fillProfileProperties(YggdrasilMinecraftSessionService.java:135) [YggdrasilMinecraftSessionService.class:?]
[B#347] 2016/06/06 08:25:26 [INFO] at net.minecraft.client.resources.SkinManager$3.run(SourceFile:106) [brm.class:?]
[B#347] 2016/06/06 08:25:26 [INFO] at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) [?:1.8.0_73]
[B#347] 2016/06/06 08:25:26 [INFO] at java.util.concurrent.FutureTask.run(Unknown Source) [?:1.8.0_73]
[B#347] 2016/06/06 08:25:26 [INFO] at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) [?:1.8.0_73]
[B#347] 2016/06/06 08:25:26 [INFO] at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) [?:1.8.0_73]
[B#347] 2016/06/06 08:25:26 [INFO] at java.lang.Thread.run(Unknown Source) [?:1.8.0_73]
[B#347] 2016/06/06 08:25:36 [INFO] [08:25:36] [pool-5-thread-102/ERROR]: Signature is missing from textures payload`
Answers:
username_1: http://paste.ubuntu.com/
Status: Issue closed
username_1: Closed due to missing information. |
cakephp/chronos | 293722979 | Title: Problems with testing with microseconds
Question:
username_0: I'm running a PHP 7.2 server.
And although I have set the "testNow" variable:
```
Chronos::setTestNow(Chronos::now());
```
I'm having problems when testing:
```
'created' => Authored Object (
'by' => 6eeaf6b5-ce76-4d15-b370-5e148b93c8db
- 'on' => 2018-02-01T12:31:39.260470+0000
+ 'on' => 2018-02-01T12:31:38.889742+0000
)
```
Both dates where generated with the Chronos::now() after the "setTestNow".
Might I be doing something wrong?
Answers:
username_1: I'll try to reproduce this. What version of chronos are you using?
username_0: @username_1 I am currently on version ```1.1.4```.
The code is running in a docker with the image ```php:7.2-fpm-alpine```.
username_0: Apparently this problem has to do with the fact the I'm calling the "setTestNow" in several tests.
I tried moving the "setTestNow" to the test case constructor and it seems to work fine.
Of course this would not be the best solution because I will eventually need to set some custom date in a test.
username_1: Each time you setTestNow() you'll likely get new microseconds. In your tests where you set a custom date how are you doing it?
username_0: I'm using (for now):
```
Chronos::setTestNow(Chronos::now());
```
username_0: I think I figured it out.
The fixture system I'm using was creating the fixture set on object instantiation. And this happens inside the container, well before I call the "setTestNow".
By tweaking the fixtures to load only when requested I am able to run everything without any strange date behaviour.
Status: Issue closed
|
Reading-eScience-Centre/edal-java | 168825370 | Title: default namespace in GetCapabilities 1.1.1
Question:
username_0: i'm using the owslib.wms to access the latest ncWMS2 2.2.1. The GetCapabilities xml document defines a default namespace. This troubles the python lxml parser in owslib, so i provided a workaround:
https://github.com/geopython/OWSLib/pull/338
The owslib people say there is no default namespace defined for wms 1.1.1. So, can it be removed in ncWMS2?
Answers:
username_1: OK, I've done this - it'll be in the next release.
username_0: thanks :)
Status: Issue closed
|
TobitSoftware/chayns-js | 243650057 | Title: no chayns environment found
Question:
username_0: After setup node server and clone the git repository.
run the command npm run start
the index.html i tried
<html>
<body>
<script src="chayns.js"></script>
<!-- api -->
<!--<script src="https://chayns-res.tobit.com/API/V3.1/js/chayns.min.js"></script>-->
<script>
console.log(chayns.env);
chayns.ready
.then(function resolved(){
console.log("chayns is ready, environment is loaded", chayns.env);
})
.catch(function rejected(){
console.log("no chayns environment found");
})
.then(function always(){
console.log("Will always be executed");
});
</script>
</body>
</html>
in console it showing "no chayns environment found".
I tried <script src="https://chayns-res.tobit.com/API/V3.1/js/chayns.min.js"></script> but same issue.
Please assist me .
Answers:
username_1: Did you run the code in a valid chayns environment such as chayns-web, chayns-web-light or in a chayns app? Otherwise the chayns.ready-promise will always be rejected.
username_0: I am using chayns-js enviornment
username_1: Chayns-js is no chayns-environment. Chayns-js is an api to control an chayns-environment.
Have a look at our getting started: https://github.com/TobitSoftware/chayns-js/wiki/Getting-Started#step-3---add-the-tapp
Status: Issue closed
|
flutter/flutter | 553360258 | Title: StreamBuilder returning passing wrong data to list sometimes causing brief error
Question:
username_0: This is my code
Why does the error `Another exception was thrown: FormatException: Invalid number (at character 1)` occur on my screen for a few microseconds before all is back to normal. Sometimes it doesn't even occur. Below is my StreamBuilder function:
```
_delivered() {
print('In the delivered function:${resId},${customerId}, ');
return StreamBuilder<QuerySnapshot>(
stream: Firestore.instance
.collection('restaurants')
.document(resId)
.collection('customers')
.document(customer)
.collection('orders')
.where('deliveryTime', isGreaterThan: '')
.snapshots(),
builder: (context, snapshot) {
print('Does snapshop have data? ${snapshot.hasData}');
if (!snapshot.hasData) return Container();
List deliveredListFromServer = snapshot.data.documents;
return Expanded(
child: ListView(
shrinkWrap: true,
children: deliveredListFromServer.map((item) {
print('document id: ${item.documentID}');
return InkWell(
child: SizedBox(
height: 50,
child: Row(
crossAxisAlignment: CrossAxisAlignment.center,
children: <Widget>[
SizedBox(
width: 80,
child: Text(
item['orderBy']['username'],
textAlign: TextAlign.center,
overflow: TextOverflow.ellipsis,
style: TextStyle(fontWeight: FontWeight.bold),
),
),
SizedBox(
width: 5,
),
Expanded(
child: ListView(
scrollDirection: Axis.horizontal,
children: item['order'].map<Widget>((item) {
return SizedBox(
width: 80,
child: Align(
alignment: Alignment.centerLeft,
child: Text(
'${item['qty']} ${item['drinkName']}',
overflow: TextOverflow.ellipsis,
),
),
);
}).toList(),
), //
[Truncated]
I/flutter (11506): #158 Element.updateChild (package:flutter/src/widgets/framework.dart:2977:15)
I/flutter (11506): #159 ComponentElement.performRebuild (package:flutter/src/widgets/framework.dart:4243:16)
I/flutter (11506): #160 Element.rebuild (package:flutter/src/widgets/framework.dart:3947:5)
I/flutter (11506): #161 StatefulElement.update (package:flutter/src/widgets/framework.dart:4413:5)
I/flutter (11506): #162 Element.updateChild (package:flutter/src/widgets/framework.dart:2977:15)
I/flutter (11506): #163 ComponentElement.performRebuild (package:flutter/src/widgets/framework.dart:4243:16)
I/flutter (11506): #164 Element.rebuild (package:flutter/src/widgets/framework.dart:3947:5)
I/flutter (11506): #165 BuildOwner.buildScope (package:flutter/src/widgets/framework.dart:2432:33)
I/flutter (11506): #166 WidgetsBinding.drawFrame (package:flutter/src/widgets/binding.dart:773:20)
I/flutter (11506): #167 RendererBinding._handlePersistentFrameCallback (package:flutter/src/rendering/binding.dart:283:5)
I/flutter (11506): #168 SchedulerBinding._invokeFrameCallback (package:flutter/src/scheduler/binding.dart:1102:15)
I/flutter (11506): #169 SchedulerBinding.handleDrawFrame (package:flutter/src/scheduler/binding.dart:1041:9)
I/flutter (11506): #170 SchedulerBinding._handleDrawFrame (package:flutter/src/scheduler/binding.dart:957:5)
I/flutter (11506): #174 _invoke (dart:ui/hooks.dart:259:10)
I/flutter (11506): #175 _drawFrame (dart:ui/hooks.dart:217:3)
I/flutter (11506): (elided 3 frames from package dart:async)
I/flutter (11506):
I/flutter (11506): ════════════════════════════════════════════════════════════════════════════════════════════════════
I/flutter (11506): Another exception was thrown: A RenderFlex overflowed by 99750 pixels on the bottom.
```
Status: Issue closed
Answers:
username_1: Hi @username_0
From what I can see, the issue is related to a FlutterFire plugin rather than to Flutter itself. Please open the issue in the dedicated [repository](https://github.com/FirebaseExtended/flutterfire/issues).
Closing, as this isn't an issue with Flutter itself.
If you disagree, please write in the comments and provide a minimal reproducible code sample that does not use 3rd party plugins, and I will reopen it.
Thank you |
quasarframework/quasar-testing | 456668304 | Title: .gitignore artifacts
Question:
username_0: <!-- BUG REPORT TEMPLATE -->
### Software version
OS: Ubuntu 18
Node: 10
NPM: yarn
Any other software related to your bug:
### What did you get as the error?
I noticed after adding this extension and playing with it, git wanted to commit all it's artifacts.
### What were you expecting?
`.gitignore` artifacts
```
ZAP_WEEKLY*.zip
test/jest/coverage/*
test/lighthouse/reports/*
test/cypress/videos/*
test/cypress/screenshots/*
test/audits/*
```
### What steps did you take, to get the error?
```
quasar ext add @quasar/testing
yarn [run some stuff]
git status
```
Answers:
username_1: Noticed this too.
.gitignore should probably be updated, but I dunno if directly on the original one (the one you get with a fresh install without anything) or upon installation of test options.
How does this relate with Jest "golden standard" for snapshots testing? Those must be kept in the repo
username_1: Local gitignore for jest and cypress has been added
Will add others as we proceed updating existing harnesses
username_0: warning " > [email protected]" has unmet peer dependency "webpack@^4.0.0".
[5/5] Building fresh packages...
success Saved lockfile.
Done in 8.04s.
App · Quasar App Extension "@quasar/testing-e2e-cypress" successfully installed.
App · Quasar App Extension "@quasar/testing" successfully installed.
$ cat .gitignore
.DS_Store
.thumbs.db
node_modules
# Quasar core related directories
.quasar
/dist
# Cordova related directories and files
/src-cordova/node_modules
/src-cordova/platforms
/src-cordova/plugins
/src-cordova/www
# Capacitor related directories and files
/src-capacitor/www
/src-capacitor/node_modules
# BEX related directories and files
/src-bex/www
/src-bex/js/core
# Log files
npm-debug.log*
yarn-debug.log*
yarn-error.log*
# Editor directories and files
.idea
*.suo
*.ntvs*
*.njsproj
*.sln
$ git status
On branch testing
Changes not staged for commit:
(use "git add <file>..." to update what will be committed)
(use "git restore <file>..." to discard changes in working directory)
modified: .vscode/settings.json
modified: babel.config.js
modified: package.json
modified: quasar.extensions.json
modified: yarn.lock
Untracked files:
(use "git add <file>..." to include in what will be committed)
.babelrc
cypress.json
jest.config.js
quasar.testing.json
test/
```
username_1: Your root .gitignore shouldn't be modified, check inside your harness folder
Status: Issue closed
username_0: Hi @username_1 , I tried this again and ran into some issues with XWindows on WSL. My previous config was there, but not the new stuff.
I'll try this again on my native linux laptop in a week or two and if there's a problem I'll open it again.
Thanks again for looking at it! |
canjs/can-reflect | 260237876 | Title: Typeof returns a string
Question:
username_0: We're seeing a lot of errors in our environment with IE11 users.
This line is the issue:
Answers:
username_1: closed by https://github.com/canjs/can-reflect/pull/45.
Status: Issue closed
username_1: This is published in [[email protected]](https://github.com/canjs/can-reflect/releases/tag/v1.4.1). Thanks for the fix! |
jlippold/tweakCompatible | 413647162 | Title: `Artsy` working on iOS 12.1.1
Question:
username_0: ```
{
"packageId": "ch.mdaus.artsy",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "ch.mdaus.artsy",
"deviceId": "iPhone10,5",
"url": "http://cydia.saurik.com/package/ch.mdaus.artsy/",
"iOSVersion": "12.1.1",
"packageVersionIndexed": true,
"packageName": "Artsy",
"category": "Tweaks",
"repository": "Packix",
"name": "Artsy",
"installed": "1.2",
"packageIndexed": true,
"packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.",
"id": "ch.mdaus.artsy",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.1.0",
"shortDescription": "Colorize and customize your music players",
"latest": "1.2",
"author": "<NAME>",
"packageStatus": "Unknown"
},
"base64": "<KEY>",
"chosenStatus": "working",
"notes": ""
}
```<issue_closed>
Status: Issue closed |
aws/aws-sam-cli | 1024790286 | Title: --no-progressbar option does not suppress ECR image upload progress
Question:
username_0: <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed).
If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. -->
### Description:
<!-- Briefly describe the bug you are facing.-->
When using `sam deploy --image-repository [ecr_uri]` and the `--no-progressbar` option, the progress bar shows when uploading images to ECR.
### Steps to reproduce:
<!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) -->
Use `sam deploy` CLI, providing `--image-repository` or `--resolve-image-repos` with `--no-progressbar` and view output which shows progress for uploading image to ECR.
### Observed result:
<!-- Please provide command output with `--debug` flag set. -->
### Expected result:
<!-- Describe what you expected. -->
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS:
2. `sam --version`: SAM CLI, version 1.30.0
3. AWS region:
`Add --debug flag to command you are running`
Status: Issue closed
Answers:
username_1: The PR is released in SAM CLI V 1.37.0 |
foam-framework/foam | 32636580 | Title: TableView 'rows' property ignored when embedded in ScrollPane.
Question:
username_0: _From [k...@<EMAIL>](https://code.google.com/u/101260952816761938100/) on November 13, 2013 09:26:01_
The 'rows' property is not honoured. Instead, the TableView will have its layout() method called which will size the table "appropriately" even if the size has been explicitly set.
_Original issue: http://code.google.com/p/foam-framework/issues/detail?id=72_
Answers:
username_0: Closed as obsolete. `foam.ui.FlexTableView` is more commonly used now in any case.
Status: Issue closed
|
DvdChe/social-pipe | 470654106 | Title: Missing LogLevel directive in configuration
Question:
username_0: When trying to run social-pipe, I got the following error message :
```
Traceback (most recent call last):
File "social-pipe.py", line 61, in <module>
LogLevel = int(conf['OPTIONS']['LogLevel'])
File "/opt/pkg/lib/python3.6/configparser.py", line 1233, in __getitem__
raise KeyError(key)
KeyError: 'LogLevel'
```
It turns out the `LogLevel` directive is missing from the `social-pipe.conf` file. I also noticed it is also missing in the `social-pipe.conf.tpl` file.
This was tested with the following platform :
- OS : macOS Mojave 10.14.5 (18F132)
- Python 3.6.9 from pkgsrc
- py36-virtualenv-16.6.1 (from pkgsrc)
- py36-setuptools-41.0.1 (from pkgsrc)
- requirements installed from requirements.txt
Answers:
username_1: Hey,
this is fixed in the following commit :
https://github.com/username_1/social-pipe/commit/7a156e2a31a5cf8d9fbcf149334ee0ce923432c5
Status: Issue closed
|
rpush/rpush | 782342717 | Title: ruby 3.0.0
Question:
username_0: Hi guys,
first of all, thanks for the great work. This gem really is a gem ;)
I noticed that `rpush` doesn't work with ruby 3.0.0. The problem is already fixed in the underlying gems (`net-http2` and `http2`). The `net-http2` commit is here [here](https://github.com/ostinelli/net-http2/issues/44). Can we update `rpush` to the newest `net-http2` version?
Thanks!
Answers:
username_1: I believe we don't restrict the `net-http2` version in a way that would prevent an update to the latest version. Are you sure it's `rpush` preventing you from updating this dependency?
Status: Issue closed
|
fogfish/httpbin | 476553683 | Title: Release 0.0.1
Question:
username_0: build is [pending](https://console.aws.amazon.com/codebuild/home?region=eu-west-1#/builds/username_0-httpbin:fc66064b-059d-4c02-bb41-4a23942400ce/view/new)
Answers:
username_0: build is [pending](https://console.aws.amazon.com/codebuild/home?region=eu-west-1#/builds/username_0-httpbin:fc66064b-059d-4c02-bb41-4a23942400ce/view/new)
username_0: [Build logs](https://console.aws.amazon.com/codebuild/home?region=eu-west-1#/builds/username_0-httpbin:fc66064b-059d-4c02-bb41-4a23942400ce/view/new)
```javascript
[
{
"phase-context": [],
"start-time": "Aug 4, 2019 1:28:53 PM",
"end-time": "Aug 4, 2019 1:28:54 PM",
"duration-in-seconds": 0,
"phase-type": "SUBMITTED",
"phase-status": "SUCCEEDED"
},
{
"phase-context": [],
"start-time": "Aug 4, 2019 1:28:54 PM",
"end-time": "Aug 4, 2019 1:28:55 PM",
"duration-in-seconds": 0,
"phase-type": "QUEUED",
"phase-status": "SUCCEEDED"
},
{
"phase-context": [
": "
],
"start-time": "Aug 4, 2019 1:28:55 PM",
"end-time": "Aug 4, 2019 1:30:08 PM",
"duration-in-seconds": 73,
"phase-type": "PROVISIONING",
"phase-status": "SUCCEEDED"
},
{
"phase-context": [
": "
],
"start-time": "Aug 4, 2019 1:30:08 PM",
"end-time": "Aug 4, 2019 1:30:11 PM",
"duration-in-seconds": 2,
"phase-type": "DOWNLOAD_SOURCE",
"phase-status": "SUCCEEDED"
},
{
"phase-context": [
": "
],
"start-time": "Aug 4, 2019 1:30:11 PM",
"end-time": "Aug 4, 2019 1:30:11 PM",
"duration-in-seconds": 0,
"phase-type": "INSTALL",
"phase-status": "SUCCEEDED"
},
{
"phase-context": [
": "
],
"start-time": "Aug 4, 2019 1:30:11 PM",
"end-time": "Aug 4, 2019 1:30:11 PM",
"duration-in-seconds": 0,
"phase-type": "PRE_BUILD",
"phase-status": "SUCCEEDED"
[Truncated]
"phase-type": "UPLOAD_ARTIFACTS",
"phase-status": "SUCCEEDED"
},
{
"phase-context": [
": "
],
"start-time": "Aug 4, 2019 1:30:11 PM",
"end-time": "Aug 4, 2019 1:30:15 PM",
"duration-in-seconds": 4,
"phase-type": "FINALIZING",
"phase-status": "SUCCEEDED"
},
{
"start-time": "Aug 4, 2019 1:30:15 PM",
"phase-type": "COMPLETED"
}
]
```
Status: Issue closed
|
caddyserver/caddy | 625738555 | Title: TLS Off
Question:
username_0: # version
v2.0.0 h1:pQSaIJGFluFvu8KDGDODV8u4/QRED/OPyIR+MWYYse8=
# Config
```
0.0.0.0:8090
tls off
file_server {
root /var/www/html
}
encode gzip
```
# Result
`run: adapting config using caddyfile: parsing caddyfile tokens for 'tls': /etc/caddy/Caddyfile:3 - Error during parsing: single argument must either be 'internal' or an email address`
# Expect
TLS off
# Question
Is real need force enable tls?
Answers:
username_1: Thanks for your question, and I'm thrilled that you're using Caddy! This looks more like a question about how to use Caddy rather than a bug report or feature request. You can find the docs for the tls directive here: https://caddyserver.com/docs/caddyfile/directives/tls
To disable HTTPS in the Caddyfile, specify `http://` in your site address directly.
Since this issue tracker is reserved for actionable development items, I'm going to close this, but we have a [community forum](https://caddy.community) where more people will be exposed to your question, including people who may be more expert or experienced with the specific question you're facing. I hope you'll ask your question there, and thanks for understanding!
Status: Issue closed
|
flutter/flutter | 803461583 | Title: TestSpan Null check operator used on a null value
Question:
username_0: ## Steps to Reproduce
```dart
import 'package:flutter/painting.dart';
void main() => TextSpan(semanticsLabel: '').computeSemanticsInformation([]);
```
**Expected results:**
Should work
**Actual results:**
```console
E/flutter (30289): [ERROR:flutter/lib/ui/ui_dart_state.cc(177)] Unhandled Exception: Null check operator used on a null value
E/flutter (30289): #0 TextSpan.computeSemanticsInformation (package:flutter/src/painting/text_span.dart:318:13)
E/flutter (30289): #1 main (package:bug/main.dart:3:45)
E/flutter (30289): #2 _runMainZoned.<anonymous closure>.<anonymous closure> (dart:ui/hooks.dart:231:25)
E/flutter (30289): #3 _rootRun (dart:async/zone.dart:1190:13)
E/flutter (30289): #4 _CustomZone.run (dart:async/zone.dart:1093:19)
E/flutter (30289): #5 _runZoned (dart:async/zone.dart:1630:10)
E/flutter (30289): #6 runZonedGuarded (dart:async/zone.dart:1618:12)
E/flutter (30289): #7 _runMainZoned.<anonymous closure> (dart:ui/hooks.dart:223:5)
E/flutter (30289): #8 _startIsolate.<anonymous closure> (dart:isolate-patch/isolate_patch.dart:301:19)
E/flutter (30289): #9 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:168:12)
```
<details>
<summary>Logs</summary>
```
[✓] Flutter (Channel stable, 1.22.6, on macOS 11.1 20C69 darwin-x64, locale en-BY)
• Flutter version 1.22.6 at /Users/username_0d/development/flutter
• Framework revision 9b2d32b605 (2 weeks ago), 2021-01-22 14:36:39 -0800
• Engine revision 2f0af37152
• Dart version 2.10.5
```
</details>
This is not a valid null check
https://github.com/flutter/flutter/blob/02d441ea55b328133c266991f43b0a1148edb63f/packages/flutter/lib/src/painting/text_span.dart#L280-L288
Answers:
username_1: Hi @username_0
Thanks for filing the issue, I can reproduce it on all channels
<details>
<summary>logs</summary>
```bash
[ ]
{"CFBundleName":"triage","DTXcode":"1240","DTSDKName":"iphonesimulator14.4","UILaunchStoryboardName":"LaunchScreen","CFBundleIcons~i
pad":{"CFBundlePrimaryIcon":{"CFBundleIconFiles":["AppIcon20x20","AppIcon29x29","AppIcon40x40","AppIcon60x60","AppIcon76x76","AppIco
n83.5x83.5"],"CFBundleIconName":"AppIcon"}},"DTSDKBuild":"18D46","CFBundleDevelopmentRegion":"en","CFBundleVersion":"1","BuildMachin
eOSBuild":"20D64","DTPlatformName":"iphonesimulator","CFBundlePackageType":"APPL","UIMainStoryboardFile":"Main","CFBundleSupportedPl
atforms":["iPhoneSimulator"],"CFBundleShortVersionString":"1.0.0","CFBundleInfoDictionaryVersion":"6.0","CFBundleExecutable":"Runner
","DTCompiler":"com.apple.compilers.llvm.clang.1_0","UISupportedInterfaceOrientations~ipad":["UIInterfaceOrientationPortrait","UIInt
erfaceOrientationPortraitUpsideDown","UIInterfaceOrientationLandscapeLeft","UIInterfaceOrientationLandscapeRight"],"MinimumOSVersion
":"9.0","CFBundleIdentifier":"com.nevercode.triage","UIDeviceFamily":[1,2],"DTPlatformVersion":"14.4","CFBundleSignature":"????","CF
BundleIcons":{"CFBundlePrimaryIcon":{"CFBundleIconFiles":["AppIcon20x20","AppIcon29x29","AppIcon40x40","AppIcon60x60"],"CFBundleIcon
Name":"AppIcon"}},"DTXcodeBuild":"12D4e","LSRequiresIPhoneOS":true,"UISupportedInterfaceOrientations":["UIInterfaceOrientationPortra
it","UIInterfaceOrientationLandscapeLeft","UIInterfaceOrientationLandscapeRight"],"UIViewControllerBasedStatusBarAppearance":false,"
NSBonjourServices":["_dartobservatory._tcp"],"DTPlatformBuild":"18D46","NSLocalNetworkUsageDescription":"Allow Flutter tools on your
computer to connect and debug your application. This prompt will not appear on release builds."}
[ +2 ms] executing: xcrun simctl launch DB6F8402-E89F-4E94-9644-151970DBE05B com.nevercode.triage --enable-dart-profiling
--enable-checked-mode --verify-entry-points --observatory-port=0
[ +293 ms] com.nevercode.triage: 54833
[ +1 ms] Waiting for observatory port to be available...
[ +466 ms] Observatory URL on device: http://127.0.0.1:56032/rPVxOgE1ZSY=/
[ +4 ms] Caching compiled dill
[ +54 ms] Connecting to service protocol: http://127.0.0.1:56032/rPVxOgE1ZSY=/
[ +63 ms] [VERBOSE-2:ui_dart_state.cc(186)] Unhandled Exception: Null check operator used on a null value
#0 TextSpan.computeSemanticsInformation (package:flutter/src/painting/text_span.dart:284:13)
#1 main (package:triage/main.dart:3:45)
#2 _runMainZoned.<anonymous closure>.<anonymous closure> (dart:ui/hooks.dart:146:25)
#3 _rootRun (dart:async/zone.dart:1354:13)
#4 _CustomZone.run (dart:async/zone.dart:1258:19)
#5 _runZoned (dart:async/zone.dart:1789:10)
#6 runZonedGuarded (dart:async/zone.dart:1777:12)
#7 _runMainZoned.<anonymous closure> (dart:ui/hooks.dart:139:5)
#8 _delayEntrypointInvocation.<anonymous closure> (dart:isolate-patch/isolate_patch.dart:283:19)
#9 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:184:12)
[ +223 ms] Launching a Dart Developer Service (DDS) instance at http://127.0.0.1:0, connecting to VM service at
http://127.0.0.1:56032/rPVxOgE1ZSY=/.
[ +63 ms] DDS is listening at http://127.0.0.1:56035/vIjxhtUqor0=/.
[ +59 ms] Successfully connected to service protocol: http://127.0.0.1:56032/rPVxOgE1ZSY=/
[ +20 ms] DevFS: Creating new filesystem on the device (null)
[ +12 ms] DevFS: Created new filesystem on the device
(file:///Users/tahatesser/Library/Developer/CoreSimulator/Devices/DB6F8402-E89F-4E94-9644-151970DBE05B/data/Containers/Data/Applicat
ion/55813AE1-B457-4E26-8C44-3F1A7645D31C/tmp/master_flutterjP3II7/master_flutter/)
[ +3 ms] Updating assets
[ +112 ms] Syncing files to device iPhone 12 Pro Max...
[ +2 ms] <- reset
[ ] Compiling dart to kernel with 0 updated files
[ +3 ms] <- recompile package:triage/main.dart 7d34d5e3-b0e2-4eee-96d0-cf0686eccd5c
[ ] <- 7d34d5e3-b0e2-4eee-96d0-cf0686eccd5c
[ +20 ms] Updating files.
[ ] DevFS: Sync finished
[ +1 ms] Syncing files to device iPhone 12 Pro Max... (completed in 26ms)
[ +3 ms] Synced 0.0MB.
[ +1 ms] <- accept
[ +5 ms] Connected to _flutterView/0x7f8aca819020.
[ +1 ms] Flutter run key commands.
[Truncated]
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin can be installed from:
🔨 https://plugins.jetbrains.com/plugin/9212-flutter
• Dart plugin can be installed from:
🔨 https://plugins.jetbrains.com/plugin/6351-dart
• Java version OpenJDK Runtime Environment (build 1.8.0_242-release-1644-b3-6915495)
[✓] VS Code (version 1.53.0)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.19.0
[✓] Connected device (4 available)
• Taha’s iPad (mobile) • 00008020-000255113EE8402E • ios • iOS 14.4
• iPhone 12 Pro Max (mobile) • DB6F8402-E89F-4E94-9644-151970DBE05B • ios • com.apple.CoreSimulator.SimRuntime.iOS-14-4 (simulator)
• macOS (desktop) • macos • darwin-x64 • macOS 11.2 20D64 darwin-x64
• Chrome (web) • chrome • web-javascript • Google Chrome 88.0.4324.146
! Doctor found issues in 2 categories.
```
</details>
Thank you
username_2: Do you have a real application use case for this?
I agree this is a bug, but I think it would be helpful to see a "real life" example rather than the constructed one.
username_0: Hi @username_2
I have a lot of nested `TextSpan` and some of them should be clickable
```dart
TextSpan(
recognizer: recognizer,
children: [
// TextSpans
],
);
```
But due to the condition
```dart
if (text != null || semanticsLabel != null) {
```
recognizer doesn't work
I decided to use `semanticsLabel` so as not to create unnecessary `text` and got the error that I created here
username_2: We should definitely fix the logic around the null check there. But it sounds like we're missing a way to add recognizers there that we shouldn't be. You shouldn't need an empty semantics label to get the recognizer to work
username_0: I agree
Need to think of something
username_2: As a work around, you might bea ble to try using `''` for the text - we shouldn't require that but I think it'll work. Can you try that?
username_0: There is a small problem
Gestures will only be recognized on the `TextSpan` to which they are bound
```dart
import 'package:flutter/gestures.dart';
import 'package:flutter/material.dart';
void main() {
runApp(_Test());
}
class _Test extends StatelessWidget {
final TapGestureRecognizer tapGestureRecognizer = TapGestureRecognizer()
..onTap = () {
print(DateTime.now());
};
@override
Widget build(BuildContext context) => Container(
color: Colors.white,
child: Center(
child: RichText(
textDirection: TextDirection.ltr,
text: TextSpan(
text: 'root text ',
recognizer: tapGestureRecognizer,
style: TextStyle(
color: Colors.black,
),
children: [
TextSpan(
text: 'Test text',
),
],
),
),
),
);
}
```
https://user-images.githubusercontent.com/16593680/107416558-ad46ea80-6b25-11eb-9f70-d7af8990ce4d.mov
Status: Issue closed
username_0: @username_2
Should I create a separate issue for this?
https://github.com/flutter/flutter/issues/75622#issuecomment-776184709
username_2: ## Steps to Reproduce
```dart
import 'package:flutter/painting.dart';
void main() => TextSpan(semanticsLabel: '').computeSemanticsInformation([]);
```
**Expected results:**
Should work
**Actual results:**
```console
E/flutter (30289): [ERROR:flutter/lib/ui/ui_dart_state.cc(177)] Unhandled Exception: Null check operator used on a null value
E/flutter (30289): #0 TextSpan.computeSemanticsInformation (package:flutter/src/painting/text_span.dart:318:13)
E/flutter (30289): #1 main (package:bug/main.dart:3:45)
E/flutter (30289): #2 _runMainZoned.<anonymous closure>.<anonymous closure> (dart:ui/hooks.dart:231:25)
E/flutter (30289): #3 _rootRun (dart:async/zone.dart:1190:13)
E/flutter (30289): #4 _CustomZone.run (dart:async/zone.dart:1093:19)
E/flutter (30289): #5 _runZoned (dart:async/zone.dart:1630:10)
E/flutter (30289): #6 runZonedGuarded (dart:async/zone.dart:1618:12)
E/flutter (30289): #7 _runMainZoned.<anonymous closure> (dart:ui/hooks.dart:223:5)
E/flutter (30289): #8 _startIsolate.<anonymous closure> (dart:isolate-patch/isolate_patch.dart:301:19)
E/flutter (30289): #9 _RawReceivePortImpl._handleMessage (dart:isolate-patch/isolate_patch.dart:168:12)
```
<details>
<summary>Logs</summary>
```
[✓] Flutter (Channel stable, 1.22.6, on macOS 11.1 20C69 darwin-x64, locale en-BY)
• Flutter version 1.22.6 at /Users/username_0d/development/flutter
• Framework revision 9b2d32b605 (2 weeks ago), 2021-01-22 14:36:39 -0800
• Engine revision 2f0af37152
• Dart version 2.10.5
```
</details>
This is not a valid null check
https://github.com/flutter/flutter/blob/02d441ea55b328133c266991f43b0a1148edb63f/packages/flutter/lib/src/painting/text_span.dart#L280-L288
username_2: We can use this one |
jamesaoverton/clovertonemusic | 675554554 | Title: Email an admin on serious errors
Question:
username_0: Please update the system (the logging system I guess) to email a configurable admin (just one address is fine) on the most serious errors. One error I'm worried about in particular is rclone failures.<issue_closed>
Status: Issue closed |
MarcGiffing/wicket-spring-boot | 128427352 | Title: Autoconfigure for Wicketstuff Restannoations
Question:
username_0: Couldn't inject spring beans when subclassing AbstractRestResource. Injecter.get().inject(this).
Throws error:
Caused by: java.lang.IllegalStateException: Concrete bean could not be received from the application context for class: com.giffing.wicket.spring.boot.example.repository.services.customer.CustomerRepositoryService.
at org.apache.wicket.spring.SpringBeanLocator.lookupSpringBean(SpringBeanLocator.java:275) ~[wicket-spring-7.1.0.jar:7.1.0]
at org.apache.wicket.spring.SpringBeanLocator.locateProxyTarget(SpringBeanLocator.java:198) ~[wicket-spring-7.1.0.jar:7.1.0]
at org.apache.wicket.spring.injection.annot.AnnotProxyFieldValueFactory.getFieldValue(AnnotProxyFieldValueFactory.java:148) ~[wicket-spring-7.1.0.jar:7.1.0]
at org.apache.wicket.injection.Injector.inject(Injector.java:111) ~[wicket-ioc-7.1.0.jar:7.1.0]
at org.apache.wicket.spring.injection.annot.SpringComponentInjector.inject(SpringComponentInjector.java:124) ~[wicket-spring-7.1.0.jar:7.1.0]
at com.giffing.wicket.spring.boot.example.web.pages.customers.CustomerResource.<init>(CustomerResource.java:31) ~[classes/:na]
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[na:1.8.0_45]
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) ~[na:1.8.0_45]
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[na:1.8.0_45]
at java.lang.reflect.Constructor.newInstance(Constructor.java:422) ~[na:1.8.0_45]
at java.lang.Class.newInstance(Class.java:442) ~[na:1.8.0_45]
at org.wicketstuff.rest.utils.mounting.PackageScanner.mountAnnotatedResource(PackageScanner.java:92) ~[wicketstuff-restannotations-7.1.0.jar:7.1.0]
at org.wicketstuff.rest.utils.mounting.PackageScanner.scanPackage(PackageScanner.java:72) ~[wicketstuff-restannotations-7.1.0.jar:7.1.0]
... 18 common frames omitted
Answers:
username_0: Couldn't inject spring beans when subclassing AbstractRestResource. Injecter.get().inject(this).
Throws error:
Caused by: java.lang.IllegalStateException: Concrete bean could not be received from the application context for class: com.giffing.wicket.spring.boot.example.repository.services.customer.CustomerRepositoryService.
at org.apache.wicket.spring.SpringBeanLocator.lookupSpringBean(SpringBeanLocator.java:275) ~[wicket-spring-7.1.0.jar:7.1.0]
at org.apache.wicket.spring.SpringBeanLocator.locateProxyTarget(SpringBeanLocator.java:198) ~[wicket-spring-7.1.0.jar:7.1.0]
at org.apache.wicket.spring.injection.annot.AnnotProxyFieldValueFactory.getFieldValue(AnnotProxyFieldValueFactory.java:148) ~[wicket-spring-7.1.0.jar:7.1.0]
at org.apache.wicket.injection.Injector.inject(Injector.java:111) ~[wicket-ioc-7.1.0.jar:7.1.0]
at org.apache.wicket.spring.injection.annot.SpringComponentInjector.inject(SpringComponentInjector.java:124) ~[wicket-spring-7.1.0.jar:7.1.0]
at com.giffing.wicket.spring.boot.example.web.pages.customers.CustomerResource.<init>(CustomerResource.java:31) ~[classes/:na]
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[na:1.8.0_45]
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) ~[na:1.8.0_45]
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[na:1.8.0_45]
at java.lang.reflect.Constructor.newInstance(Constructor.java:422) ~[na:1.8.0_45]
at java.lang.Class.newInstance(Class.java:442) ~[na:1.8.0_45]
at org.wicketstuff.rest.utils.mounting.PackageScanner.mountAnnotatedResource(PackageScanner.java:92) ~[wicketstuff-restannotations-7.1.0.jar:7.1.0]
at org.wicketstuff.rest.utils.mounting.PackageScanner.scanPackage(PackageScanner.java:72) ~[wicketstuff-restannotations-7.1.0.jar:7.1.0]
... 18 common frames omitted
username_0: The problem is that Wicket is initialized before Spring Data JPA. Wicket should be configured at last.
username_0: Even if the order changed the bean was not found. The bean was found if the name is provided. (But it couldn't be autowired cause of proxy type problems)
username_0: Couldn't even inject the spring application context.
username_0: Same issue with SpringReference:
Zero or more than one spring bean candidates. Type: interface com.giffing.wicket.spring.boot.example.repository.services.customer.CustomerRepositoryService, candidates: []
Status: Issue closed
|
jacktan1991/compose-scaling | 581482853 | Title: swarm服务发现
Question:
username_0: # 默认discovery方案
https://rollout.io/blog/docker-machine-compose-and-swarm-how-they-work-together/
- docker-machine create -d virtualbox local
- docker run -it --rm swarm create
- docker-machine create -d virtualbox --swarm --swarm-master --swarm-discovery token://YOURTOKENHERE swarm-master
- 但已经 Deprecated: https://docs.docker.com/swarm/discovery/#docker-hub-as-a-hosted-discovery-service#docker-hub-as-a-hosted-discovery-service
# consul方案改进
https://dzone.com/articles/docker-swarm-cluster-using-consul
docker-machine create -d virtualbox consul
docker run -d -p 8301:8301 -p 8302:8302 -p 8400:8400 -p 8500:8500 -p 53:8600/udp consul
docker-machine create -d virtualbox --swarm --swarm-master --swarm-discovery="consul://$(docker-machine ip consul):8500" --engine-opt="cluster-store=consul://$(docker-machine ip consul):8500" --engine-opt="cluster-advertise=eth1:2376" node-master
docker-machine create -d virtualbox --swarm --swarm-discovery="consul://$(docker-machine ip consul):8500" --engine-opt="cluster-store=consul://$(docker-machine ip consul):8500" --engine-opt="cluster-advertise=eth1:2376" node-slave01
eval $(docker-machine env --swarm node-master)
另一种方案
https://lastviking.eu/install_docker_swarm.html |
CCS-Lab/easyml | 207811497 | Title: Random forest on a continuous dependent measure --> replace RMSE with correlation coefficients
Question:
username_0: With easy_random_forest, currently I see histograms of RMSE across X number of training/test divisions. Just like easy_glmnet, it should plot histograms of correlation coefficients (perhaps along with RMSE?).
Answers:
username_1: The easy_random_forest() function will output histograms of R^2 metrics if the `measure` argument is set to `measure_r2_score` (i.e. `measure = measure_r2_score`). If this is made the default, I think it would solve the issue.
username_2: See e94013dc8bee9a8f8189b14669ba8713555c4646.
Status: Issue closed
|
jquery/jquery | 162334440 | Title: Can't retrieve the <body> or <title>, etc., of a document loaded with $.ajax()
Question:
username_0: As a corollary of bug #3203, you cannot retrieve the contents of the `<title>` tag of a document loaded with `$.ajax`.
```
$.ajax({
url: 'somePageWithATitle.html',
success: function(data) {
// NONE of these works:
console.log("The title of the loaded page is (not): "+$("title", data).text());
console.log("The title of the loaded page is (not): "+$(data).find("title").text());
}
});
```
You can't get the contents of the `<body>` either, nor the `<script>` or `<meta>` tags in the `<head>`, nor anything like that.
Answers:
username_1: You're using the API incorrectly, see https://github.com/jquery/jquery/issues/3203#issuecomment-228609314.
Status: Issue closed
|
home-assistant/frontend | 606448696 | Title: Color battery icon red when battery level under 20%
Question:
username_0: <!--
DO NOT DELETE ANY TEXT from this template!
Otherwise, your request may be closed without comment.
-->
## The request
<!--
Describe to our maintainers, the feature you would like to be added.
Please be clear and concise and, if possible, provide a screenshot or mockup.
-->
Currently Home Assistant displays all battery icons the exact same. I think that when your battery level is getting critically low, you want that to be more obvious at a glance so you can either charge the device or purchase new batteries. Especially since a lot of these Z-Wave sensors use less common batteries like CR123A that aren't always readily available at your local store.
Many apps and operating systems already do something like this, and the standard usually seems to be when <20%.

I think that it definitely makes sense on the Devices areas (listing and details) as those sections are more management/maintenance related in nature. It makes sense on the more-info dialogs as well.
Some might be a little bit more picky about it being in Lovelace, but with the `state_color:` card options they could override whatever default is set and turn the icon coloring on or off as they wish.
## The alternatives
<!--
Are you currently using, or have you considered alternatives?
If so, could you please describe those?
-->
Could do green (high), orange (medium), red (low) but that is overkill IMO. Most people will only care when it needs attention, so just red is fine.
## Additional information
Answers:
username_1: This is kinda difficult. My phone is indeed fine to show red < 20% because it will die soon. But my sensor might work just fine for 2 months on 20% and I only want to replace it when it is below 5%.
username_0: That's a very good point, there's not a perfect one size fits all criteria since battery life varies greatly by device. But even if criteria was changed to be below 5% though I think this would still be helpful.
I rarely pay attention to my sensor battery levels because they last for months. But if I'm scrolling through my Devices list and spot a red icon in the corner of my eye, and see that it's just 3% remaining, that can make the difference between spotting it in time and replacing it or noticing after it's too late.
username_2: Setup some alerts on that as well 👍 |
demisto/content-docs | 688092043 | Title: Issue with "AssignAnalystToIncident" in @site/docs/reference/scripts/assign-analyst-to-incident.md
Question:
username_0: <!--
Thank you for taking the time to help us improve our documentation! Please describe the problem and a suggested fix below and we'll get back to you as soon as we can.
-->
## Describe the problem
* Page: [AssignAnalystToIncident](https://xsoar.pan.dev/docs/reference/scripts/assign-analyst-to-incident)
* Source: https://github.com/demisto/content/blob/master//Packs/CommonScripts/Scripts/script-AssignAnalystToIncident_README.md
<!--- Is this a typo, stale information, request for improvement, inaccuracy? -->
<!--- Clearly and concisely describe the problem with the documentation -->
Documentation is not complete : the onCall parameter is not explained
## Screenshots
<!-- If applicable, add screenshots to help explain your problem. -->
## Environment
- OS: any
- Browser: any
- Browser Version: any
## Suggested fix
<!--- If possible, help us by offering a suggested fix to the problem. If you know the fix, you may also submit a PR to fix the issue if you like! -->
Write the documentation :)<issue_closed>
Status: Issue closed |
WizardFactory/TodayWeather | 271588630 | Title: 특정 버전 이상에서 차트내에서 상하 스크롤이 안되는 문제
Question:
username_0: 새로 빌드하는 경우에 WKWebview를 사용했던 것처럼 차트내에서 상하스크롤이 안됨
Answers:
username_0: https://forum.ionicframework.com/t/when-using-horizontal-scroll-ion-scroll-page-cant-be-scrolled-down-in-y-direction/3833/46
#68
username_0: https://codepen.io/jcsmith1859/full/dIHyK/ -> iPad browser에서 정상동작하지만, 따라해도 해결되지 않음.
$ionicConfigProvider.platform.ios.scrolling.jsScrolling(true); -> native scroll를 off해도 동작하지 않음. -> 완전히 off된건지는 확실치 않음.
username_0: sample code에서 아래 코드 동작확인함.
```
<ion-content class="padding" overflow-scroll="true">
<div style="overflow: auto;white-space: nowrap; overflow-x:scroll; overflow-y:hidden;">
<div style="width: 2400px; background-color: #009689">
<p>
test
</p>
<p>
test
</p>
</div>
</div>
</ion-content>
```
username_0: #1811 에서 들어간 아래 내용때문에 안되는 문제였음
.has-header.overflow-scroll일 때 -webkit-overflow-scrolling를 auto로 변경
username_1: iOS에서 시작페이지, 업데이트 주기 변경 후에 흰색 화면으로 나오는 문제 수정 작업으로 추가된 코드입니다. WKWebview에서 제거 후 정상동작하면 빼셔도 됩니다
username_0: 먼저 UIWebview에서 동작해야 하니 수정방법 궁리해봐야 할듯하네요.
username_0: - [ ] 광고 그리는 방식 변경되어, webkit-overflow-scrolling: auto가 삭제 가능할지도 모름
Status: Issue closed
|
Azure/azure-cli | 687582252 | Title: az error
Question:
username_0: ### **This is autogenerated. Please review and update as needed.**
## Describe the bug
**Command Name**
`az role assignment list`
**Errors:**
```
400 Client Error: Bad Request for url: http://localhost:50342/oauth2/token
Traceback (most recent call last):
python3.6/site-packages/knack/cli.py, ln 215, in invoke
cmd_result = self.invocation.execute(args)
cli/core/commands/__init__.py, ln 654, in execute
raise ex
cli/core/commands/__init__.py, ln 718, in _run_jobs_serially
results.append(self._run_job(expanded_arg, cmd_copy))
...
python3.6/site-packages/msrestazure/azure_active_directory.py, ln 486, in get_msi_token
result.raise_for_status()
python3.6/site-packages/requests/models.py, ln 940, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 400 Client Error: Bad Request for url: http://localhost:50342/oauth2/token
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az role assignment list`
## Expected Behavior
## Environment Summary
```
Linux-4.15.0-1092-azure-x86_64-with-debian-stretch-sid (Cloud Shell)
Python 3.6.10
Installer: DEB
azure-cli 2.11.0
Extensions:
resource-graph 0.1.10
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
Answers:
username_1: add to S175
username_2: known issue in #11749 . Pls try `az login` to workaround. thanks
Status: Issue closed
|
stfc/PSyclone | 498808165 | Title: [NEMO] Insertion of profiling around a CASE block fails
Question:
username_0: Since a CASE has no direct representation in the PSyIR the business of getting a pointer back into the parse tree is a bit ropey and in some circumstances we fail to get that location right.
Answers:
username_0: Unfortunately, my 'fix' breaks the test for profiling of case statements. I need to re-visit the actual bug I found when processing NEMO source...
username_1: Fixed in #522 .
Status: Issue closed
|
samp-incognito/samp-streamer-plugin | 723838166 | Title: Get random position in circle
Question:
username_0: Hello
First of all, **I apologize if this post has nothing to do with the section**.
SA-MP forum is unfortunately closed ...
I want a function to create random points in the circle.
I have this but it does not work, I got it from stackoverflow:
```
GetRandomPosInCircle(Float: X, Float: Y, Float: Radius, &Float: fX, &Float: fY)
{
new theta = random(2);
fX = X + Radius * floatcos(theta);
fY = Y + Radius * floatsin(theta);
return 1;
}
```
Can someone give me a function works?
Thanks<issue_closed>
Status: Issue closed |
leo424y/heysiri.ml | 863539553 | Title: 蜜蜂算法(Bees algorithm)
協同過濾
協同人力翻譯(Collaborative human interpreter)
合作創新網絡(Collaborative innovation network)
協同智能(Collaborative intelligence)
群件和Wiki
集體行動(Collective action)
集體意識(Collective consciousness)
集體決策(Collective decision-making)
集體興奮(Collective eff
Question:
username_0: 蜜蜂算法(Bees algorithm)
協同過濾
協同人力翻譯(Collaborative human interpreter)
合作創新網絡(Collaborative innovation network)
協同智能(Collaborative intelligence)
群件和Wiki
集體行動(Collective action)
集體意識(Collective consciousness)
集體決策(Collective decision-making)
集體興奮(Collective effervescence )
連通圖
從眾心理(Crowd psychology)
眾包
客戶互動(Customer engagement)
控制論
分布式認知(Distributed cognition)
企業書籤(Enterprise bookmarking)
全球腦
全球意識項目(Global Consciousness Project)
群體行為(Group behaviour)
群體心理(Group mind)
主持討論(Facilitation)和促進者(Facilitator)
基於人員的計算(Human-based computation)
百猴效應(Hundredth Monkey)
信息路由組(Information Routing Group)
攀比(Keeping up with the Joneses)
知識生態系統(Knowledge ecosystem)
迷因
智慧圈(Noosphere)
開放空間會議(Open-space meeting)
公開來源情報(Open source intelligence)
預測市場
偏好誘導(Preference elicitation)
推薦系統(Recommendation system)
聰明行動族
社會化商務(Social commerce)
社會信息處理(Social information processing)
社會認知優化(Social cognitive optimization)
共識主動性(Stigmergy)
超有機體
系統智能(Systems intelligence)
集體智慧
群體的智慧 |
Fonsart/NHdesign | 129118481 | Title: Search in the header
Question:
username_0: We need a bigger field in order to write the text and to see it. What should we do ?
1. just make it larger ?
2. add an animation when active = larger ?
3. something else ?
@Yago + @nsz + @guillaumeberry + @username_1<issue_closed>
Status: Issue closed |
schickling/dockerfiles | 1166473264 | Title: postgres-backup-s3 not updated + CD pipeline
Question:
username_0: Hey, nice repo!
It seems that postgres-backup-s3 is not up to date on Docker Hub. The last commit was just two months ago, but [Docker Hub says "Last pushed 4 years ago"](https://hub.docker.com/r/schickling/postgres-backup-s3/tags). Any chance it's been missed?
Also, I saw in https://github.com/schickling/dockerfiles/issues/57#issuecomment-271148152 that you were looking into setting up a CD pipeline pipeline to automatically release new versions of your images. Is this something that you've gotten around to do, or are you open to solutions? I'd be happy to take a look at it, but don't want to do double work in case it's already done :smile:
Cheers!
Malcolm
Answers:
username_0: @username_1 Are you maintaining the images on Docker Hub? As mentioned I'd be willing to set up a CD pipeline.
username_1: @username_0 if you could set up a CD pipeline, that would be amazing. Yup I'm helping with maintenance here!
username_0: @username_1 I'll mock something up and raise a PR. Going on vacation at the end of the week, so we'll see when I manage to get it done! :)
username_0: @username_1
I’ve thought a little bit about how this would best be done and have come up with three alternatives, each building upon the previous.
Just for clarity, when I say “package” I’m talking about one of the directories residing in the root of this repository.
## Alternative 1
It seems that currently all Docker Images in this repo is pushed to the `:latest` tag on Docker Hub. This would be simple to recreate in a CD pipeline as we could just check which package have been modified in a certain commit and push a new version of `:latest` for that package. While that’s very easy to implement the big downside is that usage of `:latest` is generally discouraged and makes it impossible to pin dependencies in a production environment.
## Alternative 2
So we have to solve the issue of maintaining multiple versions of packages in a monorepo. For tracking which version each package is on I can see three viable options:
1. Each package has a `version.txt` file (the name doesn’t matter of course) whose sole purpose is to track the version of the package.
2. There is a `versions.json` file in the root of the project containing the versions of all packages.
3. We track versions of each package using Git tags formatted like `postgres-backup-s3/v1.2.3`. An issue with this approach is that the latest version displayed on GitHub will always be the latest version of the last updated package, as there is no way to display versions for multiple tags.
Personally I prefer option 3 as it’s Git-native and makes it easier for the developer to push a new release as the CD flow could be triggered when a new tag is created and pushed: `git tag -a postgres-backup-s3/v1.2.3 -m "postgres-backup-s3 version 1.2.3"`. For options 1 and 2 we’d probably have to check whenever the version files are modified and check if the version has changed.
Regardless of the approach we would then push the following tags to Docker Hub:
- `postgres-backup-s3:v1.2.3`
- `postgres-backup-s3:v1.2`
- `postgres-backup-s3:v1`
- `postgres-backup-s3:latest`
By pushing different semver “levels” we allow users to receive only patch or minor updates while staying on the same major/mino version.
The added amount of tags would not increase the amount of images being built or the amount of storage being used on Docker Hub (even though I’m pretty sure they don’t limit storage) as each tag is just a reference to the same image, just like Git tags.
## Alternative 3
Having versioned, up to date images on Docker Hub is already great, but there is still one issue that persists. Some images target specific versions of software by nature. Take `postgres-backup-s3` for example. It uses `postgresql-client` whose version must align with the version of Postgres being targeted, but to use this backup image you must just hope that the currently released version uses the same version of `postgresql-client` as your database.
So ideally to solve this we would release tags like this for each image: `postgres-backup-s3:{postgresVersion}-v{imageVersion}`. This would allow the user to specify which version of Postgres to target.
I haven’t played around with this yet, but I’d imagine we’d have some form of manifest file in each package that needs this feature. That file would contain a list of arguments to use when building the Dockerfile. So the manifest might look something like this:
```json
{
"postgresVersion": [14, 13, 12]
}
```
And the Dockerfile would then be built three times with that arg being passed:
```bash
docker build --build-arg postgresVersion=14 .
docker build --build-arg postgresVersion=13 .
docker build --build-arg postgresVersion=12 .
```
---
I’d love to hear your thoughts and discuss which direction you’d like to go. As someone using dockerfiles from this repo myself I’d very much like to see alternative 3 being implemented, but I fully understand if it adds too much complexity to the project.
Cheers,
Malcolm |
LiskHQ/lisk-sdk | 488100703 | Title: The rate limiter is not banning peers
Question:
username_0: ### Expected behavior
The rate limiter logic should ban malicious peers for a short time.
### Actual behavior
Currently, it applies a 10-point ban score and then disconnects the peer. Disconnecting the peer resets their ban score to 100.
### Steps to reproduce
Testnet
### Which version(s) does this affect? (Environment, OS, etc...)
SDK 2.3<issue_closed>
Status: Issue closed |
istio/test-infra | 1102237631 | Title: Jobs failing with [Errno 30] Read-only file system: '/etc/rel-pipeline-docker-config/tmp9pdcb698
Question:
username_0: I see that two jobs:
https://prow.istio.io/job-history/gs/istio-prow/logs/build-base-images_release-builder_release-1.12_periodic
and
https://prow.istio.io/view/gs/istio-prow/logs/build-base-images_release-builder_periodic/1481702742933114880
have started failing with:
```
gcloud auth configure-docker -q
45
Adding credentials for all GCR repositories.
46
WARNING: A long list of credential helpers may cause delays running 'docker build'. We recommend passing the registry name to configure only the registry you are using.
47
ERROR: (gcloud.auth.configure-docker) Error writing Docker configuration to disk: [Errno 30] Read-only file system: '/etc/rel-pipeline-docker-config/tmp9pdcb698'
```
I'm not sure, but I do notice there was a Prow update PR, https://github.com/istio/test-infra/pull/3740, merged a few hours ago.
Answers:
username_1: Its a false positive highlighted by regex. The real error is
```
Error: failed image scan: base image scan of istio/base:2021-10-06T19-01-15 failed. Unable to process exit code:
```
its happened for years. Would be great to fix it
Status: Issue closed
|
carpenike/k8s-gitops | 558752539 | Title: Enable NVIDIA GPU Operator
Question:
username_0: After moving R720 as a worker.
https://github.com/NVIDIA/gpu-operator
Answers:
username_0: Look at building nvidia image with patch built in, or use another container to patch onces the driver is installed.
https://github.com/keylase/nvidia-patch/blob/master/patch.sh
Status: Issue closed
username_0: Appears that this requires a custom nvidia-docker instance. not going to move forward with it. |
ianhattendorf/autocomplete-ruby | 216068388 | Title: Error starting rsense
Question:
username_0: I've updated my gems; and i'm still getting this error.
Even after setting the `rsense` path (mine is `/Users/username_0/.rbenv/shims/rsense`
```
autocomplete-ruby: exec error: Error: Command failed: /Users/username_0/.rbenv/versions/2.3.3/bin/rsense start --port 47367 --path /Users/username_0/___masked___/___masked___
/Users/username_0/.rbenv/versions/2.3.3/lib/ruby/2.3.0/rubygems/specification.rb:2285:in `raise_if_conflicts': Unable to activate rsense-server-0.5.18, because jruby-jars-9.1.8.0 conflicts with jruby-jars (~> 1.7.4) (Gem::ConflictError)
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/2.3.0/rubygems/specification.rb:1408:in `activate'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/2.3.0/rubygems.rb:196:in `rescue in try_activate'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/2.3.0/rubygems.rb:193:in `try_activate'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/2.3.0/rubygems/core_ext/kernel_require.rb:125:in `rescue in require'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/2.3.0/rubygems/core_ext/kernel_require.rb:40:in `require'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/gems/2.3.0/gems/rsense-0.5.18/lib/rsense/client/daemon.rb:4:in `<top (required)>'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/gems/2.3.0/gems/rsense-0.5.18/bin/rsense:3:in `require_relative'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/gems/2.3.0/gems/rsense-0.5.18/bin/rsense:3:in `<top (required)>'
from /Users/username_0/.rbenv/versions/2.3.3/bin/rsense:22:in `load'
from /Users/username_0/.rbenv/versions/2.3.3/bin/rsense:22:in `<main>'
(You might need to set the rsense path, see the readme)
```
Using `rbenv 1.1.0`
I've never had this working before. I've been relying on the core implementation. `autocomplete-plus`
Answers:
username_1: This looks like an error with RSense, or possibly rbenv. Are you able to start rsense from the command line in the directory of one of your ruby projects?
username_0: Nope.; It actually spits out the same error 😟 :
```
/Users/username_0/.rbenv/versions/2.3.3/lib/ruby/2.3.0/rubygems/specification.rb:2285:in `raise_if_conflicts': Unable to activate rsense-server-0.5.18, because jruby-jars-192.168.3.11 conflicts with jruby-jars (~> 1.7.4) (Gem::ConflictError)
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/2.3.0/rubygems/specification.rb:1408:in `activate'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/2.3.0/rubygems.rb:196:in `rescue in try_activate'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/2.3.0/rubygems.rb:193:in `try_activate'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/2.3.0/rubygems/core_ext/kernel_require.rb:125:in `rescue in require'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/2.3.0/rubygems/core_ext/kernel_require.rb:40:in `require'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/gems/2.3.0/gems/rsense-0.5.18/lib/rsense/client/daemon.rb:4:in `<top (required)>'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/gems/2.3.0/gems/rsense-0.5.18/bin/rsense:3:in `require_relative'
from /Users/username_0/.rbenv/versions/2.3.3/lib/ruby/gems/2.3.0/gems/rsense-0.5.18/bin/rsense:3:in `<top (required)>'
from /Users/username_0/.rbenv/versions/2.3.3/bin/rsense:22:in `load'
from /Users/username_0/.rbenv/versions/2.3.3/bin/rsense:22:in `<main>'
```
I haven't jumped back in to find the culprit; but as a Rubyist and a veteran Rails engineer, I MUST FIND A SOLUTION 😝
username_1: I would try uninstalling all ruby versions + rbenv, cleaning all ruby related caches (gems/dotfiles/etc.), and seeing if it works with just the system ruby. If that doesn't work, try reinstalling rbenv with just the latest ruby version, and setting that as the global default version.
If that doesn't work, then I'm out of ideas unfortunately.
username_0: I'll try my best to do that; but at the risk of borking my multiple local
dev stacks, may wait until the weekend. Or give up on this, as rsense is
dead. Pity, bummer, shame. :(
username_1: I'm going to go ahead an close this, feel free to reply if you have any updates.
Status: Issue closed
username_0: All good! Thanks Ian.
username_2: I don't know if it related with this issue, but i realized that rsense was not starting recently, so I solved uninstalling everything and install in the follow order, forcing sinatra version:
```
$ gem install sinatra:1.4.8 rsense
```
That all.
-------
My current environment:
rbenv 1.1.0-2-g4f8925a
ruby 2.4.1 |
roboconf/roboconf-platform | 50216564 | Title: Support place holders for environment switch
Question:
username_0: Some properties, like credentials, must be kept secret.
They are also dependent of the environment. Developement, tests, pre-production, production, production site... can be various environments in which a same application can run.
Roboconf **must** provide a mechanism to handle these cases.
Answers:
username_0: Applications templates whose configuration relies on properties should put a **env.properties** file under the **descriptor** directory. When an application is created from this template, Roboconf will copy this file into Karaf's **etc/roboconf** directory.
Environment properties follow this convention:
* **etc/roboconf/global.properties** contains global properties.
* **etc/roboconf/app.\<app-name\>.properties** contains properties for a specific application.
They should be edited manually.
They will be injected as environment variables in plug-ins (on the agent's side). And they will be injected through string replacement in target properties. `${var_name}` is the notation to follow. By convention, such variables should have their name starting with `env_`.
There should a new parameter in the DM to customize the location of these parameters.
```properties
env.location = ${KARAF_ETC}/roboconf
```
username_0: OK, that's fine for storage.
However, we should identify all the cases we want to handle.
## Target configurations
These files specify identifiers for cloud infrastructures. They are used by the DM. So, what we want is to replace variables in these files. We do not need *real* environment variables here. It is text replacement from our properties files.
## Recipes
Let's consider the credentials to access a database. Let's assume they are exported as variables by a Roboconf component.
```
MySQL {
installer: script;
export: ip, port = 3306, user = a23df, pwd = <PASSWORD>;
}
```
With the previous post, it would look like...
```
MySQL {
installer: script;
export: ip, port = 3306, user = ${MYSQL_USER}, pwd = ${MYSQL_PASS};
}
```
... with the following properties (e.g. in **global.properties**).
```
MYSQL_USER = a23df
MYSQL_PASS = <PASSWORD>
```
This solution avoids to put credentials in the graph (sources) files.
But so that it works on the agent side, the agent needs this information too. The first option would be to update the variable values in the instances when we send them to the agent. Since the agent passes component variables as environment variables, no issue. And agents do not persist their model anywhere. So, these variables will only be persisted in one location.
Since this solution works and is easy to implement, no need to look for another one. :smiling_imp:
## Important Notice
There are [many articles](http://movingfast.io/articles/environment-variables-considered-harmful/) now to warn about using environment variables to store passwords. Some solution, like [Hashicorp Vault](https://github.com/hashicorp/vault) or [Ansible Vault](http://docs.ansible.com/ansible/playbooks_vault.html) (for Ansible) may be used. But this does not depend on Roboconf. It depends on the application architecture and developer works (to query sensitive information from these tools). So, these tools are alternative solution to environment variables and should be mentionned on our web site.
username_0: The mechanism can be used for both secrets (credentials) and environment settings (e.g. static information such as IP addresses).
username_0: I keep on thinking about it...
In fact, we should not replace properties in instances when we send the model to agents. Otherwise, everytime agents exchange information, these properties will go through the messaging server. We should limit their traffic.
Environment variables should be sent to agents only once, when their model is sent.
And recipe information should be replaced only when they are about to be executed.
username_0: Guess what... We will postpone this issue once again.
There is no more question, we just have to do it. But we would like to release at the beginning of January.
username_0: Properties for applications could also be written in a file, rather than being passed as environement variables. |
ambarket/botbattle | 63974506 | Title: Logout causing session folder issues
Question:
username_0: If you have a tab open (multiple for easier testing) then each tab will be under one session folder that is created and then each tab will have it's own folder as desired. Currently, if you log in this all remains. The issue is when you log out. Upon logging out it seems a new session is created, this makes a new session folder and when each tab is opened or refreshed it will cause new tab folder to be created under the new session folder all the while the old session folder and tabs is not deleted. We could solve this by just removing the old session folder and the subsequent data in it, but this will force the user to upload new bots. Another solution could be to not create a new session on log out since they don't get a new one if logged in or not. The session is just to track the browser not the user. So no two people will ever be able to be logged in on the same browser? Would this matter? Eitherway, this is bad for the system as it stands now.
Answers:
username_0: This also causes extra session objects to be made in testArenaInstances that are empty, but with no corresponding folder being created. It gets a little wacky.
username_0: Irrelevant now because we are going to change our approach.
Status: Issue closed
|
dariosalvi78/cordova-plugin-health | 184358756 | Title: Support queries for nutrition (dietary/food) data types in Google Fit
Question:
username_0: We are currently using the Telerik HealthKit plugin to get nutrition data types from HealthKit. The cordova-plugin-health plugin does not (yet) support this.
To get nutrition data from Google Fit, query for `DataType.TYPE_NUTRITION`. For each data point, nutrients can be retrieved via `datapoint.getValue(Field.FIELD_NUTRIENTS)` and then specifiec nutrients can be retrieved via `nutrients.getKeyValue(Field.NUTRIENT_TOTAL_CARBS)`.
Documentation on `Field.FIELD_NUTRIENTS`:
https://developers.google.com/android/reference/com/google/android/gms/fitness/data/Field.html#FIELD_NUTRIENTS
List of nutrients (constants prefixed with `NUTRIENT_`):
https://developers.google.com/android/reference/com/google/android/gms/fitness/data/Field#constant-summary
Tasks:
- [ ] include all nutrient types in `nutritiondatatypes`
- [ ] support retrieving samples for each nutrient type in `query()` with appropriate unit value for each type
- [ ] support aggregated queries for each nutrient type in `queryAggregated()`
Eventually, support should also be added for mapping and querying the equivalent data types in HealthKit (see [this list for reference](https://developer.apple.com/reference/healthkit/1627060-healthkit_constants/1664847-nutrition_identifiers?language=objc)). However, I don't personally need HealthKit support at the moment, so I will start with just the Google Fit side.
Status: Issue closed
Answers:
username_1: Thanks very much for working on this.
I have a question about Telerik's plugin:
I merged their repo with mine on Sept. 27 2016. It may already work with the nutrition quantity samples, or am I missing something?
In the future, it would be nice to have a "nutrition" datatype that returns all the known nutrients of a given sample (or aggregated) in an object (eg { fat.total: 50, cholesterol: 0.03 }), but that would probably be more difficult to implement with HealthKit.
username_2: Is this also working on the iOS part?
In that case I can prepare a release.
username_0: @username_1 I suspect that the nutrition samples could be queried via the separate HealthKit API (we are doing this using the Telerik plugin itself) but not via the health plugin API using the platform-agnostic type identifiers.
Telerik healthkit plugin:
```
window.plugins.healthkit.querySampleType(
{
'startDate' : new Date(new Date().getTime()-2*24*60*60*1000), // two days ago
'endDate' : new Date(), // now
'sampleType': 'HKQuantityTypeIdentifierDietarySugar',
'unit' : 'g'
},
onSuccess,
onError
);
```
cordova-plugin-health
```
window.navigator.health.query(
{
'startDate' : new Date(new Date().getTime()-2*24*60*60*1000), // two days ago
'endDate' : new Date(), // now
'dataType': 'nutrition.sugar',
'unit' : 'g'
},
onSuccess,
onError
);
```
So there are multiple outstanding issues:
1. translate from the nutrition data types, such as "nutrition.sugar" to the HealthKit sample types, such as "HKQuantityTypeIdentifierDietarySugar"
2. correlation queries, including support for a "nutrion" correlation
username_2: the health plugin uses the teleriks' plugin. AFAIK they are completely in sync, so there should be no missing features.
1: that's easy enough, I can do that
2: not sure why we would need this. As far as I have seen, the Android part only queries for specific substances
what about aggregated values, have you tried with Telerik's plugin?
username_0: Yes, we are using the Telerik plugin for aggregated values (by day) via `window.plugins.healthkit.querySampleTypeAggregated()`.
As for the correlation queries, Google Fit does not actually allow individual nutrients (substances) to be queried directly. It only supports queries for `DataType.TYPE_NUTRITION` and returns data points with all of the available nutrient fields. I implemented the query in the plugin to pull out the appropriate field (such as `Field.NUTRIENT_SUGAR`) and return just that one value. A correlation query should be fairly straight forward to implement in Google Fit, and can be done in a way that matches up closely to the `window.plugins.healthkit.queryCorrelationType()` from the Telerik plugin.
username_1: I have extended your work, now nutrition is supported with the same interface on both Fit and HK. I have also added the data type "nutrition" that computes a summary of all known nutrients.
Still to be done: store() of nutrients. |
umbraco/Umbraco-CMS | 482231237 | Title: When publishing content, Postsave throws an error
Question:
username_0: When publishing content PostSave stays pending a long time until it throws an 500 error. Reading the error log shows an error where it seems like it comes from the cache (see error below). When trying to reload the nucache I also get a timeout exception.
```
System.IO.IOException: An unexpected network error occurred.
at System.IO.__Error.WinIOError(Int32 errorCode, String maybeFullPath)
at System.IO.FileStream.WriteCore(Byte[] buffer, Int32 offset, Int32 count)
at CSharpTest.Net.IO.TransactedCompoundFile.fput(Int64 position, Byte[] bytes, Int32 offset, Int32 length)
at CSharpTest.Net.IO.TransactedCompoundFile.FileSection.Write(BlockRef block, FPut fput, Byte[] bytes, Int32 offset, Int32 length)
at CSharpTest.Net.IO.TransactedCompoundFile.Write(UInt32 handle, Byte[] bytes, Int32 offset, Int32 length)
at CSharpTest.Net.Storage.BTreeFileStoreV2.Update[T](IStorageHandle handleIn, ISerializer`1 serializer, T node)
at CSharpTest.Net.Collections.BPlusTree`2.StorageCache.OnItemRemoved(KeyValuePair`2 item)
at CSharpTest.Net.Collections.LurchTable`2.TryDequeue(Predicate`1 predicate, KeyValuePair`2& value)
at CSharpTest.Net.Collections.BPlusTree`2.StorageCache.Flush()
at CSharpTest.Net.Collections.BPlusTree`2.StorageCache.Commit()
at CSharpTest.Net.Collections.BPlusTree`2.CommitChanges(Boolean requiresLock)
at Umbraco.Web.PublishedCache.NuCache.ContentStore.Release(WriteLockInfo lockInfo, Boolean commit)
at Umbraco.Web.PublishedCache.NuCache.ContentStore.SetBranch(Int32 rootContentId, IEnumerable`1 kits)
at Umbraco.Web.PublishedCache.NuCache.PublishedSnapshotService.NotifyLocked(IEnumerable`1 payloads, Boolean& draftChanged, Boolean& publishedChanged)
at Umbraco.Web.PublishedCache.NuCache.PublishedSnapshotService.Notify(JsonPayload[] payloads, Boolean& draftChanged, Boolean& publishedChanged)
at Umbraco.Web.Cache.ContentCacheRefresher.Refresh(JsonPayload[] payloads)
at Umbraco.Core.Sync.ServerMessengerBase.Deliver[TPayload](ICacheRefresher refresher, TPayload[] payload)
at Umbraco.Web.Cache.DistributedCache.RefreshByPayload[TPayload](Guid refresherGuid, IEnumerable`1 payloads)
at Umbraco.Web.Cache.DistributedCacheExtensions.RefreshContentCache(DistributedCache dc, TreeChange`1[] changes)
at Umbraco.Core.Events.EventDefinition`2.RaiseEvent()
at Umbraco.Core.Events.QueuingEventDispatcher.ScopeExitCompleted()
at Umbraco.Core.Events.QueuingEventDispatcherBase.ScopeExit(Boolean completed)
at Umbraco.Core.Scoping.Scope.<>c__DisplayClass72_0.<RobustExit>b__1()
at Umbraco.Core.Scoping.Scope.TryFinally(Int32 index, Action[] actions)
at Umbraco.Core.Scoping.Scope.TryFinally(Int32 index, Action[] actions)
at Umbraco.Core.Scoping.Scope.RobustExit(Boolean completed, Boolean onException)
at Umbraco.Core.Scoping.Scope.Dispose()
at Umbraco.Core.Services.Implement.ContentService.SaveAndPublish(IContent content, String[] cultures, Int32 userId, Boolean raiseEvents)
at Umbraco.Web.Editors.ContentController.PublishInternal(ContentItemSave contentItem, Boolean& wasCancelled, String[]& successfulCultures)
at Umbraco.Web.Editors.ContentController.PostSaveInternal(ContentItemSave contentItem, Func`2 saveMethod)
at lambda_method(Closure , Object , Object[] )
at System.Web.Http.Controllers.ReflectedHttpActionDescriptor.ActionExecutor.<>c__DisplayClass6_2.<GetExecutor>b__2(Object instance, Object[] methodParameters)
at System.Web.Http.Controllers.ReflectedHttpActionDescriptor.ExecuteAsync(HttpControllerContext controllerContext, IDictionary`2 arguments, CancellationToken cancellationToken)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Web.Http.Controllers.ApiControllerActionInvoker.<InvokeActionAsyncCore>d__1.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Web.Http.Filters.ActionFilterAttribute.<CallOnActionExecutedAsync>d__6.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Web.Http.Filters.ActionFilterAttribute.<CallOnActionExecutedAsync>d__6.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Web.Http.Filters.ActionFilterAttribute.<ExecuteActionFilterAsyncCore>d__5.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Web.Http.Filters.ActionFilterAttribute.<CallOnActionExecutedAsync>d__6.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Web.Http.Filters.ActionFilterAttribute.<CallOnActionExecutedAsync>d__6.MoveNext()
[Truncated]
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Web.Http.Controllers.ExceptionFilterResult.<ExecuteAsync>d__6.MoveNext()
Timestamp | 2019-08-19T09:18:33.8071321+00:00
-- | --
Unhandled controller exception occurred for request '{RequestUrl}'
https://cms.floriade.com/umbraco/backoffice/UmbracoApi/Content/PostSave
Umbraco.Web.Editors.ContentController
5848
w3wp
29
2
LMW3SVC1811932842ROOT
RD281878EEAE2A
ERROR
42
49b29c52-fa66-454b-84d1-7b746d3ed387
```
Answers:
username_1: Hi @username_0 - that looks very ugly. First off, I would recommend trying to increase the request time-out in web.config for now and see if you can make it work that way.
After that.. we'll need a lot more information to be able to give any kind of answer, obviously (I hope) this doesn't happen for everybody so this seems limited to your implementation.
Please always mention the version you're using. And.. be prepared for the question: is this still an issue when you upgrade to the latest version? (does it..?)
Since this is a time-out maybe something about the amount of content you have could be important and more info about event handlers you've used.
For now, I will refer you to the forum where you could post a lot more detail to see if you can get to the bottom of the cause of this together with some community members. If you manage to find something that could be an actual bug in Umbraco then make sure to report back here so we can have a look, but first we need some kind of way to get closer to replicating the problem.
Status: Issue closed
|
dotnet/runtime | 1065374063 | Title: [API Proposal]: Add ZlibEncoder and ZlibDecoder to System.IO.Compression.
Question:
username_0: ### Background and motivation
Currently there is no non-stream based apis to zlib compression. As such I feel like an encoder / decoder implementation is needed similar to the Brotli implementation.
The brotli implementation also uses the encoder / decoder internally in the streams, which can help make the implementations of the zlib based streams (GZipStream, DeflateStream, and ZLibStream) better.
A single ZlibEncoder and ZlibDecoder that takes a class of values (ZlibOptions), where ZlibOptions also has subclasses named DeflateOptions, and GZipOptions where only the window bits are different.
Dependency issues that will need to be addressed before this:
https://github.com/dotnet/runtime/issues/42820
Implementing this issue *should* also resolve this one as well:
https://github.com/dotnet/runtime/issues/39327
I currently have a baseline implementation locally of this (except for the stream changes that would need to be done), and it *should* be ready by the time .NET 7 goes into an api freeze (until .NET 8's development starts).
### API Proposal
```C#
public enum ZlibOperationStatus
{
VersionError = -6,
BufError,
MemError,
DataError,
StreamError,
ErrorNo,
Ok,
Done = Ok,
StreamEnd,
NeedDictionary,
DestinationTooSmall,
OperationNotCompression,
OperationNotDecompression,
Disposed,
}
public struct ZlibEncoder : System.IDisposable
{
public ZlibEncoder(ZlibOptions options) { }
public bool IsDisposed { get { throw null; } }
public void Dispose() { throw null; }
#pragma warning disable CS3001 // Argument type is not CLS-compliant
public bool TryCompress(System.ReadOnlySpan<byte> source, System.Span<byte> dest, out uint adler32, out uint bytesWritten) { throw null; }
#pragma warning restore CS3001 // Argument type is not CLS-compliant
#pragma warning disable CS3001 // Argument type is not CLS-compliant
public ZlibOperationStatus Compress(System.ReadOnlySpan<byte> source, System.Span<byte> dest, out uint adler32, out uint bytesWritten) { throw null; }
#pragma warning restore CS3001 // Argument type is not CLS-compliant
}
public struct ZlibDecoder : System.IDisposable
{
public ZlibDecoder(ZlibOptions options) { }
public bool IsDisposed { get { throw null; } }
public void Dispose() { throw null; }
#pragma warning disable CS3001 // Argument type is not CLS-compliant
public bool TryDecompress(ReadOnlySpan<byte> source, Span<byte> dest, out uint adler32, out uint bytesWritten, out uint avail) { throw null; }
#pragma warning restore CS3001 // Argument type is not CLS-compliant
#pragma warning disable CS3001 // Argument type is not CLS-compliant
public ZlibOperationStatus Decompress(ReadOnlySpan<byte> source, Span<byte> dest, out uint adler32, out uint bytesWritten, out uint avail) { throw null; }
#pragma warning restore CS3001 // Argument type is not CLS-compliant
}
[Truncated]
using var zlibDecoder = new ZlibDecoder(new ZlibOptions());
result = zlibDecoder.TryDecompress(dest, decSource, out adler, out bytesWritten, out uint avail);
if (!result)
{
Console.WriteLine("Decompression failed.");
}
else
{
Console.WriteLine($"Adler-32: {adler}, Bytes Decompressed: {bytesWritten}, Available: {avail}.");
}
```
### Alternative Designs
I wanted to use the System.Buffers.OperationStatus enum, but I felt it lacked enough members to denote the zlib specific status codes (where they would be used for zlib, deflate, and gzip compression / decompression). Likewise adler32's and the total_out member from zlib are represented as unsigned, I do not know if I should have the encoder / decoder output signed, or leave them unsigned and keepthe CLS-Compliancy issues suppressed from it.
### Risks
Minimal
Answers:
username_0: I have modified the proposal somewhat to add a few things I missed to ensure that TryDecompress will never throw or get exceptions that it would need to catch.
username_1: @username_0 thank you so much for the detailed proposal.
This proposal would partially address both https://github.com/dotnet/runtime/issues/42820 and https://github.com/dotnet/runtime/issues/39327 . I don't think this PR depends on the first one, we can consider that one the _"uber issue"_, because it was attempting to address the same you're addressing, but for all compression stream classes.
I see you added your own ZLib-specific operation status enums. Contrary to `BrotliDecoder`/`BrotliEncoder`, which reused `System.Buffers.OperationStatus`: https://github.com/dotnet/runtime/blob/57bfe474518ab5b7cfe6bf7424a79ce3af9d6657/src/libraries/System.Private.CoreLib/src/System/Buffers/OperationStatus.cs#L10
Some feedback on the new enum values:
- `Ok` and `Done` seem redundant. We should simply keep `Done`, just like `OperationStatus.Done`.
- `DestinationTooSmall` seems fine. It's aligned to `OperationStatus.DestinationTooSmall` already.
- I'm not sure we need `OperationNotCompression`, `OperationNotDecompression`, `Disposed`, `InitError`, `ErrorNo`. Shouldn't those throw instead?
- What does `VersionError` mean?
- `DataError`, `StreamEnd`, `StreamError`, `BufError`, `MemError`, `NeedDictionary`, could be aligned to `OperationStatus.InvalidData`.
- I think `DataError` could be `OperationStatus.NeedMoreData`.
What's preventing us from just reusing `OperationStatus`?
username_0: We would need these for ``TryDecompress`` / ``TryCompress`` where it would be expansive to throw and then "catch" the exception within the ``Try`` version.
VersionError is an zlib error code for when the expected zlib version and the runtime zlib version does not match (on the major version I think).
According to the Manual DataError is for when "inflate detects an error in the zlib compressed data format, which means that either the data is not a zlib stream to begin with, or that the data was corrupted somewhere along the way since it was compressed."
Also in that manual StreamEnd is returned when inflate is at the end of the zlib/deflate/gzip stream.
https://zlib.net/zlib_how.html is where I get some of the details as well too.
username_0: I think this should be ready to review.
username_0: cc: @dotnet/area-system-io-compression I think this is ready for review.
username_0: Anything else stalling this? |
websockets-rs/rust-websocket | 393343279 | Title: Imports doesn't work on edition=2018
Question:
username_0: The `async-server` example cannot compile with editon 2018 because it has `async` in path.
Code:
`use websocket::async::Server;`
Error:
```
error: expected identifier, found reserved keyword `async`
--> src/main.rs:16:16
|
16 | use websocket::async::Server;
| ^^^^^ expected identifier, found reserved keyword
```
Answers:
username_1: [`r#async`](https://github.com/rust-lang/rfcs/blob/master/text/2151-raw-identifiers.md)?
Status: Issue closed
username_0: Works thanks :)
username_2: While the r# syntax does work, that doesn't really seem like a solution
username_1: What do you suggest instead?
Migration to new hyper 0.12 may be a rather breaking change, so it may be a good point to rename `async` to something else.
username_1: Depending on acceptance of usage of r#raw r#identifier syntax for things other than compatibility.
```rust
let fk = 0.5;
let fm = 0.1;
let r#fn = (fk + fm)/2.0;
let Fn = 1.0 / r#fn;
``` |
jlippold/tweakCompatible | 359699384 | Title: `Lithium Ion` working on iOS 11.3.1
Question:
username_0: ```
{
"packageId": "com.tweakingsdev.lithiumion",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "com.tweakingsdev.lithiumion",
"deviceId": "iPhone8,1",
"url": "http://cydia.saurik.com/package/com.tweakingsdev.lithiumion/",
"iOSVersion": "11.3.1",
"packageVersionIndexed": true,
"packageName": "Lithium Ion",
"category": "Tweaks",
"repository": "BigBoss",
"name": "Lithium Ion",
"installed": "1.3",
"packageIndexed": true,
"packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 86% with 25 working reports.",
"id": "com.tweakingsdev.lithiumion",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.1.0",
"shortDescription": "Ultimate Battery Customization",
"latest": "1.3",
"author": "creativeBeing",
"packageStatus": "Working"
},
"base64": "<KEY>
"chosenStatus": "working",
"notes": "Flawless"
}
``` |
DDMAL/xml-parse-glyph-img | 446829630 | Title: Check for overlap in the bounding box coordinates for each neume (component)
Question:
username_0: Based on my current implementation, symbols such as clefs are always separated into two separate bounding boxes based on my erosion methodology.
I have the coordinates available for the bounding box of each neume (component), but I would like to automatically only save one image in the output for those that significantly overlap in regards to their x-coordinates.<issue_closed>
Status: Issue closed |
metoppv/improver | 238874805 | Title: Documentation of the demonstration Improver/Ver suite.
Question:
username_0: A component of Epic #141
As a user of the Improver/Ver suite I need documentation that details the operational use of the suite, making clear available configuration options. It would also be desirable to have associated scientific documentation/links to papers that detail the verification options that are available.
Work with <NAME> to add any scientific content.
Acceptance criteria:
* Document configurable options of Improver/Ver suite.
* Detail features and provide usage examples.
* Add to Confluence in location to be agreed with Jonathan Flowerdew.<issue_closed>
Status: Issue closed |
n8henrie/fauxmo | 557641579 | Title: asyncio:1608 ERROR Fatal error: protocol.data_received() call failed.
Question:
username_0: get every 10 secs these entries in syslog:
rpi4 with Linux pump 4.19.93-v7l+ #1290 SMP Fri Jan 10 16:45:11 GMT 2020 armv7l GNU/Linux
Jan 30 17:01:37 pump fauxmo[5337]: 2020-01-30 17:01:37 asyncio:1608 ERROR Fatal error: protocol.data_received() call failed.
Jan 30 17:01:37 pump fauxmo[5337]: protocol:
Jan 30 17:01:37 pump fauxmo[5337]: transport: <_SelectorSocketTransport fd=18 read=polling write=>
Jan 30 17:01:37 pump fauxmo[5337]: Traceback (most recent call last):
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/urllib/request.py", line 1317, in do_open
Jan 30 17:01:37 pump fauxmo[5337]: encode_chunked=req.has_header('Transfer-encoding'))
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/http/client.py", line 1229, in request
Jan 30 17:01:37 pump fauxmo[5337]: self._send_request(method, url, body, headers, encode_chunked)
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/http/client.py", line 1275, in _send_request
Jan 30 17:01:37 pump fauxmo[5337]: self.endheaders(body, encode_chunked=encode_chunked)
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/http/client.py", line 1224, in endheaders
Jan 30 17:01:37 pump fauxmo[5337]: self._send_output(message_body, encode_chunked=encode_chunked)
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/http/client.py", line 1016, in _send_output
Jan 30 17:01:37 pump fauxmo[5337]: self.send(msg)
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/http/client.py", line 956, in send
Jan 30 17:01:37 pump fauxmo[5337]: self.connect()
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/http/client.py", line 928, in connect
Jan 30 17:01:37 pump fauxmo[5337]: (self.host,self.port), self.timeout, self.source_address)
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/socket.py", line 727, in create_connection
Jan 30 17:01:37 pump fauxmo[5337]: raise err
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/socket.py", line 716, in create_connection
Jan 30 17:01:37 pump fauxmo[5337]: sock.connect(sa)
Jan 30 17:01:37 pump fauxmo[5337]: ConnectionRefusedError: [Errno 111] Connection refused
Jan 30 17:01:37 pump fauxmo[5337]: During handling of the above exception, another exception occurred:
Jan 30 17:01:37 pump fauxmo[5337]: Traceback (most recent call last):
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/asyncio/selector_events.py", line 813, in _read_ready__data_received
Jan 30 17:01:37 pump fauxmo[5337]: self._protocol.data_received(data)
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/site-packages/fauxmo/protocols.py", line 64, in data_received
Jan 30 17:01:37 pump fauxmo[5337]: self.handle_action(msg)
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/site-packages/fauxmo/protocols.py", line 143, in handle_action
Jan 30 17:01:37 pump fauxmo[5337]: state = self.plugin.get_state()
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/site-packages/fauxmo/plugins/simplehttpplugin.py", line 179, in get_state
Jan 30 17:01:37 pump fauxmo[5337]: with self.urlopen(req) as resp:
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/urllib/request.py", line 222, in urlopen
Jan 30 17:01:37 pump fauxmo[5337]: return opener.open(url, data, timeout)
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/urllib/request.py", line 525, in open
Jan 30 17:01:37 pump fauxmo[5337]: response = self._open(req, data)
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/urllib/request.py", line 543, in _open
Jan 30 17:01:37 pump fauxmo[5337]: '_open', req)
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/urllib/request.py", line 503, in _call_chain
Jan 30 17:01:37 pump fauxmo[5337]: result = func(*args)
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/urllib/request.py", line 1345, in http_open
Jan 30 17:01:37 pump fauxmo[5337]: return self.do_open(http.client.HTTPConnection, req)
Jan 30 17:01:37 pump fauxmo[5337]: File "/opt/pyenv/versions/3.7.3/lib/python3.7/urllib/request.py", line 1319, in do_open
Jan 30 17:01:37 pump fauxmo[5337]: raise URLError(err)
Jan 30 17:01:37 pump fauxmo[5337]: urllib.error.URLError:
Answers:
username_1: I can't help much based on the information provided. I'm guessing your issue is based on a problem in your SimpleHTTPPlugin configuration, based on this: `Jan 30 17:01:37 pump fauxmo[5337]: ConnectionRefusedError: [Errno 111] Connection refused`
Please edit your issue to include the information requested in the issue template and I'll be happy to re-open.
Status: Issue closed
|
microsoft/Azure_Kinect_ROS_Driver | 532485164 | Title: image_proc bug for rectified depth images
Question:
username_0: Hi, I'm using ROS Melodic on Ubuntu 18.04, running a Kinect DK.
I am encountering this very strange bug that I am able to reproduce. On commit 60789f0, when I run slam_RTABMap.launch, the rectified depth cloud that I am able to get is normal. I have attached a picture of the normal looking RTABMap cloud here. As you can see, it is relatively clean.
However, when I use the latest version(commit d761221) of the Azure Kinect ROS Driver, I am getting a very weird rectified depth image.
I am wondering if this is an expected outcome, or which commit caused this error? As there are a lot of commits that happened between the two, I was hoping that someone here could guide me or tell me if this is normal.
Answers:
username_1: That looks like a linear interpolation bug. `image_proc/rectify` seems to be using `interpolation=1` (linear interpolation) when it should be using `interpolation=0` (nearest neighbor).
Try changing this chunk in the ROS Launch file:
```
<!-- Spawn an image_proc/rectify nodelet to rectify the depth image -->
<node pkg="nodelet" type="nodelet" name="rectify_depth"
args="load image_proc/rectify manager --no-bond"
respawn="true">
<remap from="image_mono" to="depth/image_raw" />
<remap from="image_rect" to="depth/image_rect" />
</node>
```
to this:
```
<!-- Spawn an image_proc/rectify nodelet to rectify the depth image -->
<node pkg="nodelet" type="nodelet" name="rectify_depth"
args="load image_proc/rectify manager --no-bond"
respawn="true">
<remap from="image_mono" to="depth/image_raw" />
<remap from="image_rect" to="depth/image_rect" />
<param name="interpolation" value="0" />
</node>
```
and see if that makes things better.
username_0: Yep, added interpolation=1 in the launch file and it seems to work fine! Thanks for the help. Just wondering what between the two commits caused this?
Thanks so much for the help!
username_2: Hello everyone I am using rectify_test.launch and I have same result in pointcloud. Some ghost appear in final pointcloud and there are more when sun light is present. @username_0 interpolation = 0. helped you to decrease noisy points?



username_0: Hi @username_2, your issue seems slightly different from mine. My random dots seem to be concentrated from my camera. I'm not sure about yours? They seem to be more distributed. I have noticed that the Kinect DK does seem to have a few random points past the stated range though.
Good luck, anyhow! |
games647/ChangeSkin | 309839816 | Title: Bukkit end not saving preferences
Question:
username_0: [//]: # (Lines in this format are considered as comments and will not be displayed.)
[//]: #
[//]: # (Before reporting an issue make sure you are running the latest build and checked for duplicate issues!)
### What behaviour is observed:
Setting a skin with /setskin <name> while on the Spigot server will only set the skin if instant skin is enabled, and will NOT save the preferences. Once you logout and rejoin the skin will be gone.
### What behaviour is expected:
For /setskin to change skins and save the desired skin
### Steps/models to reproduce:
Join the Bungeecord server and be placed into one of the Spigot servers containing ChangeSkin. Then execute /setskin <name>. If you have instant skin enabled, it'll set a skin temporally. Rejoining the Bungee instant gives you the previous skin you had.
### Plugin list:
ChangeSkin
### Environment description
Bungeecord, and only one of the spigot servers in the network use this plugin. The server version is 1.8, and the Bungeecord version is always within a week of the latest bungeecord release. The server is hosted on a dedicated machine, with a local MySQL instance. Server is also in Offline mode
### Plugin version or build number (don't write latest or the build number with #):
100
### Configuration:
https://pastebin.com/mHu6vw52 |
Blackmill/book-club | 832603123 | Title: The making of a manager: Mar 23: Ch 5
Question:
username_0: Current book is The Making of a Manager by <NAME> ([Readings](https://www.readings.com.au/products/25282474/the-making-of-a-manager-what-to-do-when-everyone-looks-to-you) or [Amazon](https://www.amazon.com/Making-Manager-What-Everyone-Looks-ebook/dp/B079WNPRL2/))
Aiming to read/discuss:
- Chapter 5: managing yourself
MC: @chaplonglau
Notes: John
See you 12 pm Tuesday, March 23rd @ https://whereby.com/blackmill
Ping <EMAIL> if you want a calendar invite and access to the low-volume Slack beforehand.
Answers:
username_1: Chapter 5 Managing yourself
Executive coach
• <NAME> - Hired executive coach - remarkable little internet presence - palette 17,
• made 20 page report on her - tldr, if you don't have a good handle on yourself, you don't have a good handle on how to best support your team
• Picture yourself 80, sitting on a beach and looking back on your life. What did I want to remember?
• Julie starts the chapter with a story on hiring an executive coach that gave a 20-page daa see ei on her strengths, weaknesses, management style and everything in between. What are everyone's thoughts and experiences on executive coaching?
• (Blackmill costs ~$2000/ 6 1-hour sessions) (Stacy McCarthy has a firm called Palette17 which has a remarkably small internet footprint) On their site, they state that [We] help leaders identify their core strengths and build a solid foundation and "centre" from which they can lead more effectively. Tell me about you, your past the road you have take to get here, the future.
• How do you work perceived strengths and weaknesses.
Discussion
Elle : Research states that you will be a better leader with an executive coach. People who are new to management
Adam: path would be shorter with a coach
Chap: Stacy Macara leaders identify core strengths
Lachlan: never had a formal mentor, get feedback from the people your advise affects, we are all approaching the text from different angles. There is value in formalising the mentoring relationship.
John: gain an informal mentor for a brood range of experience technical skills and presentations.
Imposter syndrome
Julie then talks about imposter syndrome. She stresses every manager feels like an imposter sometimes. It's totally normal and you get it because
- you're often looked to for answers ( and that's pressure)
- you often get put in the position of doing thing you haven't done before like firing someone
How long did it take for everyone here (if ever) to stop feeling imposter syndrome?
What are your experiences with it?
• Every manager feels like an imposter sometimes
• Checking your emails 3 times
• Teetering on the edge of a sheer cliff the world watching you waiting to fall
• The early days of being a boss, disorientation and over whelming.
• Why does this happen? You are looked to for answers. You are the first person called on.
• Management isn’t an innate skill
Discussion
Everyone: agrees that you never overcome imposter syndrome
John: Imposter syndrome leads to checking emails 3 times and not taking risks.
Lachlan: fixed and growth mindsets. Growth mindsets makes imposter syndrome easier to deal with. Writing draft emails in a text doc. Unneeded anxiety and tenancy to do things and not taking risks. concerns about trying and failing.
Adam: we spend more time on these books, does the sample size of your own experience. Re-evaluate the assumptions. Difficult to manage. Can confirm your ideas.
Elle: has been working on a draft bog post on piggies and imposter syndrome, psychological safety. Praising team mates rather that chastising them. Don’t focus on what needs to be fixed.
How do you get out of the pit
• Celebrate the wins
• Visualise
Chap: on the scientific method for determining the power of visualisation.
Adam: “what I tell myself is” As a way to get to the point. Use internal monolog to talk yourself through the problem. Pro Mindfulness but getting better with visualisations is BS. CrossFit training improves when preparing brain to get past the pain barrier.
Lachlan: Don’t beat yourself up for feeling bad. Talking over the problem with a college or partner.
Elle: Depression and Anorexia misscalabration with reality. Growth v Fixed as Fear, five good things that have happed today.
John: The benefit of self visualisation comes from taking away the negative thoughts.
Lachlan: Lean in circles(Cheryl Sanburg) what would that be like at Facebook?
Chap: Leaning in is not a solo sport. In a world where women still face bias and other barriers at work, Circles are a safe space to share your struggles, give and get advice, and celebrate each other’s wins. Whether you need help navigating your new normal at work or are looking for support from women who understand what you're going through, your Circle will give you the boost we all need to take on 2021.
Get to brutal honesty with yourself
• Who are you as a manager
• Strengths and weaknesses directly effect how you manage
• Strategic prowess, multitasking
• Facets of our personality
• What works best with what you have
Top leaders come from different moulds
• Extrovert <NAME>
• Introvert <NAME>
• Demanding <NAME>
• Remind you of a favourite relative Mother Theresa
• Leave a room breathless with their vision <NAME>
[Truncated]
• If you are not learning from your manager what can you do to fix this.
• Make a mentor out of everyone
Who do you run meetings
• Running across multiple locations
Reflection
• 6 month evaluations
• If you have the option to take formal training you should take it
• Invest in your personal learning and growth
Discussion
Chap: Do you take the time to reflect on goals? What's your method? Do you have a similar timeframe? Personally, every six months, or you know major life episode, I sit down for a couple hours and fill out [https://yearcompass.com/](https://yearcompass.com/). It's a collection of questions that I found very helpful for reflecting and planning. I encourage everyone to check it out.
Elle: Personal Kanban with a weekly retrospective with herself.
Lachlan: Daily list of tasks. Backlog for parked tasks. Asking for feedback making a journal. People don’t like looking into their skills.
Adam: I work with some people who don’t take the time to be overworked.
Lachlan: some people only focus on the code. Other processes need to be factored in.
Chap: personal development and the Metrix for measuring how you achieve your goals.
John: planning for skills gaps and factoring in time for PD.
Next time: chapter 6 Running Meetings and 7 Hiring Well.
Lachlan to run Adam taking notes
Status: Issue closed
|
jlippold/tweakCompatible | 424143165 | Title: `Dimension` working on iOS 12.1.1
Question:
username_0: ```
{
"packageId": "com.yourepo.dizzy.dimension",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "com.yourepo.dizzy.dimension",
"deviceId": "iPhone8,1",
"url": "http://cydia.saurik.com/package/com.yourepo.dizzy.dimension/",
"iOSVersion": "12.1.1",
"packageVersionIndexed": false,
"packageName": "Dimension",
"category": "Themes",
"repository": "Dizzy Source",
"name": "Dimension",
"installed": "1.1.3",
"packageIndexed": false,
"packageStatusExplaination": "This tweak has not been reviewed. Please submit a review if you choose to install.",
"id": "com.yourepo.dizzy.dimension",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.1.5",
"shortDescription": "Dimension-A-Cool-Theme-for-anemone",
"latest": "1.1.3",
"author": "Dizzy",
"packageStatus": "Unknown"
},
"base64": "<KEY>
"chosenStatus": "working",
"notes": ""
}
``` |
sile/hb | 456037117 | Title: Performance issue when a response body of a GET request is large(> MiB)
Question:
username_0: ## Overview
`hb run` seems to consume most of execution time to decode a response body when a response body of a GET request is large.
The following line may be a root cause of this issue:
https://github.com/username_1/hb/blob/10b2a8a41a8df07ef72236d7268defecb118e5e9/src/run.rs#L166
## How to reproduce
Suppose that you have a frugalos cluster, run the following commands:
```console
$ jo -a `for n in $(seq 0 1000); do jo method=PUT content=3145728 url=http://localhost:3000/v1/buckets/test/objects/$n; done` > /tmp/req1.json
$ jo -a `for n in $(seq 0 1000); do jo method=GET url=http://localhost:3000/v1/buckets/test/objects/$n; done` > /tmp/req2.json
$ hb run -i /tmp/req1.json | hb summary
$ hb run -i /tmp/req2.json | hb summary
```
`hb run -i /tmp/req2.json | hb summary` will take very long time.
## Reference
The performance difference is like below:
```console
$ time ~/.cargo/bin/hb run -i ./req.json | hb summary
{
"count": {
"total": 1001,
"ok": 1001,
"error": 0
},
"status": {
"200": 1000,
"404": 1
},
"duration": 8.034492451,
"rps": 124.58783253638033,
"latency": {
"min": 0.05501402,
"median": 0.245011632,
"mean": 0.2533950855684318,
"max": 0.563494181,
"var": 0.004940090512700124,
"sd": 0.07028577745675241
}
}
~/.cargo/bin/hb run -i ./req.json 29.36s user 2.95s system 48% cpu 1:06.85 total
hb summary 57.26s user 1.32s system 87% cpu 1:06.87 total
$ time /usr/local/bin/hb run -i ./req.json | hb summary
{
"count": {
"total": 1001,
"ok": 1001,
"error": 0
},
"status": {
"200": 1000,
[Truncated]
@@ -1,3 +1,5 @@
+# This file is automatically @generated by Cargo.
+# It is not intended for manual editing.
[[package]]
name = "ansi_term"
version = "0.11.0"
diff --git a/src/run.rs b/src/run.rs
index eb861fa..facff6d 100644
--- a/src/run.rs
+++ b/src/run.rs
@@ -163,7 +163,7 @@ impl Future for RunRequest {
let response = Response {
status: response.status_code().as_u16(),
content_length: response.body().len() as u64,
- content: String::from_utf8(response.into_body()).ok(),
+ content: None,
};
Ok(Async::Ready(response))
} else {
```
Answers:
username_1: Thank you for reporting this problem.
Because the `content` field of `Response` isn't so useful, so I think that removing the field is sufficient as a solution of this problem.
What do you think? > @username_0
username_0: I agree with you.
Status: Issue closed
username_1: Released: https://crates.io/crates/hb/0.0.13 |
Azure/azure-sdk-for-js | 777038743 | Title: RestError: AKV10032: Invalid issuer
Question:
username_0: - **Package Name**: @azure/keyvault-secrets, @azure/identity
- **Package Version**: ^4.1.0, ^1.2.0
- **Operating system**: windows 10
- [ ] **nodejs**
- **version**: 14.15.3
- [ ] **browser**
- **name/version**:
- [ ] **typescript**
- **version**: 3.9.6
- Is the bug related to **documentation** in
- [ ] README.md
- [ ] source code documentation
- [ ] SDK API docs on https://docs.microsoft.com
**Describe the bug**
similar to the bug: https://github.com/Azure/azure-sdk-for-js/issues/12539
I was following the docs how to reading key-valut in NodeJs+TS
https://docs.microsoft.com/en-gb/javascript/api/overview/azure/keyvault-keys-readme?view=azure-node-latest
**To Reproduce**
Steps to reproduce the behavior:
1. follow - https://docs.microsoft.com/en-gb/javascript/api/overview/azure/keyvault-keys-readme?view=azure-node-latest - 'Getting a key' section
2. run
```
const { DefaultAzureCredential } = require("@azure/identity");
const { KeyClient } = require("@azure/keyvault-keys");
const credential = new DefaultAzureCredential();
const vaultName = "<YOUR KEYVAULT NAME>";
const url = `https://${vaultName}.vault.azure.net`;
const client = new KeyClient(url, credential);
const keyName = "MyKeyName";
async function main() {
const latestKey = await client.getKey(keyName);
console.log(`Latest version of the key ${keyName}: `, latestKey);
const specificKey = await client.getKey(keyName, { version: latestKey.properties.version! });
console.log(`The key ${keyName} at the version ${latestKey.properties.version!}: `, specificKey);
}
main();
```
I did follow the hints from bug https://github.com/Azure/azure-sdk-for-js/issues/12539
1. this works for me fine - I'm able to get secrets:
```
az login
az account set --subscription SUBSCRIPTION_ID
az keyvault secret list --vault-name KEY_VAULT_NAME
```
2. I did change to use the `ClientSecretCredential` but the same result:
```
(node:13300) UnhandledPromiseRejectionWarning: RestError: AKV10032: Invalid issuer. Expected one of https://sts.windows.net/72f988bf..., https://sts.windows.net/f8cdef31-a31e-4b4a-93e4-5f571e91255a/, https://sts.windows.net/e2d54eb5..., https://sts.windows.net/33e01921-4d64-4f8c-a055-5bdaffd5e33d/, https://sts.windows.net/975f013f..., found https://sts.windows.net/e46bc88e....
```
I have noticed that the found issuer match the tenantId I have provided but none of the expected issuers is in my subscription.
**Expected behavior**
Login and display KV details
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Additional context**
Add any other context about the problem here.
Answers:
username_1: When you say that, what do you mean?
The error shows a list of issuers, do you see your tenant in that list of issuers?
username_0: Hi, this was just an observation - I have no clue what is going on behind.
What I want to say that in my account I do have few tenants but none of those tenants match any of the listed issuers - no idea whether this can be related.
anyway, any idea what can be wrong, there? or whether I should try anything?
thx a lot,
Roman
username_1: @username_0
Hello again! Let's see how we can dig through and solve this.
`DefaultAzureCredential` may be using an authentication method related to a different account or subscription based on the environment where you're executing this code. To narrow down the possible issues, let's use a different credential!
Let's try with `ClientSecretCredential`, where you can pass a tenant ID (subscription), a client ID and a client secret.
In case it helps, here's how to retrieve them:
- Follow [Documentation to register a new application](https://docs.microsoft.com/azure/active-directory/develop/quickstart-register-app) in the Azure Active Directory (in the Azure portal).
- Note down the `CLIENT_ID` and `TENANT_ID`.
- In the "Certificates & Secrets" tab, create a secret and note that down.
Please also consider trying the `InteractiveBrowserCredential` or the `DeviceCodeCredential`.
Our goal is to see if by specifying the account more directly, we are able to get the access you're expecting. Once we have a working approach, we can backtrack into what the original issue might be.
Thank you for your time! I'll be monitoring this issue to answer as soon as possible.
username_0: Hi,
I did give a try the device code credential l and have this error message - but still do not get what I do wrong..
keyvault details:

the assigned application (permissions are get and list):

application details:

source code:

error message:

I guess do miss something - any ideas?
username_1: @username_0
Thank you so much for your answer! I'll be back as soon as possible. Based on all you've shared so far, I'll be trying different things on my side to build up an answer.
username_1: @username_0
The Device Code Credential requires special configuration on your Azure Active Directory application. This should help:
- Go to the Azure Active Directory page in the Azure portal.
- Go to "App Registrations" on the left, then find your app on the list that will load in the main area of the screen.
- There, go to your application. This will open a page that will show an "Authentication" section on the left pane. Similar to the following screenshot:

Click on "Add a Platform", and something similar to the screenshot below will appear:

Select all of the checkboxes available, then click on "Configure". Once the portal says the configuration has finished successfully (it shouldn't last more than a minute), please run your code again! Let me know how it looks after that.
Thank you for your time.
username_0: Ok, I tried but this does not help :(
I have modified the app as you mentioned, also tried to check the `Allow public client flows` checkbox

but have still the same error message:

what is strange about that error message is the KeyVault address - it should be the general address or there should be the address of my KeyVault in that error message? Is the KeyVault address passed properly?
```
const url = 'https://kvgraphqldevwe.vault.azure.net/';
const credential = new DeviceCodeCredential(
'e46bc88e-......',
'55f130f1-...',
);
const client = new SecretClient(url, credential);
export async function getSecret(secretName: string): Promise<string> {
const secret = await client.getSecret('ApplicationInsights--InstrumentationKey');
return secret.value;
}
```
I have tried now to get rid of all variables and pass all thru string constants.
username_1: @username_0 hello hello!
The vault URL `https://kvgraphqldevwe.vault.azure.net/` is correctly passed on that snippet.
Would you mind trying with `ClientSecretCredential`? Sending a tenant ID, the client ID of your AAD application (not the object ID), and the secret of your AAD application?
In case it helps:
- Follow [Documentation to register a new application](https://docs.microsoft.com/azure/active-directory/develop/quickstart-register-app) in the Azure Active Directory (in the Azure portal).
- Note down the `CLIENT_ID` and `TENANT_ID`.
- In the "Certificates & Secrets" tab, create a secret and note that down.
The idea is to have a working path, so that we can debug more clearly what is going wrong with the DeviceCodeCredential.
Thank you for your time!
username_0: Hi,
I have uncommented the code I did use the 1st time and all works fine. Even the device credentials approach works fine now - I have no clue what has happened and what is different now :(
So thx for your time and I'm going to investigate what has been changed
Regards,
R.
Status: Issue closed
username_1: @username_0 glad to know things are working now! Please let us know if anything else happens in the future. Have a good time!
username_2: @username_1 I tried all the above steps and I am still getting same error. Also, I have three subscriptions in which I am facing problem with one subscription and it is working fine in other 2 subscription by using 'DefaultAzureCredential'. I tried using 'DeviceCodeCredential', 'ClientSecretCredential' by following the steps you mentioned as well. I am not sure what is the problem. Please let me know if you need more details.
username_1: @username_2 hello, I’m Daniel! I’ll be doing my best to help you.
Do you mind describing how is your Azure App Registration configured? What permissions does it have? how have you assigned permissions to those app registrations though your key vault? |
skylot/jadx | 626408494 | Title: [core] Code restructure failed for a small method
Question:
username_0: Hi, there is a small method with error. Class `net.sqlcipher.AbstractWindowedCursor`
```java
/* JADX WARNING: Code restructure failed: missing block: B:10:0x0016, code lost:
return r3;
*/
@Override // net.sqlcipher.AbstractCursor
public boolean isNull(int i) {
checkPosition();
synchronized (this.mUpdatedRows) {
if (!isFieldUpdated(i)) {
return this.mWindow.isNull(this.mPos, i);
}
boolean z = getUpdatedField(i) == null;
}
}
```
APK: https://drive.google.com/file/d/1Umqt8wXZTm4sBqsvzjEFq-97qC-i8l-b/view?usp=sharing
Answers:
username_0: And without `synchronized` block, class `net.sqlcipher.database.SQLiteCompiledSql`
```java
/* JADX WARNING: Code restructure failed: missing block: B:11:0x002d, code lost:
return true;
*/
public synchronized boolean acquire() {
if (this.mInUse) {
return false;
}
this.mInUse = true;
if (net.sqlcipher.database.SQLiteDebug.DEBUG_ACTIVE_CURSOR_FINALIZATION) {
android.util.Log.v("SQLiteCompiledSql", "Acquired DbObj (id#" + this.nStatement + ") from DB cache");
}
}
```
username_0: @username_1
Can you look at this one, please?
username_1: @username_0 fixed!
Status: Issue closed
username_0: @username_1
Thanks! No bugs created after this fix! |
senecajs/seneca | 139530843 | Title: EventEmitter memory leak detected seems to stem from seneca
Question:
username_0: ```
api_1 | [ Nodemon ][ Error ] (node) warning: possible EventEmitter memory leak detected. 11 clear listeners added. Use emitter.setMaxListeners() to increase limit.
api_1 |
api_1 | [ Nodemon ][ Error ] Trace
api_1 | at GateExecutor.addListener (events.js:252:17)
api_1 | at GateExecutor.once (events.js:278:8)
api_1 | at Seneca.api_ready [as ready] (/var/app/current/api/node_modules/seneca/seneca.js:923:30)
api_1 | at meta.export (/var/app/current/api/node_modules/seneca-web/web.js:661:14)
api_1 | at Layer.handle [as handle_request] (/var/app/current/api/node_modules/express/lib/router/layer.js:95:5)
api_1 | at trim_prefix (/var/app/current/api/node_modules/express/lib/router/index.js:312:13)
api_1 | at /var/app/current/api/node_modules/express/lib/router/index.js:280:7
api_1 | at Function.process_params (/var/app/current/api/node_modules/express/lib/router/index.js:330:12)
api_1 | at next (/var/app/current/api/node_modules/express/lib/router/index.js:271:10)
api_1 | at Object.exports.log_api_request (/var/app/current/api/common/utils/api_response_handler.js:510:27)
api_1 | at /var/app/current/api/app.js:105:23
api_1 | at Layer.handle [as handle_request] (/var/app/current/api/node_modules/express/lib/router/layer.js:95:5)
api_1 | at trim_prefix (/var/app/current/api/node_modules/express/lib/router/index.js:312:13)
api_1 | at /var/app/current/api/node_modules/express/lib/router/index.js:280:7
api_1 | at Function.process_params (/var/app/current/api/node_modules/express/lib/router/index.js:330:12)
api_1 | at next (/var/app/current/api/node_modules/express/lib/router/index.js:271:10)
api_1 |
```
Status: Issue closed
Answers:
username_0: ```
api_1 | [ Nodemon ][ Error ] (node) warning: possible EventEmitter memory leak detected. 11 clear listeners added. Use emitter.setMaxListeners() to increase limit.
api_1 |
api_1 | [ Nodemon ][ Error ] Trace
api_1 | at GateExecutor.addListener (events.js:252:17)
api_1 | at GateExecutor.once (events.js:278:8)
api_1 | at Seneca.api_ready [as ready] (/var/app/current/api/node_modules/seneca/seneca.js:923:30)
api_1 | at meta.export (/var/app/current/api/node_modules/seneca-web/web.js:661:14)
api_1 | at Layer.handle [as handle_request] (/var/app/current/api/node_modules/express/lib/router/layer.js:95:5)
api_1 | at trim_prefix (/var/app/current/api/node_modules/express/lib/router/index.js:312:13)
api_1 | at /var/app/current/api/node_modules/express/lib/router/index.js:280:7
api_1 | at Function.process_params (/var/app/current/api/node_modules/express/lib/router/index.js:330:12)
api_1 | at next (/var/app/current/api/node_modules/express/lib/router/index.js:271:10)
api_1 | at Object.exports.log_api_request (/var/app/current/api/common/utils/api_response_handler.js:510:27)
api_1 | at /var/app/current/api/app.js:105:23
api_1 | at Layer.handle [as handle_request] (/var/app/current/api/node_modules/express/lib/router/layer.js:95:5)
api_1 | at trim_prefix (/var/app/current/api/node_modules/express/lib/router/index.js:312:13)
api_1 | at /var/app/current/api/node_modules/express/lib/router/index.js:280:7
api_1 | at Function.process_params (/var/app/current/api/node_modules/express/lib/router/index.js:330:12)
api_1 | at next (/var/app/current/api/node_modules/express/lib/router/index.js:271:10)
api_1 |
```
username_0: Definitely from seneca - confirmed
username_1: 2016-03-30 12:33:54 +00:00: Trace
at GateExecutor.addListener (events.js:239:17)
at GateExecutor.once (events.js:265:8)
at Seneca.api_ready [as ready] (/home/ubuntu/hardwell/hardwell-inbound-api/node_modules/chairo/node_modules/seneca/seneca.js:923:30)
at Seneca.api_close [as close] (/home/ubuntu/hardwell/hardwell-inbound-api/node_modules/chairo/node_modules/seneca/seneca.js:875:12)
at Seneca.die (/home/ubuntu/hardwell/hardwell-inbound-api/node_modules/chairo/node_modules/seneca/lib/common.js:350:18)
at act_done (/home/ubuntu/hardwell/hardwell-inbound-api/node_modules/chairo/node_modules/seneca/seneca.js:1185:29)
at /home/ubuntu/hardwell/hardwell-inbound-api/node_modules/chairo/node_modules/seneca/node_modules/gate-executor/gate-executor.js:155:20
at /home/ubuntu/hardwell/hardwell-inbound-api/lib/services/config.js:31:17
at Query.<anonymous> (/home/ubuntu/hardwell/hardwell-inbound-api/node_modules/mongoose/lib/query.js:2149:28)
at /home/ubuntu/hardwell/hardwell-inbound-api/node_modules/mongoose/node_modules/kareem/index.js:177:19
at /home/ubuntu/hardwell/hardwell-inbound-api/node_modules/mongoose/node_modules/kareem/index.js:109:16
at nextTickCallbackWith0Args (node.js:415:9)
at process._tickDomainCallback (node.js:385:13)
2016-03-30 12:33:55 +00:00:
username_2: Hey @username_0 Can you show the code you are running? I've looked over the tests and I'm not getting leaks.
username_0: It spans multiple files but the GateExecutor.addListener does not seem to remove the listener in a timely fashion.
username_2: @username_0 Thanks a mill for the feedback. I'm down in the bowls of it for the next two weeks so I've marked this as being investigated. Let me see what I find.
username_2: cc/ @mcollina
username_2: @username_0 We may have resolved this. Once 3.0 drops we will know for sure
username_3: Hi @username_2 i'm still having this issue with [email protected]
any updates?
Thanks
username_4: @username_3 do you have code that can reproduce this? Thanks :)
username_3: Hi @username_4
well i think this issue is no longer happening to us with [email protected] after switching from redis to rabbitmq as our transport, so it could be a transport problem not in Core seneca itself.
Regards. :)
Status: Issue closed
username_4: @username_3 good to hear! Thank you :)
username_0: @username_3 I do believe that after upgrading to 3.x this has been fixed especially since the cache is no longer automatically enabled |
pypa/pip | 288412998 | Title: extensive search option
Question:
username_0: /request-review
I would like a feature where there is an extensive search result. For example searching pip
pip search pip results in
```...
pip (9.0.1) - The PyPA recommended
tool for installing
Python packages.
INSTALLED: 9.0.1 (latest)
...
```
Whereas on the pypi page there is a LOT more info. Information like python version, upload dates, uploads, licenses, install or docs links, etc. It doesnt even give the link back to pypi, you have to go and make a separate search in pypi to get there. Whats the point of using pip search, i might as well always just go to pypi and search. It would be nice to have the information from pypi in the terminal from pip search.
I am not saying to add it to default pip search because that is a lot of info.
But even Ubuntu's package manager 'apt' allows the option for more information of their repos
```
username_0@ubuntu:~$ sudo apt-cache show python3-bs4
Package: python3-bs4
Priority: optional
Section: python
Installed-Size: 341
Maintainer: Ubuntu Developers <<EMAIL>>
Original-Maintainer: Debian Python Modules Team <<EMAIL>>
Architecture: all
Source: beautifulsoup4
Version: 4.4.1-1
Depends: python3:any (>= 3.3.2-2~)
Recommends: python3-chardet, python3-lxml
Filename: pool/main/b/beautifulsoup4/python3-bs4_4.4.1-1_all.deb
Size: 64080
MD5sum: d6973f42208ce358e6a0fb1c7407f4f1
SHA1: 0826fc1b486766746c0c1d3356d71734434e2675
SHA256: a766e2905bc9cb4e757a14d5dae6f8b9a1aa871b72fa12141bf22793ba8c93a6
Description-en: error-tolerant HTML parser for Python 3
Beautiful Soup is a Python library for pulling data out of HTML and XML files.
It works with your favorite parser to provide idiomatic ways of navigating,
searching, and modifying the parse tree. It commonly saves programmers hours
or days of work.
.
This package contains BeautifulSoup 4, for Python 3.x.
Description-md5: fb6bbcbf2610881d02572826329c9b94
Homepage: http://www.crummy.com/software/BeautifulSoup
Bugs: https://bugs.launchpad.net/ubuntu/+filebug
Origin: Ubuntu
Supported: 5y
Task: ubuntu-desktop, ubuntu-usb, edubuntu-desktop, edubuntu-usb, xubuntu-desktop, ubuntustudio-desktop, ubuntu-gnome-desktop, ubuntu-sdk, ubuntukylin-desktop
```
Answers:
username_1: Hi @username_0!
`pip show pip` might be what you're looking for. Linking back to PyPI isn't really a definite thing since pip doesn't really care about where you fetched stuff from _after_ they have been installed (generally). The rest of the information is exposed by `pip show`.
Does the above resolve your concern? If not, what information are you looking for exactly?
As for pip search not being very helpful, that's a known issue and, as I understand, it'll improve... When Warehouse gets deployed, pip search will start giving saner results and there's also other issues with the actual implementation of search that need to be improved. |
SimulaVR/Simula | 1000130148 | Title: AMD uProf data
Question:
username_0: # Steps to generate
```
git clone https://github.com/SimulaVR/uprof
cd uprof
nix-build
./result/bin/AMDuProf-bin
# ..launch Simula..
# Find the running `godot` process in uprof and analyze it for 20 seconds
```
# AMD uProf data



# TLDR
`acquire_spin_lock_slow_path` in `libHSrts_thr-ghc8.10.4.so` is the most hogging function.
Answers:
username_0: The above data was *with* VR running. If instead we profile a running Simula *without* VR running, we get:

Some noticeable functions are absent, including `strcmp`. Looks like some sort of string comparison operation is hogging resources every frame when VR is running. Possibly related to the
```
OpenVR reports: 102
```
being spammed to the console with our current half-working SteamVR on NixOS? I checked SimulaVR/godot and godot-openvr for usage of `strcmp` and didn't see anything, so whatever it is is at a lower level most likely. |
tmdgusya/roach-web-server | 842989626 | Title: ReturnValueHandler 를 만든다.
Question:
username_0: 앞서 MappingUrlHandler 를 사용했을때 returnValue 가 나오게 되는데, 해당 Return 값을 처리해줄 필요가 있다.
따라서 ReturnValueHandler 를 통해서 요구하는 방식으로 처리할 수 있게 한다.
일단은 간단하게 어떤 View 를 불러와야 하는지만 알 수 있도록 해보자.
Answers:
username_0: RetrunType 이 String 일 경우 StringViewResolver ..? => 또 이러면 Model 이 필요한데..
이 구조는 조금 생각을 해보고 짤 필요가 있다.
객체일 경우 Json Type?
Status: Issue closed
|
helpshift/kafboy | 110677160 | Title: Errors on startup
Question:
username_0: When i try and fire up the kafboy demo i am getting the error
bash> erl -pa deps/*/ebin -pa ebin -s kafboy_demo
Erlang/OTP 18 [erts-7.1] [source] [64-bit] [smp:2:2] [async-threads:10] [kernel-poll:false]
Eshell V7.1 (abort with ^G)
1>
1 => stopped:{error,not_found} started:{ok,<0.53.0>}
On the first attempted request i get the error
=ERROR REPORT==== 9-Oct-2015::16:02:27 ===
Ranch listener http had connection process started with cowboy_protocol:start_link/4 at <0.183.0> exit with reason: {cowboy_handler,[{class,error},{reason,undef},{mfa,{kafboy_http_handler,init,2}},{stacktrace,[{kafboy_http_handler,init,[{http_req,#Port<0.2386>,ranch_tcp,keepalive,<0.183.0>,<<"POST">>,'HTTP/1.1',{{127,0,0,1},34492},<<"localhost">>,undefined,9903,<<"/batch/async/ekaf">>,undefined,<<>>,[{topic,<<"ekaf">>}],[{<<"user-agent">>,<<"curl/7.22.0 (x86_64-pc-linux-gnu) libcurl/7.22.0 OpenSSL/1.0.1 zlib/1.2.3.4 libidn/1.23 librtmp/2.3">>},{<<"host">>,<<"localhost:9903">>},{<<"accept">>,<<"*/*">>},{<<"content-length">>,<<"7">>},{<<"content-type">>,<<"application/x-www-form-urlencoded">>}],[],waiting,<<"hello=a">>,undefined,false,waiting,[],<<>>,undefined},{kafboy_http,0,false,true,false,undefined,{kafboy_demo,massage_json},undefined}],[]},{cowboy_handler,execute,2,[{file,"src/cowboy_handler.erl"},{line,41}]},{cowboy_protocol,execute,4,[{file,"src/cowboy_protocol.erl"},{line,470}]}]},{req,[{socket,#Port<0.2386>},{transport,ranch_tcp},{connection,keepalive},{pid,<0.183.0>},{method,<<"POST">>},{version,'HTTP/1.1'},{peer,{{127,0,0,1},34492}},{host,<<"localhost">>},{host_info,undefined},{port,9903},{path,<<"/batch/async/ekaf">>},{path_info,undefined},{qs,<<>>},{bindings,[{topic,<<"ekaf">>}]},{headers,[{<<"user-agent">>,<<"curl/7.22.0 (x86_64-pc-linux-gnu) libcurl/7.22.0 OpenSSL/1.0.1 zlib/1.2.3.4 libidn/1.23 librtmp/2.3">>},{<<"host">>,<<"localhost:9903">>},{<<"accept">>,<<"*/*">>},{<<"content-length">>,<<"7">>},{<<"content-type">>,<<"application/x-www-form-urlencoded">>}]},{meta,[]},{body_state,waiting},{buffer,<<"hello=a">>},{multipart,undefined},{resp_compress,false},{resp_state,waiting},{resp_headers,[]},{resp_body,<<>>},{onresponse,undefined}]},{opts,{kafboy_http,0,false,true,false,undefined,{kafboy_demo,massage_json},undefined}}]}
I am fairly new to erlang so apologies if the reason for this problem should be obvious.
Thanks |
kubernetes/kubernetes | 245174752 | Title: Refactor the code related to VerifyVolumesAreAttached
Question:
username_0: Refactor the code related to VerifyVolumesAreAttached in operation_executor and operation_generator. The code is a little bit hard to read after a few rounds of changes. This function has two different ways of checking the volume attachment status, one is BulkVerifyVolumes and the other is VerifyVolumesAreAttachedPerNode. Some for loops are performed in operation_executor and some are in operation_generator. They are not consistent in these two VerifyVolume functions. |
samvera/hyrax | 376073812 | Title: Edit Work buttons layout re-factor
Question:
username_0: ### Descriptive summary
The layout of main buttons for Edit Work page is confusing, and could be better organized. Investigate alternative layout options.
<img width="1033" alt="work-buttons-edit-page" src="https://user-images.githubusercontent.com/3020266/47806128-fed95b80-dd06-11e8-918b-654dae1b82d6.png">
Link to related tickets or prior related work here.
Answers:
username_0: @chrisdaaz @username_1 2 options for how to clean this display up a bit. Preference? I also fixed the label sizing next to the Title, and made the buttons the primary button color to follow suit on Edit Collections page. Trying to get some consistency.


username_1: i prefer the second, as it places the buttons in the same horizontal space. i'm assuming there is a hard line break between the two rows to prevent the icky re-stacking effect we're currently seeing.
also, how does this do on mobile?
username_1: @samvera/hyrax-repo-managers may want to have a peek at this
username_0: @username_1 Sounds good. Here's how it looks on mobile:
After:

Before:

username_2: looks much better!
username_0: @jcoyne How about these two options? The 2nd option, of putting the buttons in their own row, is how Github might approach this. Maybe Github should just be the style guide for how to implement Bootstrap in Hyrax when in doubt.
If I ran further with this, I'd go buttons above title, then break out "Relationships" and "Items" in their own panel groupings, with a lower subhead `<h3>`. To make the content separation more distinguishable maybe?
#### Button Group


#### Button in own row
##### Above title

##### Below title


username_0: Last screenshot, then I'll give it a rest. Something like this...

username_1: of those, i like the buttons above the title. that way they don't invade the metadata box. not sure there is a "best" approach, though.
username_0: Great. @username_1 Are you ok with 'panelizing' Relationships and Items sections in this page as well?
username_0: My idea is that maybe we're currently using panels as a container to shove all the page content in them, instead of using them to visually group UI segments.
username_3: I think the last screenshot is the most well organized for a variety of reasons:
- similar content is grouped
- tertiary content (relationships, Thumbnails) are cleanly broken out
- user actions are grouped by similar tasks (e.g. Edit, attach child, delete)
The work type ( "Generic Work") looks to be the Header of the page. It seems that should be lower on the page hierarchy than the title of the work. Even if it's above the Title it should be obvious it's of lower importance visually.
username_0: Here's another version, with the following updates:
- Uses the work title(s) as `<h1>` tag(s) which may be better for seo and accessibility?
- Swaps the "CRUD" type function buttons (Edit, .... Delete), to the right which is more typical in the application.

Status: Issue closed
|
migurski/Dymo | 110868666 | Title: Quick link swap in README
Question:
username_0: Hi there!
You're link to Acetate in the readme is not right.
It links to <https://github.com/fortiusone/acetate>
Which 404s

It should like to <https://github.com/geoiq/acetate> I think.
Good luck with the project, seems so cool :)!<issue_closed>
Status: Issue closed |
scalameta/scalameta | 223570328 | Title: Parentheses on constructor annotation
Question:
username_0: Question: Why the scalameta parser ignores parentheses appended to constructor annotations?
Use case: In a playframework-2.6 application, I'd like to expand some controllers which use dependency injection, i.e.
```scala
@someMacro class FooController @Inject() (cc: ControllerContexts) extends AbstractController(cc) { ??? }
// ^ those parentheses are indispensable
```
However, it will be expanded into:
```scala
class FooController @Inject (cc: ControllerContexts) extends AbstractController(cc) { ??? }
// ^ parentheses are GONE
```
and this results in a compilation error.
Cause: In https://github.com/scalameta/scalameta/blob/master/scalameta/parsers/shared/src/main/scala/scala/meta/internal/parsers/ScalametaParser.scala#L3367-L3370 , parentheses are removed if `allowArgss` is `false`, and it is actually `false` when parsing constructor annotations: https://github.com/scalameta/scalameta/blob/master/scalameta/parsers/shared/src/main/scala/scala/meta/internal/parsers/ScalametaParser.scala#L2714
Can I remove those lines? If not, I wonder why.
Thanks
Answers:
username_1: Are you using inline/meta macro annotations? If so, the problem is probably in LogicalTrees https://github.com/scalameta/scalameta/blob/master/scalahost/core/src/main/scala/scala/meta/internal/scalahost/converters/LogicalTrees.scala
LogicalTrees may be going away soon soon (see https://github.com/scalameta/paradise/issues/194 and #808) in favor of a more robust approach that makes it easier to preserve all details in the original source
username_0: Indeed I use inline/meta macro, but the cause is really in LogicalTrees? I just read the source now, but got no idea about where the parentheses is removed.
username_1: Inline/meta macro annotations don't use the scala.meta parser, they use LogicalTrees to convert between scala-reflect trees and scala.meta trees. The motivation behind https://github.com/scalameta/paradise/issues/194 to get rid of the inherent complexity of LogicalTrees. I spent a significant amount of time making sure the converter preserves all syntactic details in the original source and we have a large test suite to check this. However, despite our efforts, the converter is still incomplete.
Is this a blocker? You can try to add a test case [here](https://github.com/scalameta/scalameta/blob/master/scalahost/nsc/src/test/scala/scala/meta/tests/scalahost/converters/Syntactic.scala) and see if you can track down the cause. However, note that LogicalTrees may be going away in the next few weeks in favor of using the scala.meta parser directly.
Status: Issue closed
username_0: Thanks for comprehensive explanation. I've try that test (`syntactic("class Foo @Inject() (cc: ControllerContexts) extends AbstractController(cc)")`). It passes when `allowArgss == false` (as in this repository), and actually, it fails when `allowArgss == true`. My suggestion was completely wrong, sorry about that.
I decided to use some work-around implementations and wait for you to move to scalameta's parser from LogicalTrees.
Thanks again. |
EkkoLin/twitterDemo | 210622447 | Title: Project Feedback!
Question:
username_0: Looks like you are missing the following required user stories:
- Retweeting and favoriting should increment the retweet and favorite count.
Your assignment is incomplete until all the [required user stories](https://courses.codepath.com/snippets/ios_university/readme_templates/assignment_4_readme.md) are complete. Once completed, please push your updates, update your README, update your gif and [submit your assignment](https://courses.codepath.com/courses/ios_university/pages/submitting_assignments) again using the submit button on the appropriate assignment page so we can regrade it.
Whenever you make updates to your project that require re-grading, you need to **re-submit** your project using the submit button on the associated assignment page in the course portal. This will flag your project as “updated” on our end and we know to re-grade.
You should re-submit your assignment anytime you:
- Update a previously incomplete assignment
- Add optional and additional features to an already completed assignment
Answers:
username_0: :+1: Nice work! The point of this homework was to explore an example of a full MVC application with a RESTful API.
We have a detailed [Project 4 Feedback Guide](http://courses.codepath.com/snippets/ios_university/feedback_guides/project_4_feedback.md) which covers the best practices for implementing this assignment. Read through the feedback guide point-by-point to determine ways you might be able to improve your submission. You should consider going back and implementing these improvements as well. Keep in mind that one of the most important parts of iOS development is learning the correct patterns and conventions.
Check out the [project rubric](https://courses.codepath.com/snippets/ios_university/project_rubric.md) for a breakdown of how submissions are scored.
If you have any particular questions about the assignment or the feedback, feel free to reply here or email us at <<EMAIL>>.
username_0: :+1: Nice work! We added this homework so that you'll have the experience building an app that's more similar in complexity to a production app. The optional account switching feature is interesting because you have to think about how TwitterClient is caching the access token to make sure it works correctly for multiple accounts.
We have a detailed [Project 5 Feedback Guide](http://courses.codepath.com/snippets/ios_university/feedback_guides/project_5_feedback.md) which covers the best practices for implementing this assignment. Read through the feedback guide point-by-point to determine ways you might be able to improve your submission. You should consider going back and implementing these improvements as well.
Check out the [project rubric](https://courses.codepath.com/snippets/ios_university/project_rubric.md) for a breakdown of how submissions are scored.
If you have any particular questions about the assignment or the feedback, feel free to reply here or email us at <<EMAIL>>. |
angular/angular | 188992099 | Title: Property binding to the object does not work!
Question:
username_0: I've described the issue in this stackoverflow post: http://stackoverflow.com/questions/40578305/property-binding-to-the-object-does-not-work-in-angular-2
Hope I'm not violating rules by linking it here.
It seems either a significant bug or significant documentation issue. If such behavior is expected, it should be documented.
angular-cli: 1.0.0-beta.19-3
angular: 2.1.2
Answers:
username_1: you're right, this issue tracker is not suitable for support requests.
Try to use `ngValue` instead of `value`
here's an example https://plnkr.co/edit/HS9yDCxSGni9sziexu9u?p=preview
username_0: @username_1 thank you, [ngValue] worked instead of [value]. But why? Where (in the docs) can I find more about ngValue vs value differences, and when to use which? In tour of herous tutorial there was binding to value (not ngValue), and I wouldn't think that ngValue would be required. I think this behaviour needs to be explicitly documented. (Or if it is, could you please point me to it?)
username_1: Feel free to file docs issues on https://github.com/angular/angular.io/issues if anything is not clear for you.
Status: Issue closed
username_2: I'n not sure this is the right fix.
Anyway reported issues must provide the information asked in the template. Re-open if still an issue. |
hexojs/hexo | 317926388 | Title: github build failure
Question:
username_0: 2. The way to reproduce
you can download my github.io repo
https://github.com/username_0/username_0.github.io.git
there is two branch, hexo branch is for hexo
<!-- If you find that markdown files are not rendered as expected, please go to https://marked.js.org/demo/ to see if it can be reproduced there. If it can be reproduced, please file a bug to https://github.com/markedjs/marked.
-->
## For question
<!-- Question description -->
## For feature request
<!-- Feature description -->
Answers:
username_1: I also met exactly same problem.
username_1: I just start from scratch over from 'blog init', and fix the problem.
username_2: Hi,
I was suffering from this issue as well.
I followed deployment guide from the [doc.](http://hexo.io/docs/deployment)
Today I realize that Github Pages could only be generated from either master or gh-pages branch as we can see in repository settings.
So I changed branch from `published`to `master`in deploy section in _config.yml file.
However this caused my regular workflow's change.
I have to use different branch to work and make `master`branch as publicaction branch.
Hope it helps. |
ParadoxGameConverters/Vic2ToHoI4 | 392869052 | Title: Impassable states
Question:
username_0: We should somehow manage to replicate the vanilla impassable states, to avoid AI destroying their troops somewhere in the Sahara.
The line that should be added in the state history is `impassable = yes`, and I think we should add that also in the province mappings so that it'll generate states that are impassable with all the provinces of a state X that are counted so, and another with the provinces of the same state X that are not, calling the first one "Further X".
I'll take care of adding it in the province mappings if the idea gets added in the code. |
mermaid-js/mermaid | 849918875 | Title: Ability to disable the source map?
Question:
username_0: **Describe the solution you'd like**
In the current webpack configuration, the generated JavaScript file will always have a source map included.
I think it would be great if we can specify an environment variable to indicate that we don't want the source map to be generated and included.
If the development team thinks this is a good idea, I can send a pull request to help you guys get this feature in. |
DeokyunKim/Progressive-Face-Super-Resolution | 757663745 | Title: THCudaCheck FAIL
Question:
username_0: I tried to run the single-image demonstration on one image, using pytorch 1.0.0 and cud 10.0, but it gives me this output:
```
pre-trained model is loaded step:3, alpha:1 iteration:80000
THCudaCheck FAIL file=/opt/conda/conda-bld/pytorch_1544174967633/work/aten/src/THC/THCGeneral.cpp line=405 error=11 : invalid argument
```
I mean, I am using pytorch 1.0.0 as you are, so I don't see where the problem could be. Any idea?
Answers:
username_1: Hi, username_0.
Thanks for your interest out work.
Please refer the [link](https://github.com/pytorch/pytorch/issues/15797#issuecomment-472167754). |
aws/amazon-ecs-agent | 205039834 | Title: Feature Request: Stop task interval
Question:
username_0: When I set a min deployment configuration to less than 100 (say 50), I allow my tasks to go down and rolling reboot in order to update to a newer task definition. The problem is that process is a little bit too quick and sometimes my tasks take say 2 minutes to boot.
I'd like to have an environment variable to have a timeout in between two tasks being brought down in case of a rolling upgrade. Thoughts?
Answers:
username_1: @username_0 Are your tasks using load balancers? If you use load balancers and set your container to fail the load balancer health check until it becomes healthy, you can use the `minimumHealthyPercent` and `maximumPercent` parameters to control how quickly your service deploys.
username_2: @username_0: have you tried the suggestion above? |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.