repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
kubernetes/kubernetes | 47992422 | Title: Bedrock auth sources
Question:
username_0: The APIserver can authenticate users using a Bearer Token.
Currently, these can only be provided in a file, via --token_auth_file.
We should provide a way to get a token from the apiserver based on an external ("bedrock") authentication source.
Steps:
- add endpoint, e.g. "/login" to apiserver which serves a form which allows you to select a bedrock auth provider, click something, be redirected to that provider's site, be establish identity and allow apiserver to learn that identity
- maybe a similar flow, but for automated consumption by kubectl, etc.
- add "authentication plugins" for a couple authn providers, such as google.com and github.com.
- generate token (with some expiry) and give it to the user to cut from the UI and paste it into their .kubernetes_auth (or return as a parsable body for automated flow)
- make sure user records from different auth sources can't collide.
References:
- #443
- #1430
- [https://github.com/GoogleCloudPlatform/kubernetes/blob/master/docs/authentication.md]
Answers:
username_1: I'm interested in something along these lines. This ticket is a bit old and I'm not sure how it relates to the auth provider framework introduced in #23066.
My use case has Kerberos as the bedrock authentication source. I have written a bearer token issuing/verification service, and point kube-apiserver at it via the webhook token authentication feature. I'm currently wrapping kubectl with a script that authenticates to my service with Kerberos, gets a bearer token, runs kubectl config set-credentials, then execs kubectl. It would be nice if I could eliminate that wrapper script somehow.
username_2: We integrated kerberos into openshift via the challenge client concept, which is vaguely on the roadmap to add to Kube and the eventual `kubectl login`. In theory we could allow kubectl login to exec down to a child process that can negotiate for your bearer token. |
home-assistant/android | 612493579 | Title: Norwegian language/translation for Companion app
Question:
username_0: <!-- READ THIS FIRST:
- Make sure you've checked existing feature requests to make sure you aren't opening a duplicate. If you do open a duplicate issue because you didn't check existing requests, we will close your issue and laugh at you privately.
DO NOT DELETE ANY TEXT from this template! All requested information is important.
-->
**Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
No
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
Add the Norwegian language (nb-NO) for translation at Lokalise
**Describe alternatives you've considered, if any**
A clear and concise description of any alternative solutions or features you've considered.
N/A
**Additional context**
Add any other context or screenshots about the feature request here.
Should be pretty straight forward.
Answers:
username_1: Added and you should have access now.
Status: Issue closed
username_0: Has there been any change to this? I can't seem to find it anymore.
username_1: @username_0 there were duplicate languages in Lokalise, you have been added to Norwegian again. |
AdguardTeam/AdguardForAndroid | 443986532 | Title: HTTPS Filtering is "Unavailable for your Device"
Question:
username_0: <!--- Help us to avoid duplicate reports, make sure you have searched through existing issues before submitting a new one -->
<!--- If you are requesting a new feature, tell us how it should work in free form -->
<!--- If you are reporting a bug, submit the detailed description using the template below -->
### Steps to reproduce
<!--- Provide a link to a live example or a clear set of steps to reproduce the issue -->
1. Use MIUI Version: MIUI Global 10.2 - Stable [10.2.1.0 (MHOMIXM)]
2. Launch AdGuard
3. Update to (3.0.349) Nightly
4. Check Settings > Network
### Expected behavior
<!--- Tell us what should happen -->
Enable filtering should be available.
### Actual behavior
<!--- Tell us what happens instead -->
<details><summary>Screenshot:</summary>
<!--- Drag and drop, upload or paste your screenshot to this area -->

</details>
### Customer ID
<!--- Send us a diagnostic report through the application by choosing "Send Feedback" in the side menu of the app -->
<!--- You will get an autoreply with you Customer ID after sending a report -->
### Your environment
<!--- Please include all relevant details about the environment you experienced the bug in -->
* Adguard version: (3.0.349) Nightly
* Device model: MIUI Version: MIUI Global 10.2 - Stable [10.2.1.0 (MHOMIXM)]
* Operating system and version: Android Version: 6.0.1 MMB29M<issue_closed>
Status: Issue closed |
ConsumerDataStandardsAustralia/future-plan | 700907567 | Title: DSB Item - Enhanced standards publishing pipeline
Question:
username_0: # Problem Statement
As the CDR standards expand to multiple sectors and begin to incorporate explanatory statements and implementation guidelines the need for a review and revamp of the standards publishing tools and pipeline is growing.
The intent of this review would be as follows:
- To ensure the standards are understandable as they grow in complexity
- To enhance the standards with non-normative explanations that aid understanding
- To accommodate iterative change, including versioned end points
- To incorporate enhanced development tools and demonstration mechanisms to aid development<issue_closed>
Status: Issue closed |
google/ExoPlayer | 310381062 | Title: DynamicConcatenatingMediaSource removeMediaSource
Question:
username_0: Before filing an issue:
-----------------------
- Search existing issues, including issues that are closed.
- Consult our FAQs, supported devices and supported formats pages. These can be
found at https://google.github.io/ExoPlayer/.
- Rule out issues in your own code. A good way to do this is to try and
reproduce the issue in the ExoPlayer demo app.
- This issue tracker is intended for bugs, feature requests and ExoPlayer
specific questions. If you're asking a general Android development question,
please do so on Stack Overflow.
When reporting a bug:
-----------------------
Fill out the sections below, leaving the headers but replacing the content. If
you're unable to provide certain information, please explain why in the relevant
section. We may close issues if they do not include sufficient information.
### Issue description
Describe the issue in detail, including observed and expected behavior.
### Reproduction steps
Describe how the issue can be reproduced, ideally using the ExoPlayer demo app.
### Link to test content
Provide a link to media that reproduces the issue. If you don't wish to post it
publicly, please submit the issue, then email the link to
<EMAIL> using a subject in the format "Issue #1234".
### Version of ExoPlayer being used
Specify the absolute version number. Avoid using terms such as "latest".
### Device(s) and version(s) of Android being used
Specify the devices and versions of Android on which the issue can be
reproduced, and how easily it reproduces. If possible, please test on multiple
devices and Android versions.
### A full bug report captured from the device
Capture a full bug report using "adb bugreport". Output from "adb logcat" or a
log snippet is NOT sufficient. Please attach the captured bug report as a file.
If you don't wish to post it publicly, please submit the issue, then email the
bug report to <EMAIL> using a subject in the format "Issue #1234".
hi:
when I call removeMediaSource of DynamicConcatenatingMediaSource , second I call setPlayWhenReady of SimpleExoPlayer, it can not play, please help
Thanks
Jacky
Answers:
username_1: It's a bit unclear what your actual issue is. Could please fill out the [issue template ](https://github.com/google/ExoPlayer/blob/release-v2/ISSUE_TEMPLATE) and especially provide example media and reproduction steps.
Status: Issue closed
username_1: Closing due to inactivity. |
ISISComputingGroup/IBEX | 548148527 | Title: genie_python/DAE: Fix inconsistent period PV alarm
Question:
username_0: As a result of https://github.com/ISISComputingGroup/IBEX/issues/4992, genie python now shows an exception to the screen when it can not the total number of periods using g.change_number_soft_periods () or the period number using g.change_period().
The exception is thrown based whether or not the appropriate DAE PV (`DAE:HARDWAREPERIODS:SP` or `DAE:PERIOD:SP`) is in an INVALID alarm after you have set the pv to the new value.
Changing the period number should succeed unless the argument is 0 or less or unless it is larger than the total period count. Therefore, if you set the number of periods to 1000 and then try to set the period to 21, it should work. Instead, genie python throws an exception:
```
g.change_number_soft_periods(20)
g.change_period(13)
g.change_number_soft_periods(1000)
g.change_period(19)
g.change_period(20)
g.change_period(21)
Traceback (most recent call last):
File "C:\Instrument\Apps\Python\lib\site-packages\genie_python\genie.py", line 1852, in change_period
__api.dae.set_period(period)
File "C:\Instrument\Apps\Python\lib\site-packages\genie_python\genie_dae.py", line 574, in set_period
raise IOError("You are trying to set an invalid period number! The number must be between 1 and the "
IOError: You are trying to set an invalid period number! The number must be between 1 and the number of periods!
g.change_period(999)
Traceback (most recent call last):
File "C:\Instrument\Apps\Python\lib\site-packages\genie_python\genie.py", line 1852, in change_period
__api.dae.set_period(period)
File "C:\Instrument\Apps\Python\lib\site-packages\genie_python\genie_dae.py", line 574, in set_period
raise IOError("You are trying to set an invalid period number! The number must be between 1 and the "
IOError: You are trying to set an invalid period number! The number must be between 1 and the number of periods!
```
As you can see, the DAE seems to remember the old number of periods values and not actually update to 1000, making setting the period later fail.
Interestingly, if you try the above commands in a scripting console in the GUI one by one you will see that although trying to set the period to 21 or 999 throws an exception, the period number will change to the new value in the dashboard. If you open an EPICS terminal you can see that the setpoint PV is in an INVALID alarm, but that the readback was also changed to the new value. However, if you try to set the period to 1001, it will throw an exception and the dashboard period number will stay the same, and the readback PV will not change.
Freddie said that this issue might be because the DAE hardware has one period limit, and the software has a different limit, so in this case the operation fails in one place but succeeds in the other so this leads to an inconsistent error. If you try to change the number of periods to an even larger number like 1000000 then it will throw an error from the very beginning.
Answers:
username_0: As of now, https://github.com/ISISComputingGroup/IBEX/issues/4992 has not been completed yet, so in order to see the problem in the GUI as I have described it you need to checkout the https://github.com/ISISComputingGroup/genie_python/tree/Ticket4992_raise_invalid_period_count_exception branch.
username_0: When fixing this bug, we should also add a system test that checks that for medium total number of periods like 1000, change_number_soft_periods and change_period behave correctly.
username_1: pair with @FreddieAkeroyd ? or @DominicOram |
ryouchinsa/Rectlabel-support | 459230089 | Title: minor-aesthetic: double arrow mouse point orientation on edit-points
Question:
username_0: All edit points show the horizontal double arrow.
Corner points should display appropriate diagonal double arrow mouse pointer.
Mid points should display appropriate horizontal or vertical double arrow mouse pointer.
It's only an aesthetic "bug".
Status: Issue closed
Answers:
username_1: Thanks for writing the issue.
Looking at Adobe Illustrator, as you said, they consider the editing direction to show the edit icon.
At this moment, editable or not is important for us.
When we could understand the feature is necessary for RectLabel, we would implement. |
openSUSE/osem | 288202837 | Title: NoMethodError in ConferencesController#show
Question:
username_0: **I'm submitting a ..**
- [ ] Bug Report
- [ ] Feature Request
**Current behavior:**
Getting error when splash page is not present.


**Expected correct behavior:**
Clicking on splash page link should redirect user to create splashpage.
**Steps to reproduce:**
1. Create conference
2. Go to splashpage
3. Click on splash page link (Build a "splash page" with all the information for your conference )
Answers:
username_1: just for good ux ,may be instead of taking user to create splashpage , that link should direct user to some info page describing what splash page is, in a new tab . Like markdown Syntax link under every desc field :

Or even we can remove that link since theres already a bigger button on the page :thinking:
username_2: @username_0 it would be useful to know that you need to not be an admin user, but not be a simple user either. To reproduce this you need to be allowed to access the splashpage, so your user needs to be the organizer of the conference (or similar user access to the action conferences#show even when there is no splashpage).
username_3: @username_2 can i claim this issue??
username_2: sure @username_3
username_2: The error still exists for show template

username_4: Can I work on this issue?
Status: Issue closed
|
jumpinjackie/jsdoc-typescript-plugin | 139609954 | Title: Nested generic types not properly parsed
Question:
username_0: The current regex used for parsing generic types
`/(.+)(.\<)(.+)\>/`
Currently breaks down on nested types
eg. `Foo.<Array.<Bar>>`
Answers:
username_1: I don't think a regex can do this, you need a real parser. I can write this if you want.
username_1: Here's a pegjs grammar that will parse the types
``` pegjs
Type
= GenericType
/ UnionType
/ SimpleType
NonUnionType
= GenericType
/ SimpleType
UnionType
= first:NonUnionType _ rest:UnionTypeRest+ {
return {
kind: "union",
types: [first].concat(rest)
};
}
UnionTypeRest
= "|" _ type:NonUnionType {
return type;
}
GenericType
= name:Ident "."? "<" _ params:GenericTypeParams _ ">" {
return {
kind: "generic",
name: name,
parameters: params
}
}
GenericTypeParams
= first:Type? _ rest:GenericTypeParamsRest* {
return first ? [first].concat(rest) : rest;
}
GenericTypeParamsRest
= "," _ type:Type {
return type;
}
SimpleType
= name:Ident {
return {
kind: "simple",
name: text()
};
}
Ident
= [a-zA-Z$_-] [a-zA-Z0-9$_-]* {
return text();
}
_ "whitespace"
= [ \t\n\r]*
```
username_1: parsing `Entity|Entity[]|EntityCollection|DataSource|ImageryLayer|Promise.<Entity|Entity[]|EntityCollection|DataSource|ImageryLayer>` gives:
``` json
{
"kind": "union",
"types": [
{
"kind": "simple",
"name": "Entity"
},
{
"kind": "generic",
"name": "Array",
"params": [
{
"kind": "simple",
"name": "Entity"
}
]
},
{
"kind": "simple",
"name": "EntityCollection"
},
{
"kind": "simple",
"name": "DataSource"
},
{
"kind": "simple",
"name": "ImageryLayer"
},
{
"kind": "generic",
"name": "Promise",
"parameters": [
{
"kind": "union",
"types": [
{
"kind": "simple",
"name": "Entity"
},
{
"kind": "generic",
"name": "Array",
"params": [
{
"kind": "simple",
"name": "Entity"
}
]
},
{
"kind": "simple",
"name": "EntityCollection"
},
{
"kind": "simple",
"name": "DataSource"
},
{
"kind": "simple",
"name": "ImageryLayer"
}
]
}
]
}
]
}
```
username_0: Hi @username_1 could you provide a pull request for this? PegJS sounds very useful.
username_1: Looks like #83 closes this. |
s-p-a-r-k/Jacket-Tracker | 292583243 | Title: Encrypt all data sent to DB
Question:
username_0: We are storing cat 3 so we should abide by OIT standards
Answers:
username_0: Since we allow unauthenticated users to create data, and we allow any authenticated user to read the data there is no one-to-one relationship so private-public key encryption won’t really work.
Plus user data is being created by unauthenticated users so they don't have any persistence with our application.
Holding off on this until we know more about the data handling requirements and how many lieutenant accounts we will be using. |
aws/aws-codedeploy-agent | 200434532 | Title: Allow * wildcard in source instruction of files section in appspec.yml
Question:
username_0: Thanks.
Answers:
username_1: We don't support wildcard for files names in the appspec file. As the doc describes: http://docs.aws.amazon.com/codedeploy/latest/userguide/app-spec-ref-files.html, either you can specify the whole folder or you may want to give the path of specific file.
Thanks,
Binbin
username_0: I sorted my problem by replacing the wildcard with the required value by using sed utility in the post-build shell execution action on Jenkins. Please feel free to close this ticket if you don't plan to implement this feature (I still believe it can be handy in some cases though).
Thanks,
Stan
username_1: I'll forward the feature request to our product manager, thanks
Status: Issue closed
|
Exploratorio-DCC-PUC/Syllabus | 723773054 | Title: Parte 2 pregunta 1 y 2, duda id museo y ciudad draw.io
Question:
username_0: Hola, en la 1ra consulta de la parte 2, debo agregar también el nombre del artista? En la 2da consulta, debo agregar una columna del país también para que haga más sentido? Ahora, en el draw.io, para la ciudad debo elegirle yo un id único, o puedo poner "cid"? Lo mismo para museos, debo poner un id que los identifique o puedo poner "mid"?
Answers:
username_1: Hola!, para las consultas basta con que cumplas lo que pide el enunciado, si quieres agregar otra columna pq crees que así se entiende mejor, no hay problema. Sobre los id en el diagrama, si no hay algo que te demos que pueda usarse como id, debes crearlos tu, la forma en que lo indicas es lo más común: cid y mid.
Status: Issue closed
|
styled-components/styled-components | 598146374 | Title: ThemeProvider
Question:
username_0: I can't use the Theme Provider
```
<ThemeProvider theme={light}>
<>
<GlobalStyle />
<Main />
</>
</ThemeProvider>
```
But I can't access $ {props => props.theme.primary} for example in the other style files.<issue_closed>
Status: Issue closed |
go-gitea/gitea | 317837338 | Title: an error is logged when the ldap admin search filter returns no results
Question:
username_0: <!--
1. Please speak English, this is the language all of us can speak and write.
2. Please ask questions or configuration/deploy problems on our Discord
server (https://discord.gg/NsatcWJ) or forum (https://discourse.gitea.io).
3. Please take a moment to check that your issue doesn't already exist.
4. Please give all relevant information below for bug reports, because
incomplete details will be handled as an invalid report.
-->
- Gitea version (or commit ref): 1.4.0
- Git version: n/a
- Operating system: ubuntu 17.10
- Database (use `[x]`):
- [ ] PostgreSQL
- [x] MySQL
- [ ] MSSQL
- [ ] SQLite
- Can you reproduce the bug at https://try.gitea.io:
- [ ] Yes (provide example URL)
- [ ] No
- [x] Not relevant
- Log gist: https://gist.github.com/username_0/b5abbf7ec4ee1f975a6a423b87c51d37
## Description
when the ldap search is performed to determine if a user is an admin, if there is no result [meaning the user in question is not an admin], an error is logged. this is a little misleading - if there is no result, because the user is not in the admin group, then things are working exactly as they should, and there is no error. i think it would be helpful if this outcome were handled and logged differently than an error<issue_closed>
Status: Issue closed |
Java-Bom/ReadingRecord | 671049666 | Title: [아이템 88] ObjectInputValidation
Question:
username_0: 질문 : 해당 부분의 ObjectInputValidation 에 대한 예제가 있음 좋을것 같아용!
Status: Issue closed
Answers:
username_1: `readObject` 에서 `ObjectInputStream`에 `registerValidation`를 이용해서 validation 로직과 우선순위를 설정해서 사용할 수 있어!
아래와 같은 클래스가 있다고 했을때 (주석된 부분이 주석이 되어있지 않다고 가정할게!)
```java
public class ObjectInputValidation implements Serializable {
private Date start;
private Date end;
public ObjectInputValidation(final Date start, final Date end) {
this.start = new Date(start.getTime());
this.end = new Date(end.getTime());
// 아래와 같은 코드가 주석처리가 되어 있지 않다고 가정할 때
// if (start.getTime() > end.getTime()) {
// throw new IllegalArgumentException();
// }
}
private void readObject(ObjectInputStream objectInputStream) throws IOException, ClassNotFoundException {
objectInputStream.defaultReadObject();
objectInputStream.registerValidation(
() -> {
if (start.getTime() > end.getTime()) {
throw new IllegalArgumentException();
}
}
, 0);
this.start = new Date(start.getTime());
this.end = new Date(end.getTime());
}
}
```
테스트코드에서 생성된 직렬화된 바이트 코드는 사실 생성될 수 없는 직렬화 스트림일거야(시작시간이 종료시간보다 더 뒤에 있기 때문인데 실제로 생성자에서는 그 부분을 validate 로직을 통해 방어하고 있으니까!)
이렇게 잘못 생성된 직렬화 스트림을 역직렬화 하는 과정에서 readObject를 사용하는 순간 우리가 설정해둔 부분에서 Exception이 발생하면서 방어를 할 수 있어
```java
@DisplayName("ObjectInputValidation을 이용해서 잘못된 직렬화 값을 필터링할 수 있다.")
@Test
void name() throws IOException {
byte[] serializedValidationBytes;
//start가 end 보다 더 뒤에 있는 잘못된 객체의 직렬화 바이트가 있다고 가정한다면
ObjectInputValidation objectInputValidation = new ObjectInputValidation(new Date(2), new Date(1));
try (ByteArrayOutputStream baos = new ByteArrayOutputStream()) {
try (ObjectOutputStream oos = new ObjectOutputStream(baos)) {
oos.writeObject(objectInputValidation);
serializedValidationBytes = baos.toByteArray();
}
}
byte[] deserializedValidationBytes = Base64.getDecoder().decode(Base64.getEncoder().encodeToString(serializedValidationBytes));
try (ByteArrayInputStream bais = new ByteArrayInputStream(deserializedValidationBytes)) {
try (ObjectInputStream ois = new ObjectInputStream(bais)) {
// readObject 하는 단계에서 register 한 validation에 걸린다.
assertThatThrownBy(ois::readObject)
.isInstanceOf(IllegalArgumentException.class);
}
}
}
``` |
ncss-tech/soilDB | 162043715 | Title: automatic conversion of `texture_class` to factor
Question:
username_0: It would be useful to set the factor levels according to a logical sequence.
Answers:
username_1: Could you post a (mockup) use case?
Gut feeling is that it could be pretty difficult to implement given the "intelligence" required to support different texture description systems.
One way to address this could be to:
- Leave to the user to order the levels of their texture data in the `SoilProfileCollection` *before* plotting
- We could facilitate this using a dedicated function
Eg:
```
spc <- order_texture(spc$texture, ...)
plot(spc)
```
username_0: Good point. This would initially be used by the (NRCS-centric) set of functions: `fetchNASIS`, etc. You are correct, there are too many possible classes and several ways of organizing them into a 1D ordering.
I like the idea of a more generic approach--should probably be a function in the aqp package, with options to support multiple sorting schemes.
Status: Issue closed
|
FasterXML/jackson-datatypes-collections | 296841291 | Title: Null is being deserialized into a ListMultimap<String, Optional<Double>>
Question:
username_0: I'm using jackson, jackson-data-type-jdk8, and jackson-datatype-guava at version 2.8.8
It looks like nulls in my JSON object are being deserialized as null instead of Optional.empty().
I wrote a little test that shows what's happening:
```java
@Test
public void testNullDeserialization() throws JsonParseException, JsonMappingException, IOException {
ObjectMapper objectMapper = new ObjectMapper()
.registerModule(new Jdk8Module())
.registerModule(new GuavaModule());
String json = "{\"a\" : [5.0, null, 6.0]}";
ListMultimap<String, Optional<Double>> obj = objectMapper.readValue(
json,
new TypeReference<ListMultimap<String, Optional<Double>>>() {});
assertThat(
obj,
is(ImmutableListMultimap.of(
"a", Optional.of(5.0),
"a", Optional.empty(),
"a", Optional.of(6.0))));
}
```
The output of the junit test is:
java.lang.AssertionError:
Expected: is <{a=[Optional[5.0], Optional.empty, Optional[6.0]]}>
but: was <{a=[Optional[5.0], null, Optional[6.0]]}>
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:8)
Note the middle value is `null` instead of `Optional.empty`.
Looking at the code in [GuavaMultimapDeserializer.java](https://github.com/FasterXML/jackson-datatypes-collections/blob/jackson-datatypes-collections-2.8.8/guava/src/main/java/com/fasterxml/jackson/datatype/guava/deser/multimap/GuavaMultimapDeserializer.java#L161), it looks like line 161 can be replaced with
```java
value = elementDeserializer.getNullValue(ctxt);
```
Or maybe I'm just missing something?
Answers:
username_1: Thank you for reporting this. You may be right here, Guava deserializers are sometimes bit behind standard JDK handlers. One way to see is to have a look at `jackson-databind` equivalents.
I will do that when I have time to work on this, but if you want to try out fix and matching unit test for PR, that'd be great too.
username_0: Yup looks like there's a bug. In ObjectArrayDeserializer I see exactly the line I was referring to
https://github.com/FasterXML/jackson-databind/blob/jackson-databind-2.8.8/src/main/java/com/fasterxml/jackson/databind/deser/std/ObjectArrayDeserializer.java#L194
username_1: Sounds reasonable then. Similar problem may well be present in other Guava container deserializers as well; in many cases new handling has been introduced in minor versions (of `jackson-databind`), but retrofitting all datatype modules may be incomplete.
username_1: Hmmh. Need to change test slightly, to use `AtomicReference` as I don't want to add dependency to java8 module. But should be equivalent as both are reference types, but `AtomicReference` supported by core databind.
username_0: I encountered the same issue using guava’s optional too.
username_1: @username_0 Ah. Good point, yes. Not sure how I missed that one :)
Status: Issue closed
username_0: any chance this could be backported to the 2.7/2.8 branches too?
username_1: @username_0 Unfortunately no, for couple of reasons, mostly because 2.7 and 2.8 are closed for general releases (although micro-patches for critical issues can be made, on per-component basis). But partly since code would have to differ a bit as use of null value handler is new in 2.9.
Although if you really wanted to backport for 2.8, I could help with PR.
username_0: 2.8 would be great if possible
username_1: @username_0 What I mean is that I can help if anyone else wants to work on that -- at this point I do not have time or interest myself.
username_0: ah ok fair enough :). We have a couple older projects that need to be updated from 2.7. I'll work on upgrading and if it's too much of a pain to get it to 2.9 I'll come back and ask you for pointers
username_1: Sounds good. I just didn't want to give impression I'll work on this in case someone else had a chance & need to do that. Good luck and feel free to ask questions on mailing list wrt upgrades, as necessary.
username_2: GuavaImmutableCollectionDeserializer also suffers from this issue - see the following failing test
```java
public void testImmutableListOfOptionals() throws IOException
{
ImmutableList<OptionalInt> list = MAPPER.readValue("[1,null,3]", new TypeReference<ImmutableList<OptionalInt>>() { });
assertEquals(3, list.size());
assertEquals(OptionalInt.of(1), list.get(0));
assertEquals(OptionalInt.empty(), list.get(1));
assertEquals(OptionalInt.of(3), list.get(2));
}
```
username_1: @username_2 could you please file a new issue, adding that test case? This makes it easier to keep release notes clear when one issue tracks one specific problem, fixed in specific version. |
robs23/TestZXing-master | 1021797594 | Title: Fix System.Threading.SemaphoreFullException: Adding the specified count to the semaphore would cause it to exceed its maximum count. in QrHandler+<>c__DisplayClass13_1.<Scan>b__1 ()
Question:
username_0: ### Version 2.1.2(2) ###
### Stacktrace ###
TestZXing.Classes.QrHandler+<>c__DisplayClass13_1.<Scan>b__1 () [0x00052] in <3021c5f1ea784987befa15574c609ff3>:0;TestZXing.Classes
### Reason ###
System.Threading.SemaphoreFullException: Adding the specified count to the semaphore would cause it to exceed its maximum count.
### Link to App Center ###
* [https://appcenter.ms/users/username_0/apps/JDE-Scan/crashes/errors/1638700201u](https://appcenter.ms/users/username_0/apps/JDE-Scan/crashes/errors/1638700201u) |
igvteam/igv | 487693590 | Title: Can add a chevron to the start of the read?
Question:
username_0: right now it's `|---->` but it'd be nice if it was `>---->` instead. The former for some reason has never been intuitive to me when I only see `|---...`, but it would be if it was `>---...`. Let me know if I am not making sense.
Answers:
username_1: @username_0 how are you? The time for this was 10 years and probably around a million screenshots ago. Not only from IGV, this basic look has been copied all over the place. Seriously we would not entertain changing the default look of alignments short of some massive outcry to do so. Perhaps as some optional mode triggered by a preference? Who knows it might catch on.
username_0: One can always dream... thanks for responding.
Status: Issue closed
username_1: @username_0 hey I was serious about a special mode, just not the default view. If its important to you reopen.
username_0: right now it's `|---->` but it'd be nice if it was `>---->` instead. The former for some reason has never been intuitive to me when I only see `|---...`, but it would be if it was `>---...`. Let me know if I am not making sense.
username_0: Sorry, I didn’t realize a special option was on the table. Thanks!
Status: Issue closed
|
ulimartinez/movie-night-front | 685334993 | Title: issueBot
Question:
username_0: The following issue has been updated:
The following issue has been updated:
The following issue has been updated:
The following issue has been updated:
The following issue has been updated:
The following issue has been updated:
The following issue has been updated:
The following issue has been updated:
The following issue has been updated:
Add default image when image not found labeled https://github.com/username_0/movie-night-front/issues/11 labeled https://github.com/username_0/movie-night-front/issues/12 labeled https://github.com/username_0/movie-night-front/issues/14 assigned https://github.com/username_0/movie-night-front/issues/18 labeled https://github.com/username_0/movie-night-front/issues/27 labeled https://github.com/username_0/movie-night-front/issues/101 labeled https://github.com/username_0/movie-night-front/issues/129 labeled https://github.com/username_0/movie-night-front/issues/156 labeled https://github.com/username_0/movie-night-front/issues/180<issue_closed>
Status: Issue closed |
EDCD/EDDI | 513149722 | Title: EDDI doesn't tell body details anymore after ED Update
Question:
username_0: ## EDDI version in which issue is found
EDDI v.3.4.3
## VoiceAttack version in which issue is found (as applicable)
VoiceAttack v1.7.8
Singularity v2.2.4
HCS Plugins v3.3.2
## Steps to reproduce
Fly into a system that has already been discovered by another cmdr and use discovery scanner.
## Expected
If earthlikes, terraformables, ammonia or waterworlds are detected, EDDI should read them out as "mapping candidates". Likewise, ringed planets should also be described for their ring-type.
## Observed
EDDI remains silent throughout the entire scan and doesn't announce mapping-candidates after the system has been fully scanned.
Otherwise, EDDI seems to function normal, announcing system-details while in hyperspace etc.
## Investigation
Behaviour startet after switching ships from a Krait Phantom designated in EDDI as "exploration" to a KRAIT MK II designated as "combat"
Answers:
username_1: This is intended behavior. The script only triggers the additional dialog if you are in a ship with the assigned role of "exploration" or "multipurpose".
Status: Issue closed
username_0: Wouldn't it be nice, if it was also triggered in the "mining" role. Information about ringtypes is relevant. |
mozillazg/python-pinyin | 883464311 | Title: 符号 "|"
Question:
username_0: ## 运行环境
* 操作系统(Linux/macOS/Windows):Linux
* Python 版本:3.7.10
* pypinyin 版本:0.41.0
## 问题描述
符号 "|" 读 gun 无法去除
Answers:
username_1: @username_0 是想覆盖这个字的读音吗?可以参考 https://pypinyin.readthedocs.io/zh_CN/master/usage.html#custom-dict 这个文档里的 `load_single_dict` 函数的用法进行覆盖。
Status: Issue closed
|
ARM-software/armnn | 644248714 | Title: Could you share the test model and Data?
Question:
username_0: I add backend support and i want to test it with the tests in test forder. I build the test and get the test case, but I can find all the model and test data about the test cases.
So ,
Could you share the test model and data ? or Share the link about the model and data?
Thanks.
Answers:
username_0: Hi @FrancisMurtagh-arm ,
Thanks for your feedback,i will test later soon.
Lincoln |
guestful/module.amazon-client | 463345780 | Title: Misuses of cryptographic APIs
Question:
username_0: Hi
The following lines have cryptographic API misuses. <br> File name => src/main/java/com/guestful/client/amazon/AmazonS3Client.java: Line number => 125: API name => SecretKeySpec:<BR> File name => src/main/java/com/guestful/client/amazon/AmazonS3Client.java: Line number => 125: API name => SecretKeySpec:<BR> |
pytorch/xla | 610012838 | Title: multicore training on kaggle gives error
Question:
username_0: When I tried to run this code
```
def simple_map_fn(rank, flags):
torch.manual_seed(1234)
device = xm.xla_device()
print("Process", rank ,"is using", xm.xla_real_devices([str(device)])[0])
xm.rendezvous('init')
flags = {}
xmp.spawn(simple_map_fn, args=(flags,), nprocs=8, start_method='fork')
```
I am getting the below error on kaggle.
```
Exception in device=TPU:3: tensorflow/compiler/xla/xla_client/mesh_service.cc:247 : Failed to meet rendezvous 'init': Socket closed (14)
Exception in device=TPU:4: tensorflow/compiler/xla/xla_client/mesh_service.cc:247 : Failed to meet rendezvous 'init': Socket closed (14)
Traceback (most recent call last):
File "/opt/conda/lib/python3.6/site-packages/torch_xla/distributed/xla_multiprocessing.py", line 225, in _start_fn
Exception in device=TPU:6: tensorflow/compiler/xla/xla_client/mesh_service.cc:247 : Failed to meet rendezvous 'init': Socket closed (14)
Exception in device=TPU:7: tensorflow/compiler/xla/xla_client/mesh_service.cc:247 : Failed to meet rendezvous 'init': Socket closed (14)
File "<ipython-input-11-bda6862c29b1>", line 5, in simple_map_fn
xm.rendezvous('init')
Traceback (most recent call last):
Traceback (most recent call last):
File "/opt/conda/lib/python3.6/site-packages/torch_xla/distributed/xla_multiprocessing.py", line 225, in _start_fn
File "/opt/conda/lib/python3.6/site-packages/torch_xla/distributed/xla_multiprocessing.py", line 225, in _start_fn
Traceback (most recent call last):
File "/opt/conda/lib/python3.6/site-packages/torch_xla/core/xla_model.py", line 614, in rendezvous
return torch_xla._XLAC._xla_rendezvous(get_ordinal(), tag, payload)
RuntimeError: tensorflow/compiler/xla/xla_client/mesh_service.cc:247 : Failed to meet rendezvous 'init': Socket closed (14)
File "/opt/conda/lib/python3.6/site-packages/torch_xla/distributed/xla_multiprocessing.py", line 225, in _start_fn
File "<ipython-input-11-bda6862c29b1>", line 5, in simple_map_fn
xm.rendezvous('init')
File "/opt/conda/lib/python3.6/site-packages/torch_xla/core/xla_model.py", line 614, in rendezvous
return torch_xla._XLAC._xla_rendezvous(get_ordinal(), tag, payload)
File "<ipython-input-11-bda6862c29b1>", line 5, in simple_map_fn
xm.rendezvous('init')
File "<ipython-input-11-bda6862c29b1>", line 5, in simple_map_fn
xm.rendezvous('init')
File "/opt/conda/lib/python3.6/site-packages/torch_xla/core/xla_model.py", line 614, in rendezvous
return torch_xla._XLAC._xla_rendezvous(get_ordinal(), tag, payload)
File "/opt/conda/lib/python3.6/site-packages/torch_xla/core/xla_model.py", line 614, in rendezvous
return torch_xla._XLAC._xla_rendezvous(get_ordinal(), tag, payload)
RuntimeError: tensorflow/compiler/xla/xla_client/mesh_service.cc:247 : Failed to meet rendezvous 'init': Socket closed (14)
```
Answers:
username_1: I have tried this in Google Cloud TPU and it worked.
Is that repeatable every time you run it?
Did you try to restart the runtime?
username_0: For me, it works perfectly in colab also. But when I tried to run this in Kaggle I am getting this error.
Tried nightly, 20200415, and even 20200325 all have same errors but strangely sometimes the code runs fine ( 1 out of 10 times) in Kaggle.
username_1: I tried on Kaggle, and yes, I can repro that.
username_0: A couple of days back Kaggle updated their python from 3.6 to 3.7 at that time xla was not working. Now they rolled back to 3.6 but I am not sure what else they changed.
username_1: Oh, I think it's a timing issue.
Rank 0 (the mesh master) exits first and closes the master service socket.
If you add something like this, it will fix it (of course, this is not a real fix):
```Python
if rank == 0:
time.sleep(1)
```
username_0: Yes, it fixes the issue for now.
Status: Issue closed
|
tomolimo/formvalidation | 170498391 | Title: Tickets restrict tree structure creation
Question:
username_0: Question...
Good afternoon.
How could avoid creating a ticket in a wrong entity?
It happens that I have a tree or entities for which users have no permissions to a particular profile.
However if changed profile (users who have two or more profiles) lets you create tickets in a different entity. The application displays and it indicates that the entity is in tree structure but users generally are not fixed and if they find the selected category. That's when the ticket is in the state (tree structure) but the category is another entity. Therefore creation was wrong.
If I do not explain adequately. Basically I need to know if I can restrict all entities that have access to a user profile to see.
Cordially, Thanks
Answers:
username_0: 
username_0: En la imagen anterior como indico muestra todas las categorias aunque cada categoría está creada en su respectiva entidad.
Como quisiera que fuese sería así:

This means that only list what each entity and not recursively.
The categories are created in each entity and users have permission for each entity. Recursive do not have permission.
Status: Issue closed
username_1: Hello
As far as I have understood your question: it has nothing to do with the firm validation plug-in?
Feel free to re-open the issue if I am wrong.
Thank you
Regards
Tomolimo
username_0: Hello :)
Maybe I did not explain properly ...
Daughters have created categories as appropriate entities. That is, for each sub-entity I create categories corresponding to the sub-entity.
The problem is that if a user has access to several sub-entities and have multiple profiles, changing profile appears to another entity (entity "x" - tree structure). This property allows you to view all categories whether or not that entity in which it is located, allowing you to select category of another entity.
By doing this, the ticket is created in another entity, dragging the new category. For me this is a big problem.
I would like to know if I can disable this option (tree structure) for a particular user or profile .... or any idea how to make this not allowed. The best thing for me would automatically remain in the entity corresponding to the selected category.
The functionality of the tree structure or use is required for the super-admin profile that has all these possibilities, but do not know how to control this comportamieto for other users or profiles or finally works.
It can be very simple but I have not found how or what to do ... :(
I hope I have explained myself better.
username_1: Hello,
This issue list is dedicated to FormValidation plugin issues, not for GLPi issues. You may post to GLPi issues, as this question seems to be related to GLPi and NOT to FormValidation plugin.
Thank you,
Regards,
Tomolimo
username_0: Ok, thanks..
Excuse me... :)
username_1: no problemo |
yiisoft/yii2 | 243198987 | Title: HttpBearerAuth filter is not calling validateAuthKey() of the IdentityInterface
Question:
username_0: ### What steps will reproduce the problem?
Create an API and protect a controller using the HttpBearerAuth filter. Use a custom User Model as Identity:
```php
'user' => [
'class' => 'common\components\UserComponent',
'identityClass' => 'common\models\User',
'enableSession' => false,
'loginUrl' => null,
],
```
Actually I moved the auth_key to a separate table, which is why I defined the `findIdentityByAccessToken()` function as follows.
```php
public static function findIdentityByAccessToken($token, $type = null)
{
return Token::findOne(['auth_key' => $token])->user;
}
```
I did not check the validity of the auth_token at this point, as the name is `findIdentity...` and there is the `validateAuthKey()` function as well, which is implemented as follows:
```php
public function validateAuthKey($authKey)
{
$auth_key = Token::findOne(['auth_key' => $authKey, 'user_id' => $this->id, 'active' => true]);
return $auth_key && $auth_key->valid_until >= time();
}
```
However during my tests it looks like the validateAuthKey is never called and evaluated. If `findIdentityByAccessToken` returns a identity, the login using the auth_token is accepted.
### What is the expected result?
Only looking at the name of the functions, I would have assumed that `validateAuthKey` is used, well, to validate the auth key before it is accepted. Hence I would accept an access request auth an expired auth key to be rejected.
### What do you get instead?
When using an expired auth key, the request is accepted nevertheless. Looking at the actual implementation of `yii\web\User::loginByAccessToken', it assumes the class to be implementing the `IdentityInterface`interface. Hence we can safely assume that the IdentityClass implements the `validateAuthKey()` function, so I would expect this to be used.
### Additional info
| Q | A
| ---------------- | ---
| Yii version | 2.0.12
| PHP version | 7.0.14
| Operating system | Debian 8.6
Answers:
username_0: As a workaround, I created a custom `UserComponent` class extending `yii\web\User` and changing the `loginByAccessToken` method as follows:
```php
public function loginByAccessToken($token, $type = null)
{
/* @var $class IdentityInterface */
$class = $this->identityClass;
$identity = $class::findIdentityByAccessToken($token, $type);
if ($identity && $this->login($identity) && $identity->validateAuthKey($token)) {
return $identity;
} else {
return null;
}
}
```
I simply added ``` && $identity->validateAuthKey($token)```, my suggestion would be to add this to the base class.
username_1: So I guess they where only meant to be used with cookie's related token. Their naming was confusing to me too but I don't think they should be involved with rest because there still many implementations not respecting the recommended stainless nature of it. Also in theory HttpBearerAuth should involve 2 different kind of tokens: refresh and access tokens. Having the same method to validate different kind of tokens for OAuth2 and cookies doesn't look correct.
In such case of not involving cookies I'd suggest disabling them like I have done [here](https://github.com/username_1/yii2-app-builder/blob/master/auth/models/User.php#L66-L80) the same way it was done for `findIdentityByAccessToken()` in the web interface of the advanced template, then different validators should be introduced for OAuth2 or bearer related tokens.
Status: Issue closed
username_2: Access token is not the same as auth key so it doesn't make sense. |
wso2/product-microgateway | 1137162515 | Title: Refactor error logs in enforcer into formalized format
Question:
username_0: ### Describe your problem(s)
Since we introduced a new formalized logging format for error logs in https://github.com/wso2/product-microgateway/issues/2422, we need to refactor some(important) error logs in enforcer as an initial step.
Going forward, with developing new features, we will be gradually converting all error logs into formalized form.
### Describe your solution
Use the following format for errors..
Ex:
```
logger.error("Test error msg", errorLog(LoggingConstants.Severity.CRITICAL, 349000));
```
#### Related Issues:
#2422<issue_closed>
Status: Issue closed |
ClickHouse/ClickHouse | 1009479312 | Title: uniqExact very slow at small table
Question:
username_0: clickhouse-server version: 172.16.31.10
describe:
Millisecond responses are expected, but take tens of seconds
agg_table Distributed engine, agg_table_shard AggregatingMergeTree engine
```sql
-- mv
CREATE MATERIALIZED VIEW xxx.view TO xxx.agg_table_shard on cluster xxxx
AS
SELECT
toDate(ts) AS time,
uniqExactState(uid) AS uniq_uid,
uniqExactState(mid) AS uniq_mid,
e,
s,
tag
FROM xxx.ods_table_shard
GROUP BY
e,
s,
time,
tag
SELECT count()
FROM agg_table;
┌─count()─┐
│ 4222 │
└─────────┘
SELECT
uniqExactMerge(uv_unionid) AS Count,
user_tag AS userTag,
time
FROM xxx.agg_table
WHERE (e = 'p') AND (s = 'g') AND (time >= '2021-09-28') AND (time <= '2021-09-28')
GROUP BY
time,
user_tag
ORDER BY time ASC
127 rows in set. Elapsed: 42.209 sec. Processed 2.56 thousand rows, 1.18 MB (60.72 rows/s., 27.90 KB/s.)
```
logs
i find this line 38.455015483 sec. But I don't know what this line stands for. Why does it take 38 seconds
```shell
[ip] 2021.09.28 16:17:34.461438 [ 116399 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Debug> AggregatingTransform: Aggregated. 234 to 125 rows (from 463.27 KiB) in 38.455015483 sec. (6.085 rows/sec., 12.05 KiB/sec.)
```
detail log:
```shell
clickhouse-client --port 9000 --user default --password default -m -h ip --send_logs_level=trace <<< "query" > /dev/null
[ip] 2021.09.28 16:16:56.001679 [ 40491 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Debug> executeQuery: (from ip:port) SELECT
uniqExactMerge(uv_unionid) AS Count,
user_tag AS userTag,
time
[Truncated]
[ip] 2021.09.28 16:16:56.300727 [ 116399 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Trace> AggregatingTransform: Aggregating
[ip] 2021.09.28 16:16:56.300756 [ 116399 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Trace> Aggregator: Aggregation method: serialized
[ip] 2021.09.28 16:16:56.365159 [ 115043 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Trace> Aggregator: Converting aggregation data to two-level.
[ip] 2021.09.28 16:16:56.367507 [ 115043 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Debug> AggregatingTransform: Aggregated. 118 to 118 rows (from 3.66 KiB) in 0.361097206 sec. (326.782 rows/sec., 10.15 KiB/sec.)
[ip] 2021.09.28 16:16:56.408522 [ 115583 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Trace> Aggregator: Converting aggregation data to two-level.
[ip] 2021.09.28 16:16:56.410620 [ 115583 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Debug> AggregatingTransform: Aggregated. 118 to 118 rows (from 3.68 KiB) in 0.404202936 sec. (291.933 rows/sec., 9.10 KiB/sec.)
[ip] 2021.09.28 16:16:56.434947 [ 116399 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Trace> Aggregator: Converting aggregation data to two-level.
[ip] 2021.09.28 16:16:56.470950 [ 116321 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Debug> AggregatingTransform: Aggregated. 119 to 116 rows (from 15.69 KiB) in 0.464542052 sec. (256.166 rows/sec., 33.78 KiB/sec.)
[ip] 2021.09.28 16:16:56.887515 [ 110867 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Trace> AggregatingTransform: Aggregating
[ip] 2021.09.28 16:16:56.887572 [ 110867 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Trace> Aggregator: Aggregation method: serialized
[ip] 2021.09.28 16:16:56.979097 [ 110867 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Debug> MemoryTracker: Current memory usage (for query): 1.01 GiB.
[ip] 2021.09.28 16:16:57.368816 [ 110867 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Trace> Aggregator: Converting aggregation data to two-level.
[ip] 2021.09.28 16:16:57.372620 [ 110867 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Debug> AggregatingTransform: Aggregated. 118 to 118 rows (from 3.67 KiB) in 1.366195456 sec. (86.371 rows/sec., 2.69 KiB/sec.)
[ip] 2021.09.28 16:17:12.237889 [ 116399 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Debug> MemoryTracker: Current memory usage (for query): 2.00 GiB.
[ip] 2021.09.28 16:17:30.051880 [ 116399 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Debug> MemoryTracker: Current memory usage (for query): 3.02 GiB.
[ip] 2021.09.28 16:17:34.461438 [ 116399 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Debug> AggregatingTransform: Aggregated. 234 to 125 rows (from 463.27 KiB) in 38.455015483 sec. (6.085 rows/sec., 12.05 KiB/sec.)
[ip] 2021.09.28 16:17:34.461453 [ 116399 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Trace> Aggregator: Merging aggregated data
[ip] 2021.09.28 16:17:34.597947 [ 40491 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Information> executeQuery: Read 2657 rows, 1.12 MiB in 38.596216048 sec., 68 rows/sec., 29.78 KiB/sec.
[ip] 2021.09.28 16:17:34.598013 [ 40491 ] {fef17038-660d-41f8-9abd-31f364f6e341} <Debug> MemoryTracker: Peak memory usage (for query): 3.28 GiB.
```
Answers:
username_1: This is the expected behavior of uniqExactState.
In case of uniqExactState you store all data of uid column but in horizontal state instead of column.
There is no sense to use uniqExactState in materialized views.
Try to use uniqState / uniqHLL12State instead of uniqExactState.
If you need exact numbers, don't use materialized views, but query xxx.ods_table_shard directly.
uniqExactMerge can be slower than query uniqExact from the original table (ods_table_shard) because uniqExactStates are huge and your query reads a lot of excessive data because of default index_granularity. There is no DDL for agg_table_shard it hard to say more.
Status: Issue closed
username_0: OK thank you.
DDL:
```sql
CREATE TABLE xxx.agg_table_shard
(
`time` Date COMMENT 'time',
`uid` AggregateFunction(uniqExact, String),
`mid` AggregateFunction(uniqExact, String),
`e` String COMMENT 'String',
`s` String COMMENT 'String',
`tag` String
)
ENGINE = ReplicatedAggregatingMergeTree('/clickhouse/tables/{cluster}/xxx/agg_table_shard/{shard}', '{replica}')
PARTITION BY toYYYYMMDD(time)
PRIMARY KEY time
ORDER BY (time, tag)
TTL time + toIntervalDay(30)
SETTINGS index_granularity = 8192
```
username_1: Your ReplicatedAggregatingMergeTree is incorrect and corrupts data for `e`, `s` columns because they are not in ORDER BY
Also you don't need `time` in ORDERBY/PRIMARY KEY because you have only one value of time in partition.
https://username_1.github.io/Everything_you_should_know_about_materialized_views_commented.pdf
```
CREATE TABLE xxx.agg_table_shard
(
`time` Date COMMENT 'time',
`uid` AggregateFunction(uniqExact, String),
`mid` AggregateFunction(uniqExact, String),
`e` String COMMENT 'String',
`s` String COMMENT 'String',
`tag` String
)
ENGINE = ReplicatedAggregatingMergeTree('/clickhouse/tables/{cluster}/xxx/agg_table_shard/{shard}', '{replica}')
PARTITION BY toYYYYMMDD(time)
PRIMARY KEY (tag,e,s)
ORDER BY (tag,e,s)
TTL time + toIntervalDay(30)
SETTINGS index_granularity = 8192
``` |
tensorflow/tensorflow | 367316609 | Title: Tensorboard slow with S3
Question:
username_0: TensorBoard version - 1.5.1
TensorFlow version if different from TensorBoard - No
OS Platform and version (e.g., Linux Ubuntu 16.04) - Tensorflow installed on EC2 instance of AWS
Description of issue:
I have placed all event files in S3 bucket of AWS cloud. And its content size will be around 6 GB. The problem which I am facing is that the tensorboard is very slow in scanning all the event files. It is taking more than an hour. Is there a way to reduce the processing time.
Answers:
username_1: Thank you for your post. We noticed you have not filled out the following field in the issue template. Could you update them if they are relevant in your case, or leave them as N/A? Thanks.
Have I written custom code
OS Platform and Distribution
TensorFlow installed from
Bazel version
CUDA/cuDNN version
GPU model and memory
Exact command to reproduce
Mobile device
username_2: I am closing this issue since it belongs to Tensorboard related query. I would encourage you to post it on [Tensorboard repository](https://github.com/tensorflow/tensorboard/issues) if you haven't already.
Status: Issue closed
|
jedisct1/libhydrogen | 405724540 | Title: Minimising header size
Question:
username_0: Taking a glance at the secret box code, I didn't figure out if `msg_id` and the context are actually taking up 8 bytes of message header.
I'm designing for a situation with very small messages, so header size matters.
Answers:
username_1: `msg_id` is not present in the header. The context isn't either. These are supposed to be kept in the session state.
Not being present in the header doesn't prevent decryption from failing if incorrect values are being used.
Hope it helps!
Status: Issue closed
|
kiwi-cam/homebridge-broadlink-rm | 748035432 | Title: RangeError, getting SIGTERM...
Question:
username_0: after running for a bit, Hoobs will output this log:
2020-11-21, 4:27:27 PM RangeError [ERR_INVALID_OPT_VALUE]: The value "-47" is invalid for option "size"
at Function.alloc (buffer.js:373:3)
at Socket.<anonymous> (/Users/erezlandau/.hoobs/node_modules/kiwicam-broadlinkjs-rm/index.js:293:39)
at Socket.emit (events.js:315:20)
at UDP.onMessage [as onmessage] (dgram.js:919:8)
2020-11-21, 4:27:27 PM Got SIGTERM, shutting down Bridge...
Iv'e attached my config
[config.json.zip](https://github.com/username_1/homebridge-broadlink-rm/files/5577810/config.json.zip)
Answers:
username_1: This is odd. Basically the plugin is getting a packet that is smaller than expected. I've added some error handling in the latest beta which will ignore these invalid packets.
username_1: This should now be fixed in 4.3.8
Status: Issue closed
username_0: I'm running 4.3.8
still getting the same error.
username_0: I've edited the source code and added the line "if (response.length <= 0x38) return;" at index.js:293, seems to work for now, will report back if it fails again.
username_1: Haha, you found my error (no HEX conversion 🤦♂️) I've just done the same in Git - I'll push a new BETA soon.
username_0: Nice, great to have been of help 👍 |
mob-sakai/ParticleEffectForUGUI | 701740295 | Title: Simulation Space : "World" not working as expected
Question:
username_0: **Describe the bug**
Particle follows root transform when i set Simulation Space to **World**
**To Reproduce**
1. Add UI Particle Script to particle system
2. Change Simulation Space to **World**
**Expected behavior**
Particle stays at emitted position
**Screenshots**

**Environment (please complete the following information):**
- Version 3.0.0-preview.33
- Platform: Editor(Windows)
- Unity version: 2018.4.12f1 [e.g. 2018.2.8f1]
**Additional context**
Add any other context about the problem here.
Answers:
username_1: Could you please attach a minimal project (included `Assets`, `Packages` and `ProjectSettings` directories) that reproduces the issue?
username_0: @username_1
[New Unity Project.zip](https://github.com/username_1/ParticleEffectForUGUI/files/5223797/New.Unity.Project.zip)
username_1: @username_0
Thanks!
username_1: The issue has been reproduced. 👍
username_1: Perhaps it is a problem with the extra world simulation correction in UIParticleUpdater.
username_1: WIP: The issue is fixed.

username_1: @username_0
It is fixed.
Thank you for your help! 👍
Status: Issue closed
username_0: 👍 |
rust-lang/rust | 449767937 | Title: ICE in NLL borrow_check/universal_regions when bounding a trait object with nothing
Question:
username_0: Found as a byproduct of work on https://github.com/rust-lang/rust/issues/61311.
```rust
struct A {}
trait T {}
impl A where dyn T: {
fn f() {}
}
```
backtrace:
```
error: internal compiler error: src/librustc_mir/borrow_check/nll/universal_regions.rs:741: cannot convert `ReEmpty` to a region vid
thread 'rustc' panicked at 'Box<Any>', src/librustc_errors/lib.rs:637:9
note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace.
stack backtrace:
0: std::sys::unix::backtrace::tracing::imp::unwind_backtrace
at src/libstd/sys/unix/backtrace/tracing/gcc_s.rs:39
1: std::sys_common::backtrace::_print
at src/libstd/sys_common/backtrace.rs:71
2: std::panicking::default_hook::{{closure}}
at src/libstd/sys_common/backtrace.rs:59
at src/libstd/panicking.rs:197
3: std::panicking::default_hook
at src/libstd/panicking.rs:211
4: rustc::util::common::panic_hook
5: std::panicking::rust_panic_with_hook
at src/libstd/panicking.rs:478
6: std::panicking::begin_panic
7: rustc_errors::Handler::bug
8: rustc::util::bug::opt_span_bug_fmt::{{closure}}
9: rustc::ty::context::tls::with_opt::{{closure}}
10: rustc::ty::context::tls::with_context_opt
11: rustc::ty::context::tls::with_opt
12: rustc::util::bug::opt_span_bug_fmt
13: rustc::util::bug::bug_fmt
14: rustc_mir::borrow_check::nll::universal_regions::UniversalRegionIndices::to_region_vid::{{closure}}
15: rustc_mir::borrow_check::nll::type_check::free_region_relations::create
16: rustc_mir::borrow_check::nll::type_check::type_check
17: rustc_mir::borrow_check::nll::compute_regions
18: rustc_mir::borrow_check::do_mir_borrowck
19: rustc::ty::context::GlobalCtxt::enter_local
20: rustc_mir::borrow_check::mir_borrowck
21: rustc::ty::query::__query_compute::mir_borrowck
22: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors for rustc::ty::query::queries::mir_borrowck>::compute
23: rustc::dep_graph::graph::DepGraph::with_task_impl
24: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt>::get_query
25: rustc::ty::<impl rustc::ty::context::TyCtxt>::par_body_owners
26: rustc::util::common::time
27: rustc_interface::passes::analysis
28: rustc::ty::query::__query_compute::analysis
29: rustc::ty::query::<impl rustc::ty::query::config::QueryAccessors for rustc::ty::query::queries::analysis>::compute
30: rustc::dep_graph::graph::DepGraph::with_task_impl
31: rustc::ty::query::plumbing::<impl rustc::ty::context::TyCtxt>::get_query
32: rustc::ty::context::tls::enter_global
33: rustc_interface::passes::BoxedGlobalCtxt::access::{{closure}}
34: rustc_interface::passes::create_global_ctxt::{{closure}}
35: rustc_interface::interface::run_compiler_in_existing_thread_pool
[Truncated]
query stack during panic:
#0 [mir_borrowck] processing `A::f`
#1 [analysis] running analysis passes on this crate
end of query stack
error: aborting due to previous error
note: the compiler unexpectedly panicked. this is a bug.
note: we would appreciate a bug report: https://github.com/rust-lang/rust/blob/master/CONTRIBUTING.md#bug-reports
note: rustc 1.36.0-beta.1 (3a5d62bd5 2019-05-22) running on x86_64-unknown-linux-gnu
note: compiler flags: -C codegen-units=1 -C debuginfo=2 --crate-type lib
note: some of the compiler flags provided by cargo are hidden
error: Could not compile `playground`.
```
Compiles on stable 1.35.0.
Answers:
username_1: triage: Looks like matthewjasper is on top of this. Marking P-high just so I don't have to think about it further right now. Removing nomination tag.
Status: Issue closed
|
ManageIQ/manageiq | 121803632 | Title: EVM Server Main Configuration are not being passed to refreshers.
Question:
username_0: When defaults are changed in the EVM Server Main Configuration file via
`Configure/Advanced` the changes are not being passed to the Refreshers.
This can be observed using the rails console on an appliance with:
```ruby
e = ManageIQ::Providers::Amazon::CloudManager.first
r = ManageIQ::Providers::Amazon::CloudManager::Refresher.new([e])
r.refresher_options
# => nil
```
That should not be nil.
Something is preventing the options from being passed to the refreshers.
This is related to the issue reported [1] where gathering EC2 public images fails even
after the default is changed from:
`get_public_images: false` to:
`get_public_images: true`
http://talk.manageiq.org/t/ec2-public-images-capablanca/1093
Answers:
username_0: @miq-bot assign @username_0
Status: Issue closed
|
Pottus/Texture-Studio | 120650265 | Title: Flymode Speed
Question:
username_0: Hi, could you add the ability to increase/lower the flymode camera speed? Like in this video
http://forum.sa-mp.com/showthread.php?t=568931 (check 0:20)
Maybe with the + and - Keys.
Answers:
username_1: Perhaps I could make it toggleable. I love the acceleration, but of course there are those of you that like static speeds.
username_0: Yeah right, sometimes it's just necessary to make it very slow when you want to work very accurate, or increase the speed if you wanna go from Place A to Place B ^_^
username_1: `if you wanna go from Place A to Place B` just use /setloc. It's a command from fsdebug, the filterscript that comes with SA-MP which is loaded by the default .cfg that comes with TS.
username_0: Yeah I know already about setloc but it is based on coordinates - which is useful but not always. Anyways, I can wait till you added my suggestion, ty.
username_1: I'll add three commands, one to set the max speed, one to change the acceleration level, and one to toggle acceleration.
```
/fmspeed
/fmaccel
/fmtoggle
```
Status: Issue closed
|
pyca/cryptography | 394450854 | Title: Potential Security Threat: NVD CVE-2018-10903
Question:
username_0: NVD CVE-2018-10903 published Jul 30, 2018
Because of an exact directory match of phython3-cryptography version 2.3 inside cryptography/docs/ the currently identified risk stems from a flaw that I found in python-cryptography versions between >=1.9.0 and <2.3. The finalize_with_tag API did not enforce a minimum tag length. If a user did not validate the input length before passing it to finalize_with_tag, an attacker could craft an invalid payload with a shortened tag (e.g., 1 byte) such that they would have a 1 in 256 chance of passing the MAC check. GCM tag forgeries can cause key leakage.
- [ ] **Confirmed Patch** https://github.com/pyca/cryptography/pull/4342/commits/688e0f673bfbf43fa898994326c6877f00ab19ef
Answers:
username_1: I don't understand this bug report. You seem to be reporting a CVE to us that we ourselves discovered and requested?
Can you explain what we should be doing with this bug report?
username_2: I think he means that you are using a version of a component that has a known vulnerability. I think he means you should be updating the components in your repo.
Now it looks like the component may have been patched. I think that’s what he means by having a patch. He asking to confirm the patch to mitigate the vulnerability in the CVE.
username_1: We _are_ the project that discovered and filed that CVE. We're not _using_ pyca/cryptography, we _are_ pyca/cryptography.
username_0: Yes, that is our mistake. Keep up the hard work. |
spatie/calendar-links | 608992087 | Title: Allow to use custom formatters/generators
Question:
username_0: Currently it's not possible to use custom generator. An idea that we can add a public method to use any custom generator:
```php
// Link.php
public function formatWith(Generator $generator): string
{
return $generator->generate($this);
}
```<issue_closed>
Status: Issue closed |
valor-software/ng2-file-upload | 223078560 | Title: Migration Angular4
Question:
username_0: Is this lib dead ?
It could really do with fixing :
npm WARN [email protected] requires a peer of @angular/common@^2.3.0 but none was installed.
npm WARN [email protected] requires a peer of @angular/core@^2.3.0 but none was installed.
Answers:
username_1: It was updated. Using it in an ng 4 project and it gets the job done.
Status: Issue closed
|
Azure/MachineLearningNotebooks | 381720365 | Title: All recipes in /training should have covered all scenario
Question:
username_0: All samples should cover uploading training data to datastore and writing output models to ./outputs. Otherwise there will be a few blind spots for users.
I'd be happy to help modify the existing samples to include those.<issue_closed>
Status: Issue closed |
rgrove/rawgit | 318790438 | Title: Rawgit redirects to raw.githubusercontent.com
Question:
username_0: Good chance it's something wrong that I've done here. Why is `rawgit.com` redirecting to `raw.githubusercontent.com` in this case?
```
❯❯❯ curl https://rawgit.com/ruby/ruby/commit/b3dbeb6e90f316584f70e33f6bfb9d83fa5f30d3.patch
<p>Moved Permanently. Redirecting to <a href="https://raw.githubusercontent.com/ruby/ruby/commit/b3dbeb6e90f316584f70e33f6bfb9d83fa5f30d3.patch">https://raw.githubusercontent.com/ruby/ruby/commit/b3dbeb6e90f316584f70e33f6bfb9d83fa5f30d3.patch</a></p>
```
Answers:
username_1: There are several things happening here:
1. `/ruby/ruby/commit/b3dbeb6e90f316584f70e33f6bfb9d83fa5f30d3.patch` doesn't actually point to a file in a repo, it points to a dynamic endpoint on GitHub that generates a patch file for a specific commit. RawGit only supports fetching actual repo files.
2. Normally RawGit would just return a 404 for paths it doesn't support. But in this case the path actually _looks_ like it could be a file path, so RawGit tries to process it as one.
3. `.patch` isn't a [whitelisted file extension](https://github.com/username_1/rawgit/blob/master/conf/index.js#L50-L55), so rawgit.com assumes there's no benefit to proxying it and redirects you directly to GitHub to load it. But the redirect assumes the path is an actual repo file. raw.githubusercontent.com only knows how to handle repo files, so it returns a 404 because the path we're requesting is a dynamic endpoint, not a repo file.
4. cdn.rawgit.com doesn't use the extension whitelist, so it just tries to load the file from GitHub. But once again, since the path doesn't point to an actual file, we get a 404.
tl;dr: RawGit only supports repo files, not dynamic GitHub endpoints, and `ruby/ruby/commit/b3dbeb6e90f316584f70e33f6bfb9d83fa5f30d3.patch` isn't a repo file. Sorry.
We could do a better job of recognizing that `/:user/:repo/commit/*` isn't a file URL though, so I'll leave this open as a bug.
username_0: Thanks for the explanation, Ryan. Gisting the patch first and then using RawGit with the gist worked just fine. |
Quantum-Cloud/PurpleAmerica | 151768578 | Title: Color control
Question:
username_0: R, G, B color indicator should properly be found using the given statistics.
Other necessary color functions should be working independent of other classes.
Answers:
username_1: I have the color function working. I'm going to add a function so that the user can pick different color gradients for the states.
Status: Issue closed
|
django-compressor/django-compressor | 212407408 | Title: Should collectstatic be run before offline compress?
Question:
username_0: As the title really. I can't see in the documentation if `manage.py compress` does the job of `manage.p collectstatic` implicitly, or if one should be run before the other.
Answers:
username_1: Hi @username_0,
`compress` does not run `collectstatic` so a deploy script needs to do both. My playbook for this runs `collectstatic` first, but (IIRC—it's late) compress uses `STATICFILE_FINDERS` so the order shouldn't matter.
Status: Issue closed
|
Beckhoff/ADS | 329339282 | Title: Create handle failed with: 0x745 (1861)
Question:
username_0: I am trying to run the example.cpp provided by the repository on beagle bone black, where Linux is installed. But I get the error code 0x745 (1861). As far as I understand, this repository is used to communicate with Beckhoff device through ADS without TwinCat installed. All I need is to add a route before I start communicating. But how can I add a route without TwinCat installed on my Linux?
Thanks a lot!
Henry
Answers:
username_1: For details please read my responses to #14. Here is a short summary:
1. Make sure to assign a deterministic AmsNetId on your AdsLib client -> AdsSetLocalAddress()
2. In your AdsLib client add a route to your TwinCAT system -> AdsAddRoute()
3. On your TwinCAT system add a route for your AdsLib client
Status: Issue closed
|
grunwaldlab/metacoder | 328095658 | Title: A warning message from 'parse_mothur_tax_summary'
Question:
username_0: <Taxmap>
1673 taxa: aab. Root, aac. Bacteria ... cmi. Roseibacillus, cmj. WPS-2_ge
1673 edges: NA->aab, aab->aac, aac->aad, aac->aae ... bcy->cmg, bcz->cmh, bcz->cmi, bda->cmj
1 data sets:
tax_data:
# A tibble: 1,673 x 16
taxon_id taxlevel rankID taxon daughterlevels total TAS1 TAS10 TAS2 TAS3 TAS4 TAS5
<chr> <int> <chr> <chr> <int> <int> <int> <int> <int> <int> <int> <int>
1 aab 0 0 Root 1 61163 4699 6993 4105 5969 7104 3154
2 aac 1 0.1 Bacteria 31 61163 4699 6993 4105 5969 7104 3154
3 aad 2 0.1.1 Acidoba~ 8 890 4 4 0 3 14 639
# ... with 1,670 more rows, and 4 more variables: TAS6 <int>, TAS7 <int>, TAS8 <int>,
# TAS9 <int>
0 functions:
Warning message:
Naming table columns/vectors/lists the same name as functions can sometimes interfere with non-standard evaluation. The following data shares names with functions:
data$tax_data$taxon
Answers:
username_1: Hello @username_0, thanks for the report. You don't need to worry about that warning. At one point we had a problem with non-standard evaluation confusing external variables with column names, but I think that is fixed now. We will be removing removing that warning from the next release.
username_0: Oh, I see. Thanks for your reply.
Status: Issue closed
|
Esri/arcgis-osm-editor | 173292029 | Title: Missing "Repair Geometry" step in OSM File Loader (Load only) tool
Question:
username_0: Hi @username_1 ,
I have now attempted to load both the Netherlands and Denmark with the latest build of the OSM File Loader. However, when I rendered the data, I noticed issues with the display of some of the Polygon geometries in the Denmark dataset.
I then ran the Repair Geometry tool against the data, and this revealed close to 10000 faulty geometries on the Denmark dataset, primarily, but not exclusively, "ring order issues". The latter not surprisingly, as they directly relate to the non-compulsory ring order in OSM datasets and multipolygons or closed ways defined as "area" via the area key. I don't know yet how many geometries are bad in the Netherlands import, because the Repair Geometry step is only half way, but I do some issues listed in the progress dialog (which only shows a fraction, the rest is written to a text file at the end of the process).
Did you accidentally comment out a Repair Geometry step in the latest build, that needs to be activated again?
Answers:
username_1: The step is still in there, but one of the helper classes had expired.
Status: Issue closed
username_0: OK, thanks, I'll close this then. |
guofei9987/scikit-opt | 981326681 | Title: Google colab GPU环境下导包报错
Question:
username_0: RuntimeError Traceback (most recent call last)
<ipython-input-55-23e00a0a7d32> in <module>()
----> 1 from sko.PSO import PSO
2 from sko.GA import GA
4 frames
/usr/local/lib/python3.7/dist-packages/sko/__init__.py in <module>()
1 __version__ = '0.6.5'
2
----> 3 from . import DE, GA, PSO, SA, ACA, AFSA, IA, tools
4
5
/usr/local/lib/python3.7/dist-packages/sko/DE.py in <module>()
9 from abc import ABCMeta, abstractmethod
10 from .operators import crossover, mutation, ranking, selection
---> 11 from .GA import GeneticAlgorithmBase, GA
12
13
/usr/local/lib/python3.7/dist-packages/sko/GA.py in <module>()
7 import numpy as np
8 from .base import SkoBase
----> 9 from sko.tools import func_transformer
10 from abc import ABCMeta, abstractmethod
11 from .operators import crossover, mutation, ranking, selection
/usr/local/lib/python3.7/dist-packages/sko/tools.py in <module>()
7
8 if sys.platform != 'win32':
----> 9 multiprocessing.set_start_method('fork')
10
11
/usr/lib/python3.7/multiprocessing/context.py in set_start_method(self, method, force)
240 def set_start_method(self, method, force=False):
241 if self._actual_context is not None and not force:
--> 242 raise RuntimeError('context has already been set')
243 if method is None and force:
244 self._actual_context = None
RuntimeError: context has already been set
Answers:
username_1: 我没有这个环境,能否帮忙看一下 `sys.platform` 返回的是什么吗? |
StirfireStudios/Jacquard | 334779055 | Title: Nodes: Use screen space more efficiently
Question:
username_0: Currently, even on large screens, Jacquard's Node list has a lot of wasted whitespace. It would be more efficient to decrease the default padding, and to only increase it when required.
Status: Issue closed
Answers:
username_1: Check release [0.6.0](https://github.com/StirfireStudios/Jacquard/releases/edit/0.6.0) |
kimtm60/SW2016 | 154397368 | Title: index.jsp 수정
Question:
username_0: colgroup을 이용하여 가운데 정렬 과
H_LIB 가운데로 크게 정렬
Answers:
username_0: **hi**
username_0: 1. Make my changes
1. Fix bug (스페이스 두 번)
2. Improve formatting
* Make the headings bigger
2. Push my commits to GitHub
3. Open a pull request
* Describe my changes
* Mention all the members of my team
* Ask for feedback
username_0: 1.ho
1.kim
2.kim
username_0: 1.ho
1.kim
2.ho
username_0: 1.ho
1.kim
2.ho
username_0: hoho
hoho
hoho
hoho
username_0: ho
ho
ho
ho
username_0: commit ho |
tylerbrawl/Galaxy-Plugin-Rockstar | 507153174 | Title: Refreshing authentication, research share
Question:
username_0: Hey,
i would like to share with you my research regarding possible refreshing of credentials, i couldn't progress with it any further but it might be of help to you.
Regarding the fingerprint i've had yet to decipher which information in it is required so there is probably a lot of unnecessary stuff passed there.
I don't remember most of the specifics right now of why i've did something like this and not differently but i will come back to you in case i remember/research anything else.
```python
@dataclass
class Token(object):
_token: str = None
_expires: float = None
def set_token(self, token, expiration_time):
self._token, self._expires = token, expiration_time
def get_token(self):
return self._token
@property
def expired(self):
return self._expires <= time()
```
```python
import logging as log
import aiohttp
import asyncio
from base64 import b64encode
from galaxy.http import create_client_session
from yarl import URL
from time import time
from galaxy.api.errors import (
InvalidCredentials, AuthenticationRequired
)
from definitions import Token
import pickle
class CookieJar(aiohttp.CookieJar):
def __init__(self):
super().__init__()
self._cookies_updated_callback = None
def set_cookies_updated_callback(self, callback):
self._cookies_updated_callback = callback
def update_cookies(self, cookies, url=URL()):
super().update_cookies(cookies, url)
if cookies and self._cookies_updated_callback:
self._cookies_updated_callback(list(self))
class AuthenticatedHttpClient:
def __init__(self, store_credentials_callback):
self._refresh_token = Token()
self._access_token = Token()
[Truncated]
}
response = await self._session.request('GET', "https://www.rockstargames.com/auth/get-user.json", headers=headers)
response = await response.json()
if not response['loggedIn']:
log.error(f'loggedIn: {response["loggedIn"]}')
raise InvalidCredentials()
access_token = response['user']['auth_token']['access_token']
expires = (response['user']['auth_token']['expires_in'] * 60) + time()
self._access_token.set_token(access_token, expires)
refresh_token= response['user']['auth_token']['refresh_token']
expires = (response['user']['auth_token']['refresh_expires_in'] * 60) + time()
self._refresh_token.set_token(refresh_token, expires)
self.user = response['user']
```
Hope this is of any help
Answers:
username_1: Thank you for this write-up. As you mentioned, the main issue here is finding out what, if any, parts of the fingerprint can be "faked." I will see what I can do as time goes on, but I expect this to take some time to figure out.
username_1: It appears as though the Rockstar website is using the JavaScript module fingerprintjs2 for browser fingerprinting. Since the Python Requests module does not support JavaScript, we would have to try making a default fingerprint. If that does not work, then we would have to fall back to other options (such as Selenium).
username_1: I found out some more information regarding fingerprint generation. The encrypted values within the fingerprint (user_agent, regular_plugins, canvas, webgl, and js_fonts) are found by returning their corresponding x64 128-bit MurmurHash3 value using the optional seed of 31. I will see what more I can find later on.
username_1: I believe the remember me (or refresh) token is bound in some way to the generated fingerprint, so it seems like faking the fingerprint is not an option. Instead, I am going to use the JavaScript functionality of NextStep to replicate as best as possible the fingerprint generation that Rockstar uses.
username_1: Great news, everybody: I found out how to replicate the generation of the fingerprint. It worked fine on my system, but it would be appreciated if others could test it, as well. I will include a .zip file containing the needed test files, if you have a Windows computer and would like to try it for yourself. Here are the steps for proper testing:
1. Extract the contents of fingerprint-test.zip (located at the bottom of this comment) into one single folder.
2. Open JS-Test.html **with Google Chrome** and take note of the fingerprint information that appears shortly after opening it.
3. With the Google Chrome developer tools open, visit https://www.rockstargames.com and inspect the network traffic at https://signin.rockstargames.com/connect/cors/check/rsg. Specifically, check the Request Payload information that is sent during the POST request to this endpoint.
4. Verify that the fingerprint that appears within the Request Payload matches that generated by JS-Test.html.
The reason that this needs to be done on Google Chrome with a Windows PC is because I explicitly set the device_name key to be "Chrome on Windows." (This is acceptable, since Galaxy 2.0 uses a version of Google Chrome during its NextStep call, and because the Rockstar Games Launcher is only available on Windows. If the launcher ever releases on Mac, or if the demand for a Social Club-only integration is high enough, then I can change the plugin.)
[**fingerprint-test.zip**](https://github.com/username_1/Galaxy-Plugin-Rockstar/files/3742389/fingerprint-test.zip)
username_0: yup, they are the same, except on testjs there is no "device_name":"Chrome on Windows".
Good job!
username_1: That's great to hear. I will begin integrating the JavaScript into the plugin.
username_1: I believe that I have finally fixed re-authentication. If it is working correctly, then the user should remain authenticated for 20 years, but I would request some additional testing before releasing it fully. I have attached the latest binaries below. If the re-authentication works correctly for you, then I will send out this build as the final release of v0.2.
[plugin.zip](https://github.com/username_1/Galaxy-Plugin-Rockstar/files/3753659/plugin.zip)
Status: Issue closed
username_1: I have just confirmed for myself that the user is indeed successfully re-authenticated after the 10-hour access token expires. As such, I will go ahead and close this issue and release v0.2 to the public as a final release. Thanks again for all of your help!
username_0: I've played for some time with 0.2 version and unfortunately it's still loosing it's credentials. This time takes longer -> ~24h or so. Attached logs:
[plugin-rockstar-774732b5-69c4-405c-b6c9-92cd55740cfe.zip](https://github.com/username_1/Galaxy-Plugin-Rockstar/files/3767978/plugin-rockstar-774732b5-69c4-405c-b6c9-92cd55740cfe.zip)
username_1: Hey,
i would like to share with you my research regarding possible refreshing of credentials, i couldn't progress with it any further but it might be of help to you.
Regarding the fingerprint i've had yet to decipher which information in it is required so there is probably a lot of unnecessary stuff passed there.
I don't remember most of the specifics right now of why i've did something like this and not differently but i will come back to you in case i remember/research anything else.
```python
@dataclass
class Token(object):
_token: str = None
_expires: float = None
def set_token(self, token, expiration_time):
self._token, self._expires = token, expiration_time
def get_token(self):
return self._token
@property
def expired(self):
return self._expires <= time()
```
```python
import logging as log
import aiohttp
import asyncio
from base64 import b64encode
from galaxy.http import create_client_session
from yarl import URL
from time import time
from galaxy.api.errors import (
InvalidCredentials, AuthenticationRequired
)
from definitions import Token
import pickle
class CookieJar(aiohttp.CookieJar):
def __init__(self):
super().__init__()
self._cookies_updated_callback = None
def set_cookies_updated_callback(self, callback):
self._cookies_updated_callback = callback
def update_cookies(self, cookies, url=URL()):
super().update_cookies(cookies, url)
if cookies and self._cookies_updated_callback:
self._cookies_updated_callback(list(self))
class AuthenticatedHttpClient:
def __init__(self, store_credentials_callback):
self._refresh_token = Token()
self._access_token = Token()
[Truncated]
}
response = await self._session.request('GET', "https://www.rockstargames.com/auth/get-user.json", headers=headers)
response = await response.json()
if not response['loggedIn']:
log.error(f'loggedIn: {response["loggedIn"]}')
raise InvalidCredentials()
access_token = response['user']['auth_token']['access_token']
expires = (response['user']['auth_token']['expires_in'] * 60) + time()
self._access_token.set_token(access_token, expires)
refresh_token= response['user']['auth_token']['refresh_token']
expires = (response['user']['auth_token']['refresh_expires_in'] * 60) + time()
self._refresh_token.set_token(refresh_token, expires)
self.user = response['user']
```
Hope this is of any help
username_1: That's strange. I'll be sure to take a look at that.
username_1: Are you using two-factor authentication? Logging in without it, it seems that the RMT is only needed for people using it (as the value is blank for users not using two-factor authentication), while the rsso cookie is needed for every account.
username_1: Assuming that the issue was because of two-factor authentication being disabled, I have created a fix for that. The below build incorporates the proposed fix. If you still have errors with this build, then please let me know.
[plugin-auth-test.zip](https://github.com/username_1/Galaxy-Plugin-Rockstar/files/3769731/plugin-auth-test.zip)
username_2: What I have noticed with this issue is that my credentials appear to be stored in GOG 2.0, but as soon as I log in to the Rockstar Social Club website, my session in GOG 2.0 is expired and I have to re-scan my library for any additional games.
I don't have two factor enabled on my Rockstar account yet. Let me know how I can help test this issue (I can hold off on enabling two factor if you want me to try to run that zip file).
username_1: Did you download the new v0.2.1 release? Either download that or the plugin-auth-test.zip file above (although I suggest that you instead download the v0.2.1 release from [here](https://github.com/username_1/Galaxy-Plugin-Rockstar/releases/tag/v0.2.1)).
username_2: I downloaded that after I saw your comment and it fixed the issue, thanks. It still seems to not be tracking any of my playtime whereas when I manually added the .exe before using the plugin, it did, but I can make a separate issue for that.
Status: Issue closed
username_1: Yes, that should be on a separate issue. Still, I appreciate your testing regarding the re-authentication. Since you confirm that the issue is fixed, I will close this issue once more. |
Eomys/SciDataTool | 774740594 | Title: get DataND object as return value
Question:
username_0: Hello,
it seems that there is no equivalent method, e.g. to get_along or interpolate, that return DataND object.
I think this could be handy, e.g. to do a FFT on a reduced set of indices. At the moment one has to extract field and axes data, slice the data and axes and setup a new DataND object to do that.
Best regards, Sebastian
Answers:
username_1: Hello Sebastian,
Indeed, such methods had not been implemented yet. It is not such a simple issue in the case of periodicity. However, as a first step, the new DataND object could be rebuilt ignoring symmetries, simply using the axis values extracted during the `get_along` call. Of course one loses the compacity of linspace and periodic objects, but it should work as a first work around.
I created a new `get_data_along` method to do so, using the exact same syntax as `get_along` (see commit [#3ec612a](https://github.com/Eomys/SciDataTool/commit/3ec612a2e36dbd1175a6def76ebcfd0a924a42a4)). It should also work with Fourier transforms. If you want to improve it to take periodicities and linspaces into account, feel free to modify this method (I do not have much time to work on this issue for the moment).
Let me know if this new method meets your needs.
Best regards, Helene
username_0: Hello Helene,
thank you, thats okay for now. I wanted to use this method in pyleecans Solution class for the same purpose.
BTW: There is no class generator. Is it possible to use pyleecans gen. 'out of the box'?
BR Sebastian
username_1: Yes, I use pyleecan's class generator (by copy/pasting the csv files into the ClassesRef folder). I then have to manually remove from the class files the lines concerning the logger, which is not available in SciDataTool, and I replace all occurences of "pyleecan" by "SciDataTool".
Status: Issue closed
|
binary-com/binary-bot | 666869365 | Title: Can anybody help me to achieve maxloss
Question:
username_0: I made a bot from reading all here, I have little issue, I will really appreciate if someone help me in this
I have achieved
if there is 4 consecutive losses then run every 100 seconds and loss counter to 0
I want to achieve
if after above condition and loss continues and reach to 8 consecutive losses then bot stop and print Stop loss hit

Answers:
username_0: I have figured it out. I am posting if someone else might get an idea


Status: Issue closed
|
vasturiano/force-graph | 440247387 | Title: Reheating the simulation
Question:
username_0: Hi Vasturiano!
Thanks for your amazing library, super useful!
In some use cases I want to reheat the force simulation after it has started or change some settings on the fly. Often this requires access to the main simulation object, e.g. by saying simulation.alpha(1).restart(), as per d3.
I guess in general it's very useful to be able to access the main simulation object in case you need it. Is this possible or is there a way this can be done or to expose it?
Any help is much appreciated :-)
Answers:
username_1: Thanks for reaching out! Using the latest version (`1.15.0`) you can simply do `myGraph.refresh()` to reheat the force simulation engine.
As for the other configuration parameters, there are exposed properties that lets you manipulate this, such as `d3AlphaDecay`, `d3VelocityDecay`, `d3Force`, `warmupTicks` and `cooldownTicks`. If there's somehow an operation you're not able to control using these methods, please let me know and I'll consider adding additional support for.
username_0: Hi,
thanks for the update! With respect to the simulation, it would be nice to be able to manually control the tick function e.g. to say `simulation.tick(20)`. It would also be handy to redraw the canvas, say because you've changed the color for a node. Currently I do this by using a function inside `.nodeCanvasObject` that itself depends on a settings object. Hence if I change the settings object, the function changes the color. In that scenario, what would be the best way to redraw the canvas (without changing node positions etc.)?
PS: I'm using your package in vue.js, which works quite well. No fancy binding, inside a component after mount I just create an instance of `ForceGraph` and define a few helper methods on the vue object that essentially call a few methods on a `ForceGraph` instance. For instance, now there is no zoom to fit function that zooms the canvas such that all nodes are inside the view port, so I created my own. A function like that would be useful to others as well I guess.
Anyways, thanks for you help, much appreciated! :-)
username_2: Hi @username_0, I'm looking into using this with vue as well, and was just wondering if you might be interested in sharing your own vue implementation for this?
username_1: The issue with exposing the context is that you wouldn't be able to do much manipulation with it, since it is cleared at every raf tick. Anything you'd write to it would be wiped almost immediately in the following tick. This is why it's not exposed, and all the relevant draws need to happen in the `nodeCanvasObject` method. If you'd like to have some other persistent visual elements I'd suggest to have another canvas that you maintain externally that aligns exactly underneath (or above it) to this dom element. That way you're just layering the multiple canvases and you get better update performance by not repetitively drawing the same things.
As for `vue.js` I can't be of much help because I haven't attempted the integration, but feel free to post examples here for the benefit of other users that may want to do the same thing. :)
username_0: @username_1 Thanks for your comments!
I'm happy to offer some bigger, more worked out examples for others, that also include [vue](https://vuejs.org/) and [vuetify](https://vuetifyjs.com/en/) so others can see how your awesome package can be applied in those contexts. One nice thing of the vuetify docs is that they have a codepen for all the examples, which is especially nice if you want to play around with the code and experiment. If you want I can help in setting up a bunch of them for your package based on your existing code examples.
With respect to the `ctx` canvas context, previously I was not aware the `requestAnimationFrame` loop keeps running at all times unless you explicitly tell it not to via `pauseAnimation`.
Especially if you don't use the edge particles and you have largely unchanging data there is little need to keep the event loop running after the simulation freezes. For instance, currently on a graph of e.g. 200 nodes, even after the simulation freezes, the `nodeCanvasObject` callback will be executed approximately 200 * 60 = 12.000 per second, even when the data and the layout, i.e. the visual, do not change.
It's a bit unfortunate that currently when you apply `pauseAnimation` all event propagation always stops. I can see that in the generic scenario that needs to be done. However, in the above scenario, instead of wiping the canvas, it would be nice to be able to keep it alive after the layout freezes. That way you can still get the colors from the canvas to be able to generate hover events. When applying a drag move one could then quickly call `resumeAnimation` to facilitate the drag move and automatically call `pauseAnimation` after the layout stabilizes. Doing so will be less taxing for the browser, which helps in making additional transitions less choppy, especially when there are a lot of nodes in the network.
username_1: @username_0 good points. It's a little difficult to keep track of the variety of scenarios that the simulation could go dormant, even though there is certainly room for performance improvement.
Currently the decision is to let the consumer handle that state by choosing when to call `pauseAnimation` and `resumeAnimation`. A possible improvement is to decouple that from the pointer interaction events, so that one could still detect hover interactions and resume/pause the animation engine to permit node dragging. Actually the fact that dragging is not possible when the animation engine is off is the main reason why the pointer interactions are also disabled on `pauseAnimation`.
As for vue.js I would love to see a bindings repo, similar to what there is for React at [react-force-graph](https://github.com/username_1/react-force-graph).
username_0: @username_1 I think it would be really helpful to decouple the pointer interaction events from the `pauseAnimation` and `resumeAnimation` functionality. I was thinking, wouldn't it be possible to use a `mousedown` or say an invisible svg element at the location of the mouse cursor to signal the start of a drag move in case the rendering cycle of the component is paused?
username_3: @username_1 I would also be interested in decoupling the pointer events from pause and resumeAnimation functions.
Is there any (planned) progress on this?
Mainly I would like to render the graph only once it's done with positioning, allow the user to add nodes through click (triggering a rerender) and only ever show the result of the finished positioning and not showing the steps to reach the new position.
username_1: @username_3 you can achieve this by setting `cooldownTicks(0)` and `warmupTicks(80)` (adjust value to your case). This will essentially compute the final node positions in the warmup phase, and bypass the cooldown phase where the graph build-up is normally being animated.
username_3: @username_1 great thanks!
And as for the pointer events. I would like to pauseAnimations onEngineStop. Such that neither the linkCanvasObject nor the nodeCanvasObject get called after the enginestop but click/hover interactions can still be caught. (on clicking/dragging a node animations should resume)
I tried
`this.graph.onEngineStop(() => {
this.graph.pauseAnimation()
})`
but the linkCanvasObject and nodeCanvasObject callbacks keep being called and interactions were not deactivated.
username_1: @username_3 if your overall intent is to improve the graph performance by bypassing rerenders after the simulation engine has stopped, note that there's been a flag added on `v1.38.0` that addresses exactly this.
The option is `autoPauseRedraw` and it's enabled by default, so you should be automatically getting the performance improvement, if you upgrade your component version.
More details about it in [this issue](https://github.com/username_1/force-graph/issues/169).
username_3: @username_1 Oh perfect!! I saw that option but thought my current version already had that. Now that I upgraded it works really smoothly. Thanks so much, Great library! |
brownplt/pyret-lang | 145530395 | Title: "checking/synthesis for s-extend not implemented"
Question:
username_0: It appears as though type checking has not been implemented for s-extend expressions. For example,
```
{}.{a: 1}
```
produces an "checking for s-extend not implemented" error.
I believe extending the type checker to s-extend expressions should be relatively painless. At least synthesis is trivial: to synthesize a type for `A.{a: e1, b: e2, ...}` we simply synthesize a type for A, and then synthesize types for each `e1`, `e2`, and finally stitch everything together. Checking a type against an s-extend should simply amount to a subtype rule.
Answers:
username_1: That makes a few assumptions that aren't necessarily true: whether we support this kind of subtyping by default in synthesis or not, whether stitching is that straightforward (what about the self parameters of methods?), whether the benefits outweigh the costs (do we expect students to typically want to add new fields to an object they've already typed?), etc. @username_2 should chime in here.
username_2: There are a couple reasons s-extend isn't implemented yet. One is that the story for objects isn't done, namely in terms of methods. Because of the `self` argument we need to be careful about overriding field values.
However, this should be coming soon as I've almost figured out typing for objects with methods.
username_0: Ah, yes. Methods do mess things up a bit.
Still, objects are just big `letrec`s so it shouldn't be too tricky. How about synthesizing as follows,
* create an existential variable â for the object,
* desugar each of the methods into lambdas, and annotate the `self` argument to each of the lambdas with type â,
* synthesize all of the fields including the methods, producing an object type A,
* apply `A <: â`
Status: Issue closed
|
beingmrkenny/ListHighlighterForTrello | 301565621 | Title: Activating this extension overrides title card formatting
Question:
username_0: I am using an userscript to be able to format the title of my cards with markdown. If this extension is active, the formatting appears correctly and instantly it is formatted back as normal text. The userscript I am using is https://gist.github.com/gorbiz/6062481
Answers:
username_1: Hi Adrián,
Thanks for emailing me. I’m working on an update which should hopefully fix this issue. I hope to release it tomorrow.
Thanks,
Mark
Status: Issue closed
username_1: Hi, 3.2.2 is out now and resolves this issue :)
username_0: Thanks a lot! It is indeed fixed. |
libp2p/rust-libp2p | 313698978 | Title: Additional transports
Question:
username_0: - [ ] IPC
- [ ] Bluetooth
- [ ] NFC
These are all interesting because they have their own discovery mechanisms that don't require Kademlia and in fact discovered addresses should not be propagated over "globally addressable" transports like IP since they would be almost entirely false positives. We could propagate bluetooth addresses to other bluetooth devices (to prevent an `identify` call) but I don't think that's necessary since we already narrow down the number of devices that we need to connect to by having the concept of "pairing", where the user selects which devices to trust.
I think that a good API would be for these to each have a `Controller` struct that has a `fn scan` taking a `Fn(impl Stream<Item = Client>) -> impl Stream<Item = Client>`, where `Client` is a struct containing information on a discovered client and the output is a list of clients to attempt connection to. For a chat app, this function could hand off control to a UI dialog and then return the clients that the user has chosen.
Answers:
username_1: Bluetooth: https://people.csail.mit.edu/rudolph/Teaching/Articles/BTBook-march.pdf chapter 3.2
username_2: NFC would be pretty cool!
Combined with libp2p compiled to WASM, I guess you could then somehow pair two mobile phones via NFC and use bluetooth as the transport?
username_3: I think ipc is a good idea. |
WikiEducationFoundation/WikiEduDashboard | 258909013 | Title: Article titlle should not say 'null' for Wikidata items, etc
Question:
username_0: When an article is shown that has a wiki with no language attribute, such as wikidata or multilingual wikisource, the dashboard shows 'null' as the language code. Instead, should omit the language code in this case.

Answers:
username_1: I would like to pick this as my first issue :)
username_1: Hi! I would like to pick this as my first contribution to the program but unfortunately cannot reproduce the bug. Could someone help me? Thanks.
username_0: You can reproduce it by creating a course, adding a user who has been active in Wikidata (e.g., Ragesoss if you set the course to start about a month ago), adding an available article from Wikidata to connect the course to that wiki, and then visiting [course page]/manual_update.
username_1: I created a PR for a possible fix.
username_2: @username_1 Hi
Could it be simple check for `null `or `undefined ` in `shouldShowLanguagePrefix()`?
seems like it is the place responsible for "should omit the language code in this case"
`if (!this.props.article.language) { return false; }`
or
`if (this.props.article.language == undefined) { return false; }`
`== null` returns `true` if `this.props.article.language` is also `undefined`
`== undefined` returns `true` if `this.props.article.language` is also `null`
username_1: @username_2 indeed that was a better solution as it's more consistent with the logic of the whole article.jsx and prevents for false positives. I already did a commit for it and it's already added to my PR :) https://github.com/WikiEducationFoundation/WikiEduDashboard/pull/1413
username_2: @username_1 there could be a bug introduced with your previous commits - so Travis CI build failed.
I'm not a specialist in git PRs: i would get rid of previous commits and left only the last meaningfull one.
And seems like `!this.props.article.language` does that check for `null` and `undefined`. Try to play in dev cosole in a browser.
username_0: You can look at the CI build log to see the error. In this case, it's a javascript eslint error: `Unexpected whitespace before semicolon semi-spacing`.
username_1: Thanks a lot you both for the feedback! I will try the solution @username_0 proposed in https://github.com/WikiEducationFoundation/WikiEduDashboard/pull/1413#issuecomment-332241174 and if not fix this another approach. I will also check the tests to make sure nothing breaks 😅
Status: Issue closed
|
SeleniumHQ/selenium | 194613472 | Title: 12344
Question:
username_0: ## Meta -
OS:
<!-- Windows 10? OSX? -->
Selenium Version:
<!-- 2.52.0, IDE, etc -->
Browser:
<!-- Internet Explorer? Firefox?
Since Firefox version 48, Mozilla requires all add-ons to be signed. Until
recently, Firefox support in Selenium was exclusively provided by an add-on.
As this add-on is not currently signed, this solution does not work with the
latest Firefox releases. As an alternative, Mozilla are working on a WebDriver
specification compliant implementation named GeckoDriver. Please note that the specification is not complete, and that Selenium itself does not comply with
the specification at this time. This means that features previously available
through Selenium will not be available using GeckoDriver.
Any issue logged here for Firefox 48 or later will be closed as a duplicate of
#2559. Our recommendation is to switch to GeckoDriver, or to continue testing
on Firefox 45 until GeckoDriver is a viable option for you. If you are
interested in helping us to sign the add-on to restore support for later
Firefox versions, please see the following comment for what's needed:
https://github.com/SeleniumHQ/selenium/issues/2942#issuecomment-259717567
If the issue is with Google Chrome consider logging an issue with chromedriver instead:
https://sites.google.com/a/chromium.org/chromedriver/help
If the issue is with Microsoft Edge consider logging an issue with Microsoft instead:
https://developer.microsoft.com/en-us/microsoft-edge/platform/issues/
If the issue is with Firefox GeckoDriver (aka Marionette) consider logging an issue with Mozilla:
https://bugzilla.mozilla.org/buglist.cgi?product=Testing&component=Marionette
If the issue is with Safari, only Safari 10+ is supported. Please log any Safari issue with Apple:
https://bugreport.apple.com/
If the issue is with PhantomJS consider logging an issue with Ghostdriver:
https://github.com/detro/ghostdriver
-->
Browser Version:
<!-- e.g.: 49.0.2623.87 (64-bit) -->
## Expected Behavior -
## Actual Behavior -
## Steps to reproduce -
<!--
Please be sure to include an SSCCE (Short, Self Contained, Correct [compilable] example) http://sscce.org/
If you can't provide a link to the page, consider creating a reproducible page on https://jsfiddle.net/
--><issue_closed>
Status: Issue closed |
ninedraft/enum2go | 619784325 | Title: Use all fields as enum values spec in enum specification
Question:
username_0: version: v0.1.1
tool ignoring all fields except first one int enum specs
```go
type (
Foo int
_ struct {
Enum struct {
A, B, C Foo // will emit enum values
D, E Foo // will be ignored
}
}
)
```<issue_closed>
Status: Issue closed |
wbthomason/packer.nvim | 1057683327 | Title: [packer.nvim] [WARN 02:39:05] c...ite/pack/packer/start/packer.nvim[packer.nvim] [ERROR 02:39:05] a...local 'display_win' (a nil value)[packer.nvim] [ERROR 02:39:05] a...local 'display_win' (a nil value)
Question:
username_0: <!-- Before creating an issue, please search the issue tracker and make sure packer.nvim is up to date -->
<!-- If your issue is a general usage question, please create a GitHub discussions thread: https://github.com/username_5/packer.nvim/discussions -->
- `nvim --version`:
```
NVIM v0.5.1
Build type: Release
LuaJIT 2.1.0-beta3
Compilation: clang -U_FORTIFY_SOURCE -D_FORTIFY_SOURCE=1 -DNVIM_TS_HAS_SET_MATCH_LIMIT -O2 -DNDEBUG -Wall -Wextra -pedantic -Wno-unused-parameter -Wstrict-prototypes -std=gnu99 -Wshadow -Wconversion -Wmissing-prototypes -Wimplicit-fallthrough -Wvla -fstack-protector-strong -fno-common -fdiagnostics-color=auto -DINCLUDE_GENERATED_DECLARATIONS -D_GNU_SOURCE -DNVIM_MSGPACK_HAS_FLOAT32 -DNVIM_UNIBI_HAS_VAR_FROM -DMIN_LOG_LEVEL=3 -I/tmp/neovim-20211103-7276-u4ywwo/neovim-0.5.1/build/config -I/tmp/neovim-20211103-7276-u4ywwo/neovim-0.5.1/src -I/opt/homebrew/include -I/tmp/neovim-20211103-7276-u4ywwo/neovim-0.5.1/deps-build/include -I/Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk/usr/include -I/opt/homebrew/opt/gettext/include -I/tmp/neovim-20211103-7276-u4ywwo/neovim-0.5.1/build/src/nvim/auto -I/tmp/neovim-20211103-7276-u4ywwo/neovim-0.5.1/build/include
Compiled by [email protected]
Features: +acl +iconv +tui
See ":help feature-compile"
system vimrc file: "$VIM/sysinit.vim"
fall-back for $VIM: "/opt/homebrew/Cellar/neovim/0.5.1_1/share/nvim"
Run :checkhealth for more info
```
- `git --version`:
- Operating system/version:

- Terminal name/version:
### Steps to reproduce
### Actual behaviour
### Expected behaviour
### packer files
<details>
<summary>Plugin specification file(s)</summary>
init.lua
```
require('plugins')
```
plugins.lua
```
local fn = vim.fn
local install_path = fn.stdpath('data')..'/site/pack/packer/start/packer.nvim'
if fn.empty(fn.glob(install_path)) > 0 then
packer_bootstrap = fn.system({'git', 'clone', '--depth', '1', 'https://github.com/username_5/packer.nvim', install_path})
end
return require('packer').startup(function(use)
-- My plugins here
-- use 'foo1/bar1.nvim'
-- use 'foo2/bar2.nvim'
-- Automatically set up your configuration after cloning packer.nvim
-- Put this at the end after all plugins
[Truncated]
```
❯ cat ~/.cache/nvim/packer.nvim.log
[WARN Fri Nov 19 02:44:02 2021 2.8174368419658e+14] .../site/pack/packer/start/packer.nvim/lua/packer/clean.lua:75: Could not remove /Users/lidashuang/.local/share/nvim/site/pack/packer/start/packer.nvim
[ERROR Fri Nov 19 02:44:02 2021 2.8174368444758e+14] .../site/pack/packer/start/packer.nvim/lua/packer/async.lua:20: Error in coroutine: ...e/nvim/site/pack/packer/start/packer.nvim/lua/packer.lua:570: attempt to index local 'display_win' (a nil value)
[ERROR Fri Nov 19 02:44:02 2021 2.8174368453246e+14] .../site/pack/packer/start/packer.nvim/lua/packer/async.lua:20: Error in coroutine: ...e/nvim/site/pack/packer/start/packer.nvim/lua/packer.lua:570: attempt to index local 'display_win' (a nil value)
[WARN Fri Nov 19 02:44:45 2021 2.8178698160542e+14] .../site/pack/packer/start/packer.nvim/lua/packer/clean.lua:75: Could not remove /Users/lidashuang/.local/share/nvim/site/pack/packer/start/packer.nvim
[ERROR Fri Nov 19 02:44:45 2021 2.8178698237663e+14] .../site/pack/packer/start/packer.nvim/lua/packer/async.lua:20: Error in coroutine: ...e/nvim/site/pack/packer/start/packer.nvim/lua/packer.lua:570: attempt to index local 'display_win' (a nil value)
[ERROR Fri Nov 19 02:44:45 2021 2.8178698245017e+14] .../site/pack/packer/start/packer.nvim/lua/packer/async.lua:20: Error in coroutine: ...e/nvim/site/pack/packer/start/packer.nvim/lua/packer.lua:570: attempt to index local 'display_win' (a nil value)
~/.config/nvim
```
</details>
<details>
<summary>packer compiled file</summary>
Post the contents of `packer_compiled.vim` here
</details>
Answers:
username_1: I'm also getting this on a clean install. Just neovim 0.5 and so far packer (which is failing to initialise).
Do you have any progress? I'm using the code in `bootstrapping` in the Readme.
username_2: I am an echoing voice to this choir - I'm using more or less the same setup as username_0 (Monterey, nvim 5.1, etc...) and am receiving the same error.
In the same vein like username_1, I've just grabbed the `bootstrapping` code assuming that would work. Any help would be definitely appreciated!
username_3: Also hitting this issue on Neovim nightly version 0.6
```
NVIM v0.6.0-dev+625-gc132144d2
Build type: Release
LuaJIT 2.1.0-beta3
Compiled by <EMAIL>
Features: +acl +iconv +tui
See ":help feature-compile"
system vimrc file: "$VIM/sysinit.vim"
fall-back for $VIM: "/usr/local/Cellar/neovim/HEAD-c132144_1/share/nvim"
Run :checkhealth for more info
```
username_4: I recompiled from latest source and all is well.
username_5: I haven't been able to reproduce this yet. @username_4, what commit of upstream fixed this for you?
username_4: @username_5
`04c7b550384b2246d37c1a893c0f858bc332cb8b
`
```
NVIM v0.6.0-dev+650-g04c7b5503
Build type: RelWithDebInfo
LuaJIT 2.1.0-beta3
Compilation: /usr/bin/cc -U_FORTIFY_SOURCE -D_FORTIFY_SOURCE=1 -DNVIM_TS_HAS_SET_MATCH_LIMIT -O2 -g -Og -g -Wall -Wextra -pedantic -Wno-unused-parameter -Wstrict-prototypes -std=gnu99 -Wshadow -Wconversion -Wmissing-prototypes -Wimplicit-fallthrough -Wvla -fstack-protector-strong -fno-common -fdiagnostics-color=always -DINCLUDE_GENERATED_DECLARATIONS -D_GNU_SOURCE -DNVIM_MSGPACK_HAS_FLOAT32 -DNVIM_UNIBI_HAS_VAR_FROM -DMIN_LOG_LEVEL=3 -I/home/sib/source_neovim/build/config -I/home/sib/source_neovim/src -I/home/sib/source_neovim/.deps/usr/include -I/usr/include -I/home/sib/source_neovim/build/src/nvim/auto -I/home/sib/source_neovim/build/include
Compiled by <EMAIL>
Features: +acl +iconv +tui
See ":help feature-compile"
system vimrc file: "$VIM/sysinit.vim"
fall-back for $VIM: "/home/sib/dotfiles/bin/share/nvim"
Run :checkhealth for more info
```
username_0: 
username_6: Hit this issue on latest nightly.
```
NVIM v0.7.0-dev+839-g09d270bce
Build type: RelWithDebInfo
LuaJIT 2.1.0-beta3
```
username_5: @username_8 (or others) do you have more details on the error? In particular, a line number in the error message, or steps to reproduce the error?
username_7: @username_5 Not sure if this is exactly the same as everyone else or the right place but I've been getting the same error message when it attempts to remove the start/packer.nvim directory upon first startup.
```
nvim -v
shell-init: error retrieving current directory: getcwd: cannot access parent directories: No such file or directory
NVIM v0.6.1
Build type: RelWithDebInfo
LuaJIT 2.1.0-beta3
```
The exact error line is as follows:
`[packer.nvim] [ERROR 01:49:47] async.lua:20: Error in coroutine: ...e/nvim/site/pack/packer/start/packer.nvim/lua/packer.lua:570: attempt to index local 'display_win' (a nil value)`
That said, I've been having overall permission issues with my neovim/packer.nvim config on a fresh install w/o any plugins so it could be related. I've had to create the opt directory myself to even get to this point because Packer could not do it automatically via the bootstrapping method and outputting permission denied errors.
This is with Ubuntu 20.04 on WSL using the latest prebuilt nvim.appimage on the Neovim Github Repo.
username_8: I don't have precise steps to reproduce right now unfortunately, sorry about that. IIRC the error message was the same as OP has provided:
```
.../site/pack/packer/start/packer.nvim/lua/packer/async.lua:20: Error in coroutine: ...e/nvim/site/pack/packer/start/packer.nvim/lua/packer.lua:570: attempt to index local 'display_win' (a nil value))
```
I've been able to make it work. What I did is I've added nvim-lua/popup.nvim plugin as the second plugin after packer.nvim and I've spammed PackerSync until everything stabilized.
username_7: For anyone that is experiencing similar issues as I was, I resolved this behavior by changing ownership of the ~/.local directories, as well as ~/.config/nvim/plugin if permission denied errors were mentioned for that folder in the log output, using the following example command(s). I wasn't able to figure out why I needed to do this on my installation when I didn't need to on another machine but this should workaround the error.
```
sudo chown -R "${USER}" ~/.local
sudo chown $USER ~/.config/nvim/plugin
```
username_5: @username_7 Who owned those directories before the `chown`?
The line in `packer.lua` giving an error seems to imply that there are no plugins to install or update that aren't frozen (by tracing the conditions for initializing the `display_win` object). I'm not sure why that would happen for the cases you describe, but it's easy enough to add a check for that edge case.
Status: Issue closed
username_5: If we still see this behavior (or any regressions) after 11df2e3, we'll reopen this.
username_7: If my memory serves me right, it was showing as root, but I don't recall this ever causing me errors previously. That said, I try to bootstrap/install my dotfiles/packages via shell script so could be related.
username_9: Clearing cache (`~/.cache/nvim`) fixed it for me.
username_10: i had the same problem, i realized that i did not add the following let packer manage itself:
``` lua
local fn = vim.fn
local install_path = fn.stdpath('data')..'/site/pack/packer/start/packer.nvim'
if fn.empty(fn.glob(install_path)) > 0 then
packer_bootstrap = fn.system({'git', 'clone', '--depth', '1', 'https://github.com/username_5/packer.nvim', install_path})
end
return require('packer').startup(function(use)
-- make sure to add this line
use 'username_5/packer.nvim'
-- Automatically set up your configuration after cloning packer.nvim
-- Put this at the end after all plugins
if packer_bootstrap then
require('packer').sync()
end
end)
``` |
fabric/fabric | 60337387 | Title: UnicodeDecodeError: 'ascii' codec can't decode byte 0xc4 in position 20: ordinal not in range(128)
Question:
username_0: On Ubuntu, when trying to recursively `put()` a directory with a file which contains accented characters *in the filename* (specifically, the filename is: `Fiat 500 - ciesz się małymi rzeczami!-720.mp4`), I got an exception like below when run with `--show=debug`:
````
Fatal error: put() encountered an exception while uploading 'salt_master/root'
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/fabric/operations.py", line 395, in put
mirror_local_mode, mode, temp_dir)
File "/usr/local/lib/python2.7/dist-packages/fabric/sftp.py", line 317, in put_dir
n = posixpath.join(rcontext, f)
File "/usr/lib/python2.7/posixpath.py", line 80, in join
path += '/' + b
UnicodeDecodeError: 'ascii' codec can't decode byte 0xc4 in position 20: ordinal not in range(128)
Aborting.
Disconnecting from localhost... done.
put() encountered an exception while uploading 'salt_master/root'
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/fabric/operations.py", line 395, in put
mirror_local_mode, mode, temp_dir)
File "/usr/local/lib/python2.7/dist-packages/fabric/sftp.py", line 317, in put_dir
n = posixpath.join(rcontext, f)
File "/usr/lib/python2.7/posixpath.py", line 80, in join
path += '/' + b
UnicodeDecodeError: 'ascii' codec can't decode byte 0xc4 in position 20: ordinal not in range(128)
````
Additional info:
````
$ echo $LANG
en_US.UTF-8
````
Answers:
username_1: Could you add 2 debug lines in `/usr/local/lib/python2.7/dist-packages/fabric/sftp.py` before line 317:
```
print type(rcontext), repr(rcontext)
print type(f), repr(f)
```
and paste the output here?
username_2: Gross. I'm guessing this is because `rcontext` (containing your filename w/ non-ASCII-friendly chars) is being added to the regular string `'/'` within `posixpath` and that's triggering an attempt at encoding it (Python 2 has a number of spots where Unicode strings will be automatically encoded using the 'ascii' default encoding and this is likely one of them).
Offhand I don't see a great way to handle this on our end besides possibly trying a more lenient, explicit encoding step before calling `posixpath.join`, but that has its own issues (i.e. your filename would get munged unexpectedly).
Google finds that this is a widespread problem with `posixpath.join`, including but not limited to sphinx-doc/sphinx#1163
username_0: FYI: As of now, after some googling, I've applied a "workaround hack" as mentioned in http://stackoverflow.com/questions/2276200/changing-default-encoding-of-python (putting `reload(sys)` and stuff in first lines of my fabfile.py). Seems to "Work For Me Now(tm)", but unfortunately it's reported as **"dangerous, may break basic stuff in language"** according to [a comment in the stackoverflow thread](http://stackoverflow.com/questions/2276200/changing-default-encoding-of-python#comment2240875_2276222). This particular script is currently non-critical for me, so I can live with the risk.
@username_1 :
````
<type 'unicode'> u'/home/CENSOREDX/FOOBAR/salt_master/root/./CENSORED/fiat'
<type 'str'> 'Fiat 500 - ciesz si\xc4\x99 ma\xc5\x82ymi rzeczami!-720.mp4'
Fatal error: put() encountered an exception while uploading 'salt_master/root/.'
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/fabric/operations.py", line 395, in put
mirror_local_mode, mode, temp_dir)
File "/usr/local/lib/python2.7/dist-packages/fabric/sftp.py", line 319, in put_dir
n = posixpath.join(rcontext, f)
File "/usr/lib/python2.7/posixpath.py", line 80, in join
path += '/' + b
UnicodeDecodeError: 'ascii' codec can't decode byte 0xc4 in position 20: ordinal not in range(128)
Aborting.
Disconnecting from localhost... done.
put() encountered an exception while uploading 'salt_master/root/.'
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/fabric/operations.py", line 395, in put
mirror_local_mode, mode, temp_dir)
File "/usr/local/lib/python2.7/dist-packages/fabric/sftp.py", line 319, in put_dir
n = posixpath.join(rcontext, f)
File "/usr/lib/python2.7/posixpath.py", line 80, in join
path += '/' + b
UnicodeDecodeError: 'ascii' codec can't decode byte 0xc4 in position 20: ordinal not in range(128)
````
username_1: From what I tested with `os.path.join`, I think that this will work in the files loop (https://github.com/fabric/fabric/blob/master/fabric/sftp.py#L318):
```
try:
f = f.decode('utf-8')
except:
pass
```
username_1: @username_2 I believe the problem starts with `os.walk`, normally it returns `str` but if `unicode_literals` are imported, file & directory names are converted to type `unicode`, at least from what I tested. Which as you note in your comment is perfectly fine, until you face an implicit encoding/decoding when attempting to concatenate str and unicode.
Perhaps importing `unicode_literals` in the fabfile (or file using Fabric as a library) has this effect? I think it might.
username_0: @username_1 Yes, seemed to work for me ok with [the patch](https://github.com/fabric/fabric/issues/1292#issuecomment-78136344)!
username_1: Thanks @username_0 for testing!
@username_2 what do you think? Wouldn't this solve all cases?
I would be happy to submit a pull request (with some tests of course) if you think this is a proper solution.
username_2: My understanding is that trying to use UTF-8 as a catchall will still cause issues elsewhere (e.g. folks using other encodings, such as UK Windows users) and possibly even for folks who wouldn't otherwise have encountered this.
Also really not a fan of bare `except` - I think I get what you're going for but it still makes me squeamish (I've encountered so many issues where well-meaning "try a thing, bare except and continue" setups caused lots of debugging pain :()
I think what would be ideal is to merge the two situations together:
``` python
for f in files:
try:
n = posixpath.join(rcontext, f)
except UnicodeDecodeError:
n = posixpath.join(rcontext, f.decode('utf-8')
# ...
```
This way we ensure the `decode` only fires in the presence of an otherwise unrecoverable error.
If that change still works for @username_0 and/or others, I'll go ahead and merge a copy of it.
username_1: @username_2 I agree on the bare `except` remark, it may indeed be far more difficult to debug, since it is too generic. Your solution is certainly better :) .
Perhaps a more broad approach could be using [getlocale](https://docs.python.org/2/library/locale.html#locale.getlocale) to get the remote locale settings and use it instead of defaulting it to `utf-8` as you pointed out? Not sure how reliable would that be though, especially cross-platform.
I still do not get entirely though the part that `unicode_literals` modifies the `os.walk` returning type from `str` to `unicode`.
username_3: Hi,
Same issue occurs simply when trying to run remote command, having output with some french characters:
`
Traceback (most recent call last):
File "/usr/local/lib/python2.7/threading.py", line 801, in __bootstrap_inner
self.run()
File "/usr/local/lib/python2.7/threading.py", line 754, in run
self.__target(*self.__args, **self.__kwargs)
File "build/bdist.linux-x86_64/egg/fabric/network.py", line 350, in outputter
out_stream.write("%s: %s\n" % (prefix, line)),
UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 32: ordinal not in range(128)
`
As an example, we were trying to run testng (6.11) from ant, on a remote machine.
[Here](https://github.com/cbeust/testng/blob/ff464f1102a99418ac1e33f79c57ba0657344600/src/main/resources/org/testng/internal/VersionTemplateJava) is the source doe which logs the character which causes trouble.
`
[testng] ...
[testng] ... TestNG 6.11 by <NAME> (<EMAIL>)
[testng] ...
[testng]
`
To reproduce, simply execute command on file with above content:
`
run('cat file.txt')
`
Thanks
username_4: Is it possible that your terminal LANG or related vars are not UTF-8?
```
$ echo LANG=$LANG
LANG=en_US.UTF-8
```
username_3: Hi,
It's UTF-8.
After more research, concluded I have to upgrade fabrics first.
Thanks a lot! |
rossfuhrman/_why_the_lucky_markov | 509543747 | Title: #<FakeRabbit says 'Thusly and thusly and thusly...'> The thing is: irb is talking back. Sieging of Elmer Lake.
Question:
username_0: Toot: #<FakeRabbit says 'Thusly and thusly and thusly...'> The thing is: irb is talking back. Sieging of Elmer Lake.
One comment = 1 upvote. Sometime after this gets 2 upvotes, it will be posted to the main account at https://mastodon.xyz/@_why_toots |
jesus-collective/mobile | 858865458 | Title: Response assignments go to blank screen on beta
Question:
username_0: I went to the course edit page to see if maybe someone forgot to link lessons to "Response" type lessons. Turns out some were not linked to anything. So I linked them and synced the lessons, but they're still giving me a blank page.<issue_closed>
Status: Issue closed |
FourierFlows/GeophysicalFlows.jl | 644095265 | Title: Stratified Flows module
Question:
username_0: We've discussed this before, but now I've submitted my thesis (yay!), I actually have time to start working on a module for a vertical plane with an imposed mean stratification. There are a couple of options for the implementation of this:
- Vorticity-streamfunction formulation. Should be relatively simple to implement given its similarity to the existing modules.
- Primitive variable formulation for velocity (_u_, _w_) and buoyancy perturbation _θ_ with a pressure solve through a projection method. I saw you have a 3D box as a potential future development on FourierFlows, so implementing a pressure solver may be useful to lead in to that.
Any thoughts?
Once this is working it should be relatively simple to include multiple scalars and make some cool examples of salt fingering and diffusive convection.
I'll be using the existing modules as a starting point and will let you know how I get on. Any tips for developing these modules are much appreciated!
Answers:
username_1: hey @username_0 congrats on submitting thesis!
A stratified flows module would be great.
FourierFlows.jl already includes a `ThreeDGrid` but we don't have any module in GeophysicalFlows.jl using it.
Can you point to a pdf/notes with the equations written down so I think about it a bit?
username_0: Thanks @username_1 !
Here's a document going over the two options in a bit more detail:
[stratified_periodic_equations.pdf](https://github.com/FourierFlows/GeophysicalFlows.jl/files/4824380/stratified_periodic_equations.pdf)
The method I suggested for the pressure is what Diablo uses, and may come in useful if you want to make a 3DNavierStokes module.
I've also outlined a couple of extensions that it would be good to have in mind while developing this.
username_2: @username_0 nice! Congrats!
I think a 2D vorticity-streamfunction formulation might be a good place to start because of its simplicity.
However, one consideration is that a 3D implementation could be easily reduced to 2D by setting, for example, `ny=1`. The downside is the added complexity. The upside is that we immediately obtain efficient 3D and 2D solvers.
As for the pressure solve, we do not need to use pressure projection or fractional time stepping methods with a 3D spectral method. Instead, we can obtain the pressure directly from the transformed equations in Fourier space, and substitute the result into the RHS. This approach is described in [Winters et al 2004](https://watermark.silverchair.com/1520-0426(2004)021_0069_asmfps_2_0_co_2.pdf?token=<KEY>) (see equations 33-35).
Looking at those equations, I do think it probably makes sense to start with the vorticity-streamfunction formulation. It's much simpler.
I'd suggest using `b` rather than `θ` for buoyancy. `θ` is often used for "temperature".
Finally, we should probably also have rotation! This requires an additional equation for `v` in the vorticity-streamfunction formulation. The equations are given in [Xie and Buhler (2019)](https://www.researchgate.net/profile/Jin_Han_Xie/publication/333795799_Two-dimensional_isotropic_inertia-gravity_wave_turbulence/links/5d75b9ef4585151ee4a8bf02/Two-dimensional-isotropic-inertia-gravity-wave-turbulence.pdf). We can probably implement a rotating version as an "option" in the new module (we don't need to put rotation in at the outset). |
deviceinsight/kafkactl | 674772412 | Title: Dynamic Completion not working when installed via snap
Question:
username_0: When kafkactl is installed via snap bash auto-completion is packaged inside the snap and should work out of the box.
Static autocompletion is working but dynamic auto-completion not.
Type e.g.:
```
kafkactl config use-context <TAB> <TAB>
```
Debugging the auto-completion shows that it should be working:
```
kafkactl __complete config use-context ""
default
:4
Completion ended with directive: ShellCompDirectiveNoFileComp
```
It probably has something to do with prefixing of the binary by snap. |
electron/electron | 1106924283 | Title: [Bug]: Media Information not showing on Windows 10
Question:
username_0: ### Preflight Checklist
- [X] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/main/CONTRIBUTING.md) for this project.
- [X] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/main/CODE_OF_CONDUCT.md) that this project adheres to.
- [X] I have searched the [issue tracker](https://www.github.com/electron/electron/issues) for a bug report that matches the one I want to file, without success.
### Electron Version
13.6.7, 14.2.4, 15.3.5, 16.0.7, 17.0.0-beta.5
### What operating system are you using?
Windows
### Operating System Version
Windows 10 (Pro 20H2 and others)
### What arch are you using?
x64
### Last Known Working Electron version
12.2.3
### Expected Behavior
When using keyboard media keys a OS Media Information overlay should pop up (showing track information, album cover and play control buttons). Worked in Electron 12 and works in current Chrome (97).
### Actual Behavior
From Electron 13 and up the Media Information overlay is not displayed. (The keyboard media keys do work though.) Only tested last patch of each major version.
### Testcase Gist URL
https://gist.github.com/b19bc801bd50597c8eeedaf29f0efa2a
### Additional Information
Steps to reproduce:
1. Click "Play" on media in webpage (using Chrome example in the Gist)
2. Trigger Pause using keyboard media keys (Next and Previous have same problem)
This looks very similar to: #30860, there the example URL loaded is for video, here it is audio, but otherwise it looks the same. Created new issue to add details as that one is closed.
(Checking "Obsolete" in Fiddle Electron Settings is needed to load the last working 12 version.)
Answers:
username_1: I'd like to expand on this:
1. This didn't use to happen (it worked completely fine a few months ago, even on electron 12)
2. The media session OSD shows up when changing the volume, but NOT when using any other media key (play/pause/next/previous)
Important note: chrome/edge/spotify etc do show the media OSD when using the media keys, as it is expected
3. the OSD is really bugged on Electron fiddle when not using the latest electron version
with electron 14: (no media information including play/pause status)

with electron 17: (play/pause status and media information is updated)

// test case: https://gist.github.com/581fd5d38496621e5cd9a6020b68438f |
Mopecat/Daily-Mission-Board | 658803336 | Title: 简述下你对执行上下文的理解
Question:
username_0: 1第一阶段 代码=预编译阶段,包括内存分配,变量提升等var color changeColor 函数提升,函数内的代码提升,包括anotherColor,swapColors tempColor等
2.第二阶段代码执行,全局上下文入栈,遇到可执行代码changeColor 入栈,遇到swapColors,入栈
3.销毁 swapColors 出栈,tempColor 销毁,changeColor 出栈anotherColor 销毁
color的销毁???
Answers:
username_1: 执行上下文包含,变量和作用域链,以这个例子来说
第一阶段是预编译创建变量 `color`, `changeColor`, `anotherColor`, `swapColors`, `tempColor` 包括变量提升
第二阶段是执行代码 包括变量的赋值,在函数调用时激活对应的作用域,才可以访问对应作用域中的变量,`changeColor`执行才可以访问`anotherColor`,`swapColors`执行才可以访问`tempColor`
第三阶段是销毁,当前作用域内函数调用完毕后就会将相关的变量销毁,这里涉及函数调用,是栈型结构,所以先调用的后出栈,`swapColors`调用后,`swapColors`出栈并将不再被引用的变量`tempColor`销毁,然后`changeColor **`出栈将不再被引用的`anotherColor`、`swapColors `销毁。
完事了
username_0: 1第一阶段 代码=预编译阶段,包括内存分配,变量提升等var color changeColor 函数提升,函数内的代码提升,包括anotherColor,swapColors tempColor等
2.第二阶段代码执行,全局上下文入栈,遇到可执行代码changeColor 入栈,遇到swapColors,入栈
3.销毁 swapColors 出栈,tempColor 销毁,changeColor 出栈anotherColor 销毁
color的销毁???
username_2: 之前看标题以为是说this的(我的错),来公司早了,大概总结了一下this,回复打开发现主要是变量提升方面,上面两个讲的都很对,那我只能从this方面敷衍一下这道题了。https://catsaid.cn/2020/07/22/this/
说执行上下文除了作用域链,变量提升,this之外还有一个绕不开的话题:闭包。改天在弄吧~~先上班了 |
shellvon/zh-downloader | 352815732 | Title: 当推荐视频数量小于单屏需展示的数量时,尝试自动加载
Question:
username_0: #### 问题描述:
由于昨日 `1.0.5` 版本发布增加了采集时即删除推荐商品的功能,因此当用户点击删除/或者采集时,推荐视频数量会-1,最后会便为0,需要用户手动点击刷新按钮才可以刷出新的视频,不是很智能
#### 期望行为:
当推荐视频的数量小于单页需要展示的视频数量时,尝试获取新的视频数量。比如当用户删除/采集时需要检查recomendLst.length 以决定是否需要获取新的数据. 目前插件单页展示的视频数量是2。
#### 解决方式:
目前已知至少可以有2⃣️种方式解决此问题:
+ 使用 Vue 的 watch 机制。
+ 在deleteVideo的时候增加逻辑判断(就算采集实际上最后也会调用delete) |
yashaka/NSelene | 153936256 | Title: Enhance error messages for failed search by condition in collection
Question:
username_0: TODO: add test
Status: Issue closed
Answers:
username_0: TODO: add test
Status: Issue closed
username_0: enhanced a bit:
```
OpenQA.Selenium.WebDriverTimeoutException :
Timed out after 0.25 seconds
while waiting entity with locator: By.Selene: (By.CssSelector: #todo-list>li).FindBy(ExactText is b)
for condition: Visible
Expected : True
Actual : False
----> OpenQA.Selenium.NotFoundException : element was not found in collection by condition ExactText is b
Actual visible texts : [a,c]
Actual html elements : [<li class="active" data-index="0">
<div class="view">
<input class="toggle" type="checkbox">
<label>a</label>
<button class="destroy"></button>
</div>
<input class="edit">
</li>,<li class="active" data-index="1">
<div class="view">
<input class="toggle" type="checkbox">
<label>c</label>
<button class="destroy"></button>
</div>
<input class="edit">
</li>]
``` |
xszi/javascript-algorithms | 774990657 | Title: 字典和散列表知识小结
Question:
username_0: 之前在数据结构中获得一个值(使用get方法),需要遍历操作,如果使用散列函数,就**知道值得具体位置**,因此可以快速检索到值。
散列函数的作用就是给定一个键值,然后返回值在表中的地址。
```js
function HashTable() {
var table = []
var loseloseHashCode = function(key) {
var hash = 0
for (var i = 0; i < key.length; i++) {
hash += key.charCodeAt(i)
}
return hash % 37
}
this.put = function(key, value) {
var position = loseloseHashCode(key)
console.log(position + ' _ ' + key);
table[position] = value
}
this.get = function(key) {
return table[loseloseHashCode(key)]
}
// 不能将位置本身从数组中删除(这会改变其他元素的位置),
// 否则,当下次需要获得或移除一个元素的时候,这个元素会不在我们用散列函数求出的位置上。
this.remove = function(key){
table[loseloseHashCode(key)] = undefined
}
}
``` |
mintproject/mic | 598404343 | Title: Adding inputs is cumbersome
Question:
username_0: For each input you must: 1) Select the property to edit; 2) Select `add` 3) add the input information 4) exit 5) repeat 1.
This is slow and prone to errors. We should support not having to go back to add every time. This could be done by going back to the 'add' or 'edit' menu for each input.<issue_closed>
Status: Issue closed |
dillonkearns/elm-pages | 1185206908 | Title: A way to disable Elm Optimize Level 2
Question:
username_0: Yes, I like it. So basically one will set the optimization level by passing `--optimize=<level>` with default being `2`. And `--optimize` and `--debug` will be mutually exclusive.
Let's see how rusty my Node.js skills are :-)
Answers:
username_1: Thanks for the followup!
So one detail to consider here is what the default should be. It's also worth noting that the current default is the fully optimized build (using EOL2). I lean towards the idea of having the default be optimized. It feels safer that way since the most natural thing to do is have a production-ready build that's optimized as much as possible (instead of having to find a magic incantation to make sure you're doing your production builds correctly). And if you need to opt out of it, you can find the appropriate options to do so. That also means that it wouldn't change the existing default, so that seems like an all-around winner to me.
It would also be nice to mirror the familiar options from `elm` as much as possible. I think the `--debug` option is a pretty clear one here. I think it would also be intuitive to keep `--optimize` for option (3) here as well for that reason.
One idea would be to have a CLI flag `--optimize` that can take an argument of a number. I'm not sure if `commander` (the tool that elm-pages uses for CLI flag parsing) supports having a CLI flag that can be used with or without an argument. But if it does, then this seems like a good setup to me:
|Number|Description|Command|
| --- | --- | --- |
|1|Debug ( `--debug` )|`elm-pages build --debug`|
|2|Without `--optimize` and without EOL2 (elm-optimize-level-2)|`elm-pages build --optimize=0`|
|3|With `--optimize` but without EOL2|`elm-pages build --optimize=1`|
|4|With `--optimize` and EOL2|`elm-pages build` (or `--optimize=2`, so 2 would just be the default)|
And then maybe just an error if `--debug` is passed in with anything but `--optimize=0`. What do you think? If that sounds good to you, then I think the only thing remaining would be to see if Commander supports that.
username_0: Yes, I like it. So basically one will set the optimization level by passing `--optimize=<level>` with default being `2`. And `--optimize` and `--debug` will be mutually exclusive.
Let's see how rusty my Node.js skills are :-) |
ServiceStack/Issues | 518579388 | Title: ServiceStack OrmLite order by error
Question:
username_0: ORDER BY "SAV"."N°série") As RowNum
FROM "SAV" LEFT JOIN "Batch" ON ("SAV"."Compteur" = "Batch"."ProductId")
WHERE (("SAV"."Date garantie" >= @0) AND ("SAV"."Date garantie" <= @1))
GROUP BY "SAV"."N°série", "SAV"."Observations") AS RowConstrainedResult WHERE RowNum > 40 AND RowNum <= 80`
### Notes
In my case, adding manually `ORDER BY RowNum` does the trick but I cannot use it directly inside code.
Thank you very much for your help!
Answers:
username_1: I'd suggest you use Custom SQL or client sorting as the RowNum windowing function is a necessary hack for legacy SQL Server databases which I don't want to change to preserve backward compatibility.
Status: Issue closed
username_0: Ok. I finally ended up with the followiing method:
```
List<Product> products;
// HACK Remove this when upgrading to another SQL Server Database version
var selectStatement = ev.ToMergedParamsSelectStatement();
if (selectStatement.Contains("RowNum"))
{
selectStatement += " ORDER BY RowNum";
products = db.SqlList<Product>(selectStatement);
}
else
products = db.Select(ev);
``` |
gin-gonic/gin | 998821500 | Title: The role of NoMethod
Question:
username_0: ## Description
<!-- Description of a problem -->
What is the role of `NoMethod` , it does not seem to be effective
I found it is marked as **TODO** :
```
// NoMethod sets the handlers called when... TODO.
func (engine *Engine) NoMethod(handlers ...HandlerFunc) {
engine.noMethod = handlers
engine.rebuild405Handlers()
}
```
## How to reproduce
<!-- The smallest possible code example to show the problem that can be compiled, like -->
```
package main
import (
"fmt"
"github.com/gin-gonic/gin"
"net/http"
)
func main() {
r := gin.Default()
r.NoRoute(func(c *gin.Context) {
c.String(http.StatusOK, "NoRoute")
})
r.NoMethod(func(c *gin.Context) {
fmt.Println("NoMethod")
c.String(http.StatusOK, "NoMethod")
})
r.POST("/", func(c *gin.Context) {
c.String(http.StatusOK, "/")
})
r.Run(":8080")
}
```
## Expectations
<!-- Your expectation result of 'curl' command, like -->
```
$ curl http://localhost:8080
NoMethod
$ curl -X POST http://localhost:8080
/
$ curl -X PUT http://localhost:8080
NoMethod
```
## Actual result
<!-- Actual result showing the problem -->
```
$ curl http://localhost:8080
NoRoute
$ curl -X POST http://localhost:8080
/
$ curl -X PUT http://localhost:8080
NoRoute
```
## Environment
- go version: go version go1.17 windows/amd64
- gin version (or commit ref): v1.7.4
- operating system: windows/amd64
Answers:
username_0: I found a way to turn it on :`engine.HandleMethodNotAllowed = true`
username_0: A new problem,Can NoRoute and NoMethod take effect together?
username_0: So why is NoMethod marked as TODO?
username_0: Should HandleMethodNotAllowed be turned on when using NoMethod ?
Status: Issue closed
|
imranhsayed/gatsby-woocommerce-themes | 946348297 | Title: Vercel first build error
Question:
username_0: <img width="1110" alt="Снимок экрана 2021-07-16 в 17 29 49" src="https://user-images.githubusercontent.com/23171540/125963781-dc93a409-ac82-42ff-a50e-7bac9642357f.png">
Hi guys I want to use this starter I love much Gatsby and Wordpress but I have this error. Can you help me, please? |
xorpaul/g10k | 583666097 | Title: Raise an error on missing branch
Question:
username_0: Hello,
We currently use `warn_if_branch_is_missing: true` but when we get that warning, the rest of our workflow fails because the branch is missing. Would it be possible to have a `error_if_branch_is_missing` parameter, so our workflow would be interrupted with proper error reporting when a branch is missing?
Thanks! |
balena-io/etcher | 442981051 | Title: Etcher : invalid extension
Question:
username_0: hello,
Debian installers on a raspberry pi3, I have this:
`pi@raspberrypi:~ $ sudo apt-get install balena-etcher-electron
Lecture des listes de paquets... Fait
Construction de l'arbre des dépendances
Lecture des informations d'état... Fait
N: « etcher.lis » dans le répertoire « /etc/apt/sources.list.d/ » a été ignoré car il utilise une extension non valable
E: Impossible de trouver le paquet balena-etcher-electron
pi@raspberrypi:~ $`
In the meantime, cordially
Answers:
username_1: See #2748
Status: Issue closed
username_2: The error is in the message: the file `etcher.lis` in `/etc/apt/sources.list.d/` has an incorrect `.lis` extension, it should be `.list`. |
GilesStrong/lumin | 632420840 | Title: Add Literal types (aka move to Python 3.8)
Question:
username_0: In a few locations, string values arguments are used in LUMIN and specific value are expected. Python 3.8 introduced literal types, where the set of expected values can be stated in the method definitions. This is potentially useful, but would then require setting the minimum version of python for LUMIN from 3.6 to 3.8, which could be disruptive to our user-base. (Although, since I doubt we actually have a consistent user-base, this change may not be disruptive and should be done sooner rather than later). |
pyannote/pyannote-audio | 628439423 | Title: File contains data in an unknown format.
Question:
username_0: **The issue:**
RuntimeError: Error opening '/sample.wav': File contains data in an unknown format.
while running, **ovl_scores = overlap_detection(file)**
**Steps to reproduce the behavior:**
Installed pytorch, pyannote-audio, libsndfile1, librosa==0.7.0, pandas, scikit-learn.
Code is run in python 3.8.3 terminal
Code:
- import torch
- import os
- from pathlib import Path
- overlap_detection = torch.hub.load('pyannote/pyannote-audio', 'ovl_ami', pipeline=True)
- path_to_file=Path("/sample.wav")
- file_name=os.fspath(path_to_file)
- file={"uri":"sample","audio":file_name}
- ovl_scores = overlap_detection(file)
**Traceback (most recent call last):**
File "<stdin>", line 1, in <module>
File "/poc-env/lib/python3.8/site-packages/pyannote/audio/pipeline/overlap_detection.py", line 126, in __call__
ovl_scores = self._scores(current_file)
File "/poc-env/lib/python3.8/site-packages/pyannote/audio/features/wrapper.py", line 280, in __call__
return self.scorer_(current_file)
File "/poc-env/lib/python3.8/site-packages/pyannote/audio/features/base.py", line 149, in __call__
y, sample_rate = self.raw_audio_(current_file, return_sr=True)
File "/poc-env/lib/python3.8/site-packages/pyannote/audio/features/utils.py", line 228, in __call__
y, sample_rate = sf.read(
File "/poc-env/lib/python3.8/site-packages/soundfile.py", line 256, in read
with SoundFile(file, 'r', samplerate, channels,
File "/poc-env/lib/python3.8/site-packages/soundfile.py", line 627, in __init__
self._file = self._open(file, mode_int, closefd)
File "/poc-env/lib/python3.8/site-packages/soundfile.py", line 1181, in _open
_error_check(_snd.sf_error(file_ptr),
File "/poc-env/lib/python3.8/site-packages/soundfile.py", line 1355, in _error_check
raise RuntimeError(prefix + _ffi.string(err_str).decode('utf-8', 'replace'))
RuntimeError: Error opening '/sample.wav': File contains data in an unknown format.
Answers:
username_1: I do not know what `os.fspath` does but I think you do not need it...
`file = {"uri": "sample", "audio": path_to_file}` should be enough.
username_0: It also gave error, read that soundfile might not be able to read the path properly.
username_1: Then, it might simply be that your wav file is corrupted, as the error message suggests.
username_0: But I am able to play the audio and do preprocessing and speech-to-text conversion.
username_1: `pyannote.audio` relies on `SoundFile` to load audio files.
If `SoundFile` cannot read it, then `pyannote.audio` cannot process it.
Please check `SoundFile` [documentation](https://pysoundfile.readthedocs.io/en/latest/) and check whether it can load your file.
username_0: The library works with mono channel wav files
Status: Issue closed
|
Tomson124/SimplyJetpacks-2 | 318330640 | Title: Illegal Argument Exception, equal objects must have equal hashcodes
Question:
username_0: Don't know why, or if it is some other mod interfering with this mod, but right now on the newest version of thermal expansion and this I am experiencing a crash. Thought I should mention it here. [Here's the crash log](https://pastebin.com/9H0fMt1j). [And here's the mod list if you need it](https://pastebin.com/TVjdsh1Y).
Answers:
username_1: Nothing in that logs mentions SimplyJetpacks anywhere.
I see cofhcore and tinkers construct in there.
username_2: What you could try is reproducing the error with just SimplyJetpacks2, cofh mods and Tinkers and see if it happens then, too. |
gruntwork-io/helm-kubernetes-services | 497152878 | Title: Adding a sideCarContainer gives an error about mapping values
Question:
username_0: I'm trying to deploy an app that is build out of 2 containers. The main app I can deploy just fine, but when I add a "sideCarContainers:" section I get an error like this:
error: error converting YAML to JSON: yaml: line 47: mapping values are not allowed in this context
This is the section I added, commenting this out makes the error disappear:
sideCarContainers:
appengine:
image: eu.gcr.io/our-gce-project/container-name:0.1.0
Deploying this container using a separate helm chart works just fine.
We use this helm chart using the Terraform helm provider. I've been trying to get more debugging information, but that has failed up to now.
As a workaround/test I manually edited the resulting deployment yaml in GKE and added the 2nd container, and that also worked fine.
Let me know what I can do to further debug this or what more information you require.
Status: Issue closed
Answers:
username_1: This has been fixed in https://github.com/gruntwork-io/helm-kubernetes-services/releases/tag/v0.0.9 |
darkreader/darkreader | 757998945 | Title: [Broken Website] trezor.io
Question:
username_0: 1. Open trezor.io on mobiel
**Expected behavior**
A clear and concise description of what you expected to happen.
Must look good.
**Actual behavior**
A clear and concise description of what happened.
Looks bad.
**Screenshots**
If applicable, add screenshots to help explain your problem.

<!--
Please add a version of the browser you are using.
If you don't know how to get your browser/darkreader version please search it up online.
-->
**System info:**
- OS: [e.g. Windows, MacOS, Linux] linux
- Browser: [e.g. chrome, safari] firefox
- Darkreader Version: [e.g. 4.9.9] 4.9.26
**Additional context**
Add any other context about the problem here.
Status: Issue closed
Answers:
username_0: ack, thank you |
necolas/react-native-web | 735707073 | Title: Can't resolve 'core-js/modules/es6.typed.float32-array'
Question:
username_0: <!--
Thank you for reporting an issue. Please note that an issue must include the
information that is marked as REQUIRED below, or it may be closed.
-->
**I am trying to use react-native-web within a Gatsby site and when running gatsby develop I see
**ERROR #98124 WEBPACK**
Generating development JavaScript bundle failed
Can't resolve 'core-js/modules/es6.typed.float32-array' in '/Users/xxxxx/code/Gatsby/xxxx/node_modules/react-native-web/dist/vendor/react-native/Animated'
If you're trying to use a package make sure that 'core-js/modules/es6.typed.float32-array' is installed. If you're trying to use a local file make sure that the path is correct.
File: node_modules/react-native-web/dist/vendor/react-native/Animated/bezier.js
**
<!--
REQUIRED: A clear and concise description of the bug or problem.
-->
**How to reproduce**
<!--
REQUIRED: Create a test case by forking this template https://codesandbox.io/s/6lx6ql1w5r
Failing to include a reduced test case may result in the issue being closed,
and will delay any potential fix. Your application or GitHub project is NOT
considered a reduced test case. If the issue only affects certain browsers,
providing screenshots is also helpful.
-->
Simplified test case: <!-- add link here -->
Steps to reproduce:
1. Run gastby develop
**Expected behavior**
<!--
REQUIRED: A clear and concise description of what you expected to happen.
Please check that the behaviour is not expected React Native behaviour by
running your test case on iOS or Android using https://snack.expo.io.
-->
es6.typed.float32-array gets resolved
**Environment (include versions). Did this work in previous versions?**
* React Native for Web (version): 0.14.7
* React (version): 17.01
* Browser: Build stage
<!--
OPTIONAL:
**Additional context**
Add any other context about the problem here.
-->
Answers:
username_1: Maybe a Gatsby problem?
[Float32Array](https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Float32Array) has been supported in browsers for a long time and there is a feature test [here](https://github.com/username_1/react-native-web/blob/master/packages/react-native-web/src/vendor/react-native/Animated/bezier.js#L26)
Status: Issue closed
username_0: Thanks username_1, I think it is a typescript setup issue as I have just created a hello, world app to post here to show the problem including react-native-web with no issue. I will relook at my config |
JanOppolzer/CampusIdP | 270973866 | Title: Move container dependent variables to the end
Question:
username_0: ...so when a variable changes, the common tasks for Jetty and Shibboleth IdP are not run unnecessarily.
Jetty: `PASSWORD_CERT_KEY`, `PASSWORD_PKCS12`, `PASSWORD_KEYSTORE`
Shibboleth IdP: `SHIBBOLETH_SCOPE`, `SHIBBOLETH_ENTITYID`, `SHIBBOLETH_HOSTNAME`, `SHIBBOLETH_PASSWORD_*`<issue_closed>
Status: Issue closed |
wallabag/wallabag | 794930093 | Title: on app.wallabag.it Fetching www.contretemps.eu articles goes to error 500
Question:
username_0: Hi Guys,
(Almost) all articles on Contretemps.eu fail to be saved in wallabag, going to an Error 500
https://www.contretemps.eu/pandemopolitique-covid-triage-inegalites-sante-vaccins-alternatives/
(End on error message says "invalid date format)
I maintly use Vivaldi browser, seen even on Debian ans W7.


Thanks for watching.
Status: Issue closed
Answers:
username_1: Hello.
It works on wallabag.it now. (I did nothing). |
Roam-Research/issues | 615428350 | Title: Allow relative date queries on pages other than date pages
Question:
username_0: **Describe the problem**
I want to be able to make dynamic todo lists based on the current date, that are separating current from overdue tasks. However, currently the "{between: [[yesterday]] [[tomorrow]]" syntax only works on date pages.
This means I have to make a new query every time, instead of for instance going to my "Todo List" page and generating the query from there.
**Feature implementation**
Allow relative date queries on other pages. Today is assumed to be the current date. |
ma-ku/homebridge-elero-stick | 1061221634 | Title: Update
Question:
username_0: Any chance to update this in regard to "Lüftungsstellung" etc?
Answers:
username_1: If you could put this into context? I know that there are additional positions defined in the protocol. Unfortunately I believe they are not supported by HomeKit at least when i developed this.
```
UP: 0x20,
INTERM_POS: 0x44,
VENT_POS: 0x24,
DOWN: 0x40,
STOP: 0x10
```
One thing that could be done is to provide additional accessories (switch) to allow moving to these positions. If this is somehow supported within the context of HomeKit Coverings, we could add this.
username_2: Thanks for your reply. Unfortunately, I am not the biggest programmer. My knowledge is quite modest.
Specifically, I want to have the shutters open, for example, to circa 7.6% (ventilation position according to Homebridge status after manual operation with the hand-held radio transmitter). The percentage display in Homekit unfortunately does not work reliably.
Maybe a dummy switch could be used to send a command such as "open blinds to 7.6%"?
username_1: So you are working with blinds. I only have shutters and with these the math is a bit complicated. In addition, I do not get percentages from the motor. I can only measure the time the motor is running. And this becomes especially tricky since I do not get notified when the motor is started via the remote control. In wait mode I am only polling the motor status every five seconds so eventually a few seconds aka centimeters aka degrees are missing.
you can modify the times in the config file if you want to have more rigid and thus more often polling.
username_2: Thank you! I had already tried that, but unfortunately I did not come to a happy solution with it.
Is there then a command I could start via http, for example something like "set blinds to 7.5%"?
Many greetings
username_2: I have solved my problem:
With the Eve app I created two scenes ("blind up" and "blind stop"); then I created an automation in the Home app with both scenes, which I converted into a shortcut; between the first scene ("blind up") and the second scene ("blind stop") I inserted a few seconds pause (2 seconds), so that the blind goes into the ventilation position.
Unfortunately, the Home app does not allow so many possibilities, therefore the small detour via the Eve app.
Thanks for the quick reply and best regards.
Status: Issue closed
|
weaveworks/ignite | 464892883 | Title: Prometheus causes panic when re-starting stopped VM
Question:
username_0: Tested by calling `ignite run -i weaveworks/ignite-ubuntu` -> `reboot` and then trying to start the VM again using `ignite start -id <name>`. `ignite-spawn` crashes with the following:
```go
panic: listen unix /var/lib/firecracker/vm/6a2b6ebafcb0e75c/prometheus.sock: bind: address already in use
```
Status: Issue closed
Answers:
username_1: Fixed with ef49edd5397390c0f0606364dc608b5b3fd4a186 |
apache/servicecomb-service-center | 976033815 | Title: 【Question】What is the difference between the 2.x branch and the 1.x branch?
Question:
username_0: What is the difference between the 2.x branch and the 1.x branch? Is the 2.x branch compatible with the code implemented based on the 1.x branch? Anything to note if you're switching from 1.3.0 to 2.0? Thanks
Answers:
username_1: 2.x架构做了重构,支持多数据源(mongo、etcd)
1.x属于维护分支 |
wso2/product-apim | 838345480 | Title: [APIM-3.2.0] - Admin portal displays only 25 applications
Question:
username_0: ### Description:
Even though the devportal has more than 25 applications, when checking the applications in the admin portal it only shows 25 applications. This is because the REST call to retrieve the application has a limit set as 25.

### Steps to reproduce:
1. Log in to devportal
2. Create 30 applications excluding Default Application

3. Log in to the admin portal
4. Click on Settings --> Applications
5. You will be able to see only 25 applications

### Affected Product Version:
wso2am-3.2.0 |
bambinos/bambi | 842696543 | Title: Does bambi mess with theano.__version__?
Question:
username_0: When bambi is not installed:
<img width="275" alt="image" src="https://user-images.githubusercontent.com/674200/112747459-fe1b8100-8fb5-11eb-9884-18ba03acba10.png">
After installing bambi and importing it:
```
--------------------------------------------------------------------------
AttributeError Traceback (most recent call last)
<ipython-input-7-3dc591cc42d6> in <module>
3 import seaborn as sns
4 import numpy as np
----> 5 import bambi
6 import textwrap
7
~/miniconda3/envs/pymc3py38/lib/python3.8/site-packages/bambi/__init__.py in <module>
1 import logging
2
----> 3 from .models import Model
4 from .priors import Prior, Family
5 from .backends import PyMC3BackEnd
~/miniconda3/envs/pymc3py38/lib/python3.8/site-packages/bambi/models.py in <module>
8 import numpy as np
9 import pandas as pd
---> 10 import pymc3 as pm
11
12 from arviz.plots import plot_posterior
~/miniconda3/envs/pymc3py38/lib/python3.8/site-packages/pymc3/__init__.py in <module>
32
33
---> 34 if not semver.match(theano.__version__, ">=1.1.2"):
35 print(
36 "!" * 60
AttributeError: module 'theano' has no attribute '__version__'
```
Answers:
username_0: Curiously this only seems to happen when installing through conda. Odd.
username_1: I tried this in a new environment, PyMC3 3.11.2, bambi 0.4.0 and I can not reproduce.
Status: Issue closed
username_0: Maybe a fluke.
username_2: For the record, I can't reproduce either. PyMC3 3.11.2, Theano 1.1.2, and Bambi 0.4.0. |
jitpack/jitpack.io | 170459163 | Title: Krackle/e6447a4 build fails on Jitpack
Question:
username_0: Log https://jitpack.io/com/github/blackberry/Krackle/e6447a4/build.log
Hi,
It would be nice if Krackle was available on Jitpack. Currently it doesn't build because a repository source is missing in the pom.xml, I think.
Thanks in advance.
Status: Issue closed
Answers:
username_0: Oops, I thought the report button would report to the project not building... |
renovatebot/config-help | 417981383 | Title: update CircleCi orb when available
Question:
username_0: My current `renovate.json`:
```
{
"commitMessagePrefix": "[renovate] ",
"groupName": "all dependencies",
"labels": [
"renovate"
],
"npmToken": "<<PASSWORD>>",
"packageRules": [
{
"packagePatterns": [
"^@dmx/"
],
"schedule": "at any time"
}
],
"rangeStrategy": "bump",
"schedule": "before 3am on the first day of the month"
}
```
I want to:
- update `@dmx/` packages when available
- update other NPM packages only once a month (but if they're updated in a PR updating a `@dmx/` package, is OK)
- update CircleCI orb version when available
- update other docker dependencies (Node.js, MongoDB, Redis...) only once a month
My `.circleci/config.yml` starts with:
```
version: 2.1
orbs:
dmx_base: dmx-io/[email protected]
```
And I want this `orbs.dmx_base` to be updated each time when a new version is published.
I guess I can type something like that in `renovate.json`:
```
"docker": {
"schedule": "at any time"
},
```
But I don't know how to specify only orb to be published at any time. I cannot find such an example in https://renovatebot.com/docs/configuration-options/ .
Answers:
username_0: Btw. please let me know if the payments for the renovate app in `dmx-io` organization are configured properly.
Currently I see this:
<img width="1063" alt="screen shot 2019-03-06 at 20 37 33" src="https://user-images.githubusercontent.com/20856232/53909696-93a3eb00-4052-11e9-99a8-04cba0b31d11.png">
username_1: You should be able to do it with one extra packageRule:
```
{
"commitMessagePrefix": "[renovate] ",
"groupName": "all dependencies",
"labels": [
"renovate"
],
"npmToken": "<<PASSWORD>>",
"packageRules": [
{
"packagePatterns": [
"^@dmx/"
],
"schedule": "at any time"
},
{
"managers": [
"orb"
],
"schedule": "at any time"
}
],
"rangeStrategy": "bump",
"schedule": "before 3am on the first day of the month"
}
```
username_0: @username_1
I tried this and renovate created an issue:
There is an error with this repository's Renovate configuration that needs to be fixed. As a precaution, Renovate will stop PRs until it is resolved.
File: `renovate.json`
Error type: The renovate configuration file contains some invalid settings
Message: `packageRules:
You have included an unsupported manager in a package rule. Your list: orb.
Supported managers are: (ansible, bazel, buildkite, bundler, cargo, circleci, composer, docker-compose, dockerfile, github-actions, gitlabci, gomod, gradle, gradle-wrapper, kubernetes, maven, meteor, npm, nuget, nvm, pip_requirements, pip_setup, pipenv, terraform, travis).`
username_1: Sorry! I mixed up manager and datasource there (orb is a datasource, not manager). For now, you'll need to replace `orb` with `circleci` in the rule I gave, although that will mean any non-orb Docker images will also get updated immediately too.
username_0: @username_1
Thanks. Now it works. However it would be nice to provide a feature to specify config for only the orb.
Status: Issue closed
username_1: You can raise a feature request in the main repo to include `datasources` as a selector in `packageRules`
username_0: @username_1
I created this https://github.com/renovatebot/renovate/issues/3359 feature request. |
nikita36078/J2ME-Loader | 924461036 | Title: NokiaUI default pixel format causes issues on "Virtual Me"
Question:
username_0: **Emulator version:**
``1.0.7-open``
**Game version:**
Tested on ``v0.14`` and ``v0.17``
**Game resolution:**
``128x128``
**Device:**
``Samsung SM-G357FZ``
**Android version:**
Android 4.4.4
**Description of the issue:**
The game crashes upon entering a new game, with the following error:
```
E/ACRA (27210): ACRA caught a IllegalArgumentException for ru.playsoftware.j2meloader
E/ACRA (27210): java.lang.IllegalArgumentException: Illegal format: 8888
E/ACRA (27210): at com.nokia.mid.ui.DirectGraphicsImp.getPixels(DirectGraphicsImp.java:18)
E/ACRA (27210): at com.nokia.mid.appl.vime.b.j(Unknown Source)
E/ACRA (27210): at com.nokia.mid.appl.vime.VirtualMeMidlet.k(Unknown Source)
E/ACRA (27210): at com.nokia.mid.appl.vime.NGameMidlet.t(Unknown Source)
E/ACRA (27210): at com.nokia.mid.appl.vime.g.run(Unknown Source)
E/ACRA (27210): at java.util.Timer$TimerImpl.run(Timer.java:284)
```
The game directly uses the pixel format returned by ``com.nokia.mid.ui.getNativePixelFormat()`` when calling the ``com.nokia.mid.ui.getPixels(short[]...)`` functions, without checking whether it is a short-based format.
The TYPE_INT_8888_ARGB value returned by J2ME Loader as the native pixel format causes the game to crash, as it is not supported by the getPixels(short[]...) function.
This issue also occurs on real Nokia devices with higher screen depths.
An option to choose the default pixel format could help solving this, but it would likely bloat the menus with a feature unuseful for vast majority of games. Is this worth fixing at J2ME Loader's side or should the game just get patched instead?
Thank you!
Answers:
username_1: I will consider adding the pixel format setting, at least this is more convenient for users than jar patching.
username_2: I think better (and maybe in general enough) make one of short-based format as default - `DirectGraphics.<get/set>Pixels` use mostly the oldest games.
username_1: Now you can set the pixel format through system properties: com.nokia.mid.ui.DirectGraphics.PIXEL_FORMAT: 565
Other values can be found here: https://github.com/username_2/JL-Mod/blob/43a6f6f1b4fba7212774443637ad1563a85041fb/app/src/main/java/com/nokia/mid/ui/DirectGraphics.java#L36
Status: Issue closed
|
igorski/MWEngine | 378769990 | Title: Unit Tests - Running / Develop
Question:
username_0: Hey,
I use Windows for dev, is there any advice you can give for a setup? I am trying to figure out how to use Android Studio as the IDE with code completion/error checking and run a single test or the whole suite.
* Do you just use a text editor and "run" in place?
* How do you run 1 test at a time when you are doing TDD (mind you I do this but not as religion)?
* I have tried to understand how AS works with NDK testing but it doesn't seem to jive with what you have going.
Answers:
username_1: Currently running the suite is done by manually invoking the _test.sh_ shell script. Haven't invested any time in how this can be done directly from Android Studio as it should be possible (assumption) to run on a per-file basis, rather than the whole suite at once. For intermediate single tests, I only include the test I'm working on in _main.cpp_.
This can be improved once https://github.com/username_1/MWEngine/issues/93 completes.
username_1: This will be tackled as part of #106, together with #93 this ensures that unit tests can be written and executed side-by-side during development without requiring custom scripts or configuration changes.
Status: Issue closed
|
louimet/E-Calc | 144991499 | Title: Evaluate if "mr" and "Ans" read properly
Question:
username_0: in longer expressions with e and pi in them, mr seems to get lost. I propose M and potentially (to diferentiate from the rest of the expression given their different level of abstraction) the use of literals "[M]" and "[Ans]". Thoughts?<issue_closed>
Status: Issue closed |
notion-enhancer/notion-enhancer | 803018703 | Title: Bug emoji sets
Question:
username_0: **describe the bug**
Bug was occurred when using emoji sets. When I change the emoji using custom emoji they won't change unless I restart the app or refresh the app.
**expected behavior**
Emoji wont show unless I restart the app or refresh the app.
**screenshots**

**platform**
Windows 10 20H2<issue_closed>
Status: Issue closed |
microsoft/PowerToys | 1024200837 | Title: FancyZones - Move newly created windows to active monitor issues when switching multiple W10 Virtual Desktops
Question:
username_0: ### Microsoft PowerToys version
0.47.1
### Running as admin
- [X] Yes
### Area(s) with issue?
FancyZones
### Steps to reproduce
Enable FancyZones
Enable Move newly created windows to current active monitor (Experimental)
Create a virtual desktop using Task View > New Desktop.
Launch applications on your two "virtual" desktops starting with the first one, then the second. Upon switching back (using Ctrl + Windows + Left/Right) to the first or second, Windows that are already open are treated like they are "newly created" and as such, follow the placement of the cursor on the monitor that it's currently on aka the active monitor.
### ✔️ Expected Behavior
When using Ctrl + Windows + Left/Right or when switching to multiple desktops, there should be a way for FancyZones to not move newly created windows if you just recently performed a multiple desktop switch within the last second or so. It should just follow the behavior exhibited when turning FancyZones completely off when doing a task view desktop switch.
### ❌ Actual Behavior
After performing a switch to a different desktop within the Windows 10 Multiple Desktop feature, windows or applications that are already running on the virtual desktop are being moved to follow the cursor or the active monitor instead of staying put during the switch.
### Other Software
_No response_ |
dart-lang/dart_style | 224349817 | Title: Inconsistent indentation of list literal between variable and getter declarations.
Question:
username_0: ```dart
var x = [
"long value that won't fit everywhere",
"long value that won't fit everywhere",
"long value that won't fit everywhere"
];
get y => [
"long value that won't fit everywhere",
"long value that won't fit everywhere",
"long value that won't fit everywhere"
];
```
In the var declaration, the list is indented by two, in the getter, it's indented by six.
I'd prefer if they both used the var indentation.
Answers:
username_1: Same issue with map literals
username_2: This is (more or less) intentional. It naturally falls out of how `=>` works, but I could have made it handle this case specifically to match variables. I didn't because I think it's useful to make these visually distinct. With a variable, the collection initializer is evaluated once. With a getter, it's evaluated each time. They really are different concepts, and I think it's OK for them to format differently.
Status: Issue closed
username_3: For this case, I often find myself changing to block-style function bodies in order to increase readability, which is unfortunate (I want that nice, clean, arrow syntax 😄)
dart
```dart
get y {
return [
...
];
}
```
Or, I'm temped to use comments as a hack to force the opening bracket to match the indentation of the closing one.
Myself and other members of my team think that starting the literal on the next line makes it easier to tell at a glance that the subsequent lines are within a list (when compared to the current formatting shown in the description).
```dart
get y => //
[
...
];
```
Same for list-returning lambdas in argument lists with trailing-comma; I would prefer for the list to not have that continuation indent, similar to list arguments. Note the difference:
```dart
new FooWidget(
y: () => [
...
],
z: [
...
],
);
```
Any chance this issue could be revisited?
Also is that argument list case a different enough code path that I should open a separate issue for it?
username_2: ```dart
var x = [
"long value that won't fit everywhere",
"long value that won't fit everywhere",
"long value that won't fit everywhere"
];
get y => [
"long value that won't fit everywhere",
"long value that won't fit everywhere",
"long value that won't fit everywhere"
];
```
In the var declaration, the list is indented by two, in the getter, it's indented by six.
I'd prefer if they both used the var indentation. |
romkatv/powerlevel10k | 681171729 | Title: Appending "in" before current directory
Question:
username_0: I was looking through the `p10k.zsh` file for this option but I couldn't find what I'm looking for specifically.
Is there a way for me to add `in` before my current directory?

So the end result of the above would look something like
`as sensanaty@sensanaty in [ICON] ~/Code/Projects/vanessa on [ICON] [ICON] master `
Answers:
username_1: Open `~/.p10k.zsh` and find this:
```zsh
# Custom prefix.
# typeset -g POWERLEVEL9K_DIR_PREFIX='%fin '
```
Uncomment the second line.
Status: Issue closed
|
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.