repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
Adyen/adyen-python-api-library | 758336419 | Title: Checkout payment details : error 702
Question:
username_0: **Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
Some banks return unwanted fields after a payment (especially creditmutuel.fr). The submission to paymentDetails fail due to these:
AdyenAPIValidationError:Received validation error with errorCode: 702, message: Structure of PaymentCompletionDetails contains the following unmapped fields <unwanted arguments>
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
I think this library should include accepted fields dictionary like in https://github.com/Adyen/adyen-node-api-library/blob/develop/src/typings/checkout.ts#L2885 to help filtering.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
Answers:
username_0: @KadoBOT What do you think ?
username_0: Ok, I'll do it as it causes problems at our side
Status: Issue closed
username_1: Hello @username_0,
Thank you for opening the pr. For us it would be hard to maintain this feature so it is recommend it to implement this on the application side.
kind regards,
Alexandros
Adyen |
Duke-GCB/bespin-api | 196742359 | Title: Enable JSONAPI exception handler and fix bug with queryset
Question:
username_0: With ` 'EXCEPTION_HANDLER': 'rest_framework_json_api.exceptions.exception_handler',`, the BrowsableAPIRenderer triggers an exception on views requiring authentication when it should return an error early on.
With this handler disabled, the browsable api renders as expected and prompts for authentication.
API clients (like curl) still get a 401 response.
With exception_handler:
```
$ curl http://127.0.0.1:8000/api/jobs/
[{"status":"401","source":{"pointer":"/data"},"detail":"Authentication credentials were not provided."}]
```
Without:
```
$ curl http://127.0.0.1:8000/api/jobs/
{"detail":"Authentication credentials were not provided."}
```
Answers:
username_0: I'm working on a branch here (https://github.com/Duke-GCB/bespin-api/tree/gut-json-api) and in bespin-ui (https://github.com/Duke-GCB/bespin-ui/commit/11af649ddac0fc7471cffb3b96fdd707bf21d5b1) that will remove the JSONAPI package, making this issue moot.
The JSONAPI spec is nice but doesn't work well unless you adopt it completely. We had planned to expose two APIs via DRF, but JSONAPI doesn't contain itself very well from the browsable API, exception handler (see above), or pagination.
The primary change we need to make on the API side is to include a resource name in responses, so that the clients (e.g. ember) know what kind of objects they are receiving. This work is underway in the gut-json-api branch.
username_0: Closing as obsolete by #12
Status: Issue closed
|
prooph/event-store | 94057471 | Title: Reduce hard dependencies on ZF2?
Question:
username_0: Would it be possible to reduce the dependencies on ZF2? I am trying to integrate the event-store into Symfony2. Event-store depends on prooph/common which has hard dependencies on several ZF2 modules.
Right now, installing event-store drags in 7 zend modules as required dependencies
* zend/event-manager
* zend/json
* zend/log
* zend/math
* zend/serializer
* zend/service-manager
* zend/stdlib
Surely these aren't all required just to use the event-store in a non-ZF2 app? Can't a bunch of these be turned into "suggests"?
Answers:
username_1: You are right! We need to check the dependencies of prooph/common. prooph/event-store should only depend on the interfaces defined in prooph/common. But prooph/common ships also with default implementations for ZF2. However, composer suggestions should work fine. We will change that.
username_1: @username_0 It will take some time to get rid of the dependcies (~2-3 weeks). We already cut some hard connections in the past but not all due to time constraints. We need to prepare new major versions for all related prooph/components and align documentation. But we have this on our roadmap anyway so stay tuned!
Status: Issue closed
username_1: [5.0-beta.1](https://github.com/prooph/event-store/releases/tag/v5.0-beta.1) ships without ZF2 dependencies. |
PetalsOnWind/Statistics-and-Econometrics-for-Data-Science | 756992699 | Title: Add autocorrelation notebook
Question:
username_0: @username_1 I think Mintu had mentioned his interest in this topic in Gitter. I guess you can both coordinate and one person can focus more on autoregression
Answers:
username_1: Hi Pankhuri, Please assign me this issue.
username_0: @username_1 I think Mintu had mentioned his interest in this topic in Gitter. I guess you can both coordinate and one person can focus more on autoregression
username_2: If no updates I would like to address this issue
username_3: Hi @username_0 you can assign this issue to @username_2 . As due to my exams i wasn't able to work on it
username_2: Thank for such good gesture. All the best for exams !
username_2: @AnushkaThakkar @username_0 I am focusing on Auto regression (AR) time series model here. So please edit issue name because it might create confusion later on.
username_4: Hi @username_2 are you creating auto regression and implementing correlation in that?
username_4: Hi @username_0 , If @username_2 is working on auto regression . I would like to contribute to auto correlation.
I will create a notebook that will implement correlation. |
joshnh/CSS-Snippets | 52978033 | Title: These shortcuts only works within braces.
Question:
username_0: 
Answers:
username_1: @username_0 Would you mind writing out exactly what the issue is?
username_0: @username_1 Embarrassed,my english is not well,i'll try to expess what the issue is:
Some snippets just work in bracket.If out of brackets, the `CSS-Snippets` doesn't work,just like the picture show.
For examle,if I input `hov` in bracket and enter `Tab`,it will show the correct representation:
:hover,
:focus {
}
,but if I input `hov` out of bracket, it just show `overflow: hidden;`.
Is i expess clear?
username_1: Yeah, I understand. I'm not entirely sure what is happening. Hopefully when I get some time I can figure it out. Thanks.
username_2: If you edit the hover snippet and change it to the below then it solves the above. Same thing can be done for the active snippet
`<snippet>
<content><![CDATA[$1:hover,
$1:focus $2]]></content>
<tabTrigger>hov</tabTrigger>
<description>CSS - :hover, :focus</description>
<scope>source.css, source.less, source.scss, source.stylus</scope>
</snippet>`
username_2: [https://github.com/skuroda/PackageResourceViewer](url)
Status: Issue closed
username_1: Thanks for letting me know, Ashley. I'll make that change now to both hover and active snippets. |
coding-blocks/poker | 623839406 | Title: Add monitoring pages
Question:
username_0: Currently, all the actions in the Application are controlled using the Django Admin Panel but we need views where we can monitor the application for new logs and check the response from the server.
While creating templates please follow the practices in the current setup
- Layouts go in the root of the templates folder.
- Each application's templates are in their separate folder.
- Each application will have its own partials folder.
While creating Views please follow these practices
- Try using class-based views wherever possible.
- Don't use `View` if the customization can be done using mixins.
- Create a folder named `views` inside the cron application which have following folders.
- `pages` - This will contain all the views which we will be working on in this issue
- `api` - This will contain all the views which we will be working on in the upcoming issue. (For now just add a `.gitkeep` file in this folder
- Inside each folder add a `views.py` and `urls.py`
- [ ] Create View's folder structure
- [ ] Implement Page Views
- [ ] Application list page view with pagination - `/applications`
- [ ] Application list page template
- [ ] Application detail view which lists all the jobs paginated - `/applications/<id>`
- [ ] Application detail view template
- [ ] Jobs detail view which lists all the logs paginated - `/jobs/<id>`
- [ ] Jobs detail view template
- [ ] Should have an execute now button
- [ ] Can apply filter on Logs ( status, created_at )
- [ ] Test the application locally and attach screenshots while making the final submission.
Answers:
username_0: 
These wireframes for the pages might help although you can add more features if you want
username_1: You mean, a template folder inside application like
```
-cron
-templates
-cron
```
Am I right?
@username_0
username_0: What I mean is something like
- templates
- cron
- index.html
- oneauth
- index.html
Status: Issue closed
|
github/roadmap | 1128886751 | Title: Actions: Management experience for Cache (Server)
Question:
username_0: ### Summary
Dependency caching is one of the most effective ways to make jobs faster on GitHub Actions.
However, caches can often go corrupt, and there is no great way to clear caches or debug issues with them. Also while we evict caches at a certain threshold limit, the cache performance can take a hit when the storage is near its quota. All this generates significant support requests and prevents the broad adoption of our caching capability.
We are now enabling a management experience for users to view and delete their existing caches and monitor their current cache storage usage within a Repository.
### Intended Outcome
- GitHub Actions enables customers to cache intermediate outputs and dependencies for their workflows. Caching is an effective way to make jobs faster and a management experience for cache usage would further promote its adoption.
- Cache Management makes it easier to debug caching issues and provides greater control over cache usage by enabling users to view and delete existing caches.
- As we start supporting caching in ES and AE, a self-serve way will reduce support tickets by providing better ways to troubleshoot and debug corrupted caches.
### How will it work?
Once we enable Cache support in GHES, We will also be enabling a new Management experience (API based) to help cache users to view and monitor their cache usage across all the workflows in their repository. |
solo-io/gloo | 715732528 | Title: Support a template for customizing discovered upstreams
Question:
username_0: **Is your feature request related to a problem? Please describe.**
I'd like to be able to control the default configuration (e.g. connectionConfig, loadBalanncerConfig) of discovered upstreams through a template
**Describe the solution you'd like**
Support a template for customizing discovered upstreams
Answers:
username_1: another option: should not be a template but expand the annotation based stuff to respect all Upstream config options
username_2: Setup UDS Syncer: https://github.com/solo-io/gloo/blob/master/projects/discovery/pkg/uds/syncer/setup_syncer.go#L17
Kubernetes plugin implements Discovery: https://github.com/solo-io/gloo/blob/a9bcedb882bbb501f47f0350fb0d6abd752bd417/projects/gloo/pkg/plugins/kubernetes/plugin.go#L20
Kubernetes implementation of Upstream Discovery: https://github.com/solo-io/gloo/blob/a9bc<KEY>bd417/projects/gloo/pkg/plugins/kubernetes/uds.go#L23
Relies on Kube UpstreamConverter: https://github.com/solo-io/gloo/blob/a9bcedb882bbb501f47f0350fb0d6abd752bd417/projects/gloo/pkg/plugins/kubernetes/uds_convert.go#L37
Which uses implementations of a ServiceConverter to modify the upstream further: https://github.com/solo-io/gloo/blob/a9bcedb<KEY>0fb0d6abd752bd417/projects/gloo/pkg/plugins/kubernetes/uds_convert.go#L69
UseHttp2Converter implementation: https://github.com/solo-io/gloo/blob/a9bcedb882bbb501f47f0350fb0d6abd752bd417/projects/gloo/pkg/plugins/kubernetes/serviceconverter/use_http2_annotation_converter.go#L26
username_3: Currently we are hardcoding just a few supported annotations ([here](https://github.com/solo-io/gloo/blob/master/projects/gloo/pkg/plugins/kubernetes/serviceconverter/use_ssl_annotation_converter.go#L30) and [here](https://github.com/solo-io/gloo/blob/master/projects/gloo/pkg/plugins/kubernetes/serviceconverter/use_http2_annotation_converter.go#L15)). The plan is to change this to a single annotation , which contains a json string of all the fields from the upstream spec to apply to the discovered upstream, e.g. `{"useHttp2": true, "sslConfig": { "sds": { "targetUri": "blah" }}}`. For an example annotation with json value, see the k8s “last applied” annotation
username_4: To clarify some questions that came up around this feature:
- The documentation introduced around this feature suggests that users use `snake_case` field names to configure discovered upstreams, i.e.:
- `gloo.solo.io/upstream_config: '{"initial_stream_window_size": 2048}'`
- This is not ideal UX, as our documentation provides these values in `camelCase`
- Some manual testing on my part suggests that `camelCase` configuration is also valid
- e.g.: `gloo.solo.io/upstream_config: '{"initialStreamWindowSize": 2048}'`
I am working on creating a PR that will:
- clarify this matter in the documentation
- introduce new testing which validates this behavior |
aaronpk/indielogin.com | 373715433 | Title: Failed to Authenticate with email, Twitter on www.downes.ca
Question:
username_0: The Twitter error message was: "Your Twitter profile linked to https://www.downes.ca/ but we were expecting to see https://www.downes.ca/. Make sure you link to https://www.downes.ca/ in your Twitter profile." Not sure what more I could have done.
Trying to login using email gives me "Invalid session state..
login
logout
validate"
with the URL https://sso.indieweb.org/auth?code=91b3cb15a7c5f6fee4df1d1305e56a0f780af20d6ea937f803a1084fcad21800&state=jU1Xn9v0c18108aWWXaplYZRKbYrjPULtdD2qnAeXM0%3D
Using Firefox Quantum 63.0 (64-bit) on Windows 10 64-bit.
Answers:
username_1: Invalid session state sounds like maybe a cookie is being blocked. Do you have any extensions enabled that would block cookies?
Status: Issue closed
username_1: This is either an error with cookies being blocked, or is a duplicate of #16, so I'm going to close it. |
massgov/feature-requests | 366514721 | Title: Add an "All agency pages" option for Alert placement
Question:
username_0: The Department of Revenue (DOR) is requesting that an "All agency pages" option be added to the Alert placement section for users when creating Drupal alerts.
An example for when an All agency feature would be beneficial was recently — when DOR received a request from the Governor’s Office to place a notice/alert on all of DOR’s pages.
Another example would be when DOR’s Commissioner or Child Support’s Deputy Commissioner has a message for their viewing audience.
We think this feature will be a very helpful and reasonable addition to the alert placement section.
**Example:**
Alert placement
Site wide alerts appear at the top of every page on the site. Page alerts appear below the menu on a specific group of pages selected by the author of the alert. NOTE: Only users with the special role of Emergency Alert author are permitted to create site wide alerts.
• Site wide on all pages
• On specific pages
**• NEW FEATURE — All agency pages**
Answers:
username_1: Jira ticket DP-10118
username_1: This has been implemented. |
ANRDigital/ANR-Deck-Builder | 373850181 | Title: Handle unknown card crashes on deck load
Question:
username_0: when the Deck load encounters an unknown card it throws an exception (in Card.setCardCount) and the app crashes.
It should instead handle as an 'unrecognised card (id#)' entry in the deck - it should still have an id, and be able to set card count
Answers:
username_0: implemented as deck.hasUnknown, deck cards has dummy entry
Deck cannot be opened, displays error message - removed action from list click
Status: Issue closed
username_0: to remove error user should update card list, then kill / reopen app |
amzn/selling-partner-api-docs | 930592387 | Title: Cannot get correct API response when the system was deployed on VPS
Question:
username_0: I can get sales orders list via SP API on my pc, but after I deployed it on the VPS, it does not work. (The orders info are NULL)
The issue is that I can get the correct result with the same access credentials. (refresh_token, client_id, client_secret, access_key, secret_key, role_arn)
I do not set any ip limit on AWS role.
Is there any IP limit for the SP API?
Answers:
username_0: SP API does work on my pc but NOT work on cloud
username_0: I tried create another app in the Amazon developer console, it does not work either. |
korzio/note | 479358226 | Title: Sprint Change - p2
Question:
username_0: 
Answers:
username_0: 
username_0: - https://hangouts.google.com/call/mpHMRoWZUrsbr4y5xHGxAEEI
- https://note-edu.slack.com/archives/CKP4XAA9G/p1566649228001100
- https://medium.com/@username_0/educational-javascript-project-4th-sprint-change-58281393785d
Status: Issue closed
|
hamxabaig/new-github-pull-request-url | 583675790 | Title: Let title be optional
Question:
username_0: If you use the quick pull URL and don't specify a title, GitHub will pre-populate the last commit message in the branch. This can be a useful default if you don't have a title to insert for the user.
If `title` is not provided, it would be nice if the package omitted it from the URL instead of erroring out, allowing GitHub to use its default title.
Answers:
username_1: @username_0 makes sense, You want to give it a try and PR here? Otherwise i'll check it out on weekend.
Status: Issue closed
|
ant-design/ant-design | 301958651 | Title: onPanelChange not triggered
Question:
username_0: <!--
IMPORTANT: Please use the following link to create a new issue:
http://new-issue.ant.design
If your issue was not created using the app above, it will be closed immediately.
-->
<!--
注意:请使用下面的链接来新建 issue:
http://new-issue.ant.design
不是用上面的链接创建的 issue 会被立即关闭。
-->
The date panel has changed when I click on a date in the calendar for the last month or the following month, but the onPanelChange method is not triggered, causing the page data to be unloaded, resulting in a poor user experience, although I can use onSelect Method to solve this problem, but I think the official should pay attention to this issue! ! ! ! ! !
Answers:
username_0: <!--
IMPORTANT: Please use the following link to create a new issue:
http://new-issue.ant.design
If your issue was not created using the app above, it will be closed immediately.
-->
<!--
注意:请使用下面的链接来新建 issue:
http://new-issue.ant.design
不是用上面的链接创建的 issue 会被立即关闭。
-->
The date panel has changed when I click on a date in the calendar for the last month or the following month, but the onPanelChange method is not triggered, causing the page data to be unloaded, resulting in a poor user experience, although I can use onSelect Method to solve this problem, but I think the official should pay attention to this issue! ! ! ! ! ! |
zalando/nakadi | 227602198 | Title: last or max occured_at per partition exposed
Question:
username_0: Is it already possible (without actually consuming the last message in the nakadi stream and checking it) to check the last occured_at or the last received_at of a partition with an endpoint? we are analyzing the nakadi stream, and to check if we got all the latest events we would also need the last occured_at. this should be easily integratable with https://github.com/zalando-zmon and i would like to avoid subscriptions all the time for it.
just getting the last message with the low level api would be easier (i don't have to create a subscription), but i would avoid using this too for deprecation reasons.
this could also be extend to just get the last message of a partition, so other usecase could be covered as well ...
Answers:
username_1: Sorry for the long delay in getting back to you. Currently it is not possible, but that sounds interesting. We'll try to include it in our backlog, but of course, PRs are always welcome :) |
IHE/IT-Infrastructure | 701354070 | Title: Open Discussion IHE issue 1: This profile does not specify the internal structure of "client_id".
Question:
username_0: Issue 1: This profile does not specify the internal structure of "client_id". This is a major concern for operations and security management. But, OAuth does not provide a full specification for client_id. It just specifies its purpose. DICOM's equivalent information attributes are: Manufacturer, Model, Software Versions, and Serial Number. The OAuth client ID must identify the device, the application (including any necessary version information), the particular instance, and any other information needed to identify the client application uniquely. Registration of clients is a significant operational and security problem that is being postponed until there is more experience with problems in the field and reasonable solutions. There is known danger from spoofing of client_id. At this time, the method for assignment of client_id is not included in the profile. In the field there are a variety of methods being tried. Many depend upon physical distribution methods or out of band communications to manage the authentication problems.
Answers:
username_0: Can we not force certificate based client authentication, where the certificate could contain this app metadata?
@username_1 is there a common solution to this?
username_1: re: yes, if the authn is certificate-backed, the certificate can contain some relevant metadata about the app and its operator, and unique cert attributes can be used as a substitute for a client_id (e.g. iss+s/n). Note, however, that "Standard" X.509 extensions may not be well suited for many app-specific attributes; I like JWTs bound the the certificate for this.
re: client_id: Note that client_id is assigned by the resource server's AS and used to bind the client registration step to subsequent client authentication when token endpoint requests are submitted to the resource server's AS. IIRC, the draft IUA workflow does not require registration and the construction of the access token takes place entirely on the client side (i.e. it is not generated by the data holder's AS), so this effectively skips the workflow steps that require the client_id in the first place. Obviously, this won't work for other flows (like authorization_code).
username_0: I would like to see an option defined for certificate authentication of app registration. I would agree that this should not be minimal support.
@username_1 can you provide some links to specifications regarding your proposal for certificate based app registration? (I did say that right, correct?)
username_1: @username_0 yes UDAP Trusted Dynamic Client Registration is certificate-backed.
The specification is here: http://www.udap.org/udap-dynamic-client-registration.html.
In reviewing the updates to the IUA draft, I find that there is still a disconnect between the current IUA draft and the typical OAuth deployments for healthcare use cases that I am familiar with, especially when it comes to app registration and client_ids. The IUA actor labeled "Authorization Server" doesn't appear to be primarily representing the resource server's interests, but appears to be more closely coupled with the client. (If not, how does the AS know what data to populate in the assertion about purpose of use, consent, etc.?) It seems to me like the IUA-AS is based on the entity in XUA that generates the SAML assertions that a requestor includes in its requests. At least for the privileged app flows with client_credentials. This helps explain why typical OAuth 2 app registration and client_id management are not easily worked into this model.
UDAP works these types of assertions into OAuth 2 flows by adapting the assertion framework in RFCs 7521-7523. I'd be happy to discuss further if the IUA stakeholders are interested in aligning IUA more closely with UDAP. Though this is a non-trivial change to IUA flows, it might be possible to include by reference as an alternative flow. See, for example, section 5.2 of [UDAP Client Auth Grants](http://www.udap.org/udap-client-authorization-grants.html). Note that in UDAP, the entity that creates the Authorization Tokens for the client is called the "Client Token Service" (and is separate from the Authorization Server that represents the responder).
username_0: I am not clear what the action is that is left to do on this one. Will this become an open issue? Is there clarifications we need to address in the text? Has other pull requests resolved some of this?
Status: Issue closed
|
spring-projects/spring-boot | 937631995 | Title: Unexpected behavior while trying to persist entity with id strategy = GenerationType.SEQUENCE and existing column marked as @CreationTimestamp
Question:
username_0: I've got an unexpected behavior while I have been trying to persist entity with id
```
strategy = GenerationType.SEQUENCE
```
and existing column marked as `@CreationTimestamp`.
Assumption: there are two models `Message1` and `Message2` defined as follows:
```
@Entity
@Getter
@Setter
public class Message1 {
@Id
@GeneratedValue(strategy = GenerationType.IDENTITY)
@Column(name = "id", updatable = false)
private Long id;
@Column(name = "payload", columnDefinition = "text")
private String payload;
@Column(name = "created", columnDefinition = "timestamp with time zone")
@CreationTimestamp
private ZonedDateTime created;
@Override
public boolean equals(Object o) {
if (this == o) {
return true;
}
if (o == null || getClass() != o.getClass()) {
return false;
}
Message1 message1 = (Message1) o;
return Objects.equals(id, message1.id);
}
@Override
public int hashCode() {
return Objects.hash(41);
}
}
```
and
```
@Entity
@Getter
@Setter
public class Message2 {
@Id
@GeneratedValue(strategy = GenerationType.SEQUENCE)
@Column(name = "id", updatable = false)
private Long id;
@Column(name = "payload", columnDefinition = "text")
private String payload;
@Column(name = "created", columnDefinition = "timestamp with time zone")
@CreationTimestamp
private ZonedDateTime created;
[Truncated]
Assertions.assertNotNull(message1.getCreated());
}
```
and test for the `Model2` fails
```
@Test
@Transactional
void message2SaveCreatedTest() {
Message2 message2 = new Message2();
message2.setPayload("message2");
message2 = message2Repository.save(message2);
Assertions.assertNotNull(message2.getCreated());
}
```
Application code with configurations and tests you may find here
[https://github.com/username_0/creationtimestamp](https://github.com/username_0/creationtimestamp)
Answers:
username_1: `@CreationTimestamp` isn't honoured until the entity is first inserted into the database and the different generated value strategies change when this first insert occurs. In the case of `Message2` it is delayed beyond the `save` and `created` remains `null`. You can flush the repository to trigger the insert:
```java
@Test
@Transactional
void message2SaveCreatedTest() {
Message2 message2 = new Message2();
message2.setPayload("message2");
message2 = message2Repository.saveAndFlush(message2);
Assertions.assertNotNull(message2.getCreated());
}
```
Your test will now pass.
Status: Issue closed
username_0: @username_1 Could you please explain how `@GeneratedValue(strategy` influence on insertion to the database?
username_1: That's really a Hibernate question. https://thorben-janssen.com/jpa-generate-primary-keys/ contains quite a good explanation.
If you have any further questions, please follow up on Stack Overflow or [Gitter](https://gitter.im/spring-projects/spring-boot). As mentioned in [the guidelines for contributing](https://github.com/spring-projects/spring-boot/blob/master/CONTRIBUTING.adoc#using-github-issues), we prefer to use GitHub issues only for bugs and enhancements. |
dehli/TouchDraw | 176377851 | Title: Cocoa pod v 1.3.1 Missing Files
Question:
username_0: Hello,
When I ran pod update I noticed a new version of TouchDraw had been published.
After updated to 1.3.1 I was unable to compile due to missing files. Specifically where "Stroke" is defined.
I uninstalled the pod file and forced version 1.3.0 to be installed and the project compiled fine again.
I believe it's probably a reference issue in the pod spec file or something related to that.
Cheers,
TB
Answers:
username_1: Thanks for letting me know! I refactored the code so that there were 3 files instead of just the one. I'll try to figure out what the issue is tonight and push it. Sorry!
username_0: No problem! I am grateful that this library exists in the first place! It has saved me a lot of headache!
username_1: Hey, I just created a demo project and created a podfile referencing `TouchDraw` version `1.3.1`. It ran fine for me. Can you try again? Maybe try removing the pods folder and redownloading?
Status: Issue closed
username_0: So to fix it, I had to remove the pod by commenting out the TouchDraw. Place the TouchDraw pod back into the pod file. Run pod install and it seems to fix it.
username_1: Awesome! Glad it's working for you! |
dilshod/xlsx2csv | 107676089 | Title: Problem with conversion
Question:
username_0: I have one xlsx file that doesn't convert correctly. It appears that within a line of text there might be a line feed. This is running on Solaris.
An example is "Laura R_x000D_\n3318Park Ave" This is the only file that I have not been able to successfully convert. If I convert to xls in excel and then run xls2csv.pl it converts fine. |
dart-lang/sdk | 92235512 | Title: pkgbuild/samples/third_party/dromaeo in win-russian fails
Question:
username_0: Not sure why yet. This seems to have started with the reorganization of dart2js's libraries (see b18c5a3e7c64e1346c80466231679fd1b8db1951 and #23750)
```
pkgbuild/samples/third_party/dromaeo
Expected: Pass Slow
Actual: Fail
Could not find asset compiler_unsupported|lib/sdk/internal/compiler/js_lib/shared/async_await_error_codes.dart.
```
More details at: http://build.chromium.org/p/client.dart/builders/pkg-win-russian-be/builds/88/steps/pkgbuild_repo_pkgs%20tests/logs/stdio
```
stdout:
Loading source assets...
Loading dromaeo and smoke/src/default_transformer transformers...
Building dromaeo...
[Info from Dart2JS]:
Compiling dromaeo|web/Dromaeo.dart...
[Info from Dart2JS]:
Took 0:00:17.332933 to compile dromaeo|web/Dromaeo.dart.
[Info from Dart2JS]:
Compiling dromaeo|web/tests/dom-modify-html.dart...
[Info from Dart2JS]:
Took 0:00:07.441773 to compile dromaeo|web/tests/dom-modify-html.dart.
[Info from Dart2JS]:
Compiling dromaeo|web/tests/dom-attr-html.dart...
[Info from Dart2JS]:
Took 0:00:04.399538 to compile dromaeo|web/tests/dom-attr-html.dart.
[Info from Dart2JS]:
Compiling dromaeo|web/tests/dom-query-html.dart...
[Info from Dart2JS]:
Took 0:00:05.179599 to compile dromaeo|web/tests/dom-query-html.dart.
[Info from Dart2JS]:
Compiling dromaeo|web/tests/dom-traverse-html.dart...
[Info from Dart2JS]:
Took 0:00:04.883175 to compile dromaeo|web/tests/dom-traverse-html.dart.
stderr:
Could not find asset compiler_unsupported|lib/sdk/_internal/compiler/js_lib/shared/async_await_error_codes.dart_.
package:pub/src/barback/pub_package_provider.dart 112 PubPackageProvider._assertExists
package:pub/src/barback/pub_package_provider.dart 106 PubPackageProvider.getAsset.<async>
dart:isolate _RawReceivePortImpl._handleMessage
This is an unexpected error. Please run
pub --trace build
and include the results in a bug report on http://dartbug.com/new.
```
Answers:
username_1: Closing this out since we're removing Dromaeo anyway.
Status: Issue closed
|
clearlydefined/service | 612192426 | Title: Component URL has duplicate versions in curation URL
Question:
username_0: Looking at https://github.com/clearlydefined/curated-data/pull/3960 as an example, notice the link at the bottom of the first comment. It points to the definition on ClearlyDefined but the version number is duplicated as another segment on the end of the URL. The URL still works but a) looks funny and b) may mess up further navigation. |
yhirose/maxminddb | 134533088 | Title: MaxMind-DB-Writer-perl create mmdb can't read
Question:
username_0: ```
cat 1.rb
require 'maxminddb'
db = MaxMindDB.new('/tmp/MaxMind_DB_Writer_perl_create.mmdb')
ret = db.lookup('1.0.16.1')
print ret.found?
[zhangrui@RuiZhangdeMacBook-Air-2 ~ ]$ ruby 1.rb
false%
cat 2.rb
require 'maxminddb'
db = MaxMindDB.new('/tmp/GeoLite2-Country.mmdb')
ret = db.lookup('1.0.16.1')
print ret.found?
[zhangrui@RuiZhangdeMacBook-Air-2 ~ ]$ ruby 1.rb
true%
```
but python maxminxddb two mmdb files can read
```
import maxminddb
import json
reader = maxminddb.open_database('/tmp/MaxMind_DB_Writer_perl_create.mmdb')
result = reader.get('1.0.16.1')
reader.close()
print json.dumps(result, indent=2)
```
Answers:
username_0: @username_1 You can `git clone https://github.com/username_0/mmdb-script.git` get it or use `csv_to_mmdb.pl ` to create one. Thanks!
Status: Issue closed
username_0: ```
cat 1.rb
require 'maxminddb'
db = MaxMindDB.new('/tmp/MaxMind_DB_Writer_perl_create.mmdb')
ret = db.lookup('1.0.16.1')
print ret.found?
[zhangrui@RuiZhangdeMacBook-Air-2 ~ ]$ ruby 1.rb
false%
cat 2.rb
require 'maxminddb'
db = MaxMindDB.new('/tmp/GeoLite2-Country.mmdb')
ret = db.lookup('1.0.16.1')
print ret.found?
[zhangrui@RuiZhangdeMacBook-Air-2 ~ ]$ ruby 1.rb
true%
```
but python maxminxddb two mmdb files can read
```
import maxminddb
import json
reader = maxminddb.open_database('/tmp/MaxMind_DB_Writer_perl_create.mmdb')
result = reader.get('1.0.16.1')
reader.close()
print json.dumps(result, indent=2)
```
username_0: Hi
Any progress?
username_1: Sorry for the delay. I confirmed the bug with your test data. I'll take a look at the libmaxminddb source to understand what's missing in the maxmanddb ruby gem.
Thanks for your patience.
username_2: Without testing, I'd guess it is the 32 bit record size. Most of the databases released by MaxMind use a 24 bit or 28 bit record size.
username_1: @username_2, thank you for the nice clue.
@username_0, I fixed the problem. Could you get my commit 50ac399, and test it on your machine?
Thanks!
username_0: Work for me, Thanks a lot.
username_1: Thanks for the report!
Status: Issue closed
|
OHDSI/WebAPI | 350883956 | Title: Dummy "No matching concept" option in filters for Cost & Util reports
Question:
username_0: Today back-end returns `No matching concept` (with id = 0) among options for filters in Cost & Util reports. The option appears due to empty stratum-s, which represent "All values" rows:
```
from @results_schema.heracles_results_dist
join @vocabulary_schema.concept on concept_id = cast((CASE WHEN stratum_2 <> '' THEN stratum_2 ELSE '0' END) as INTEGER)
```
But neither name, nor the concept id (which is zero) represents the actual data in DB. What is more, since in fact there is no stratum-s with value `0` for the cases, the filter option, when applied, leads to empty set.
**Proposed solution:** the options endpoints should not return separate option for empty value - it is handled by UI already:
```
from @results_schema.heracles_results_dist
join @vocabulary_schema.concept on concept_id = CAST(stratum_2 AS INTEGER)
where stratum_2 = ''
```
Applies to both Visit and Drug.
<issue_closed>
Status: Issue closed |
rime/home | 1089212551 | Title: 如何輸入 Unicode 符號?
Question:
username_0: 平時使用的是注音輸入。
使用拼音和注音的話,如何輸入 Unicode 符號?
如:♡, ♢, ♣, ♤ (U+2661 - U+2664)
Answers:
username_1: 可以参考 Emoji 的用法:
```
五角星 五角星 🟉
爱心 爱心 ❤︎
黑桃 黑桃 ♠︎
```
username_0: 謝了,我有配置 Emoji 想打 三角形來着。。。發現原來是 🔺(小三角)
但好像還是與 Unicode 的 ▲ 不太相同
Emmmm,我再看看如何添加字典好了。
謝謝咯。 |
google/blockly | 993495065 | Title: Use `goog.require` to obtain interface types to be extended or implemented
Question:
username_0: Normally an imported name which is mentioned only in comments (i.e., used only as a type) can (and should) be obtained by `goog.requireType`d rather than `goog.require`.
Unfortunately , based on discussion in google/closure-compiler#3851, it appears that—due to a limitation in the Closure Compiler implementation—if that name is an `@interface`, then files which contain code that `@extends` or `@implements` that interface need to obtain the interface name using `goog.require` in order for the compiler to properly type check uses of the interface elsewhere.
The Closure Compiler contains code which will produce an error if `goog.requireType` is used where `goog.require` is needed, but another limitation of the compiler is that *this error is only emitted if the code in question is in a `goog.provide` file rather than a `goog.module`*.
As a result of the missing error message and my ignorance of the above facts I was able in #5343 to erroneously replace many instance of `goog.require` with `goog.requireType`. This mistake should be (partially) reverted:
* Where an interface is type is used in an `@extends` or `@implements` clause, then it should be obtained by `goog.require`.
* Where an interface is _only_ used in type declarations (`@type`, `@param`, `@returns` etc.) then it should be obtained by `goog.requireType`.
(Both of these rules apply to all interface types, not just the ones in `core/interfaces/*.js` that I 'fixed' in #5343.)
Answers:
username_0: If that's allowed, then yes!
Status: Issue closed
|
TeamSpectral/SpectralGuns | 65458713 | Title: Gun assembly station: Component crafting tab
Question:
username_0: A new tab inside the gun assembly station. It has aregular crafting window, but with a button in the right-top corner saying "Show recipes". If you click it, a menu will pop up beside the GUI (attached to it, but doesn't change the GUIs alignment) and the button will change to "Hide recipes".
The menu should be having it's custom style, including some fancy opacity and such. In the menu, we have a list of the different component types; Magazine, Barrel and so on. Clicking them will lead to a new page, containing the different types if the component; Snowball, Fireball". Clicking one of these, will show the types of that again; Wood, Iron, Gold, Diamond. Clicking one of these will paste the recipe inside the GUI.
It will only be in the background of the boxes, and you'll not be able to put in any other items than used to make the recipe. To clear it out, there's a "Clear" button in the bottom right corner.
Inside the GUI there's also a "<" button in a header. It will bring you back to
BRB, gonna eat. Will write the rest soon. |
BrailleRapSP/BrailleRapSP | 791242779 | Title: files needed for braillerap manufacturing and sales distribution
Question:
username_0: Jerry sent me the following email... I need these files to get a price quote for manufacturing and selling braillerap to the general public... Will someone from this project please respond in order to get started with mass sales distributing and marketing of the braillerap embosser?
----------------------------------------
Josh,
Please let me know when the files needed https://macrofab.com/knowledgebase/required-design-files/ are available. At that time we can assist.
Regards,
Jerry
<NAME> Director - Customer Success
MacroFab, Inc.
O: 713.300.2591 x713
M: 713-384-9911
MacroFab Podcast | Twitter | Slack |
electron/electron | 649011502 | Title: Can't access location
Question:
username_0: <!-- As an open source project with a dedicated but small maintainer team, it can sometimes take a long time for issues to be addressed so please be patient and we will get back to you as soon as we can.
-->
### Preflight Checklist
<!-- Please ensure you've completed the following steps by replacing [ ] with [x]-->
* [x] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/master/CONTRIBUTING.md) for this project.
* [x] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/master/CODE_OF_CONDUCT.md) that this project adheres to.
* [x] I have searched the issue tracker for an issue that matches the one I want to file, without success.
### Issue Details
* **Electron Version:**
* 9.0.5
* **Operating System:**
* Windows 10 1909 (64-bit)
I am developing a weather app using OpenWeatherMap and [this tutorial](https://enlight.nyc/projects/weather). I tried getting the user's location using `navigator.geolocation.getCurrentPosition` in the new Microsoft Edge and it worked fine, but when I compiled it with Electron and ran it, it seemed to automatically deny the permission. [This issue](https://github.com/electron/electron/issues/3268#issuecomment-152249479) says that Electron should automatically accept, so I'm wondering if I'm missing something. Thanks
I've put the code on a gist [here](https://gist.github.com/username_0/a00409d980c3a8182696862a8a3b12d9).
Answers:
username_1: Facing something similar
username_2: the same on mac,when i visit https://www.googleapis.com/,it came a 400 code
Status: Issue closed
|
zopefoundation/ZODB | 630096887 | Title: Asyncio support
Question:
username_0: Are there any plans for ZODB to support asynchronous reads and writes from the database with asyncio?
Answers:
username_0: Okay, I found #53. However, I'm not planning to use ZODB in a Zope project but something completely unrelated, which uses asyncio. From my understanding, in the current state this would break:
- transactions, which are thread-local (explicit declaration will bypass this)
- the async-ness, because commiting a transaction is blocking, and reading a large object from the DB could block sometimes too
username_1: #53 discusses some of the practical issues with asyncio. More fundamentally, ZODB's programming model of transparent demand-paged objects simply does not fit well with asyncio. In ZODB, any attribute access anywhere on any persistent object could potentially lead to (blocking) database calls. That doesn't work well with asycio's "all yield points must be excruciatingly annotated as such" model.
For asynchronous programming, gevent does work extremely well with ZODB (when backed by a storage like RelStorage and at least at one point, ZEO).
username_0: Perhaps when accessing these attributes, an await can be applied to the actual attribute - as though the attribute is a coroutine (using an @property), and similarly awaiting on transaction saves? Without asyncio, ZODB is useless to me, as I'm already highly invested in it.
username_1: Persistent ZODB objects work by [overridding `__getattribute__` ](https://github.com/zopefoundation/persistent/blob/master/persistent/persistence.py#L277) ( to retrieve their state when needed. (Note that this happens in a base class that all persistent objects must extend.) `__getattribute__` is implicitly called by the Python runtime. I don't know how one would make that `async`.
username_0: Yep, was reading the source and saw that. Surely we can return a coroutine from __getattribute__, which can then be awaited by user code?
username_1: Even if that were possible (I'm not sure it is) it would be **extremely unpleasant** because you'd have to do that for *every* attribute access. And like a disease, it would spread, to every consumer of any persistent object (`asyncio's` major fatal flaw). Since persistent objects are *just objects* that no one has to treat any differently from any other object, that rather defeats the point (e.g., *every* consumer of any object would have to be modified to handle a persistent object — every function everywhere would have to be `async def`; what does that even mean for C extensions?).
Hypothetical sketch of what that would look like:
```python
async def debit(account_number, amount):
p_bank_account = connection.root()[account_number]
can_debit = await p_bank_account.can_debit
if not can_debit:
frozen = await p_bank_account.frozen
act_type = await p_bank_account.type
throw AccountError(
"Cannot debit. Account might be frozen (%s) or wrong type (%s)" % (frozen, act_type))
# Calling a method first gets the attribute so we have to wait for that
has_enough_funds = await p_bank_account.has_enough_funds
# The method itself probably uses other attributes, including methods,
# so it must be declared async too...
has_enough_funds = await has_enough_funds(amount)
if not has_enough_funds:
throw AccountError("Cannot debit, insufficient funds")
# Logging is a problem. We can't just pass the object there, logging won't wait for
# anything.
logging.info("Debiting %s from account %s", amount, p_bank_account) # WRONG
# etc…
```
username_0: Regarding the way you fetch has_enough_funds, I would assume that methods aren't saved to the db (are transient) by default, so that could be safely removed.
Regarding your complaint that asyncio spreads like a disease, that's sort of the whole point of it. I'd say that it should be optional - there could he an AIODB class, perhaps?
About logging, you can probably just override __repr__ to provide the data in the str only if its available, as in it will just log "<evicted BankAccount at 0x12345678>" if the object was not loaded to RAM. For that to work you should create a new thing, _c_property which is either the property or a marker token, to mark that the property wasn't loaded to RAM.
username_1: Getting attribute method
1
```
username_1: Individual persistent objects talk to an [`IPersistentDataManager`](https://github.com/zopefoundation/persistent/blob/master/persistent/interfaces.py#L359) implementation stored in their `_p_jar` (as in "pickle jar") to save and retrieve their state. In ZODB, `IPersistentDataManager` is provided by the `Connection`, which is created by the DB with a particular `IStorage` instance that holds the data, but there are other persistent data manager implementations out there. At a minimum, one would need to create a new `IPersistentDataManager` implementation, but to work within the ZODB framework, it's probable that a new `IStorage` implementation would be the better option. To get the semantics sketched out above, one would probably also have to create a new `Persistent` base class.
Or perhaps everything async could be hidden inside the `IStorage` implementation by kicking work off to a `asyncio` cooperative thread pool and waiting for those tasks.
Happy hacking!
username_0: Of course they go through getattribute, just don't know whether ZODB pickles them... It seems wasteful to pickle the functions when they are provided by the code
username_0: Oh, okay. See you next week, when I try to start and utterly fail :P
username_1: Getting attribute method
42
```
username_0: That's true. I guess it will need an await then unless they are annotated @classmethod perhaps. I guess it would still go through getattribute, but we can detect a classmethod and skip it, perhaps? I think that wouldn't work, but perhaps a similar solution could work (applying a marker decorator to class-level functions we want to use, regardless of what's in the db)
username_1: They're you're getting into the descriptor protocol, which is somewhat complex (functions themselves are non-data descriptors, which is how they implement binding; a `@classmethod` just wraps a different descriptor around the function).
Fortunately, that complexity is not the responsibility of `Persistent` or `ZODB`. The job of `Persistent.__getattribute__` is to restore the object's data (it's `__dict__` and `__slots__`) using the storage provided by `_p_jar` (ZODB). Once that's done, actually finding the correct value, taking all the descriptor mechanics into account, is left to `object` itself.
Doing anything else would require a different implementation of `Persistent`.
username_0: From my understanding of `Persistent`, invoking `Persistent.__getattribute__` will still trigger the data storage to load the object, even if the object being fetched is a descriptor, and hence, without additional modification to either `Persistent` or the data store, the object will still be loaded even for fetching a descriptor.
[Descriptors are searched for in the instance before the class](https://docs.python.org/3/reference/datamodel.html#invoking-descriptors), so assuming that all class-level descriptors are not replaceable on instances is false.
I think descriptors don't really solve the problem, the best way to avoid it would surely be an annotation applied from Persistent to mark that the class-level attribute never requires lookup (this would presumably make the library faster too) |
mitra-social/mitra-frontend | 679549696 | Title: Accessing login page after being logged in behaves incorrect
Question:
username_0: If I'm logged in. Ans visit again the `/login` path it prompts again for credentials. Instead it should automatically redirect me to `/`. If a valid access token is available in the local store of the browser.<issue_closed>
Status: Issue closed |
TeamCOMPAS/COMPAS | 950336005 | Title: Implement ECSN fixes from my kicks paper
Question:
username_0: **Is your feature request related to a problem? Please describe.**
We should implement the ECSN fixes proposed in my NS kicks paper (arxiv:2107.04251).
**Describe the solution you'd like**
In the paper, we propose that the ZAMS mass range for ECSN progenitors should be reduced to below 0.2Msol, but this is not that practical to implement directly in COMPAS because we don't trigger ECSNe based on the ZAMS mass. The simplest reasonable solution is to adjust the trigger for ECSN to account for previous MT events (or simply H env mass).
For stripped progenitors which _should_ go ECSN, we can play with the values of e.g Mcbur1,2 and see how that affects predictions for DNS merger rates. That aspect might be worth a follow up (mini) paper, perhaps around the time of the O3b data release to better compare with any updated rates from the LVK gang.
**Describe alternatives you've considered**
Leave things as is, I guess?
Answers:
username_1: Suggestion:
add a new option (something like noECSNinSingleStars); if that option is set, check if a star that goes into the ECSN branch has a hydrogen envelope, and if it does, it stays a WD, avoiding an ECSN. |
netbox-community/netbox | 1155801789 | Title: Automatically adjust documentation theme to match user's preference
Question:
username_0: ### Change Type
Addition
### Area
Other
### Proposed Changes
Per the [mkdocs-material documentation](https://squidfunk.github.io/mkdocs-material/setup/changing-the-colors/#color-palette-toggle), it should be possible to automatically toggle light/dark mode as requested by the user's browser.
Answers:
username_1: Submitted a PR for this.
Tested it across Safari, Firefox & Chrome on macOS 12.2 successfully. Documentation theme followed my system preference for light/dark mode.
Status: Issue closed
username_0: Thanks @username_1! |
gostor/gotgt | 282514904 | Title: Potential memory leak?
Question:
username_0: I created a dummy lun using null backing store and wrote 4GB using `dd` from `/dev/zero` and resident size for `gotgt daemon` grew to 5GB even though null backing store doesn't keep track of the writes. I think this suggests a memory leak.
I had to do the following changes to get null backing store working through config file:
```
username_0@iron:~/go/src/github.com/gostor/gotgt$ git diff
diff --git a/pkg/scsi/backingstore/null.go b/pkg/scsi/backingstore/null.go
index 2cbde02..56edd93 100644
--- a/pkg/scsi/backingstore/null.go
+++ b/pkg/scsi/backingstore/null.go
@@ -40,6 +40,7 @@ func newNull() (api.BackingStore, error) {
}
func (bs *NullBackingStore) Open(dev *api.SCSILu, path string) error {
+ bs.DataSize = 1024 * 1024 * 1024 * 1024
return nil
}
@@ -56,11 +57,11 @@ func (bs *NullBackingStore) Exit(dev *api.SCSILu) error {
}
func (bs *NullBackingStore) Size(dev *api.SCSILu) uint64 {
- return 0
+ return bs.DataSize
}
func (bs *NullBackingStore) Read(offset, tl int64) ([]byte, error) {
- return nil, nil
+ return make([]byte, tl), nil
}
func (bs *NullBackingStore) Write(wbuf []byte, offset int64) error {
username_0@iron:~/go/src/github.com/gostor/gotgt$
```
Answers:
username_1: Did you log out how many bytes created by "return make([]byte, tl)" until you found memory up to 5G?
I guess you produced a memory leak by yourself while you one -call-Read, one-tl-bytes-generated.
Status: Issue closed
|
Azure/azure-xplat-cli | 163234394 | Title: ARM: vm extension set: has no way to disable extensions
Question:
username_0: CLI Version: **0.10.0**
OS Type: Mac
Installation via: npm
Mode: **ARM**
ARM mode has no way to disable the installed VM Extension via the CLI.
ASM had this:
```
➜ ~ azure vm extension set --help
...
help: -b, --disable disable extension
help: -u, --uninstall uninstall extension
...
```
but ARM does not have anything like that:
```
➜ ~ azure vm extension set -h
help: Enable/disable resource extensions for a VM in a resource group
help:
help: Usage: vm extension set [options] <resource-group> <vm-name> <name> <publisher-name> <version>
help:
help: Options:
help: -h, --help output usage information
help: -v, --verbose use verbose output
help: -vv more verbose with debug output
help: --json use json output
help: -g, --resource-group <resource-group> the resource group name
help: -m, --vm-name <vm-name> the virtual machine name
help: -n, --name <name> the extension name
help: -p, --publisher-name <publisher-name> the publisher name
help: -o, --version <version> the extension version
help: -r, --reference-name <reference-name> extension's reference name
help: -i, --public-config <public-config> public configuration text
help: -c, --public-config-path <public-config-path> public configuration file path
help: -f, --private-config <private-config> private configuration text
help: -e, --private-config-path <private-config-path> private configuration file path
help: -u, --uninstall uninstall extension
help: -t, --tags <tags> Tags to set to the resource group. Can be mutliple. In the format of 'name=value'. Name is required and value is optional. For example, -t tag1=value1;tag2
help: -q, --quiet quiet mode, do not ask for uninstall confirmation
help: -s, --subscription <id> the subscription id
help:
help: Current Mode: arm (Azure Resource Management)
```
by design or just an oversight?
Answers:
username_1: Does ```help: -u, --uninstall uninstall extension``` not work? This is also listed in the ARM mode command.
username_1: OK. This is for disabling. Looking into it.
username_1: Confirmed it with the service team -- It's by design that it's not supported. Add @sauryadas FYI.
Close.
Status: Issue closed
username_0: Thanks. |
LeetCode-Feedback/LeetCode-Feedback | 1095930950 | Title: Missing Test Case - 1094. Car Pooling
Question:
username_0: <!--
Note - Any content mention below in `<!-- ->` blocks are just comments
to help you fill-up the issue. It won't be visible in the actual issue after
you click on submit.
-->
#### Your LeetCode username
<!-- Your LeetCode username -->user5617C
#### Category of the bug
- [ ] Question carPooling
- [ ] Solution
- [ ] Language
- [#] Missing/Wrong Test Cases
#### Description of the bug
<!-- A clear and concise description of what the bug is. --> For the test case
Input --> [[2,1,5]]
Ans: 4
#### Code you used for Submit/Run operation
<!--
Please make sure you wrap your code with ``` tags.
Otherwise we may reject your request.
-->
```
// carPooling
My solution is flawed : needs further iteration but I am stuck in this testcase which seams wrong
max_capacity = 0
start, i = 0, 1
_capacity = trips[start][0]
trips = sorted(trips, key=lambda item: (item[1], item[2]))
while start < len(trips) and i < len(trips):
if trips[start][2] > trips[i][1]:
_capacity += trips[i][0]
else:
if _capacity > max_capacity:
max_capacity = _capacity
_capacity = trips[i][0]
start = i
i += 1
if _capacity > max_capacity:
max_capacity = _capacity
# print(trips[start][2], trips[i][1], max_capacity)
print(max_capacity)
if capacity == max_capacity:
return True
else:
False
```
#### Language used for code
<!-- C++ --> Python
#### Expected behavior
<!-- A clear and concise description of what you expected to happen in
contrast with what actually happened. --> Result should be 4 but the answer is 2.
#### Screenshots
<!-- If applicable, add screenshots to explain your issue. -->
#### Additional context
<!-- Add any other additional context about the bug. -->
Answers:
username_1: Problem 1094 requires returning a bool value. Why are you returning an int value?
Status: Issue closed
username_2: Hi @username_0,
Thank you for reaching out to us, the test case is correct. Please feel free to refer to the discuss section of the problem to see how other users have optimized their codes! |
box/spout | 198308663 | Title: Cannot open file.
Question:
username_0: `
$reader = new ReaderFactory(Type::XLSX);
$reader->open($this->container['path'].$r);
foreach ($reader->getSheetIterator() as $sheet) {
foreach ($sheet->getRowIterator() as $row) {
var_dump($row);
}
}
$reader->close();
`
Fatal error: Call to undefined method Box\Spout\Reader\ReaderFactory::open()
Answers:
username_0: ReaderFactory::create ..... |
facebookresearch/pytorch3d | 650013596 | Title: Installing on Windows
Question:
username_0: Traceback (most recent call last):
File "<stdin>", line 1, in <module>
ImportError: DLL load failed while importing _C: The specified module could not be found.
```
Even adding `site-packages\torch\lib` to the PATH is not enough, though according to [Dependencies](https://github.com/lucasg/Dependencies) all the required dll are in located.
### Second attempt, using local clone:
I followed the instructions from [this section](https://github.com/facebookresearch/pytorch3d/blob/master/INSTALL.md#2-install-from-a-local-clone). Code changes were roughly the same for torch 1.5 than for 1.4. The command `pip install -e .` runs correctly in the developper command prompt, running msvc for some time, then installing stuff in site-packages. But I still get the same error:
```
ImportError: DLL load failed while importing _C: The specified module could not be found.
```
## Appendices
Full log of `pip install "git+https://github.com/facebookresearch/pytorch3d.git"`:
```
Collecting git+https://github.com/facebookresearch/pytorch3d.git
Cloning https://github.com/facebookresearch/pytorch3d.git to e:\tmp\appdata-eliem\temp\pip-req-build-tznsqs6k
Running command git clone -q https://github.com/facebookresearch/pytorch3d.git 'E:\tmp\AppData-eliem\Temp\pip-req-build-tznsqs6k'
Requirement already satisfied (use --upgrade to upgrade): pytorch3d==0.2.0 from git+https://github.com/facebookresearch/pytorch3d.git in e:\python38\lib\site-packages\pytorch3d-0.2.0-py3.8-win-amd64.egg
Requirement already satisfied: torchvision>=0.4 in e:\python38\lib\site-packages (from pytorch3d==0.2.0) (0.6.1)
Requirement already satisfied: fvcore in e:\python38\lib\site-packages (from pytorch3d==0.2.0) (0.1.1.post20200630)
Requirement already satisfied: torch in e:\python38\lib\site-packages (from torchvision>=0.4->pytorch3d==0.2.0) (1.5.1)
Requirement already satisfied: numpy in e:\python38\lib\site-packages (from torchvision>=0.4->pytorch3d==0.2.0) (1.19.0)
Requirement already satisfied: pillow>=4.1.1 in e:\python38\lib\site-packages (from torchvision>=0.4->pytorch3d==0.2.0) (7.2.0)
Requirement already satisfied: yacs>=0.1.6 in e:\python38\lib\site-packages (from fvcore->pytorch3d==0.2.0) (0.1.7)
Requirement already satisfied: pyyaml>=5.1 in e:\python38\lib\site-packages (from fvcore->pytorch3d==0.2.0) (5.3.1)
Requirement already satisfied: tqdm in e:\python38\lib\site-packages (from fvcore->pytorch3d==0.2.0) (4.47.0)
Requirement already satisfied: portalocker in e:\python38\lib\site-packages (from fvcore->pytorch3d==0.2.0) (1.7.0)
Requirement already satisfied: termcolor>=1.1 in e:\python38\lib\site-packages (from fvcore->pytorch3d==0.2.0) (1.1.0)
Requirement already satisfied: tabulate in e:\python38\lib\site-packages (from fvcore->pytorch3d==0.2.0) (0.8.7)
Requirement already satisfied: future in e:\python38\lib\site-packages (from torch->torchvision>=0.4->pytorch3d==0.2.0) (0.18.2)
Requirement already satisfied: pywin32!=226; platform_system == "Windows" in e:\python38\lib\site-packages (from portalocker->fvcore->pytorch3d==0.2.0) (228)
Building wheels for collected packages: pytorch3d
Building wheel for pytorch3d (setup.py) ... done
Created wheel for pytorch3d: filename=pytorch3d-0.2.0-cp38-cp38-win_amd64.whl size=3112729 sha256=7eee268e22362299ff8bda89419d10dd72667040c78fef33a5863bff95bcac10
Stored in directory: E:\tmp\AppData-eliem\Temp\pip-ephem-wheel-cache-k4as0fbe\wheels\5e\a7\7c\814c3888eaaecaaf0f533faeb3c6812d82274484be51f7441c
Successfully built pytorch3d
```
Answers:
username_1: What happens if you do `import torch` before `from pytorch3d import _C`?
Status: Issue closed
username_0: Ok, there was two things, first I was linking against the wrong version of CUDA, and then indeed calling `import torch` first fixed the issue.
username_0: Actually not sure everything is installed correctly: following the first tutorial, about chamfer loss on dolphin, it seems that despite `torch.cuda.is_available()` is true it runs on CPU:

NB: I have a single GPU so `device = torch.device("cuda:0")` should be correct.
username_0: Traceback (most recent call last):
File "<stdin>", line 1, in <module>
ImportError: DLL load failed while importing _C: The specified module could not be found.
```
Even adding `site-packages\torch\lib` to the PATH is not enough, though according to [Dependencies](https://github.com/lucasg/Dependencies) all the required dll are in located.
### Second attempt, using local clone:
I followed the instructions from [this section](https://github.com/facebookresearch/pytorch3d/blob/master/INSTALL.md#2-install-from-a-local-clone). Code changes were roughly the same for torch 1.5 than for 1.4. The command `pip install -e .` runs correctly in the developper command prompt, running msvc for some time, then installing stuff in site-packages. But I still get the same error:
```
ImportError: DLL load failed while importing _C: The specified module could not be found.
```
## Appendix
Full log of `pip install "git+https://github.com/facebookresearch/pytorch3d.git"`:
```
Collecting git+https://github.com/facebookresearch/pytorch3d.git
Cloning https://github.com/facebookresearch/pytorch3d.git to e:\tmp\appdata-eliem\temp\pip-req-build-tznsqs6k
Running command git clone -q https://github.com/facebookresearch/pytorch3d.git 'E:\tmp\AppData-eliem\Temp\pip-req-build-tznsqs6k'
Requirement already satisfied (use --upgrade to upgrade): pytorch3d==0.2.0 from git+https://github.com/facebookresearch/pytorch3d.git in e:\python38\lib\site-packages\pytorch3d-0.2.0-py3.8-win-amd64.egg
Requirement already satisfied: torchvision>=0.4 in e:\python38\lib\site-packages (from pytorch3d==0.2.0) (0.6.1)
Requirement already satisfied: fvcore in e:\python38\lib\site-packages (from pytorch3d==0.2.0) (0.1.1.post20200630)
Requirement already satisfied: torch in e:\python38\lib\site-packages (from torchvision>=0.4->pytorch3d==0.2.0) (1.5.1)
Requirement already satisfied: numpy in e:\python38\lib\site-packages (from torchvision>=0.4->pytorch3d==0.2.0) (1.19.0)
Requirement already satisfied: pillow>=4.1.1 in e:\python38\lib\site-packages (from torchvision>=0.4->pytorch3d==0.2.0) (7.2.0)
Requirement already satisfied: yacs>=0.1.6 in e:\python38\lib\site-packages (from fvcore->pytorch3d==0.2.0) (0.1.7)
Requirement already satisfied: pyyaml>=5.1 in e:\python38\lib\site-packages (from fvcore->pytorch3d==0.2.0) (5.3.1)
Requirement already satisfied: tqdm in e:\python38\lib\site-packages (from fvcore->pytorch3d==0.2.0) (4.47.0)
Requirement already satisfied: portalocker in e:\python38\lib\site-packages (from fvcore->pytorch3d==0.2.0) (1.7.0)
Requirement already satisfied: termcolor>=1.1 in e:\python38\lib\site-packages (from fvcore->pytorch3d==0.2.0) (1.1.0)
Requirement already satisfied: tabulate in e:\python38\lib\site-packages (from fvcore->pytorch3d==0.2.0) (0.8.7)
Requirement already satisfied: future in e:\python38\lib\site-packages (from torch->torchvision>=0.4->pytorch3d==0.2.0) (0.18.2)
Requirement already satisfied: pywin32!=226; platform_system == "Windows" in e:\python38\lib\site-packages (from portalocker->fvcore->pytorch3d==0.2.0) (228)
Building wheels for collected packages: pytorch3d
Building wheel for pytorch3d (setup.py) ... done
Created wheel for pytorch3d: filename=pytorch3d-0.2.0-cp38-cp38-win_amd64.whl size=3112729 sha256=7eee268e22362299ff8bda89419d10dd72667040c78fef33a5863bff95bcac10
Stored in directory: E:\tmp\AppData-eliem\Temp\pip-ephem-wheel-cache-k4as0fbe\wheels\5e\a7\7c\814c3888eaaecaaf0f533faeb3c6812d82274484be51f7441c
Successfully built pytorch3d
```
username_1: I think you mean you are using the new commit of the deform meshes tutorial just now. I don't know what task manager is really showing. How quick is the main loop? Can you try restarting the kernel and setting device manually to `torch.device("cpu")` instead and see if it gets slower?
username_0: Oh yeah forgot to mention I am running everything local (jupyter is on http://localhost) so the task manager is the one of the machine running the ipython kernel. It takes 1min 56s with cuda and 12 min with a cpu tensor so there is a difference still.
Note that the previous screenshot was taken on an example with more points (an icosphere with n=5), with the exact tutorial code the CPU load is more around 50% (but well, it depends on the chip anyway), and the GPU load is even less, like 1 or 2 percents... feels underused. ^^
username_2: Hi,
Has anyone solved this issue for windows 10? I did everything as requested and I still get: "cannot import name '_C' from 'pytorch3d'".
I also see that there is a # pyre-fixme[21]: Could not find name `_C` in `pytorch3d`. comment.
Is this a known issue?
Thank you
username_1: @username_2 Yes, pytorch3d can be built on Windows. But it's quite fiddly. I suggest you set up a new conda environment with pytorch 1.6.0 and make the changes to the pytorch sources given in issue #323. If you have problems, please open a new issue.
Status: Issue closed
username_1: A performance improvement to the mesh_normal_consistency calculation has just landed (4bfe7158b) which you should see makes a big difference to this experience. I am therefore closing this issue. |
intelxed/xed | 680015948 | Title: the difference between xed_nonterminal_enum_t
Question:
username_0: Hi team,
I have some question about [xed_nonterminal_enum_t](https://intelxed.github.io/ref-manual/xed-nonterminal-enum_8h.html#ae0eb8f71a8f89cddf0df4b1a6a2b27af), what's the difference between `XED_NONTERMINAL_MMX_B` and `XED_NONTERMINAL_MMX_R`?
I don't really understand the meaning of suffixs here, Such as `B/SB/R/N/`.
Could you please give me some example of the different suffixs of `xed_nonterminal_enum_t`?
Answers:
username_0: And what's the defference of some similar iform, such as
Status: Issue closed
username_0: Hi team,
I have some question about [xed_nonterminal_enum_t](https://intelxed.github.io/ref-manual/xed-nonterminal-enum_8h.html#ae0eb8f71a8f89cddf0df4b1a6a2b27af), what's the difference between `XED_NONTERMINAL_MMX_B` and `XED_NONTERMINAL_MMX_R`?
I don't really understand the meaning of suffixs here, Such as `B/SB/R/N/`.
Could you please give me some example of different suffixs of `xed_nonterminal_enum_t`?
username_1: The suffixes have to do with where things are encoded. B=modrm.rm, R=modrm.reg, N=VEX.vvvv or EVEX.vvvv. Then there are some less frequent special cases SB is when we use the low 3b of the opcode byte. SE is when we use part of the imm8 to encode a register. If you look at the definitions of the associated nonterminals, you'll see all other inputs to these.
With regard to your 2nd question: there are some instructions that have multiple opcodes associated with them. Most people don't care about such things, but validation people have to test everything. So I have to support the aliases and give them names. I chose to incorporate the opcode and modrm.reg value for the ones you show do disambiguate the aliases.
username_0: Can you give me an example of `ADC_GPR8_GPR8_10` and `ADC_GPR8_GPR8_12`?
```
232 ADC ADC_GPR8_GPR8_10 BINARY BASE I86 ATTRIBUTES: BYTEOP
3
0 REG0 EXPLICIT RW NT_LOOKUP_FN INVALID GPR8_B
1 REG1 EXPLICIT R NT_LOOKUP_FN INVALID GPR8_R
2 REG2 SUPPRESSED RW NT_LOOKUP_FN INVALID RFLAGS
236 ADC ADC_GPR8_GPR8_12 BINARY BASE I86 ATTRIBUTES: BYTEOP
3
0 REG0 EXPLICIT RW NT_LOOKUP_FN INVALID GPR8_R
1 REG1 EXPLICIT R NT_LOOKUP_FN INVALID GPR8_B
2 REG2 SUPPRESSED RW NT_LOOKUP_FN INVALID RFLAGS
```
the instruction `adc dil, sil` was decoded to `ADC_GPR8_GPR8_10`.
I wonder what instruction form could be decoded to `ADC_GPR8_GPR8_12`.
username_1: Notice the operand order changes for the same MODRM byte. The way x86 encodings work we usually have a "load" form that can take either a reg and memop or a reg and another reg instead of the memop. We also have store form that can take a memop and reg or a reg and a reg. So you end up with two forms that take 2 registers.
```
% obj/wkit/bin/xed -d 10 d8
10D8
ICLASS: ADC
CATEGORY: BINARY
EXTENSION: BASE
IFORM: ADC_GPR8_GPR8_10
ISA_SET: I86
ATTRIBUTES: BYTEOP
SHORT: adc al, bl
% obj/wkit/bin/xed -d 12 d8
12D8
ICLASS: ADC
CATEGORY: BINARY
EXTENSION: BASE
IFORM: ADC_GPR8_GPR8_12
ISA_SET: I86
ATTRIBUTES: BYTEOP
SHORT: adc bl, al
```
username_0: Thanks for your apply~ |
indication/docker-cacti | 418389172 | Title: spine is not work well
Question:
username_0: It shows following:
```
FATAL: Connection to Cacti database failed. Please ensure:
- the PHP MySQL module is installed and enabled.
- the database is running.
- the credentials in config.php are valid.
```
Answers:
username_0: It got segmentation fault.
```
/var/www/html/cacti # sudo -u www-data /usr/local/spine/bin/spine -C /etc/spine.conf
SPINE: Using spine config file [/etc/spine.conf]
SPINE: Version 1.2.2 starting
Segmentation fault
```
username_0: ```
/var/www/html/cacti/spine # export CFLAGS=-g
/var/www/html/cacti/spine # make
/var/www/html/cacti/spine # ./spine
SPINE: Using spine config file [/etc/spine.conf]
SPINE: Version 1.2.2 starting
Segmentation fault (core dumped)
/var/www/html/cacti/spine # gdb ./spine core.3508
GNU gdb (GDB) 8.0.1
Copyright (C) 2017 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html>
This is free software: you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law. Type "show copying"
and "show warranty" for details.
This GDB was configured as "x86_64-alpine-linux-musl".
Type "show configuration" for configuration details.
For bug reporting instructions, please see:
<http://www.gnu.org/software/gdb/bugs/>.
Find the GDB manual and other documentation resources online at:
<http://www.gnu.org/software/gdb/documentation/>.
For help, type "help".
Type "apropos word" to search for commands related to "word"...
Reading symbols from ./spine...done.
[New LWP 3509]
[New LWP 3508]
warning: Can't read pathname for load map: No error information.
Core was generated by `./spine'.
Program terminated with signal SIGSEGV, Segmentation fault.
#0 poll_host (host_id=host_id@entry=1, host_thread=host_thread@entry=1, last_host_thread=last_host_thread@entry=1,
host_data_ids=host_data_ids@entry=0, host_time=host_time@entry=0x7f4827e88960 "2019-03-07 15:33:15",
host_errors=host_errors@entry=0x7f4827e88958, host_time_double=<unavailable>,
host_time_double@entry=1551972795.383692) at poller.c:125
125 void poll_host(int host_id, int host_thread, int last_host_thread, int host_data_ids, char *host_time, int *host_errors, double host_time_double) {
[Current thread is 1 (LWP 3509)]
(gdb) bt
#0 poll_host (host_id=host_id@entry=1, host_thread=host_thread@entry=1, last_host_thread=last_host_thread@entry=1,
host_data_ids=host_data_ids@entry=0, host_time=host_time@entry=0x7f4827e88960 "2019-03-07 15:33:15",
host_errors=host_errors@entry=0x7f4827e88958, host_time_double=<unavailable>,
host_time_double@entry=1551972795.383692) at poller.c:125
#1 0x000055919a74f9c9 in child (arg=0x55919c141b60) at poller.c:79
#2 0x00007f4827c506c2 in ?? () from /lib/ld-musl-x86_64.so.1
#3 0x0000000000000000 in ?? ()
(gdb) quit
```
It may be lack of stack...
```
/var/www/html/cacti/spine # ulimit -a
-f: file size (blocks) unlimited
-t: cpu time (seconds) unlimited
-d: data seg size (kb) unlimited
-s: stack size (kb) 10240
-c: core file size (blocks) unlimited
-m: resident set size (kb) unlimited
-l: locked memory (kb) 64
-p: processes unlimited
-n: file descriptors 1048576
-v: address space (kb) unlimited
-w: locks unlimited
-e: scheduling priority 0
-r: real-time priority 0
```
It could not increace stacksize over 10MB. |
2020PB/police-brutality | 657581387 | Title: Incident in New York City, New York
Question:
username_0: ---
## Location
New York City, New York
## Date
May 30th
## Description
Police use batons to shove protestors to the ground
## Links
https://twitter.com/DonnyMartell/status/1266943106429128711?s=20<issue_closed>
Status: Issue closed |
github-vet/rangeloop-pointer-findings | 776104244 | Title: albertrdixon/romulus: Godeps/_workspace/src/k8s.io/kubernetes/pkg/apis/extensions/types.generated.go; 5 LoC
Question:
username_0: [Click here to see the code in its original context.](https://github.com/albertrdixon/romulus/blob/5a0bf879172a8077a17d89589d4b35c81db13516/Godeps/_workspace/src/k8s.io/kubernetes/pkg/apis/extensions/types.generated.go#L17461-L17465)
<details>
<summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary>
```go
for _, yyv1408 := range v {
z.EncSendContainerState(codecSelfer_containerArrayElem1234)
yy1409 := &yyv1408
yy1409.CodecEncodeSelf(e)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 5a0bf879172a8077a17d89589d4b35c81db13516 |
mhausenblas/rback | 449913167 | Title: Error using v 0.1
Question:
username_0: Hi,
Nice project!
I was trying to use it into my environment and I'm facing the following error:
```
/usr/local/bin/kubectl get sa --all-namespaces --output json
/usr/local/bin/kubectl get roles --all-namespaces --output json
/usr/local/bin/kubectl get rolebindings --all-namespaces --output json
/usr/local/bin/kubectl get clusterroles --output json
/usr/local/bin/kubectl get clusterrolebindings --output json
panic: interface conversion: interface {} is nil, not []interface {}
goroutine 1 [running]:
main.lookupRoles(0xc0000a1650, 0x7, 0xc0000a1640, 0x7, 0xc000060150, 0xc000192270, 0xc0000ea400, 0x1c, 0x20, 0xc00023eb40, ...)
/Users/hausenbl/go/src/github.com/username_1/rback/main.go:199 +0x586
main.genGraph(0xc000060150, 0xc000192270, 0xc0000ea400, 0x1c, 0x20, 0xc00023eb40, 0xc0000eae00, 0x1c, 0x20, 0x0)
/Users/hausenbl/go/src/github.com/username_1/rback/main.go:310 +0xc01
main.main()
/Users/hausenbl/go/src/github.com/username_1/rback/main.go:27 +0x141
```
My env:
OS CentOS 7.6 64 bit
K8S version: 1.13.4
Tks
Answers:
username_0: Tested here inside a Ubuntu 18.04 and the same occurs
username_1: Interesting! Thanks for reporting this @username_0 … are you using the binaries or built it yourself?
username_0: Using the binaries :)
username_1: Thanks, and what's your `kubectl` version?
username_0: ```
Client Version: version.Info{Major:"1", Minor:"13", GitVersion:"v1.13.6", GitCommit:"<PASSWORD>", GitTreeState:"clean", BuildDate:"2019-05-08T13:53:53Z", GoVersion:"go1.11.5", Compiler:"gc", Platform:"linux/amd64"}
```
username_1: OK, thanks. I suspect the `kubectl` version is the cause (mine is 1.14). Can you paste the output of `kubectl get clusterrolebindings/cluster-admin -o yaml` please?
username_0: ```yaml
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
annotations:
rbac.authorization.kubernetes.io/autoupdate: "true"
creationTimestamp: "2017-04-18T23:50:47Z"
labels:
kubernetes.io/bootstrapping: rbac-defaults
name: cluster-admin
resourceVersion: "26105543"
selfLink: /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/cluster-admin
uid: d840db91-2491-11e7-9c82-0050568f9e1a
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: cluster-admin
subjects:
- apiGroup: rbac.authorization.k8s.io
kind: Group
name: system:masters
```
Upgraded the kubectl to 1.14.2 but the error persists
username_1: So I tried to reproduce your issue, using the [Katacoda Kubernetes playground](https://www.katacoda.com/courses/kubernetes/playground#) which has an 1.11 client installed (result see in the [examples/](https://github.com/username_1/rback/tree/master/examples) directory). Seems to work fine.
username_1: This is even stranger! Hmmmm …
username_0: Yeap. The error seems to be in genGraph.
I'll put some brake points here and try to reproduce/find what's happening.
username_1: Thanks. Yeah, it's in `lookupRoles()` but I don't get why, because I get the exact same result as you get above for the CRB. Maybe some interesting other CRB that uses no name?? :)
username_0: @username_1 could figure out the problem here: when you have a RoleBinding with no subjects it returns the error.
The following is a RoleBinding that can reproduce the error:
```json
{
"apiVersion":"rbac.authorization.k8s.io/v1",
"kind":"RoleBinding",
"metadata":{
"creationTimestamp":"2017-04-24T22:53:27Z",
"name":"admin",
"namespace":"namespace1",
"resourceVersion":"27780594",
"selfLink":"/apis/rbac.authorization.k8s.io/v1/namespaces/namespace1/rolebindings/admin",
"uid":"d487bbe2-2940-11e7-bff1-0050568f9e1c"
},
"roleRef":{
"apiGroup":"rbac.authorization.k8s.io",
"kind":"ClusterRole",
"name":"admin"
}
}
```
username_0: Made a PR to fix this ;)
username_0: Now I'm facing another error with Graphviz, but not related to rback by itself :P I've a pretty huge environment and cairo is not rendering it well.
Anyway, this is nice!
username_1: Yeah, that's what my hunch was as well (missing `name` and or even `subjects`). I really appreciate you taking the time. I do, however, wonder if such a `RoleBinding` is actually valid. I mean, what would the semantic of that be? Can you tell from where that comes from? I'm not against the patch #2 and will merge it but I really don't want to support bad practices in terms of RBAC ;)
Status: Issue closed
username_0: Yes, this was strange. Anyway it seems Kubernetes API supports RoleBinding without subjects. This should be an empty array other then a null field, but it happens this is a cluster created while it was version 1.7 and then being upgraded. Probably we've hit some old migration issue
username_1: Oho! TIL :)
Any references I can peruse? |
sublimehq/Packages | 201075579 | Title: ST3 Tcl syntax highlighting not working with regexp
Question:
username_0: Here is another bug in the syntax highlighting of regexp command:
regexp -inline -all {%[a-zA-Z_]*%} "whatever"
the -inline option appears in green (string) instead of white (option).
With:
regexp -all -inline {%[a-zA-Z_]*%} "whatever"
the -all option appears in green (string) instead of white (option).<issue_closed>
Status: Issue closed |
JosefNemec/Playnite | 619163331 | Title: Not detecting Epic Installed games
Question:
username_0: I'm still having the same issue as listed in this closed issue: https://github.com/username_2/Playnite/issues/1132
That is: Cannot identify installed epic games. I found out the reason tho.
I'm using a pt-br locale of Windows 10 and my "All Users" folder inside "Users" isn't named "All Users" but "Todos os Usuários".
I managed to band aid fix this by creating a symlink of the Epic folder inside "Todos os Usuários" inside a manually created "All Users" folder.
I guess you should be able to fix that @ the base code with this information!
Answers:
username_1: Hi username_0, I have the same problem, and I can't find the "All Users" or "Todos os Usuários" folder. If you have free time, could you help me? I would appreciate it very much, thank you!
Olá username_0, eu passo pelo mesmo problema, e não consigo achar a pasta "All Users" ou "Todos os Usuários". Se tiver um tempo livre, poderia me ajudar? Agradeceria muito, obrigado!

username_0: "Exibir -> Opções -> Modo de Exibição" then uncheck "Ocultar arquivos protegidos do sistema operacional (Recomendado)."
This should make the "Todos os Usuários" folder appear in your "Users" dir.
Then just make a symbiotic link (google it) that links the "Todos os Usuários" folder to a folder named "All Users" folder. This should band-aid fix this. At least it worked for me.
username_2: Done in `devel8`.
Status: Issue closed
|
google/flatbuffers | 409310337 | Title: [C++] nested_parser is not cleaned if error occurred
Question:
username_0: This causing memory access violation in flatbuffers::Parser next time it used.
idl_parser.cpp:
```
// Parse JSON substring into new flatbuffer builder using nested_parser
if (!nested_parser.Parse(substring.c_str(), nullptr, nullptr)) {
ECHECK(Error(nested_parser.error_));
}
```
Following clean-up is not performed:
```
// Clean nested_parser before destruction to avoid deleting the elements in
// the SymbolTables
nested_parser.enums_.dict.clear();
nested_parser.enums_.vec.clear();
```
Answers:
username_1: Can you explain the context of the issue?
Is it possible to reproduce the error as part of `tests.cpp' with help of `Monster` schema?
username_0: Yes it is. You need to add invalid nested flatbuffer field to monsterdata_test.golden, e.g:
```
testnestedflatbuffer: {
color: "NONEXISTS"
},
```
And then in ParseAndGenerateTextTest() case replace
```
TEST_EQ(parser.Parse(jsonfile.c_str(), include_directories), true);
```
with
```
TEST_EQ(parser.Parse(jsonfile.c_str(), include_directories), false);
TEST_EQ(parser.Parse(jsonfile.c_str(), include_directories), false);
```
Although the right way is probably to write a testcase for invalid input json
username_0: Yes, here is test case that crashes:
```
void InvalidNestedFlatbufferTest() {
std::string schemafile;
TEST_EQ(flatbuffers::LoadFile(
(test_data_path + "monster_test.fbs")
.c_str(),
false, &schemafile),
true);
auto include_test_path =
flatbuffers::ConCatPathFileName(test_data_path, "include_test");
const char *include_directories[] = { test_data_path.c_str(),
include_test_path.c_str(), nullptr };
flatbuffers::Parser parser1;
TEST_EQ(parser1.Parse(schemafile.c_str(), include_directories), true);
TEST_EQ(parser1.Parse("{ testnestedflatbuffer: { color: \"NONEXISTS\"}}"),
false);
}
```
Status: Issue closed
username_0: Fixed in PR #5184 |
euctrl-pru/website | 65243732 | Title: Navigation context not highlighted
Question:
username_0: Clicking on any nav bar entry but `Home` does not keep the label highlighted.
The `<li>` should get classed `active`, see js code in `_includes\_navigation.html` but does not ... always.
Status: Issue closed
Answers:
username_0: Fixed by 9eb822478ad3086f20d81dee25a1d317e8d100bd |
jasonrohrer/OneLifeData7 | 413675467 | Title: game crashes on loading
Question:
username_0: For some reason, whenever i load into the game, when it says loading sprites, it freezes half way through and crashes. My cousin is getting the same problem and it was working fine until we both got the new update. Please help !!!

Answers:
username_1: This was fixed last week. If you are off-Steam, you may need to do the following:
delete:
sprites/bin_cache.fcz
sounds/bin_cache.fcz
reverbCache/bin_cache.fcz
If you are on-Steam, the newest update will fix this for you.
Status: Issue closed
|
scls19fr/AIRAC.jl | 802666464 | Title: Implements Airac arithmetic
Question:
username_0: Instead of `move`, `next`, `previous` it could be possible to do
```julia
Airac() + 1
Airac() - 1
```
or
```
Airac() + AiracCycleNumber()
Airac() - AiracCycleNumber()
```
It could also be possible to do
```
julia> a1=Airac(2018)
Airac(1801, 2018-01-04)
julia> a2=previous(Airac(2023))
Airac(2213, 2022-12-29)
julia> a1:a2
```
```julia
using Base: -
-(::Airac, ::Airac) # need to be implemented
```
Answers:
username_0: ```julia
julia> range(a1, stop=a2, step=AiracDiff())
ERROR: MethodError: no method matching -(::Airac, ::Int64)
Closest candidates are:
-(::Base.CoreLogging.LogLevel, ::Integer) at logging.jl:117
-(::Airac, ::AiracDiff) at C:\Users\Admin-pc\.julia\dev\AIRAC\src\AIRAC.jl:110
-(::Airac, ::Airac) at C:\Users\Admin-pc\.julia\dev\AIRAC\src\AIRAC.jl:114
...
Stacktrace:
[1] steprange_last(::Airac, ::AiracDiff, ::Airac) at .\range.jl:235
[2] StepRange at .\range.jl:205 [inlined]
[3] StepRange at .\range.jl:256 [inlined]
[4] _colon at .\range.jl:46 [inlined]
[5] Colon at .\range.jl:40 [inlined]
[6] _range at .\range.jl:103 [inlined]
[7] #range#43 at .\range.jl:91 [inlined]
[8] top-level scope at REPL[91]:1
``` |
cardano-community/guild-operators | 777709850 | Title: Cant see
Question:
username_0: **Describe the bug**
With node running 1.24.2, when running Guild LiveView v1.15, I show 3 peers in and 3 peers out on my Block producer. I do have 3 relays for that Block Producer.
When I updated to LiveView v1.17, it now shows 6 peers in and 0 peers out. Its still getting txs and all, and producing blocks.
On my second Pool, which uses the same 3 relays as the other one, it shows 3 peers in and 3 out like it should when running LiveView v1.15
**Screenshots**

Answers:
username_1: Check and verify the cnode port in your env file
Status: Issue closed
username_1: Closing due to no response |
Dastnbek/movie-list | 434744623 | Title: Main info about movies and recomendation
Question:
username_0: when user clicks page full info about movie and recomendation of movies according to genre will be displayed
 |
rust-lang/rust | 819054738 | Title: regression: cannot return value referencing local variable
Question:
username_0: https://crater-reports.s3.amazonaws.com/beta-1.51-1/beta-2021-02-14/gh/bow.gtetools/log.txt regressed in the beta crater run. Unfortunately the span points into a quick_error macro call, so hard to say exactly whether this is caused by something wrong in the compiler code or just changes to inference or something like that.
Probably next steps are to bisect and/or get a MCVE of that crate.
Answers:
username_1: Reduced:
```rust
use std::error::Error;
use std::fmt;
#[derive(Debug)]
pub enum IntervalError {
InvalidRange,
}
impl fmt::Display for IntervalError {
fn fmt(&self, _fmt: &mut fmt::Formatter) -> fmt::Result {
unimplemented!()
}
}
impl Error for IntervalError {}
#[derive(Debug)]
pub enum ModelError {
InvalidInterval(IntervalError),
}
impl ModelError {
pub fn description(&self) -> &str {
match *self {
ModelError::InvalidInterval(ref err) => match err {
ref otherwise => otherwise.description(),
},
}
}
}
```
The type of `otherwise` is `&&IntervalError`. Prior to #75180, this was autoderefed; now, `&impl Error` implements `Error` so no autoderef happens and so the function returns a value that borrows `otherwise`.
Similar regressions were [observed](https://github.com/rust-lang/rust/pull/75180#issuecomment-681492119) in #75180's crater run, but this particular repository [failed to build on stable](https://crater-reports.s3.amazonaws.com/pr-75180/master%23119d2a1a98fe87d4ae6cabf12134a0ef2fb95851/gh/bow.gtetools/log.txt) in that crater run. Something similar happened to [another crate](https://github.com/rust-lang/rust/pull/75180#issuecomment-767632568), and the issue was [since fixed](https://github.com/rust-lang/crater/issues/560). It turned out to involve path dependencies, and this crate has one.
In short, this is #81460, which has been declared acceptable breakage.
Bisection points to #75180's rollup.
@rustbot label -E-needs-mcve -E-needs-bisection
username_2: as per the issue analysis (thanks @username_1 !), removing the prioritize label
@rustbot label -I-prioritize
username_3: Triage: Closing this as another instance of #81460, as mentioned above (https://github.com/rust-lang/rust/issues/82667#issuecomment-788281442).
Status: Issue closed
|
dulibrarytech/digitaldu-frontend | 386462994 | Title: download packages spec
Question:
username_0: add spec to ingest structure doc https://docs.google.com/document/d/1lQcEt3_slGvSKYmw3hKKGtm9HiLsLEytdbC7oEI0xVo/edit?usp=sharing
download package of content + metadata + derivatives
Answers:
username_1: image - file will be jp2 or jpg
audio - file will be mp3 from kaltura
video - file will be mp4 from kaltura
pdf - file will be pdf
compound - file will be a zip file of individuals (this applies for images and audio/video)
download file button
download metadata button (includes transcript)
username_0: kate: limit front end users to downloading derivative file, contact for master file. downloading metadata not a requirement.
username_0: if kaltura and cannot use api to download video, have popup in button that says you will be redirected to kaltura to download file
Status: Issue closed
|
adriangodong/githubhook | 269460699 | Title: IssuesEvent: Only one label is returned when issue is labeled/delabeled
Question:
username_0: Payload for `issues` event, action is `labeled` or `delabeled` only return one label in the top-level `label` field. The `issue.labels` field contains all the resulting label set.
Answers:
username_0: After further investigation, one webhook is delivered for each label. Closing this as by design.
Status: Issue closed
|
thiagocordeiro/laravel-translator | 735716284 | Title: Translation as Default Value
Question:
username_0: Is it possible to assign a default value to the translation somehow? Maybe a flag (suggestion: "useKeysAsDefaultValue")?
Laravel, for example, gives us the option to use keys instead of the translation itself, as keys of the JSON translation files. Because of this, libraries which automatically translate JSON files relies in the value, not the key, to then generate automatically translated files (https://github.com/fkirc/attranslate as an example).
I did not found any setting which would allow me to have the Translation as both, key and default value, so i could then use this JSON file as an entry point to generate the localised JSON's in other languages.
**TL:DR**
Is it possible to have a flag which allows me to generate a JSON:
`
{
"First Message Translated": "First Message Translated",
"Second Message Translated": "Second Message Translated",
"Third Message Translated": "Third Message Translated",
}
`
Instead of the current:
`
{
"First Message Translated": "",
"Second Message Translated": "",
"Third Message Translated": "",
}
`
Answers:
username_1: Thanks @username_0 for the suggestion.
For sure it's possible, it would require us to configure a default language, because we don't want to set the value on a different language rather than the default.
Bellow an idea of the new configuration file:
```php
return [
'languages' => ['en', 'pt-br', 'es'],
'defaultLanguage' => 'en',
'useKeysAsDefaultValue' => true, // requires default language to be set, will fill the values on the default language
...
];
```
username_0: @username_1 that sounds great, yes, i would much appreciate that functionality.
Also, thanks for your work on this package. It's very helpful.
username_1: @username_0 awesome, I'll work on this improvements in the upcoming weeks ;) |
cloudfoundry/uaa | 95474916 | Title: Cannot specify groupRoleAttribute for dynamic LDAP identity provider registration.
Question:
username_0: The documentation about (LDAP identity providers)[https://github.com/cloudfoundry/uaa/blob/master/docs/UAA-LDAP.md#ldap-groups-as-scopes-configuration] suggests that it is
possible to specify the extraction of groupRoles from an LDAP attribute via the `ldap.group.groupRoleAttribute` property.
A try to dynamically create an LDAP based identity provider via an `uaac curl POST ...` request to the `/identity-providers` endpoint will result in an internal server error, since the `groupRoleAttribute` property doesn't exist in `org.cloudfoundry.identity.uaa.ldap.LdapIdentityProviderDefinition` so it cannot be set / used.
At present it seems that one can only specify this globally in the `uaa.yml` configuration file via the `ldap.groups.groupRoleAttribute` setting which is referenced in:
https://github.com/cloudfoundry/uaa/blob/786372e012e29ec51e5d16a3971b4c04ec2e7fe0/uaa/src/main/resources/ldap/ldap-groups-as-scopes.xml#L22
It would be helpful to be able to set the groupRoleAttribute on individual LdapIdentityProvider basis.
Answers:
username_1: I believe that https://github.com/cloudfoundry/uaa/commit/750b0bb3c396e13bc76b14e6d8a01614e3b57610 added that property.
Status: Issue closed
|
appium/appium | 283236828 | Title: cannot deploy app to device / emulator since upgrading to AS 3
Question:
username_0: ## The problem
We've update to AS 3 , and since then the `adb install` command stoped working
Here [https://commonsware.com/blog/2017/10/31/android-studio-3p0-flag-test-only.html](url) you can find a great explanation for the problem (Thanks to CommonsWare)
The solution is to run `adb install -t` instead ...
Therefore this part should be fixed in appium's side as well
## Environment
* Appium version : 1.7.1
## Link to Appium logs
`Encountered internal error running command: Error: Error executing adbExec. Original error: 'Command 'Android/sdk/platform-tools/adb -P 5037 -s emulator-5554 shell pm install -r /data/local/tmp/027188cea6fb44201196dca04b3a5184.apk' exited with code 1'; Stderr: 'Failure [INSTALL_FAILED_TEST_ONLY: installPackageLI]'; Code: '1'`
Answers:
username_1: how about removing `android:testOnly="true"` parameter from the manifest?
username_0: @username_1 I cannot change the "manifest"
username_1: Ok, I'll consider it as feature request
username_2: Is there any update on this?
Status: Issue closed
username_1: Duplicate of https://github.com/appium/appium/issues/10758 |
JuliaPlots/Plots.jl | 841030019 | Title: [BUG] axis labels overlap with plots in large subplots
Question:
username_0: <!-- Please search existing issues to avoid duplicates. -->
## Details
For large subplots (i.e. > 6 subplots in one direction) the x- and y-axis labels travel into the plots.
```julia
function make_plot(N,M)
plotsize = (400,400)
plts = [plot(1:10, ylabel="Some Text", plotsize=plotsize) for i in 1:N*M]
plot(plts..., layout=grid(N,M), size=plotsize .* (M,N))
end
make_plot(1,1)
make_plot(1,6)
make_plot(1,12)
```


### Backends
This bug occurs on ( insert `x` below )
Backend | yes | no | untested
-------------|-----|-----|---------
gr (default) | x | |
pyplot | | | x
plotly | | | x
plotlyjs | | | x
pgfplotsx | | | x
inspectdr | | | x
### Versions
Plots.jl version:
[91a5bcdd] Plots v1.10.2
Backend version (`]st -m`):
[28b8d3ca] GR v0.53.0
[d2c73de3] GR_jll v0.53.0+0
Output of `versioninfo()`:
Julia Version 1.5.3
Commit 7<PASSWORD> (2020-11-09 13:37 UTC)
Platform Info:
OS: Windows (x86_64-w64-mingw32)
CPU: Intel(R) Core(TM) i7-9750H CPU @ 2.60GHz
WORD_SIZE: 64
LIBM: libopenlibm
LLVM: libLLVM-9.0.1 (ORCJIT, skylake)
Environment:
JULIA_DEPOT_PATH = C:\Users\steff\.julia;D:\JuliaPro-1.5.3-1\Julia-1.5.3\local\share\julia;D:\JuliaPro-1.5.3-1\Julia-1.5.3\share\julia
JULIA_EDITOR = "D:\JuliaPro-1.5.3-1\app-1.53.0\atom.exe" -a
JULIA_LOAD_PATH = C:\Users\steff\.julia\environments\JuliaPro_v1.5.3-1;@;@v#.#;@stdlib
JULIA_NUM_THREADS = 12
JULIA_PKG_SERVER = pkg.juliahub.com
Answers:
username_1: Affects me too: https://discourse.julialang.org/t/plots-with-many-sub-plots/62440
username_2: Same for me too.
Version:
[91a5bcdd] Plots v1.16.5
[28b8d3ca] GR v0.57.5
[d2c73de3] GR_jll v0.57.3+0
versioninfo():
Julia Version 1.6.1
Commit <PASSWORD> (2021-04-23 05:59 UTC)
Platform Info:
OS: Windows (x86_64-w64-mingw32)
CPU: Intel(R) Core(TM) i7-8565U CPU @ 1.80GHz
WORD_SIZE: 64
LIBM: libopenlibm
LLVM: libLLVM-11.0.1 (ORCJIT, skylake)
Environment:
JULIA_HOME = C:\Users\Takahiro\AppData\Local\Programs\Julia 1.6.1\bin |
DigitalPlatform/dp2 | 391926804 | Title: 如何设置超期提醒的时间
Question:
username_0: 为 dp2library 的 library.xml 配置 monitors/readersMonitor 元素的 notifyDef 属性,可以定义图书到期之前的通知参数。
举例说明一下,如果定义为 notifyDef="50%,75%"
表示每当借期到了 50% 和 75% 的时候,各提醒通知一次。因为各种读者类型,各种图书类型,定义的借期长度比较复杂多变,用百分比统一定义提醒时刻是个好办法。
但也可以用固定的时间长度来定义提醒时刻。比如 notifyDef="15day,30day" 表示按照日来提醒。
这个方法有些固有的问题。比如有一种图书的借期定义为 60day,那么上述定义提醒时刻就会太早。如果有一种图书的借期定义为 15day,则首次提醒又太晚。不如最开始介绍的比例提醒合理。
不过别着急,还有一种定义方法。例如 notifyDef="-3day"。这是表示,从超期的日子开始倒着计算,前三天的时刻进行提醒。这样的适应面就大多了。上面介绍的几种定义方法可以混用。例如 nodifyDef="15day,-3day,75%"。
最后提醒一下,dp2 系统中对于借期时间的定义,时间单位不仅只能用 day,还可以用 hour。有些图书馆的阅览室不让读者借出不好,读者总要能借出去复印什么的,但不让读者借期太长,可以用小时来定义。
Answers:
username_0: ### <monitors> -- 后台监控任务
样例:
```
<monitors>
<readersMonitor startTime="15:40" />
<messageMonitor startTime="01:00" />
<arriveMonitor startTime="01:00" />
</monitors>
```
解释:
library.xml根元素下的<monitors>元素下可以使用子元素<readersMonitor>、<messageMonitor>、<arriveMonitor>来分别定义三种后台监控任务的特性:读者超期催还任务、消息清扫任务、预约到书通知任务。
<readersMonitor>、<messageMonitor>、<arriveMonitor>这三个元素都可以用startTime属性来定义每日自动启动时刻。startTime属性值的格式为:小时:分。例如“15:40”。小时采用24小时制。
如果不配置startTime属性,缺省的效果是后台任务每小时试探启动一次。显然缺省的效果是任务启动更频繁、密集,但对服务器CPU资源占用就更大一些。建议把这些后台任务安排到夜晚图书馆不开馆的时候执行,如果图书馆服务器24小时不间断开机的话。
readersMonitor监控超期未还的情况,可及时给读者发通知。
messageMonitor监控消息库内的消息,及时清理长期积累的旧消息,以免消息库所占用的硬盘空间失控。
arriveMonitor监控预约到书情况,和通知后读者未及时来取的情况。在后者情况下,软件会接着通知后面的预约者。
对于arriveMonitor,这个任务如果配置为每日启动一次,可能会因为刻度太粗,不能做到及时通知后面的预约者,要等到第二天才通知。如果希望这个任务的灵敏度高一些,建议不要配置<arriveMonitor>小节,那么缺省的效果就是每小时启动处理一次。
数据目录中log子目录中,会自动创建一些名字模式为*_lasttime.txt的文件,这些都是有用的工作文件,记载了后台任务启动的相关信息,系统管理员不要去删除这些文件。
log子目录中的日志文件中有对后台任务启动、停止状态的记载信息。
```
2018/12/17 23:32:28 (定时)arriveMonitor 启动。
2018/12/17 23:32:28 (定时)readersMonitor 启动。
2018/12/17 23:32:28 (定时)messageMonitor 启动。
2018/12/17 23:32:29 (定时)arriveMonitor结束。共处理记录 1 个。
2018/12/17 23:32:29 (定时)messageMonitor结束。共处理记录 2 个。
2018/12/17 23:32:29 (定时)readersMonitor结束。共处理记录 2 个。
```
username_1: 测试用例TC-ReadersMonitor-01 用百分比统一定义提醒时刻,比如 notifyDef="50%,75%"
测试用例TC-ReadersMonitor-02 用固定的时间长度来定义提醒时刻,比如 notifyDef="15day,30day"
测试用例TC-ReadersMonitor-03 从超期的日子开始倒着计算,比如 notifyDef="-3day"。表示:从超期的日子开始倒着计算,前三天的时刻进行提醒。
测试用例TC-ReadersMonitor-04 几种定义方法可以混用,比如 nodifyDef="15day,-3day,75%"
测试用例TC-ReadersMonitor-05 借期时间的定义,时间单位用 hour
测试用例TC-ReadersMonitor-06 不配置startTime属性,缺省的效果是后台任务每小时试探启动一次
测试用例TC-ReadersMonitor-07 配置startTime属性,后台任务在指定时间启动一次
username_1: ## 测试用例TC-ReadersMonitor-01
### 一、测试目标
用百分比统一定义提醒时刻,比如 notifyDef="50%,75%",测试读者是否收到公众号超期提醒通知。
### 二、测试环境
dp2LibraryXE升级到最新版本3.4,dp2内务升级到最新版本3.2.6994.30865
### 三、测试步骤
1)启动dp2内务,服务器地址输入单机版服务器地址,用超级管理员supervisor帐号登录。
### 四、测试结果
20190301-lcy:测试结果。
username_1: ## 测试用例TC-ReadersMonitor-02
### 一、测试目标
用固定的时间长度来定义提醒时刻,比如 notifyDef="15day,30day",测试读者是否收到公众号超期提醒通知。
### 二、测试环境
dp2LibraryXE升级到最新版本3.4,dp2内务升级到最新版本3.2.6994.30865
### 三、测试步骤
1)启动dp2内务,服务器地址输入单机版服务器地址,用超级管理员supervisor帐号登录。
### 四、测试结果
20190301-lcy:测试结果。
username_1: ## 测试用例TC-ReadersMonitor-03
### 一、测试目标
从超期的日子开始倒着计算,比如 notifyDef="-3day",测试读者是否收到公众号超期提醒通知。
### 二、测试环境
dp2LibraryXE升级到最新版本3.4,dp2内务升级到最新版本3.2.6994.30865
### 三、测试步骤
1)启动dp2内务,服务器地址输入单机版服务器地址,用超级管理员supervisor帐号登录。
### 四、测试结果
20190301-lcy:测试结果。
username_2: ### 测试用例 4
### 测试目的
读者在借书超期状态,内务执行批处理超期通知
### 测试步骤
### 测试结果 |
frontendbr/vagas | 1065533501 | Title: [Remoto] Swift Freelancer ( Pleno / Sênior ) @ Plataforma Vibbra!
Question:
username_0: ## Somos uma **plataforma freelancer fechada e gratuita** que conecta empresas e profissionais de tecnologia de **forma inteligente e cuida da gestão do relacionamento e contratação**.. 😎
Você vai receber demandas **com segurança**, flexibilidade, **sem calotes, escopos absurdos e leilão reverso**, sem se preocupar em buscar clientes **e valorizando o seu tempo**.
Você cuida da programação e a gente cuida do resto!🤝😊
Venha ter a experiência de um **match** com uma **demanda ideal com o seu perfil** aqui na **Vibbra!** 😍
#tempoévida!
## Local
Remoto!🌎
## Requisitos
**Obrigatórios:**
- 3 anos de experiência com desenvolvimento
- Ser adepto ao trabalho remoto
- Saber trabalhar com métodos ágeis
**Diferenciais:**
- Autogestão
- Comprometimento
- Prazer pelo que faz
## Contratação
PJ Autônomo, ganhos mediante às horas trabalhadas em cada projeto;
## Como se candidatar
Comece agora mesmo! Acesse o nosso **site** e clique no botão de **participar da seleção**, depois é só seguir o processo de cadastro até a ativação do seu perfil.😊
https://bit.ly/git_vagas_ios
## Dúvidas
Caso surjam dúvidas durante o processo de cadastro, você pode entrar em contato com a equipe Vibbra! pelos seguintes canais:
- [LinkedIn](https://www.linkedin.com/in/andremacieln/);
- [Email](<EMAIL>);
- Ou falar com um dos nossos Tech Recruiters pelas redes.
## Destaques
⚖️ Freela
🤝 PJ/autônomo
🗺️ Remoto
👴 Sênior
👨 Pleno<issue_closed>
Status: Issue closed |
StarsiegePlayers/ss-rerelease-vol | 1072027158 | Title: Maximum servers (port retries) before failure is coded to 20
Question:
username_0: The default number of ports for the server to try before failure is hard-coded to 20 - this may be increased or set to a variable in the future. Not that it's needed in most circumstances, but it helps with debugging master-related issues.
https://github.com/StarsiegePlayers/ss-rerelease-vol/blob/2b787761d0879bd3068b1d8f88af59b0aeadfd2b/Scripts/scripts.vol/server.cs#L20 |
ballerina-platform/ballerina-lang | 663707183 | Title: Code Analyzer fails for incomplete listener declaration
Question:
username_0: **Description:**
Consider the following source snippet,
```
import ballerina/http;
import ballerina/lang.'object as lang;
public listener http:Listener lst = new h
```
This fails with the following exception,
```
Caused by: java.lang.NullPointerException
at org.wso2.ballerinalang.compiler.semantics.analyzer.CodeAnalyzer.visit(CodeAnalyzer.java:2261)
at org.wso2.ballerinalang.compiler.tree.expressions.BLangTypeInit.accept(BLangTypeInit.java:62)
at org.wso2.ballerinalang.compiler.semantics.analyzer.CodeAnalyzer.analyzeExpr(CodeAnalyzer.java:2862)
at org.wso2.ballerinalang.compiler.semantics.analyzer.CodeAnalyzer.visit(CodeAnalyzer.java:1500)
at org.wso2.ballerinalang.compiler.tree.BLangSimpleVariable.accept(BLangSimpleVariable.java:53)
at org.wso2.ballerinalang.compiler.semantics.analyzer.CodeAnalyzer.analyzeNode(CodeAnalyzer.java:334)
at org.wso2.ballerinalang.compiler.semantics.analyzer.CodeAnalyzer.lambda$analyzeTopLevelNodes$1(CodeAnalyzer.java:323)
at java.util.ArrayList.forEach(ArrayList.java:1257)
at org.wso2.ballerinalang.compiler.semantics.analyzer.CodeAnalyzer.analyzeTopLevelNodes(CodeAnalyzer.java:323)
at org.wso2.ballerinalang.compiler.semantics.analyzer.CodeAnalyzer.visit(CodeAnalyzer.java:318)
at org.wso2.ballerinalang.compiler.tree.BLangPackage.accept(BLangPackage.java:161)
at org.wso2.ballerinalang.compiler.semantics.analyzer.CodeAnalyzer.analyze(CodeAnalyzer.java:307)
at org.wso2.ballerinalang.compiler.CompilerDriver.codeAnalyze(CompilerDriver.java:330)
at org.wso2.ballerinalang.compiler.CompilerDriver.compile(CompilerDriver.java:261)
at org.wso2.ballerinalang.compiler.CompilerDriver.compilePackageSymbol(CompilerDriver.java:248)
at org.wso2.ballerinalang.compiler.CompilerDriver.compilePackage(CompilerDriver.java:140)
at org.wso2.ballerinalang.compiler.Compiler.compilePackages(Compiler.java:166)
at org.wso2.ballerinalang.compiler.Compiler.compilePackages(Compiler.java:137)
at org.ballerinalang.langserver.compiler.LSModuleCompiler.compilePackagesSafe(LSModuleCompiler.java:291)
... 11 more
```
Answers:
username_1: Tried these in the latest master and it's not crashing for me though.
username_2: Related issue: https://github.com/ballerina-platform/ballerina-lang/issues/25210.
NPE occurs at https://github.com/ballerina-platform/ballerina-lang/blob/v2.0.0-alpha2/compiler/ballerina-lang/src/main/java/org/wso2/ballerinalang/compiler/semantics/analyzer/CodeAnalyzer.java#L3073 due to `type.tsymbol` becoming null.
username_0: @username_2 yes, we have to handle these cases with null checks at least. Since the parser is resilient and recovers the incomplete sources, at the semantic layer however we have to handle this.
username_3: Above issue is fixed in the master now. Hence, closing the issue.
Status: Issue closed
|
advancedfx/advancedfx | 454961411 | Title: csgo crashes after executing pov cfg (latest version)
Question:
username_0: This was a previous issue (#167) and it was stated to have been fixed, however for some reason the issue has risen again. I've narrowed down the crash to the command
mirv_cam source calcVecAng awmOfs ;
Whenever this command is run (with the other part of "pov.cfg") the game always crashes. This is the only command in the entire group that causes a crahsh.
Pov.cfg: https://pastebin.com/e8dvt1gn
Answers:
username_1: I can not reproduce the issue with the latest HLAE and CS:GO version:
## 1. Make sure you load with correct AfxHookSource.dll
We recommend using File -> Launch CS:GO instead of Custom Loader, because Custom Loader tends to start in old HLAE installs and thus people select an old AfxHookSource.dll there often.
If you are unsure, you can check / provide us with the AfxHookSource version by scrolling up the console after the game has been loaded, it should look something like this:
```
|
| AfxHookSource ([....])
| http://advancedfx.org/
|
| VClient: VClient018 (CS:GO)
| VEngineClient: VEngineClient014 (CS:GO)
| VEngineCvar: VEngineCvar007 (CS:GO)
| GameDirectory: C:\Program Files (x86)\Steam\steamapps\common\Counter-Strike Global Offensive\csgo
|
```
## 2. Make sure you are launching correct and fully-updated latest CS:GO install
Some people end up with two CS:GO installs and often enough launch the older (not updated one):
Please compare the **_full_** path to csgo.exe you set in HLAE with the one when you right click CS:GO in Steam Library -> Local Files tab -> Browse Local Files button.
Also enter version into your game's console, it should be like this currently (latest CS:GO update):
```
] version
Protocol version 13701 [934/934]
Exe version 1.37.0.1 (csgo)
Exe build: 09:40:07 May 29 2019 (7508) (730)
Windows version [....]
```
username_0: I've made sure that I am using the correct AfxHookSource.dll and the latest CS:GO install by looking through the console. I've already been using the File -> Launch CS:GO method to launch the game, but the issue is still there. I've also deleted all previous versions of hlae.
Status: Issue closed
username_0: Ok for some reason the issue has resolved itself, as I was testing trying out multiple different scenarios after doing a bit with mirv_campath and making a couple smooths the command seems to be working again, I don't know if this is the cause of it but everytime I restart the game it seems to be working fine now. thanks for the help!
username_1: Consider verifying game cache (button on local files tab in properties dialogue of the game). |
heroku/heroku-cli-oauth | 216496410 | Title: clients:info -s does not show all the details
Question:
username_0: `heroku help clients:info` shows
```
-s, --shell # output in shell format
```
but it appears to only show the ID and SECRET which is undefined:
```
$ heroku clients:info 01234567-89ab-cdef-0123-456789abcdef -s
HEROKU_OAUTH_ID=<KEY>
HEROKU_OAUTH_SECRET=undefined
``` |
b3ross/dotenvi | 577173547 | Title: Strings with curly braces cause memory error
Question:
username_0: env.yaml:
```
default_env: &default_env
TEST_TPL: "a string with {something} in brackets"
```
```
❯ dotenvi -s default_env
```
runs for a while and then:
```
FATAL ERROR: Ineffective mark-compacts near heap limit Allocation failed - JavaScript heap out of memory
```
Answers:
username_1: This is almost definitely failing somewhere here:https://github.com/username_1/dotenvi/blob/master/src/rewriter.ts#L25
We should probably add some state there to check whether we were actually capturing something.
We should also add a test here:
https://github.com/username_1/dotenvi/blob/master/src/rewriter.test.js
username_1: I think I have a fix in place - will merge after https://github.com/username_1/dotenvi/pull/50 is merged down.
Status: Issue closed
|
microsoft/WebTemplateStudio | 444801949 | Title: Generate the template using typescript
Question:
username_0: It would be great to have the choice between typescript and javascript for the generated template.
Answers:
username_1: Great idea
username_2: I would like to implement that feature, it shouldn't be very hard!
Do I have the guarantee for a merge if I implement it?
username_3: Hi @username_2, we really appreciate your offer, but we are not yet ready for language selection in Web Template Studio. |
greenplum-db/gpdb | 257815677 | Title: to_json() function is not available in GPDB 5.0.0
Question:
username_0: This hint is returned.
___
HINT: No function matches the given name and argument types. You might need to add explicit type casts.
___
Casting to json data type works.
The function is documented in the GPDB 5.0.0 docs and is documented in the Postgres docs.
Answers:
username_1: This has been removed from the documentation with https://github.com/greenplum-db/gpdb/pull/3303
Status: Issue closed
|
tinymce/tinymce | 299582292 | Title: Serious bug in new version 4.7.7.
Question:
username_0: Hi,
Untill this last update this bug that I am reporting didnt happen. Test case: https://jsfiddle.net/agLjo58f/
Open the test case link and click in the BACKGROUND COLOR button. You will see the colors (colorpicker) appears offscreen and the user can only see them if he scroll. ALSO if the page is not tall enough, the user wont be able to scroll so he wont see the colors.
Also when you click the PARAGRAPH button it shows weird, it should show up without having the user to scroll down the page.
Finally the hints of the buttons in the last row (when you hover then for a few seconds) dont show (actually they show at the bottom, offscreen).
The problems above could be corrected with 2 paths: 1) improve the algorithm that decides wheter the dropdown will be shown up/down or 2) give the user the ability to set the dropdown/hint to be shown above or below, many other libraries like SEMANTIC UI offer this basic UI setting since the very first release!
I am receiving tons of users complaning of this problem, I use tinymce in at least 800 websites and custeomers are complaning colors are not showing, the dropdown menus cant be accessed when the page has no enought height to allow scroll...
I know I was stupid, instead of using "http://cdn.tinymce.com/4/tinymce.min.js" in all the 800 websites my company has been developing in the last years, I should have used somethings like "http://cdn.tinymce.com/4.5/tinymce.min.js" to avoid problems... how I am complete fuc***ed and this is not TINYMCE team fault, it's mine.
However it would be nice to have this fixed.
Answers:
username_1: I moved the fiddle to: http://fiddle.tinymce.com/Okgaab so that it's easier to flip between tinymce versions.
From what I can see there the colorpickers has always appeared offscreen in that scenario. Tested 4.7.6 the version before the latest 4.7.7 where we did some alterations to how the floating panels gets calculated.
The tooltips are also offscreen in older versions so I see no regression there either.
The only issue I could find is that the the menus appear offscreen when there isn't enough room height wise.


One of the things we changed is that it now constrains the panels to the document body/document element height not the innerHeight of the window. It seem more common to do that or even not have any constraints at all.
Would be helpful to have a fiddle at fiddle.tinymce.com that reproduces the problem so we can look into it. As I see it there is no regressions between 4.7.7 and 4.7.6 in terms of color panels or tooltips they work the same for me at least.
username_1: Rolled back the document constraint in favor of the old window viewport constraint. Updated the nightly so a.f.i.i.k. everything is now working as 4.7.6 before we did that change. We will probably release this early next week to the old cdn and community download.
username_0: @username_1 thanks! You are right, everything you said is right.
The colorpicker never worked but because of a hack in my CMS we made it work! The same happens with the tooltips. However in the newest version the "+" button (and others similar to that) does not work anymore, it worked perfectly till today but in the 4.7.7 it does not work anymore.
So my suggestion:
1) fix the tooltip cause it's great having the tooltip working. It's not veeery important but it helps a lot!
2) fix the colorpicker since it's been a long time with this bug living in the code. I managed to fix it using a custom stylesheet however I think this is a good opportunity to fix it
3) the "+" button bug also happens when using custom drop down buttons and also in the paragraph button.
I thank you for reverting the change regarding this issue in the nightly version, you saved my life!
username_0: @username_1 I used your suggestion and created another test case of this bug with custom buttons. You can see the bug also here -> http://fiddle.tinymce.com/Skgaab
Click the ACOES button at the bottom, you will see the options inside cant be seen. This is something my customers complain a lot cause this didnt happen in older version.
Here is an example of the same cobe above using tinymce 4.7.6 (previous version) and you can see it works perfectly. -> http://fiddle.tinymce.com/Tkgaab
So: the link http://fiddle.tinymce.com/Tkgaab has the old version and the link http://fiddle.tinymce.com/Skgaab has the new version, that's the only change in both cases. However the old test case works and the new does not work as expected.
username_1: Yes, the nightly seems to solve that as well.
But this is a bit weird use case since tinymce is not really designed to work with the toolbar at the bottom. The default UI we have the toolbar is always at the top and the content is below the toolbar. So we don't really support any other mode out of the box.
username_0: @username_1 yeap I understood :) I know you dont support toolbar at the bottom, however it was working well till last update. Maybe you could redesign the code that was changed in order to prevent this problem.
See, in all my websites I insert the toolbar at the bottom and center aligned so the user can edit the content and use the tools at the bottom cause when you are writing a post/blog it's much more confortable having the toolbar at the bottom so you can focus at the content at the top, where your eyes are already pointed.
The nightly worked perfectly (except the hints/tooltips and the colorpicker which already didnt work), can I expect this nightly fix to be kept later in the next release or should I expect this problem in the next release?
Status: Issue closed
username_1: This bug/issue has been fixed in the latest version. If you can still reproduce this on the latest version please let us know and we re-open this ticket. Thanks.
username_0: @username_1 the bug with the colorpicker still happens. Wouldnt it be possible to fix it? |
gobengo/distbin | 420702423 | Title: Feature Request: `docker run gobengo/distbin` should work
Question:
username_0: To prevent more issues like #22.
This was heretofore never a design goal.
But this issue would take it on.
I don't have time in the next couple weeks to do this.
Answers:
username_0: I made time.
This now works:
`docker run -p 8000:80 username_0/distbin`
Status: Issue closed
username_0: https://github.com/username_0/distbin/pull/24 |
EdenServer/community | 1027932953 | Title: Golden-Tongued Culberry issue when under 666 mp
Question:
username_0: ### Checklist
<!--
Don't edit or delete this section, but tick the boxes after you have submitted your issue.
If you want to tick boxes before you submit your issue, use an x WITHOUT spaces like this: - [x]
Make sure you comply with the checklist (as in you're not just ticking boxes you actually do the checklist) and then start writing in the details section below.
-->
- [x] I have searched for existing issues for issues like this one. The issue has not been posted. (Duplicate reports slow down development.)
- [x] I have provided reproducible steps. (No "as the title says" posts please. Provide reproducible steps even if it seems like a waste of time.)
- [ ] I have provided my client version in the details. (type /ver into your game window)
- [ ] ***I have provided proof of my claim (era videos, links to era wiki and/or forum posts)*** Lack of proof will slow down fixes and prioritization. Please provide details so that developers' research time is minimized.
### Details
I was fighting Golden-Tongued Culberry and *almost* ran him out of mp. A recently dead character used a Blood Pact (Nether Blast) and Golden-Tongued Culberry started casting Stonega 3 and completed the cast even though Diabolos got hate before the cast was off.
I think this was due to Golden-Tongued Culberry not having enough MP to actually change course and cast Death. Shortly after he stopped casting all together.
Answers:
username_1: the hate shift interrupt mechanic is separate from the pet hate mechanic, so i wonder if the hate shift was just too late in the spellcast, but before the results appeared on your screen. will leave up but that is my guess. |
didi/di18n | 709886983 | Title: 单标记标签转译之后会有问题 出现多余的嵌套标签
Question:
username_0: 
Answers:
username_1: 自闭合标签确实有 bug,解决中。。
username_2: 请问这个问题解决了吗,现在还是有这个问题

而且

这种语法不支持替换报错了
 |
Chalwk77/HALO-SCRIPT-PROJECTS | 420760481 | Title: Inventory Transactions (attractive mod) for HPC|CE
Question:
username_0: ### [link to mod](https://github.com/username_0/HALO-SCRIPT-PROJECTS/blob/master/INDEV/Inventory%20Transactions.lua) | [Project Tracker](https://github.com/username_0/HALO-SCRIPT-PROJECTS/projects/3)
- - - -
Answers:
username_1: Dude
Will it work if I add the "god" custom command and can buy 30 seconds?
(the "god" will be over after spending 30 seconds?)
Example
{ ["god2"] = { 'god', "100", "30", "Purchased (30 Seconds of God) for $%price%. New balance: $%balance%", -1 } },
Status: Issue closed
username_0: ### [link to mod](https://github.com/username_0/HALO-SCRIPT-PROJECTS/blob/master/INDEV/Upgrade%20System.lua) | [Project Tracker](https://github.com/username_0/HALO-SCRIPT-PROJECTS/projects/3)
- - - -
username_0: Sorry for the late response.
I have added that feature in the latest udpate.
username_1: @username_0 Cool, thanks
username_0: No worries. Just a heads up: This mod is still in development, which is why it's in the INDEV directory.
When it's finished, it will be moved to SAPP >> ATTRACTIVE MODS.
username_1: @username_0 If I know, I am testing it if I find an error I will report it, but I think it would be better to give 5 or 10 of money for each kill.
Note: It is only my opinion.
username_0: [new update] - > Upgrade System.
[!] A couple of bug fixes.
[+] Added stock weapons to the weapons table and options to enable/disable each one individually.
[+] Added stock grenades (Plasma / Frags) to grenade table + options to enable/disable each one.
username_1: the version released 22h ago, has a bug when the game ends and starts another on another map the script stops working, in the version released recently was this bug fixed?
username_1: Script: V2.0
Error:
stack traceback: [string "UpgradeSystem2.0"]:1252: in function <[string "UpgradeSystem2.0"]:1251> [C]: in function '__sub' [string "UpgradeSystem2.0"]:778: in function 'update' [string "UpgradeSystem2.0"]:1044: in function <[string "UpgradeSystem2.0"]:974> EventCallback >> OnPlayerKill >> Lua Error: [string "UpgradeSystem2.0"]:778: attempt to perform arithmetic on local 'balance' (a nil value)
username_0: The new /weapons command iterates through the entire weapons table and retrieves the command id, price and weapon tags automatically.
Command Entries are split into 5 columns automatically and rows are determined by the amount of entries, so it's neat and tidy.
New config options:
`local weapon_list = "weapons"`
`local weapon_list_perm = -1`
`local max_columns, max_results = 5, 15`
`local startIndex = 1` -- <<--- do not touch
`local endIndex = max_columns` -- <<--- do not touch
`local spaces = 2 -- Spaces between results`
`local output_format = "/%command% | $%price%"`
username_1: Cool. I have a request, (if possible). I could do a new version without: Combo, Streaks and Assists. And get points for kill.
modify this:
[1] = { "5", "10", "Kills: (%kills%) +%upgrade_points% Upgrade Points" },
[2] = { "10", "10", "Kills: (%kills%) +%upgrade_points% Upgrade Points" },
[3] = { "20", "10", "Kills: (%kills%) +%upgrade_points% Upgrade Points" },
[4] = { "30", "10", "Kills: (%kills%) +%upgrade_points% Upgrade Points" },
[5] = { "40", "10", "Kills: (%kills%) +%upgrade_points% Upgrade Points" },
[6] = { "50", "10", "Kills: (%kills%) +%upgrade_points% Upgrade Points" },
[7] = { "60", "10", "Kills: (%kills%) +%upgrade_points% Upgrade Points" },
[8] = { "70", "10", "Kills: (%kills%) +%upgrade_points% Upgrade Points" },
[9] = { "80", "10", "Kills: (%kills%) +%upgrade_points% Upgrade Points" },
[10] = { "90", "20", "Kills: (%kills%) +%upgrade_points% Upgrade Points" },
[11] = { "100", "30", "Kills: (%kills%) +%upgrade_points% Upgrade Points" },
},
to this:
kill = {
[1] = { "10", "KILL (+%upgrade_points% points)" },
},
(if it is possible)
username_0: # ADDED
```lua
-- Every kill will reward X amount of points.
on_kill = {
-- POINTS | MESSAGE
enabled = true, -- Set to 'false' to disable
{"10", "Kill (+%upgrade_points% Upgrade Points)"}
},
```
You can now enable or disable the `combos, streaks, assists and kills tables`.
username_1: *Bug*
Error:
stack traceback: [string "UpgradeSystem2.0"]:1383: in function <[string "UpgradeSystem2.0"]:1382> [C]: in function '__add' [string "UpgradeSystem2.0"]:902: in function 'update' [string "UpgradeSystem2.0"]:575: in function 'AddRemove' [string "UpgradeSystem2.0"]:617: in function <[string "us"]:450> LuaEventCallback >> OnServerCommand >> Lua Error: [string "UpgradeSystem2.0"]:902: attempt to perform arithmetic on a nil value -------------------------------------------------------- Please report this error on github: https://github.com/username_0/HALO-SCRIPT-PROJECTS/issues --------------------------------------------------------
username_1: *Bug*
He does not let me buy frag grenades/plasma grenades, he says "That does not command work on this map." Test map: "BloodGulch" and "RatRace"
username_1: [Bug]
Error:
stack traceback: [string "UpgradeSystem2.0"]:1383: in function <[string "UpgradeSystem2.0"]:1382> [C]: in function '__sub' [string "UpgradeSystem2.0"]:904: in function 'update' [string "UpgradeSystem2.0"]:1160: in function <[string "UpgradeSystem2.0"]:1088> EventCallback >> OnPlayerKill >> Lua Error: [string "UpgradeSystem2.0"]:904: attempt to perform arithmetic on local 'balance' (a nil value)
username_0: # BUG FIXES:
* Frag/Plasma grenade assignment fixed.
* Betray will now deduct money from the killer instead of the victim.
* Another fix for the Column sorting routines for "/weapons" command.
username_1: Ok, (HCE) Server name: "! ROCKTS »BLOODGULCH«"
the purchase of "weapons" does not work on my server. the gametype is not "all weapons"
username_0: You can enable and disable each individual weapon in the weapons table.
Set "true" to "false" to disable them and only leave index 4 (rocket launcher) enabled.
username_1: I know, I have most of the weapons disabled
username_0: Are OK. I made a video demonstrating how to do this in case you didn't know.
https://www.youtube.com/watch?v=XHtC-Cs3tgA&feature=youtu.be
username_1: Script Version: 1.26
this error was fixed in version 1.35?
stack traceback: [string "UpgradeSystem"]:1390: in function <[string "UpgradeSystem"]:1389> [C]: at 0x011d6090 [C]: in function 'update' [string "UpgradeSystem"]:1175: in function <[string "UpgradeSystem"]:1092> EventCallback >> OnPlayerKill >> Lua Error: [string "UpgradeSystem"]:1175: money:update() -> Unable to get "balance".
username_0: Do you have "save_money" set to TRUE or FALSE?
username_1: false
username_1: My name in the game is: "LilPump" or "_!GoD", I play saturdays and sundays mostly.
username_1: out of curiosity, why did I want to know my name in the game?
username_0: ### FIXED.
### Updated to version 1.42
username_1: I tried version 1.46 and it does not give errors in the console, however when doing many kills it does not give me the points.
(After about 20 kills)
username_1: I found these errors.
Bugs:
stack traceback: [string "UpgradeSystem"]:1504: in function <[string "UpgradeSystem"]:1503> [C]: at 0x011d6090 [C]: in function 'update' [string "UpgradeSystem"]:1283: in function <[string "UpgradeSystem"]:1201> EventCallback >> OnPlayerKill >> Lua Error: [string "UpgradeSystem"]:1283: money:update() -> Unable to get "balance". -------------------------------------------------------- Please report this error on github: https://github.com/username_0/HALO-SCRIPT-PROJECTS/issues Script Version: 1.46 -------------------------------------------------------- stack traceback: [string "UpgradeSystem"]:1504: in function <[string "UpgradeSystem"]:1503> [C]: at 0x011d6090 [C]: in function 'update' [string "UpgradeSystem"]:1272: in function <[string "UpgradeSystem"]:1201> EventCallback >> OnPlayerKill >> Lua Error: [string "UpgradeSystem"]:1272: money:update() -> Unable to get "balance". -----------------------------------------------,--------- Please report this error on github: https://github.com/username_0/HALO-SCRIPT-PROJECTS/issues Script Version: 1.46 --------------------------------------------------------
username_0: I was unable to reproduce the same error, as such, it's very difficult to ascertain the circumstances under which are causing this problem. I have made some changes in hopes that it will fix the issue but there are no guarantees at this point. Script version is now 1.47.
username_1: Ok I understand, if this helps, I have all the weapons disabled and also combos, streaks, assists and "total deaths"
username_0: Let me know if you continue to have the same problem.
username_1: Well, at the moment the server is full and has not presented any errors.
Status: Issue closed
username_0: Thank you to everybody who helped test and find errors this.
@RoadHog, @username_1, @VicFirth, @iAirPirate, @TheManyRandomPlayers
username_1: @username_0 The players that enter my server and some of my clan informed me that the money does not work correctly, (they say that almost nobody passes the 100 points), the console does not present any error. My question is, what can I do to prevent this from happening?
username_0: Something to note:
If 'save_money' is disabled, player points are reset to 0 (by design) when they quit.
When they rejoin the server and discover their points are gone, this is why.
Change: `local starting_balance = 0` (on or near line 34), to suit your needs.
Beyond that, I can only recommend changing the points received for kills, etc.
Everything is working fine for me. |
cs2103jan2015-w13-3j/main | 64863225 | Title: Parsing Error for Autocomplete
Question:
username_0: To replicate:
moonstars /start wed 2pm /end wed 3pm --> 1st April
modify (index) *autocomplete*
*enter* --> 4th January
Answers:
username_1: Lol, this is because the date parser use American date format as mm/dd/yyyy and there is no api to change this. I think I will autocomplete it in a less ambiguous format then.
Status: Issue closed
|
Azure/azure-sdk-tools | 723103821 | Title: Git operations pipelines
Question:
username_0: From time to time the EngSys team is asked to perform operations on the Git repo directly against the upstream. This is inherently a fairly dangerous operation since one mistake could cause potential data-loss.
I propose we create a set of pipelines for performing pre-scripted operations against the repo so that these operations can become self-service AND we can script them in away that there are certain protections in place. The example that I have in mind is deleting release tags, but there might be others to consider as well.
We could add approval workflows to these operations so particularly sensitive operations require permissions.<issue_closed>
Status: Issue closed |
mi2-warsaw/CzasDojazdu | 155021492 | Title: Czemu dane nie pobierały się na serverze
Question:
username_0: Problem był ze ścieżkami względnymi i bezwzględnymi @abrodecka @mikolajjj @michalcisek
````{R}
mkosinski@MI2:~$ Rscript /srv/shiny-server/CzasDojazdu/000_runme.R
Error in file(filename, "r", encoding = encoding) :
cannot open the connection
Calls: source -> file
In addition: Warning message:
In file(filename, "r", encoding = encoding) :
cannot open file 'Rscripts/001_pakiety.R': No such file or directory
Execution halted
````
dlatego opakujemy pobieranie danych w Docker'a.
Chociaż możnaby to było rozwiązać podając bezwzględne ścieżki w pliku 000_runme.R.<issue_closed>
Status: Issue closed |
ksAutotests/TutorialsPool_TestGreen | 549509660 | Title: Tutorial Page autotest_tutorialy568qy.md Issue. TEST GREEN
Question:
username_0: Tutorial issue found: [https://github.com/ksAutotests/TutorialsPool_TestGreen/blob/master/tutorials/dap1dajh.1nr/autotest_tutorialy568qy.md](https://github.com/ksAutotests/TutorialsPool_TestGreen/blob/master/tutorials/dap1dajh.1nr/autotest_tutorialy568qy.md) contains invalid primary tag.
Your tutorial was not created. Please double-check primary tag property.
Each tutorial md-file shall have primary tag provided above. Example:
\-\-\-
title: Text Bundles within Node.js SAP HANA applications
description: Working with text bundles in Node.js
primary_tag: products>sap\-hana
tags: [ tutorial>intermediate\, products>sap\-hana\, products>sap\-hana\-\-express\-edition ]
\-\-\-
Affected server: TEST GREEN |
bitfocus/companion-module-panasonic-ptz | 741076865 | Title: Panasonic PTZ add Preset recall Mode
Question:
username_0: Please add a Dropdown for the Recallmode to a specific PTZ Preset on panasonic cameras.
Or make it an individual command to set the recall mode.
There are Mode A / B / C wich allow to not recall everything stored in the preset. Presets include gain white ballance and collor information wich are allo stored in a preset memory.
In some cases you want only recall position zoom and focus of the camera while doing live color correction on the fly with the rcp panel.
it can bea found on the manual of the remote controller rcp 120 from panasonic;
https://pro-av.panasonic.net/manual/pdf/AW-RP120G_Operations(VQT4T93A-1(E))_E.pdf Page 21
This my also be an dynamic variable wich should be able to read out. An innstance feedback might also be handy.
I would like to use it as follows:
One Presets or multiple ones should host the presets for white ballance gain and all the colorsettings. Which should be recallable via companion. For this specific preset/s i want them to recall all values. I will recall them before show start to reload color settings or if a scenic change requires an other color set.
Then i have a lot of presets wich only exist to call camera positions and zoom values. but need to leave the color settings untouched. This is neccassry during an event over the day which is daylight affected. The color changes will drift over the day and the operator can change it like an normal with the rcp on the fly. while he can recall the stored positions without recalling the now uncorret color
@PerRoine
Answers:
username_1: I have transferred this to the Panasonic PTZ repository where someone can follow up with you.
Status: Issue closed
username_2: @username_0 Sorry about the wait and no response, This is now added to the module and will be available in the next release, as well as new v2.2.0 betas in a week or so.
although this is not added as a dropdown on the recall preset action itself, its added as a separate action and feedback, if you want to you can always stack them on the same button :) there are also premade presets available in the "Preset Recall" folder under presets. |
wildpeaks/packages-eslint-config | 586199566 | Title: Prettier 2.0
Question:
username_0: Some defaults have changed:
- `arrowParens`: `always`
- `endOfLine`: `lf`
- `trailingComma`: `es5`
Answers:
username_0: First problem: `comma-dangle: always-multiline` isn't enough because lists of parameters don't have a dangling command:
````
function myfunction5(
_long_parameter_name1,
_long_parameter_name2,
_long_parameter_name3,
_long_parameter_name4,
_long_parameter_name5,
_long_parameter_name6,
_long_parameter_name7,
_long_parameter_name8,
_long_parameter_name9,
_long_parameter_name10
) {
return 123;
}
````
username_0: So far, it appears import statements and object properties use a comma at the end, whereas export statements and function parameters don't.
This **doesn't seem consistent**, so let's use `trailingComma false` in Prettier configs instead, even if I'm not fond of having to override yet another default value.
username_0: Release 9.0.0 is compatible with Prettier 2.0
Status: Issue closed
|
bpanel-org/bpanel | 353611827 | Title: Client Unauthorized Error
Question:
username_0: There exists a condition in which a bug can happen here:
https://github.com/bpanel-org/bpanel/blob/development/server/bcoinRouter.js#L24
If the request times out on the back end, the `token` is not reset
For the wallet, we could do something like:
```js
let wallet = client.wallet(id, token)
wallet.request(...)
```
But that wouldn't work with the node client, as it wouldn't have a wallet method.
Maybe we can create a new and temporary client with something like:
```js
// inject(Client)
let newclient = client.inject(client)
newclient.token = token
newclient.request(...)
```
or
```js
// inject(Client, opts {})
let specialized = client.inject(client, { id, token })
// returns client.wallet() or client.node()
// that comes from a new client instance
```
Maybe inject is a static method that returns a new instance of the client?
Would have to reference `client.prototype` to be able to reach the method then I believe
Answers:
username_1: That would mean making a new client on every request? Does the `finally` get skipped on timeouts?
username_0: Only if the token from the front end is different than the token on the backend, because that token needs to be proxied for some requests. I think an exception is raised and it breaks the control flow and doesn't trigger the `finally` |
kulshekhar/ts-jest | 832214366 | Title: TypeError: require(...) is not a function
Question:
username_0: | ^
57 | }
58 | };
59 |
at loadServerRoutes (src/server/index.ts:56:5)
```
This is my `yarn test` command:
```json
{
"test": "jest --coverage --silent --detectOpenHandles",
}
```
And this is my `jest.config.js`:
```js
module.exports = {
clearMocks: true,
coverageDirectory: 'coverage',
preset: 'ts-jest',
testEnvironment: 'node',
coveragePathIgnorePatterns: [
// abstract class
'src/common/domain-logic.ts',
'types.ts',
'<rootDir>/dist/'
],
};
```
## To Reproduce
Steps to reproduce the behavior:
Create a function that internally calls `require`.
## Expected behavior
Tests shouldn't throw exception.
## Link to repo
You can see git diff here:
https://gitlab.com/username_0/hireme/-/merge_requests/38
## envinfo
```
System:
OS: Arch Linux 5.11.6 64 bit
Node version: 12.14.1
Npm packages:
jest: 26.6.3
ts-jest: 26.5.3
typescript: 4.1.3
babel(optional):
```
Answers:
username_1: Hi, [TypeError](https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/TypeError) is JavaScript runtime error and it has nothing to do with TypeScript (or `ts-jest`). In this case, you get the error because Jest can't execute your codes.
`ts-jest` only throws TypeScript type checking errors, not runtime errors. An example of TypeScript type checking error is
```
error TS2345: Argument of type 'string' is not assignable to parameter of type 'boolean'.
```
Notice that, TypeScript type checking error, it always contains error code in a form of `TS<number>`.
I'd suggest you to check your codes or google the keyword "nodejs TypeError: require(...) is not a function" and you will find plenty of suggested solutions for your problem.
I will close this issue as it is not `ts-jest` issue.
Status: Issue closed
|
albertogeniola/MerossIot | 699143084 | Title: Querying ON/OFF State doesn't seem to work
Question:
username_0: Hi,
I have 2 MSS310 plugs for which I would like to query if they are ON or OFF. I am using the following code based on the provided examples:
```
plugs = manager.find_devices(device_type="mss310")
if len(plugs) < 1:
print("No MSS310 plugs found...")
else:
for plug in plugs:
on = plug.is_on()
print(on)
if on == True:
print(f"{plug.name} is ON!")
else:
print(f"{plug.name} is OFF!")
```
The output is this (Washer and Dryer are the correct names of the 2 plugs):
```
None
Washer is OFF!
None
Dryer is OFF!
```
Any idea why instead of the expected status (True / False) an empty list (None) is returned?
Answers:
username_1: Hi @username_0 ,
have a closer look at the [documentation](https://username_1.github.io/MerossIot/common-gotchas.html). In particular, the first time you work with a device you need to fully update its status by invoking
`await dev.async_update()`
I will stress this in the source code of the examples.
username_0: Perfect, I didn't know that. I believe the idea of making it more visible in the examples or stessing it out somewhere else is really good ;)
username_1: Hi @username_0 ,
thanks for reporting it back. I've updated the examples to stress that out and I've also added some error logs (in the latest version) to warn the user what that situation happens.
Status: Issue closed
|
tdwg/dwc | 54784632 | Title: Normative document as standard
Question:
username_0: We would like to propose that the RDF document of current terms be made to represent the normative standard for Darwin Core, that it be in the path http://rs.tdwg.org/terms/dwc_normative.rdf. Under this proposal the information currently held in http://rs.tdwg.org/dwc/rdf/dwctermshistory.rdf would be captured only in a history document http://rs.tdwg.org/terms/history.html. All documents other than the normative document would be free of the requirements of the Executive Committee approval, leaving them much easier to adapt to broader contributions and evolving demands.
Answers:
username_1: We would like to propose that the **RDF document of current terms** be made to represent the **normative standard for Darwin Core**. Under this proposal:
1) the normative standard for Darwin Core would consist of a single document at http://rs.tdwg.org/terms/dwc_normative.rdf.
2) information currently held in http://rs.tdwg.org/dwc/rdf/dwctermshistory.rdf would be captured only in a history document http://rs.tdwg.org/terms/history.html and not be part of the normative document.
3) all documents other than the normative document would be free of the requirements of the Executive Committee approval, leaving them much easier to adapt to broader contributions and evolving demands.
username_1: @username_0: I guess we can close this with the new normative document at https://github.com/tdwg/dwc/blob/master/standard/vocabularies/term_versions.csv
username_0: Yes.
Status: Issue closed
|
Project-MONAI/MONAILabel | 943912934 | Title: AssertionError: Torch not compiled with CUDA enabled
Question:
username_0: Error encountered when running monailabel inference on a windows machine: AssertionError: Torch not compiled with CUDA enabled
To reproduce: click button "Next Sample" in 3DSlicer
Cause: CPU version of PyTorch installed when running pip install -r requirements.txt
Fix: pip3 install torch==1.9.0+cu111 torchvision==0.10.0+cu111 torchaudio===0.9.0 -f https://download.pytorch.org/whl/torch_stable.html
Answers:
username_1: For windows, the prerequisites are mentioned over here... please follow the same.
https://docs.monai.io/projects/label/en/latest/installation.html#prerequisites
Also have you tried to install monai on windows. It should be same behavior for monai and monailabel.
username_0: Thanks. I did not look at the link you provided before I installed. The link does indeed contain the information I needed. Can it be made clearer in the monailabel readme, that windows users should refer to the link for installation instructions. Currently it is not obvious:

Status: Issue closed
|
phetsims/tasks | 177017162 | Title: RC Test: Friction 1.3.0-rc.1
Question:
username_0: The Friction sim wasn't able to support the screen shot feature due to its use of an outdated approach for rendering particles, see https://github.com/phetsims/friction/issues/35. This has been resolved and a new version is ready for testing.
**Link to sim:**
http://www.colorado.edu/physics/phet/dev/html/friction/1.3.0-rc.1/friction_en.html
**Link to iFrame:**
http://www.colorado.edu/physics/phet/dev/html/friction/1.3.0-rc.1/friction_en-iframe.html
**Test Matrix:**
The code for particle rendering has changed dramatically, and the simulation will be using updated versions of all common libraries, so a pretty thorough test is required.
Also, the screenshot capability should be verified on all platforms and the issue (https://github.com/phetsims/friction/issues/35) closed if the tests pass.
Answers:
username_1: bumping to medium as the sim is pretty easy to test, and the QA backlog is quite reduced
username_2: @username_0, this sim is ready after phetsims/friction/issues/36.
username_2: This was deployed, closing.
Status: Issue closed
|
LaravelDaily/laravel-charts | 459339998 | Title: laravel 5.8 taking area but Not showing charts
Question:
username_0: Did instruction but not showing charts as istructed but no error. take chart area but showing nothing but h1 tag.
Answers:
username_1: @username_0 sorry it took almost a year, we didn't look into issues during those times. I guess it was some kind of JavaScript error, would need more info, especially from your Development Console.
Status: Issue closed
|
DataDog/dd-trace-js | 652447654 | Title: Provide dd-trace/init script
Question:
username_0: At the moment, every project I work with has to have a dummy script:
```js
// /init/dd.js
import tracer from "dd-trace";
tracer.init();
```
Then we include that script whenever launching a program, e.g.
```bash
node -r ./dist/init/dd.js ./dist/bin/server.js
```
This is done in such a way because dd must be initialised before importing any modules.
It would be nice if there was an init script that I could use directly from `dd-trace`, i.e.,
```bash
node -r 'dd-trace/init' ./dist/bin/server.js
```
This would save me from creating a redundant file.<issue_closed>
Status: Issue closed |
ivanbalic/bit-book | 386743491 | Title: Commenting Post
Question:
username_0: - As a user when I preview post content, I should be able to type in a comment about that post and send it via send button
When comment input field is empty, send button should be disabled
- As I user when I leave a comment, I should be able to see it below my post<issue_closed>
Status: Issue closed |
joshuahysong/TICompanion | 233704877 | Title: Users should be able to "clear" a player slot since deletions have been removed
Question:
username_0: Instead of deleting a player altogether a user should be able to click a button to reset that player back to defaults with no selections.
Answers:
username_0: Instead of deleting a player altogether a user should be able to click a button to reset that player back to defaults with no selections.
username_0: Initial thought is to have an "X" button/link right justified on the same line as the Player Name in the navigation menu. This X could be gray and then red on hover.
If we think resetting a player may not be a commonly used option then might be better to put it in a less visible area to reduce overall visual clutter.
Maybe in a general menu of options for each player within the player component. |
hzxiajun/xiajun | 124323360 | Title: 泰旅
Question:
username_0: 旅行时间:2015.12.20-2015.12.28
旅行者:Xiajun (LP) & Yuanlihao (LG)
出发地:南岳-🚄-长沙-✈️-曼谷
曼谷
主要有三种选择:商场购物、宫殿/寺庙观光、市场/集市/夜市淘宝
长沙-廊曼机场
bangkok
酒店:true siam hotel
旅行第一天:taxi - true saim hotel,803
king power,了解免税店购物流程,兑换泰铢攻略没做好,机场汇率5.02,取一次还有手续费
晚上路边街边摊吃粉,椰汁,橙汁
旅行第二天:central world,购买naraya包包, superrich人民币兑换泰铢5.46。坐bts(了解曼谷地铁线路)-水牛桥渡船-河滨夜市。泰式按摩700Baht
克里普索人妖秀。(不看可惜,看了后悔,近距离看着人妖怕怕的,彻底百度了解了人妖,看了心里挺心酸的,人妖不容易,同样都是人类,为了生存背负的不一样,贫穷实在是太可怕)
Vicotory monument
到达方式:乘坐BTS到Sapan Taksin站,出站前往中央码头,乘坐前往Asiatique的免费游船,在Asiatique河滨夜市Warehouse No.3(3号仓库)。 表演时间:每天20:15、21:45各一场。 门票:常规门票1200THB、网上订购900THB;泰餐+表演套票2000THB,网上订购1500THB。建议找旅行社或酒店议价,有更换。 Tips: 每场表演结束人妖会留下来和游客合影赚些小费,价格30-80THB。人妖很多是穷苦家庭出身,迫于生计在走上这条路,而且病痛多,寿命普遍不长,观看人妖秀时请记得礼貌和尊重。
7.11购买日用品(满大街遍地都是7.11,而且东西一应具全,着实带来了很大
到达方式:乘坐BTS到Saphan Taksin站,然后转乘Asiatique的免费Shuttle boat摆渡船前往。 营业时间:每天17:00-24:00,渡船运营时间18:00-23:00。
7-11便利店 因为免不了要去的,所以单独提一下。7-11在泰国遍地都是,几乎每隔100米就有一个,24小时营业,里面商品的价格基本与大超市持平,不必担心价格问题。泰国很多酒店不提供脱鞋和洗漱用品,可以在这里买到,还可以在里面买到手机卡、给手机卡充值等
研究曼谷轻轨地铁路线
旅行第三天:换Happy卡,下载Google map,Big C购买药品、Levis牛仔裤2条、nantita包包、siam paragon曼谷海洋世界暹罗海洋世界(Siam Ocean World(免费Wi-Fi),ZEN随便看看、bai yoke sky hotel餐厅82楼吃饭
游客在泰国购买了有“vat refund for tourists”退税标志的商品,并超过一定限额(同一家商店同一天消费满2000THB;网传素万那普国际机场退税还要求在泰国的总消费额满5000THB),在离境当天可到机场要求退还7%的增值税。Lee、Levi's、Topshop可以直接在专柜退税,NaRaYa和Bata不能退税。具体退税流程请百度。
旅行第四天
The grand palace:舍利塔 藏经阁 太庙
玉佛寺:
卧佛寺:chang water
tutu车
建兴酒家(咖喱蟹、冬阴功汤、芒果饭)
湄南河夜游(dance、music、drink water)
汇率、7.11
免费Wi-Fi
旅行第五天
Bts轻轨 victory monument-phloem chit
Terminal 21
Terminal 21是一座大型综合购物商场。整个商场以机场航站楼为概念,每一层则配以不同国家城市的特色,不光是售卖的产品有地域的区别,整个装修风格,包括卫生间都非常有意思。比如一层是巴黎,二层是东京,三层就是伦敦,上去就是美国圣弗朗西斯哥 ,所以每层的厕所也按照这个城市的风格设计的。
Central world. Messege
旅行第六天
Bangkok - Chiangmai
旅行第七天
Wat Phrathat Doi Suthep,Pui National Park,Old City双龙寺+素贴山+古城+清迈大学
租摩托车,晚间逛古城摩托车坏了,很淡定,忙活大半天,对面lina message 服务人员过来帮忙,看看瞅瞅,帮忙打电话找帮忙,淳朴的泰国人满心温暖
修车司机正好是租车的人,泰语听不大懂,最后把押金和租车的钱全放在我手上,心底一暖,原来师傅觉得车坏了,给我们造成不便,深感歉意不收钱,这在国内找不到的感觉,满满的淳朴善良,佛教有信仰的人们一定是这世界上最善良的人。最后还是过意不去,给了200泰铢,要是在中国那简直了。 |
sveltejs/sapper | 698113960 | Title: Document.referrer empty
Question:
username_0: Hi,
I try to access the document.referrer but the returned value is always null. An idea of why ?
Thanks.
Status: Issue closed
Answers:
username_0: Thanks for you answer. So there is no solution? It's not related to routing?
username_2: @username_0 for help and support, please [come to Discord](https://svelte.dev/chat) |
cran/WordR | 303482202 | Title: [Feature] Tracking changes
Question:
username_0: Is there a way to show changes made to the document - simmilar to using the "track changes" Option from the Word package?
If no is there an R specific Workaround to this?
I know that two Versions of the Word file can be compared to Show differences made to the newer file.
The Problem is that this process requires
1. Word
2. Manual Point and click operations
3. is not really reproducible.
Is there a way of achievieng this from within R?
Answers:
username_1: Hi, this is a read-only mirror or CRAN, please contact the package authors, see the DESCRIPTION file |
rackerlabs/encore-ui | 107253297 | Title: Nav is broken on Safari without flexbox.less
Question:
username_0: We recently stopped using our `flexbox.less` in our `mixins.less` (https://github.com/rackerlabs/encore-ui/pull/1181) assuming that browsers didn't need it any more.
Unfortunately, as of the latest release of Safari, this is still necessary. Without it, the nav takes the full width of the screen.
We don't officially support Safari, so we won't do anything _for now_. Let's check back in October, after the next major release of Safari has happened. I suspect that it'll work just fine with that. If not, we can look at putting our `flexbox.less` back in.
Answers:
username_1: Sad :panda_face:
username_2: 
username_1: FYI: Safari 9.0 works beautifully.
<img width="1265" alt="screen shot 2015-10-14 at 1 41 54 pm" src="https://cloud.githubusercontent.com/assets/545605/10493936/4b2dd958-7279-11e5-817e-3e12cc402c96.png">
Status: Issue closed
username_3: @username_1 @username_0 So do we still need this issue? Seems resolved, so I'm closing, but reopen if needed.
username_1: Yeah, nothing to fix, really. |
tmobile/sawtooth-next-directory | 457115898 | Title: Remove assumed atomicity in NEXT
Question:
username_0: Epic | Remove assumed atomicity in NEXT
----------|----------
T-shirt | XL
Description | Architecturally, NEXT assumes that the flow of data from sawtooth -> rethink is a single atomic transaction (that a successful submission to sawtooth will result in a successful transaction in rethink). This is incorrect and has begun causing side effects of varying impact. The rbac provider sync, ledger-sync, and REST APIs need to be refactored to stop assuming that this is an atomic transaction. Inserts, updates, and deletions should return a correct 202 status code rather than a 200 to reflect that the payload has been successfully accepted by NEXT but not yet worked on (details [here](https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/202)).
The inbound queue will need to be refactored so that entries are not removed until they have been confirmed written to both rethinkdb and sawtooth (at this point it may be worth looking into a solution like rabbitmq or repubsub, which is recommended by rethinkdb, rather than spend more time going against best practice to implement queue-like features in tables).
Business Need (the why): | Assumed atomicity is causing side effects in the application which impact usability, stability, and data integrity.
User Scenarios | TBD
Requirements | TBD. For now stop assuming that because a transaction has been submitted to sawtooth that it has been successfully inserted to sawtooth and will be successfully written to rethinkdb.
As a side note we should assess what impacts we can expect to see when we move to a multinode rethinkdb and sawtooth deployment, and have a delayed head in sawtooth. This could affect concurrency and cause race conditions and may affect architecture decisions later. |
clojusc/ltest | 262106119 | Title: Improve documentation & add examples
Question:
username_0: i'm having some issues with including ltest and runing it in my project
i find it hard to follow the documentation
it would help to have a sample project or simply explain how to add it to a new lein project
thanks a lot
Dan
Answers:
username_1: Yeah, there's no docs whatsoever :-(
(In fact, I'm surprised anyone's really looking at this project!)
I'll add something to the README ...
Status: Issue closed
username_1: Ah, there was more usage than I had remembered. However, I did add more. Until master or the release/0.2.x branch is updated, the task branch has the latest changes:
* https://github.com/clojusc/ltest/blob/task/add-more-usage/README.md
Note, however, this does assume knowledge of the Clojure programming language, the `lein` tool, and the `clojure.test` library.
username_0: @username_1 thanks for the quick reply
can you add a link to the simplest example project you can think off ?
(a git repo example would be perfect)
thanks!
Dan |
metaskills/minitest-spec-rails | 169018850 | Title: Breaks order setting for Rails and Minitest
Question:
username_0: No matter what you set `test_order` to on either `MiniTest::Test` or `ActiveSupport::TestCase` the tests will always run out of order with this gem.
I've created a simple example of the problem here: https://github.com/username_0/minitest_testing
Just run the tests with `TESTOPTS="-v" rake` to see the order change for the tests that rely on this gem and that they do not change for tests which only rely on `minitest/spec`. I've reviewed the code for this Gem but I can't figure out why this happens.
Answers:
username_1: When using a outer describe block that does not resolve the described class to a constant, you are indeed in raw MiniTest::Spec subclass and not a descendant of ActiveSupport::TestCase. I [highly recommend](https://github.com/username_1/minitest-spec-rails#test-styles) you stick with the outer Rails subclassing.
So, in order to get results you want... you have to use the Rails' interface. Because this project does aim to work in that framework. To do that, open up your `config/environments/test.rb` and set the following. The default in that file should be set to `:random`.
```ruby
config.active_support.test_order = :alpha
```
So I set it to :alpha and ran the tests again. And things appear to be working great.
<details>
```ruby
class MiniTestSpecRailsTest < ActiveSupport::TestCase
it('a') { assert true }
it('b') { assert true }
it('c') { assert true }
it('d') { assert true }
describe 'whatever' do
it('a') { assert true }
it('b') { assert true }
it('c') { assert true }
it('d') { assert true }
end
end
```
</details>
```
$ ./bin/rake test TESTOPTS='-v'
MiniTestSpecRailsTest::whatever#test_0001_a = 0.00 s = .
MiniTestSpecRailsTest::whatever#test_0002_b = 0.00 s = .
MiniTestSpecRailsTest::whatever#test_0003_c = 0.00 s = .
MiniTestSpecRailsTest::whatever#test_0004_d = 0.00 s = .
MiniTestSpecRailsTest#test_0001_a = 0.00 s = .
MiniTestSpecRailsTest#test_0002_b = 0.00 s = .
MiniTestSpecRailsTest#test_0003_c = 0.00 s = .
MiniTestSpecRailsTest#test_0004_d = 0.00 s = .
```
Here is the source excerpt from the AcitveSupport::TestCase source. Hope that helps!
```ruby
# Sets the order in which test cases are run.
#
# ActiveSupport::TestCase.test_order = :random # => :random
#
# Valid values are:
# * +:random+ (to run tests in random order)
# * +:parallel+ (to run tests in parallel)
# * +:sorted+ (to run tests alphabetically by method name)
# * +:alpha+ (equivalent to +:sorted+)
def test_order=(new_order)
ActiveSupport.test_order = new_order
end
```
username_0: Ah sorry that's my fault. Rails 4.2.6 is setting it to `:random` in `test.rb` when generating a new project. I understood `:sorted` was the default until Rails 5.0 (as per their docs). I guess it is the "default" which is then overridden by the generator.
This means I still have an undiagnosed issue on my actual project's test order. Sigh, thanks for the help.
Status: Issue closed
username_0: No matter what you set `test_order` to on either `MiniTest::Test` or `ActiveSupport::TestCase` the tests will always run out of order with this gem.
I've created a simple example of the problem here: https://github.com/username_0/minitest_testing
Just run the tests with `TESTOPTS="-v" rake` to see the order change for the tests that rely on this gem and that they do not change for tests which only rely on `minitest/spec`. I've reviewed the code for this Gem but I can't figure out why this happens.
username_0: @username_1 ok I dug a bit deeper. The tests are in fixed order, but the outer describes are not. With multiple describe blocks the blocks of tests are output in random order, while the inner tests are fixed.
username_1: Right... cause the describes are a new class and scope. The block it self would be randomized in relation to that scope. For example... `MiniTestSpecRailsTest::whatever#test_#{n}` in my example above could be before or after the `MiniTestSpecRailsTest#test_#{n}` group even when `:alpha` is run. I think this is the core of your issue? Seems to be the default behavior of MiniTest too.
username_0: @username_1 yeah I think it is. I'll talk to Ryan and see if it can be fixed. I'm not a big fan of the tests running in random order. Or at least the output being randomly ordered.
Status: Issue closed
username_1: Cool! Going to close this issue. If you think we need to re-open it and/or do something, please feel free to reply and let me know.
username_0: No, not much you can do sadly since Ryan decided that this was desired behaviour. |
gpvigano/AsImpL | 298862336 | Title: Support for other file formats
Question:
username_0: Even if the only supported format is currently OBJ, the idea is to create a common framework on which the support for other formats could be developed, allowing the exchange of data with other applications.
I created this issue to collect suggestions about the support of new formats, maybe with an hint about the implementation.
Answers:
username_1: Hi Giovanni,
It could be cool to implement FBX format. I'm working on low poly model for hololens dev and that's the best format we can use ! 😄
username_0: FBX support in AsImpL? It should be great! But...
You really hit a nerve, for a number of reasons.
* There are no plans to port the FBX SDK to work on Android at this time.
* FBX SDK is not open source
* It is written in C++, even if several C# wrapper exist.
* Autodesk does not provide any file format specification for FBX
* The license of FBX SDK prevents to repackage and distribute it without written consent by Autodesk.
* Alternatives to FBX SDK are incomplete and do not well support the numerous versions of FBX.
My feeling about FBX support in AsImpL is that we will never be able to develop (and maintain) a complete and full-working support for this format.
If we opt to FBX SDK we must be aware about license issues: we can't include it in AsImpL and we must handle the support for the different platforms.
We also cannot forget the lack of support for Android, thus, to prevent binding AsImpL to only the platforms supported by Autodesk FBX SDK, we should provide the support in AsImpL as a separate extension, so we will not break the Android support in AsImpL.
We could try to develop FBX support from scratch or from one of the existing "Autodesk-free" projects (like https://github.com/mpreble/UnityFBXImporter or https://github.com/izrik/FbxSharp). These projects anyway have a limited support for only a few ASCII FBX formats (this is perfectly understandable, due to the previously mentioned reasons).
This would be a huge work, but this does not prevent someone who needs it to start developing a even limited support for his/her needs, I am available to collaborate, as much as I can, from the AsImpL core library side. As you can see the current data structures cannot hold all the features coming from FBX or other formats, anyway an extension is obviously required for supporting new formats. Furthermore changes are needed to import the new data to Unity scene.
username_2: There is also an FBX importer asset that seems to support Android, though I have no tried it yet. They don't support WebGL yet. https://assetstore.unity.com/packages/tools/modeling/trilib-unity-model-loader-package-91777
username_0: Thank you for your suggestion, I'm quite busy now, but it sounds interesting, I'll have a look on it as soon as I have enough time. In the meanwhile, anyone who wants to try it and give us a feedback is welcome...
username_3: @username_0 , Could you tell me all the OS that AsImpL support?
username_0: @username_3 I tested AsImpL on Windows Standalone, Universal Windows Platform, Android and @username_2 provided PR #28 #29 #30 to make it work with iOS. I have no feedback about Linux, I think it should work (any feedback from GitHub's community?).
P.S. You have chosen this issue (Support for other file formats) to ask this question, so if your question was actually about 3D model formats the answer is: currently only OBJ. Otherwise, if you prefer (e.g., to get other feedback), you can create a new issue related to OS support.
username_4: What do you think about DAE format? It is supposed to be pretty widely used, and it is just XML, so it shouldn't be too tricky to parse?
It would mean making a `loaderDae` similar to `loaderObj` and strapping it into the menu, and then it should be good, right?
username_0: COLLADA DAE could be a choice because SDK is open source (FBX has a restrictive license) and support model hierarchy and animations (OBJ has only groups and no animation).
Anyway it's not that easy...
Some years ago I wrote a DAE loader for an OpenGL-based application (using [COLLADA-DOM](https://sourceforge.net/projects/collada-dom/)), but the implementation was really hard (and at one point I started developing with Unity, so I gave up and I never used it for my projects).
Anyway I think a problem in that case could be the mapping from the external DAE parser and AsImpL data structures. The current data structures in AsImpL are not able to hold more data than what OBJ supports, so they must be extended.
username_0: A format that is catching on is [glTF](https://www.khronos.org/gltf/) (also supported by Unreal Engine editor), available with a [Unity package](https://github.com/KhronosGroup/UnityGLTF). I already successfully [used it with AsImpL in my OpenVISDIVE project](https://github.com/username_0/OpenVISDIVE/tree/master/OpenVISDIVE_Unity/Assets/OBJ2glTF). |
jackolney/WhoCascade | 173248124 | Title: Weighting Test Two
Question:
username_0: If we pick 'Kenya' and make all HIV indicators RED except those for "PLHIV on ART" then we get the following with 'speed' =

and with 'quality' =
<issue_closed>
Status: Issue closed |
sezyl/Water-Level-Sensor | 424406340 | Title: Tracking water level during pumping out
Question:
username_0: Currently controller turns off pump few seconds after achieving low threshold level. It would be much safer if water level is tracked during pumping out to ensure that it is lowering.
This would help with opening bottom end of the sensor pipe allowing fresh air comming into it.
Suggested solution: as soon as low threshold is achieved continue pumping out as long as water level is droping.
Answers:
username_0: Feature added. #5
Status: Issue closed
|
dglo/StringHub | 629575691 | Title: [jacobsen on 2007-07-02 17:14:57] : pDAQ: StringHub - add FPGA and comstat proc file dumps when DOMs drop out
Question:
username_0: ... also reset comstat at beginning of run
Answers:
username_0: [jacobsen on 2007-07-02 17:46:13]
Sending src/main/java/icecube/daq/domapp/DataCollector.java
Sending src/main/java/icecube/daq/dor/Driver.java
Sending src/main/java/icecube/daq/dor/IDriver.java
Transmitting file data ...
Committed revision 1699.
username_0: [jacobsen on 2007-07-02 17:57:29]
Sending src/test/java/icecube/daq/sender/test/MockDriver.java
Transmitting file data .
Committed revision 1700.
username_0: [jacobsen on 2007-07-02 18:35:42]
Sending src/main/java/icecube/daq/dor/Driver.java
Transmitting file data .
Committed revision 1701.
Status: Issue closed
|
nbnuk/nbnatlas-issues | 409769692 | Title: Remove verification button on data resource edit page
Question:
username_0: On the edit page of a data resource, i.e. https://registry.nbnatlas.org/dataResource/show/dr236
please remove the text:
Verification status: Currently this data resource is marked as unverified [button] from the first panel.
Answers:
username_1: Done
Status: Issue closed
username_0: thanks |
Subsets and Splits
No saved queries yet
Save your SQL queries to embed, download, and access them later. Queries will appear here once saved.