repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
schmerl/LLStaging | 210329809 | Title: Make sure brasscomms parameter checking is right
Question:
username_0: Coordinates for place obstacle are consistent with wiki but not consistent with map (i.e., currently not possible to place an obstacle on the path)
In place_obstacle (and probably elsewhere) 65 is not considered a float, and so a malformed expression is returned. Has to be 65.0 currently - let's accept 65.
Status: Issue closed
Answers:
username_0: Should be in brasscomms. |
nestjs/nest | 836983714 | Title: can not gzip post response
Question:
username_0: ## Bug Report
## Current behavior
According to https://docs.nestjs.com/techniques/compression, I use nginx to compression my api.
The nginx.conf is below:
server {
listen 80;
server_name localhost;
gzip on;
gzip_types text/css application/javascript application/json image/svg+xml;
location / {
root /usr/share/nginx/html;
index index.html index.htm;
try_files $uri $uri/ /index.html;
if ( $uri ~ \.html$ ) {
add_header Cache-Control 'no-store';
add_header 'Access-Control-Allow-Origin' '*';
add_header 'Access-Control-Allow-Methods' 'GET, POST, OPTIONS';
add_header 'Access-Control-Allow-Headers' 'DNT,X-CustomHeader,Keep-Alive,User-Agent,X-Requested-With,If-Modified-Since,Cache-Control,Content-Type';
}
location ~* \.(js|css|gif|jpg|jpeg|png|webp|svg|ttf|woff|woff2|eot)$ {
add_header Cache-Control max-age=315360000000;
add_header 'Access-Control-Allow-Origin' '*';
add_header 'Access-Control-Allow-Methods' 'GET, POST, OPTIONS';
add_header 'Access-Control-Allow-Headers' 'DNT,X-CustomHeader,Keep-Alive,User-Agent,X-Requested-With,If-Modified-Since,Cache-Control,Content-Type';
}
}
location /api/ {
proxy_pass http://easy-doctor-api:3000/;
}
error_page 500 502 503 504 /50x.html;
location = /50x.html {
root /usr/share/nginx/html;
}
}
http://easy-doctor-api:3000/ is my backend which is nestjs.
The strange thing is my get method can be gzip, but post method can not.
## Input Code
Get request header:
<img width="529" alt="图片" src="https://user-images.githubusercontent.com/6395323/111894075-c86d1a00-8a42-11eb-93f5-55e0cd07f61d.png">
Get response header:
<img width="289" alt="图片" src="https://user-images.githubusercontent.com/6395323/111894097-e76bac00-8a42-11eb-9771-b4866e581b52.png">
Post request header:
<img width="530" alt="图片" src="https://user-images.githubusercontent.com/6395323/111894128-226ddf80-8a43-11eb-8fc1-bf19b39ea514.png">
Post response header:
<img width="294" alt="图片" src="https://user-images.githubusercontent.com/6395323/111894134-31ed2880-8a43-11eb-8f06-5d404cdbb776.png">
You can see, there's no Content-Encoding: gzip in Post response header.
## Expected behavior
Can gzip post method
## Environment
"@nestjs/common": "7.6.14",
"@nestjs/core": "7.6.14",
"@nestjs/jwt": "7.2.0",
"@nestjs/passport": "7.1.5",
"@nestjs/platform-express": "7.6.14",
"@nestjs/swagger": "4.7.16",
"@nestjs/typeorm": "7.1.5",
For Tooling issues:
- Node version: v12.21.0
- Platform: Linux
Answers:
username_1: Hi @username_0, I advise you to leave a minimal reproduction of a clonable git repository so that the core team can evaluate the problem you have reported.
If you have difficulty you can always use the [discord](https://discord.com/invite/G7Qnnhy) channel for support.
Status: Issue closed
username_2: Please, use our [Discord](https://discord.gg/G7Qnnhy) channel (support) for such questions. We are using GitHub to track bugs, feature requests, and potential improvements. |
pberrecloth/butter-sketch-plugin | 255361187 | Title: Touch Bar support
Question:
username_0: Might be nice to add touch bar support for MacBook Pro users. When selecting multiple objects the 'align' options are displayed. Butting options could be shown alongside. Not a pressing need but perhaps something to explore.
Answers:
username_1: @username_0 do you have a MacBook Pro with touch bar that you could test an implementation of this on? I do not, which could make it difficult to develop and test :P |
AnSyn/ansyn | 541464789 | Title: menus and popups gets out of the window.
Question:
username_0: 1. open edit menu for an annotation when the annotation is close to the end of the window.
or
2. choose color from the color palate for annotation when the annotation is close to the end of the window.
or
3. get the ,ouse hover an annotation in the timeline for preview it - when the annotation is close to the end of the window.



Answers:
username_0: This is still happening with context menu and color palette
<img width="243" alt="Screen Shot 2019-12-31 at 20 48 59" src="https://user-images.githubusercontent.com/57096192/71631042-267bca80-2c0f-11ea-92d1-d3186b0df7a4.png">
<img width="195" alt="Screen Shot 2019-12-31 at 20 49 04" src="https://user-images.githubusercontent.com/57096192/71631043-27146100-2c0f-11ea-9e8f-1170096f1b5a.png">
username_1: @username_0
Regarding the context menu, if we shift it, the center point of the context menu will no longer be on the point on the map that we clicked on. Won't this be a problem?
username_0: You got a good point.
At this point we can leave the context menu issue.
thank you!
username_1: Ok, was decided to leave the context menu for now.
username_2: @username_1 if the color is opened down it doesn't show
username_1: @username_2 Ok, open a new issue about this, and we'll add it. |
LaQuay/TDTChannels | 606599792 | Title: America's Voice
Question:
username_0: Internacional (US),
Entretenimiento y noticias.
web: https://americasvoice.news/
m3u8: https://cdnapi.kaltura.com/p/2377021/sp/237702100/playManifest/entryId/0_m2pjgwiu/format/applehttp/.m3u8
Answers:
username_1: Lo añado, gracias!
Status: Issue closed
|
piotrpalarz/vscode-gitignore-generator | 819393067 | Title: 🛑 Cloudflare Triggered, Not Working Anymore
Question:
username_0: I've been using your extension for over a year now, and I absolutely loved it. However, today while I was using the command palette to generate `.gitignore`, it returns code of cloudflare page as the dropdown menu and stuck there. Check out the screenshot below.

**System Info**
Windows 10
VSCode 1.53.2
.gitignore Generator v 1.0.2
Answers:
username_1: @username_0 Thanks for letting me know. It's not a problem for everyone because Cloudflare is filtering traffic based on different criteria. In your case it's probably the IP address that's causing the problem. I have an idea how to fix it and I'll try to do it this week.
username_0: @username_1 Thanks for the attention. Will be wait for the update. 🚀
username_1: @username_0 I managed to create an alternative API URL to bypass Cloudflare's protection and it should be working now. Make sure your editor downloads new version of the extension (v1.0.3) and let me know if that did the trick. If you have VS Code opened, close it, open it again and go to the Extensions panel. You should see the update to v1.0.3.
Status: Issue closed
username_0: @username_1 Woohoo! 🔥 I've tested the latest version (v1.0.3) and it works like a charm 👌.
Thanks again for the the update. Awesome work. 🤘 |
yiisoft/yii | 6746170 | Title: [ENHANCEMENT] Provide easy access to aggregation functions (max, min, avg...) the same way as count()
Question:
username_0: Just like CActiveRecord has a count() method, it should also have a max(), min(), avg() methods and/or a generic method (with the name of the SQL aggregate function as a parameter) for getting those aggregate functions in a very similar way as is currently possible with count().
I know it's not completely trivial and it requires additional parameters compared to the count() case, but this should be perfectly feasible.<issue_closed>
Status: Issue closed |
opencloset/OpenCloset-Cron | 94020715 | Title: AnyEvent::Timer::Cron 0.002000 버전 지원
Question:
username_0: 기존 AnyEvent::Timer::Cron 모듈 0.001002 버전은 시간대를 지원하지 않았기 때문에 OpenCloset::Cron::SMS 모듈에서 멍키패치(f8e18da, e3f078c)를 수행해서 시간대를 지원했습니다. 오늘 릴리스된 최신 0.002000버전은 시간대 기능이 추가되었기 때문에 OpenCloset::Cron 모듈도 이를 지원해야 합니다.
Answers:
username_0: 작업한 내역은 다음과 같습니다.
- AnyEvent::Timer::Cron 모듈을 0.002000 버전을 사용하도록 명시
- 시간대를 지원하도록 수정
- 문서 갱신
username_0: 이슈 닫습니다.
Status: Issue closed
username_0: OpenCloset::Cron 0.003 버전을 릴리스 했습니다.
다음 주소에서 확인할 수 있습니다.
https://cpan.theopencloset.net/ |
EBISPOT/efo | 819232494 | Title: New term request for Neuvonen PMID 32961594
Question:
username_0: For new term requests, please provide the following information:
## Preferred term label
glycodeoxycholate 3-O-glucuronide measurement
## Synonyms
GDCA-3G
## Textual definition
Quantification of the amount of glycodeoxycholate 3-O-glucuronide in a sample.
## Suggested parent term
measurement
## Attribution
Answers:
username_1: Added as EFO_0011010, which will be available at our next release (15th March)
Status: Issue closed
|
sonata-project/SonataMediaBundle | 812024573 | Title: Sonata media can't valid allowed extensions in formMapper
Question:
username_0: <!--
Before you open an issue, make sure this one does not already exist.
Please also read the "guidelines for contributing" link above before posting.
-->
<!--
If you are reporting a bug, please try to fill in the following.
Otherwise remove it.
-->
### Environment : LAMP ubuntu 18.04 LTS
#### Sonata packages
<details><summary>show</summary>
<p>
```
$ composer show --latest 'sonata-project/*'
# Put the result here.
```
sonata-project/admin-bundle 3.88.0 3.89.1 The missing Symfony A...
sonata-project/block-bundle 3.21.0 4.5.0 Symfony SonataBlockBu...
sonata-project/cache 2.0.1 2.1.1 Cache library
sonata-project/datagrid-bundle 3.3.0 3.3.0 Symfony SonataDatagri...
sonata-project/doctrine-extensions 1.9.1 1.11.0 Doctrine2 behavioral ...
sonata-project/doctrine-orm-admin-bundle 3.29.0 3.29.0 Integrate Doctrine OR...
sonata-project/exporter 2.5.0 2.5.1 Lightweight Exporter ...
sonata-project/form-extensions 1.6.0 1.9.0 Symfony form extensions
sonata-project/intl-bundle 2.10.0 2.10.1 Symfony SonataIntlBundle
sonata-project/media-bundle 3.28.0 3.30.0 Symfony SonataMediaBu...
sonata-project/twig-extensions 1.4.1 1.5.1 Sonata twig extensions
sonata-project/user-bundle 4.9.0 4.11.0 Symfony SonataUserBundle
</p>
</details>
#### Symfony packages
<details><summary>show</summary>
<p>
```
$ composer show --latest 'symfony/*'
# Put the result here.
```
symfony/asset v4.4.19 v5.2.3 Manages URL generation and versioning of web assets suc...
symfony/browser-kit v4.4.19 v5.2.3 Simulates the behavior of a web browser, allowing you t...
symfony/cache v4.4.19 v5.2.3 Provides an extended PSR-6, PSR-16 (and tags) implement...
symfony/cache-contracts v2.2.0 v2.2.0 Generic abstractions related to caching
symfony/config v4.4.19 v5.2.3 Helps you find, load, combine, autofill and validate co...
symfony/console v4.4.19 v5.2.3 Eases the creation of beautiful and testable command li...
symfony/css-selector v4.4.19 v5.2.3 Converts CSS selectors to XPath expressions
symfony/debug v4.4.19 v4.4.19 Provides tools to ease debugging PHP code
symfony/debug-bundle v4.4.19 v5.2.3 Provides a tight integration of the Symfony Debug compo...
symfony/dependency-injection v4.4.19 v5.2.3 Allows you to standardize and centralize the way object...
symfony/deprecation-contracts v2.2.0 v2.2.0 A generic function and convention to trigger deprecatio...
symfony/doctrine-bridge v4.4.19 v5.2.3 Provides integration for Doctrine with various Symfony ...
symfony/dom-crawler v4.4.19 v5.2.3 Eases DOM navigation for HTML and XML documents
[Truncated]
'provider'=>'sonata.media.provider.image',
'context' => 'default',
'required' => false,
))
when upload an pdf or excel insted of image(png or jpg)
## Expected results
A falsh message contain for exemple "you must upload an image"
## Actual results
an error (exception) with this message
An image could not be created from the given input
<!--
If it's an error message or piece of code, use code block tags,
and make sure you provide the whole stack trace(s),
not just the first error message you can see.
More details here: https://github.com/sonata-project/SonataMediaBundle/blob/3.x/CONTRIBUTING.md#issues
-->
Answers:
username_1: It works as expected. You just need to make sure you add a `Valid` constraint to the property. It can be done like this:
```
->add('logo', MediaType::class, [
'provider' => 'sonata.media.provider.image',
'context' => 'default',
'required' => false,
'constraints' => new \Symfony\Component\Validator\Constraints\Valid(),
])
```
Status: Issue closed
|
AutoMapper/AutoMapper | 163278116 | Title: [Feature Request] Add MemberList.Intersection
Question:
username_0: Times ago, a very popular extension method were born to make it easier to allow mapping objects with different shape and ignoring all unmapped properties. We can find [various](http://stackoverflow.com/questions/954480/automapper-ignore-the-rest) examples [around](http://stackoverflow.com/questions/4987872/ignore-mapping-one-property-with-automapper).
Those methods is supposed to be replaced by the `MemberList` enum which ATM has the `Destination` and `Source` values, but is lacking the option to only map the properties that exists on both source and destination type.
Does it make sense to add a 3rd option to the previous the `MemberList` enum to simplify mapping configuration for this scenario?
Regards, Max.
Answers:
username_1: Ignore is about configuration validation. I opened a separate issue to do
MemberList.None, which means don't validate either side. Whatever matches,
maps. No configuration validation occurs.
username_1: This one! #1416
I think both of our ideas are the same - map the matching members, but don't validate either side, right?
username_0: Sorry, for opening a dupe, closing because #1416 already covers my needs!
Status: Issue closed
|
EvotecIT/PSWriteHTML | 514877507 | Title: Out-HTMLView option to make file save without opening file to browser
Question:
username_0: Would be nice if Out-HtmlView command could have some option to just save output to disk and not to open saved html file to browser. This would help to do easily simple HTML reports with Out-HtmlView command in scripts.
Answers:
username_0: Thanks, it works too.
Status: Issue closed
|
elegantthemes/Divi-Beta | 174833307 | Title: Don't Fail Silently When Its Not Appropriate To Do So
Question:
username_0: ### Problem:
Review features/functionality that directly depend on data retrieved via AJAX and ensure they don't fail silently when its not appropriate to do so (eg. loading premade/saved layouts, saving the page, etc).
Answers:
username_1: Improved errors handling in following ajax requests:
- Saving the page. Saving will fail as it should if error in ajax request will happen
- Loading of library items/saving of library items. This operations will be failed with correct response.
- Updating computed properties. This operation will be failed without hanging the module.
Status: Issue closed
|
TIY-Durham/2015-FALL-FEE | 111316914 | Title: 25 -- <NAME>
Question:
username_0: * [ ] **Yak Shaving**
* _WIP Issue_ with links to:
* your PRs in `USERNAME.github.io`
* your PRs in `TIY-GitHub` for features and releases
* your PRs in `TIY-Catalog` for `feature--*` into `develop`
* _WIP Branch_:
* `USERNAME.github.io:journal-week-6`
* `TIY-GitHub`:
* `release--*` -- if you're ready for another release
* `feature--*` -- for any new features
* `TIY-Catalog`:
* `feature--*` -- for any feature you're working on
* [ ] **And the Beat Goes On**
* _Rough Draft_ for Reflective Journal on Week 6
* _Rough Draft_ for Tutorial Journal
* _Final Draft_ for Resource Journal, due Tuesday!
* [ ] **Reading APIs: Etsy**
* [ ] **Shaping Up with Angular JS**
* [ ] Completed Level 1!
* [ ] Completed Level 2!
* [ ] Pics or it didn't happen!
* [ ] **For Your GitHub Only**
* [ ] **Etsy Goes On**
Answers:
username_0: Journal PR:
https://github.com/username_0/username_0.github.io/pull/24#issuecomment-147597305
Reading Etsy API PR:
Shaping Up with Angular
Github PR
Etsy Goes on:
Team-Lead for the day... no PR.
Activity done: organize tasks for the night, created missing issue for a feature. Research needed API. |
symfony/symfony | 101884585 | Title: The Monolog extras should appear in the logger panel too
Question:
username_0: We current display the context of the message, which is a PSR-3 concept.
In 2.8, we also display the channel when available, which is a Monolog-only feature, recognizing that most Symfony projects are using Monolog for logging. There is another Monolog feature which is not displayed in the profiler: extras, i.e. the bucket where processors can attach extra info on the log record. It may be worth adding support for it
Answers:
username_1: :+1:
username_2: Where can I find more information about the usage of these "Monolog extras"? Thanks.
username_3: https://github.com/Seldaek/monolog/blob/master/doc/01-usage.md#adding-extra-data-in-the-records
So just the `'extra'` key in the record.
username_0: Note that the first step to display them will be to update the [Monolog DebugHandler](https://github.com/symfony/symfony/blob/2.8/src/Symfony/Bridge/Monolog/Handler/DebugHandler.php) to include extras in the data passed to the logger collector. Once this is done, the profiler panel can be updated to display this info when it is available
username_4: Closing as this is getting old, and nobody cared since then.
Ping @username_5 because I know you like these things if you want to reopen :)
Status: Issue closed
username_5: For now I'm fine with the extra. Thanks for the notif anyway. |
ionic-team/ionic-framework | 683810193 | Title: feat: Manually start IonRefresher
Question:
username_0: <!-- Before submitting an issue, please consult our docs (https://ionicframework.com/docs/). -->
<!-- Please make sure you are posting an issue pertaining to the Ionic Framework. If you are having an issue with the Ionic Appflow services (Ionic View, Ionic Deploy, etc.) please consult the Ionic Appflow support portal (https://ionic.zendesk.com/hc/en-us) -->
<!-- Please do not submit support requests or "How to" questions here. Instead, please use one of these channels: https://forum.ionicframework.com/ or http://ionicworldwide.herokuapp.com/ -->
<!-- ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. -->
# Feature Request
**Ionic version:**
<!-- (For Ionic 1.x issues, please use https://github.com/ionic-team/ionic-v1) -->
<!-- (For Ionic 2.x & 3.x issues, please use https://github.com/ionic-team/ionic-v3) -->
[x] **4.x**
**Describe the Feature Request**
<!-- A clear and concise description of what the feature request is. Please include if your feature request is related to a problem. -->
I want to be able to start the Refreshing of the **IonRefresher** automatically.
**Describe Preferred Solution**
<!-- A clear and concise description of what you want to happen. -->
I think the best Way is, if we have a **start** Function on the IonRefresher Object, that we can access via `ViewChd()` 🤔
**Additional Context**
<!-- List any other information that is relevant to your issue. Stack traces, related issues, suggestions on how to add, use case, Stack Overflow links, forum links, screenshots, OS if applicable, etc. -->
I found an old (closed) FR: #11393 🤷🏼♂️ |
dbeaver/dbeaver | 371875486 | Title: Dbeaver don't open after install ubuntu 18.10
Question:
username_0: <!--
Read contribution guide before submitting ticket:
https://github.com/dbeaver/dbeaver/blob/devel/CONTRIBUTING.md
You can submit:
- Question
- Feature request
- Bug report
-->
After clicking, dbeaver sends the error and reports the problem in the log
2018-10-19 04:20:31.306 - DBeaver 5.2.2 is starting
2018-10-19 04:20:31.306 - OS: Linux 4.18.0-10-generic (amd64)
2018-10-19 04:20:31.307 - Java version: 11 by Oracle Corporation (64bit)
2018-10-19 04:20:31.307 - Install path: '/usr/share/dbeaver'
2018-10-19 04:20:31.308 - Instance path: 'file:/home/alexis/.dbeaver4/'
2018-10-19 04:20:31.309 - Memory available 66Mb/1024Mb
2018-10-19 04:20:31.332 - Starting RMI server at 34321
2018-10-19 04:20:31.424 - Run workbench
2018-10-19 04:20:31.438 - > Start org.eclipse.e4.ui.workbench.swt [0.14.101.v20170710-1119]
2018-10-19 04:20:31.449 - > Start org.eclipse.e4.core.services [2.1.0.v20170407-0928]
2018-10-19 04:20:31.452 - > Start org.eclipse.e4.core.contexts [1.6.0.v20170322-1144]
2018-10-19 04:20:31.463 - > Start org.eclipse.e4.ui.services [1.3.0.v20170307-2032]
2018-10-19 04:20:31.501 - Internal error in workbench lifecycle
org.eclipse.e4.core.di.InjectionException: java.lang.NoClassDefFoundError: javax/annotation/PostConstruct
at org.eclipse.e4.core.internal.di.InjectorImpl.internalMake(InjectorImpl.java:410)
at org.eclipse.e4.core.internal.di.InjectorImpl.make(InjectorImpl.java:318)
at org.eclipse.e4.core.contexts.ContextInjectionFactory.make(ContextInjectionFactory.java:162)
at org.eclipse.e4.ui.internal.workbench.swt.E4Application.createDefaultHeadlessContext(E4Application.java:491)
at org.eclipse.e4.ui.internal.workbench.swt.E4Application.createDefaultContext(E4Application.java:505)
at org.eclipse.e4.ui.internal.workbench.swt.E4Application.createE4Workbench(E4Application.java:204)
at org.eclipse.ui.internal.Workbench.lambda$3(Workbench.java:614)
at org.eclipse.core.databinding.observable.Realm.runWithDefault(Realm.java:336)
at org.eclipse.ui.internal.Workbench.createAndRunWorkbench(Workbench.java:594)
at org.eclipse.ui.PlatformUI.createAndRunWorkbench(PlatformUI.java:148)
at org.jkiss.dbeaver.core.application.DBeaverApplication.start(DBeaverApplication.java:156)
at org.eclipse.equinox.internal.app.EclipseAppHandle.run(EclipseAppHandle.java:196)
at org.eclipse.core.runtime.internal.adaptor.EclipseAppLauncher.runApplication(EclipseAppLauncher.java:134)
at org.eclipse.core.runtime.internal.adaptor.EclipseAppLauncher.start(EclipseAppLauncher.java:104)
at org.eclipse.core.runtime.adaptor.EclipseStarter.run(EclipseStarter.java:388)
at org.eclipse.core.runtime.adaptor.EclipseStarter.run(EclipseStarter.java:243)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.eclipse.equinox.launcher.Main.invokeFramework(Main.java:653)
at org.eclipse.equinox.launcher.Main.basicRun(Main.java:590)
at org.eclipse.equinox.launcher.Main.run(Main.java:1499)
at org.eclipse.equinox.launcher.Main.main(Main.java:1472)
Caused by: java.lang.NoClassDefFoundError: javax/annotation/PostConstruct
at org.eclipse.e4.core.internal.di.InjectorImpl.inject(InjectorImpl.java:124)
at org.eclipse.e4.core.internal.di.InjectorImpl.internalMake(InjectorImpl.java:399)
... 23 more
Caused by: java.lang.ClassNotFoundException: javax.annotation.PostConstruct cannot be found by org.eclipse.e4.core.di_1.6.100.v20170421-1418
at org.eclipse.osgi.internal.loader.BundleLoader.findClassInternal(BundleLoader.java:433)
at org.eclipse.osgi.internal.loader.BundleLoader.findClass(BundleLoader.java:395)
at org.eclipse.osgi.internal.loader.BundleLoader.findClass(BundleLoader.java:387)
at org.eclipse.osgi.internal.loader.ModuleClassLoader.loadClass(ModuleClassLoader.java:150)
[Truncated]
2018-10-19 04:20:31.533 - null
java.lang.IllegalStateException: Todavía no se ha creado el entorno de trabajo.
at org.eclipse.ui.PlatformUI.getWorkbench(PlatformUI.java:93)
at org.jkiss.dbeaver.core.application.DBeaverApplication.shutdown(DBeaverApplication.java:290)
at org.jkiss.dbeaver.core.application.DBeaverApplication.start(DBeaverApplication.java:165)
at org.eclipse.equinox.internal.app.EclipseAppHandle.run(EclipseAppHandle.java:196)
at org.eclipse.core.runtime.internal.adaptor.EclipseAppLauncher.runApplication(EclipseAppLauncher.java:134)
at org.eclipse.core.runtime.internal.adaptor.EclipseAppLauncher.start(EclipseAppLauncher.java:104)
at org.eclipse.core.runtime.adaptor.EclipseStarter.run(EclipseStarter.java:388)
at org.eclipse.core.runtime.adaptor.EclipseStarter.run(EclipseStarter.java:243)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.eclipse.equinox.launcher.Main.invokeFramework(Main.java:653)
at org.eclipse.equinox.launcher.Main.basicRun(Main.java:590)
at org.eclipse.equinox.launcher.Main.run(Main.java:1499)
at org.eclipse.equinox.launcher.Main.main(Main.java:1472)
Thanks for the help!
Answers:
username_1: DBeaver 5.2.2 and earlier doesn't support Java 11.
Version 5.2.3 will be based on new Eclipse platform with Java 11 support.
Status: Issue closed
username_0: thanks for the info! |
swagger-api/swagger-editor | 234432338 | Title: Tag Object description not respecting multi-line literals in swagger-editor
Question:
username_0: We cannot incorporate line breaks into the description field for tags. All other description fields respect markdown and multi-line literals. This is really hampering readability of our tags for the APIs.
Example:
tags:
- name: item
description: |
To Do - Create a tag for each resource in your api and tag the operations that relate to that resource.
Provide the documentation for the resource in the description of the tag.
The following views are supported by this resource.
- complete, includes all fields.
- summary, includes only id, descriptor, name and description.
- otherView, includes only id, primaryCategory and secondaryCategories.
Answers:
username_1: Moving this to swagger-ui as this is a rendering issue.
username_1: This issue was moved to swagger-api/swagger-ui#3208
Status: Issue closed
|
kosen13s/tiny | 243156501 | Title: How construct concrete data from syntax trees
Question:
username_0: Now, `LexicalTree` express a tree of `String`.
In semantic parsing phase, probably we need to construct data from `String` tree.
Type information can be added in this phase, but it requires second scan nevertheless first scan is done in lexical parsing phase.
Type information is owned by only `Token` s that can be become `Leaf` of `Tree`. (`Literal` and `Identifier`)
The former can get a type information in lexical parsing phase, but latter cannot.
Based on the above, I'm coming up with two way.
1. Add type information to `Literal` in lexical parsing phase. To `Identifier`, adding it in semantic parsing phase.
2. Type information is added to all `Leaf` element in only semantic parsing phase.
Or do you have ideas? I need your opinion.
Answers:
username_1: the 2nd way looks better.
i think adding type information should not be split.
username_0: OK, doing it. |
jstewmc/gravity | 365147796 | Title: Add continuous integration
Question:
username_0: Obviously, we should add continuous integration to be certain the library works! It seems like Travis CI is the best option.
Answers:
username_0: Welp, I tried this a few days ago, and I botched it. I think I started at travis-ci.org, which is deprecated. And, at some point in the process, I was forwarded to travis-ci.com to activate the GitHub Application for the Gravity repository, which I did. However, according to their error message, I can't activate travis-ci.com for Gravity, even after deleting it from travis-ci.org.
I have a ticket into their support team.
username_0: Ok, I'm going to go ahead and close this one.
The Travis CI support theam helped to migrate the project from travis-ci.org to travis-ci.com.
Then, we had four checks in the pull request. We were checking a branch build (i.e., a build starts when a branch is pushed) and a pull request build (i.e., a build starts when a pull request is created or committed to). Rather than [double build](https://docs.travis-ci.com/user/pull-requests/#double-builds-on-pull-requests), I turned off the branch build.
We still have two checks, but that's because of a quirk with the [Github application versus the Github API](https://github.com/travis-ci/travis-ci/issues/9618), which they plan to [fix tomorrow](https://blog.travis-ci.com/2018-09-27-deprecating-github-commit-status-api-for-github-apps-managed-repositories).
Status: Issue closed
|
necolas/react-native-web | 563034680 | Title: Can't resolve 'react-native-web/dist/exports/ViewPropTypes' after upgrading to 0.12.1
Question:
username_0: I have tried to upgrade from `react-native-web` v0.11.7 to 0.12.1 and it breaks my app:
Module not found: Can't resolve 'react-native-web/dist/exports/ViewPropTypes' in '/home/dka/workspace/test-native/node_modules/react-native-swipe-list-view/components'
See:
https://github.com/jemise111/react-native-swipe-list-view/blob/6f22515c1d404d51bf8a6d1591ec51d241b77b9b/components/SwipeRow.js#L11
Related to:
- https://github.com/jemise111/react-native-swipe-list-view/issues/412
Answers:
username_1: It was removed as mentioned in the release notes https://github.com/username_1/react-native-web/releases/tag/0.12.0
Status: Issue closed
username_2: A lot of other packages are not able to be compiled after upgrading to 0.12.x
We use react-native-web as alias for react-native. So, if react-native still support prop types exports, but as deprecated, i think this package should behave in the same way: support them, but deprecated.
username_3: Hello.
I have the same issue like this.
My import is like this
`import {Overlay as BaseOverlay, Button, Text} from 'react-native-elements/src/index';`
And errors in terminal
`./node_modules/react-native-elements/src/config/ViewPropTypes.js
Attempted import error: 'ViewPropTypes' is not exported from 'react-native-web' (imported as 'RNViewPropTypes').`
I created the react app using `create-react-native-web-app` command.
"node": v13.6.0
"react": "^16.13.0",
"react-native": "0.61.5",
"react-native-elements": "^1.0.0-beta8",
What is the solution to fix this issue?
Thanks for your reply.
username_4: Hello @username_3. I had the same issue. Before, I manually copied the library I was using in my code, but it turns out was not the best way so I created a workaround.
**Explanation**
If you mock ViewPropTypes in node_modules/react-native-web/dist/index.js the error will go away.
So what I did is I wrote some batch file since you are on Windows, you can copy and paste.
**Workaround**
1. create a **.bat** file anywhere in your app and feel free to customize the mock. For me setting the style to null fixed the issue. In my case, I created folder called bin so my bat is located at **bin/autofix.bat**
Copy & paste the code
`@echo off
echo 'Fixing ViewPropTypes issues'
REM Fix ViewPropTypes issues
ECHO export const ViewPropTypes = { style: null };>>"PATH_TO_NODE_MODULES/react-native-web/dist/index.js"`
Please replace PATH_TO_NODE_MODULES with the path to your node_modules folder. For example, if you at your root folder it should be "node_modules"
2. (optional) Go in your **package.json**
Modify your scripts add to make sure to add this
Basically it would execute your bat script each time you hit npm or yarn install
`"postinstall": "autofix"`
Hope it would help.
It works for me. Don't hesitate to ask me anything.
Thanks!
Franz
username_5: Hello @username_4 I have the same issue of @username_3 but the script didn't fix my problem, but I'm not sure if I'm making the things good, this is the way that I put that script in the package.json :
....
"scripts": {
"web": "node scripts/start.js",
"test:web": "node scripts/test.js",
"build": "node scripts/build.js",
"start": "react-native start",
"start-clean": "rm -rf $TMPDIR/react-* && watchman watch-del-all && npm start -- --reset-cache",
"test": "npm run test:web && npm run test:native",
"android": "react-native run-android",
"ios": "react-native run-ios",
"test:native": "node_modules/.bin/jest -c ./config/jest/jest.config.native.js",
**_"postinstall": "autofix.bat"_**
},
....
and I have put the file autofix.bat in the root of my project.
than you
username_4: Hello @username_5,
it should normally work. May I please see the content of your autofix.bat?
**Possible fixes:**
1. Did you try to run the script manually via cmd? The script will only run if something is installed. For the first time, I had to run it manually, as I already installed my packages.
2. Open **node_modules/react-native-web/dist/index.js**. Can you see this at the bottom
`export const ViewPropTypes = { style: null };`
Thanks!
Franz
username_5: Ok well, I have added it manually to the file and it works, let me try to
run the script manually.
thank you so much for your response.
username_4: Hello @username_5,
You're welcome and I'm happy it works for you too.
To test the **postinstall** feature. You can install any package, let say **lodash**, then check if the script updated it. After that, you know it works and therefore free to remove **lodash** from your packages.
Thanks!
Franz
username_6: Module not found: Can't resolve 'react-native-web/dist/exports/ViewPropTypes' after update to expo sdk 39.0
username_4: Hello @username_6 . I had the same issue. Before, I manually copied the library I was using in my code, but it turns out was not the best way so I created a workaround.
Check out my comment above: https://github.com/username_1/react-native-web/issues/1537#issuecomment-619969653
It works for me. Don't hesitate to ask me anything.
Thanks!
Franz
username_7: @username_6 the same
username_4: Did you try my workaround above, this is because ViewPropTypes was removed from react native, so react native web removed it too.
username_8: Are you sure? I have the same issue, but npm package works without problems on RN.
username_8: Thank you for your answer. I am sure it can help me. Could you please explain in details what this lines do? I need create a bashscript for my project.
username_4: Hello @username_8,
When you add `export const ViewPropTypes = { style: null }` to `PATH_TO_NODE_MODULES/react-native-web/dist/index.js`. You mock **ViewPropTypes**, by exporting an object.
Check out my answer above https://github.com/username_1/react-native-web/issues/1537#issuecomment-619969653
I updated the answer, now there is a batch script, a bash script, and a python script. I recommend you to use the Python version since it works on both Unix-like OS & Windows having Python installed.
Best regards,
Franz
username_9: @username_4 have a look at https://www.npmjs.com/package/patch-package as it makes patching stuff like this much easier 😊
username_4: @username_9
Thanks, I'll check it out
username_10: Hello everyone :) I am having the same issue as @username_6. I tried adding `export const ViewPropTypes = { style: null };` to `/react-native-web/dist/index.js`as @username_4 mentioned, but it doesn't change anything. Are there any other ideas for a possible solution?
username_4: Hi @username_10 , what is the error message?
In my case **ViewPropTypes.style** was undefined, so it meant that `ViewPropTypes `was not defined that is why I exported `ViewPropTypes ` in the `/react-native-web/dist/index.js`
If for example, it says **ColorPropType** is undefined, then you have to export it too.
username_11: It didn't work for me as well unfortunately (it still wanted the export, but in /exports folder), but I've used part of that script and made myself my own workaround here (specifically for react-native-table-component): https://github.com/Gil2015/react-native-table-component/issues/131 . It's not the best solution, but will work until the package is upgraded.
Maybe in your case, you also need to add a ViewPropTypes file into /exports folder and declare that there instead (and from index.js, just export it like the rest are)
username_4: Hi @username_11,
it depends on what is in the index.js. If there is already something like this: `export { default as ViewPropTypes } from './exports/ViewPropTypes';` then you have to create that file., or simply remove the line. If not you can just mock it by exporting an empty object for example like: `export const ViewPropTypes = { style: null };`
username_10: Hey @username_4, the error message with or without exporting the empty ViewPort object is `Module not found: Can't resolve 'react-native-web/dist/exports/ViewPropTypes'`. Also thx @username_11, I will try out your idea in a minute.
username_4: @username_10 in this case @username_11's solution is what you should try.
If you don't mind me asking, could you show what is the dist/index.js
username_10: I just tried out @username_11's solution with node_modules/react-native-web/dist/exports/ViewPropTypes/index.js containing `node_modules/react-native-web/dist/exports/ViewPropTypes/index.js` - still the same error message.
@username_4 dist/index.js:
`// export const ViewPropTypes = { style: null }; // manual fix no 1
// export { default as ViewPropTypes } from './exports/ViewPropTypes'; // manual fix no 2
export { default as unstable_createElement } from './exports/createElement';
export { default as findNodeHandle } from './exports/findNodeHandle';
export { default as processColor } from './exports/processColor';
export { default as render } from './exports/render';
export { default as unmountComponentAtNode } from './exports/unmountComponentAtNode';
export { default as NativeModules } from './exports/NativeModules'; // APIs
export { default as AccessibilityInfo } from './exports/AccessibilityInfo';
export { default as Alert } from './exports/Alert';
export { default as Animated } from './exports/Animated';
export { default as Appearance } from './exports/Appearance';
export { default as AppRegistry } from './exports/AppRegistry';
export { default as AppState } from './exports/AppState';
export { default as BackHandler } from './exports/BackHandler';
export { default as Clipboard } from './exports/Clipboard';
export { default as DeviceInfo } from './exports/DeviceInfo';
export { default as Dimensions } from './exports/Dimensions';
export { default as Easing } from './exports/Easing';
export { default as I18nManager } from './exports/I18nManager';
export { default as Keyboard } from './exports/Keyboard';
export { default as InteractionManager } from './exports/InteractionManager';
export { default as LayoutAnimation } from './exports/LayoutAnimation';
export { default as Linking } from './exports/Linking';
export { default as NativeEventEmitter } from './exports/NativeEventEmitter';
export { default as PanResponder } from './exports/PanResponder';
export { default as PixelRatio } from './exports/PixelRatio';
export { default as Platform } from './exports/Platform';
export { default as Share } from './exports/Share';
export { default as StyleSheet } from './exports/StyleSheet';
export { default as UIManager } from './exports/UIManager';
export { default as Vibration } from './exports/Vibration'; // components
export { default as ActivityIndicator } from './exports/ActivityIndicator';
export { default as Button } from './exports/Button';
export { default as CheckBox } from './exports/CheckBox';
export { default as FlatList } from './exports/FlatList';
export { default as Image } from './exports/Image';
export { default as ImageBackground } from './exports/ImageBackground';
export { default as KeyboardAvoidingView } from './exports/KeyboardAvoidingView';
export { default as Modal } from './exports/Modal';
export { default as Picker } from './exports/Picker';
export { default as Pressable } from './exports/Pressable';
export { default as ProgressBar } from './exports/ProgressBar';
export { default as RefreshControl } from './exports/RefreshControl';
export { default as SafeAreaView } from './exports/SafeAreaView';
export { default as ScrollView } from './exports/ScrollView';
export { default as SectionList } from './exports/SectionList';
export { default as StatusBar } from './exports/StatusBar';
export { default as Switch } from './exports/Switch';
export { default as Text } from './exports/Text';
export { default as TextInput } from './exports/TextInput';
export { default as Touchable } from './exports/Touchable';
export { default as TouchableHighlight } from './exports/TouchableHighlight';
export { default as TouchableNativeFeedback } from './exports/TouchableNativeFeedback';
export { default as TouchableOpacity } from './exports/TouchableOpacity';
export { default as TouchableWithoutFeedback } from './exports/TouchableWithoutFeedback';
export { default as View } from './exports/View';
export { default as VirtualizedList } from './exports/VirtualizedList';
export { default as YellowBox } from './exports/YellowBox'; // compat (components)
export { default as DrawerLayoutAndroid } from './exports/DrawerLayoutAndroid';
export { default as InputAccessoryView } from './exports/InputAccessoryView';
export { default as ToastAndroid } from './exports/ToastAndroid'; // compat (apis)
export { default as PermissionsAndroid } from './exports/PermissionsAndroid';
export { default as Settings } from './exports/Settings';
export { default as Systrace } from './exports/Systrace';
export { default as TVEventHandler } from './exports/TVEventHandler'; // plugins
export { default as DeviceEventEmitter } from './exports/DeviceEventEmitter'; // hooks
export { default as useColorScheme } from './exports/useColorScheme';
export { default as useWindowDimensions } from './exports/useWindowDimensions';`
username_10: I changed the file to only contain `export default function ViewPropTypes(props) {}` and now it compiles again :) Thx everybody for your help!
username_4: @username_10 you're welcome.
Your index.js looks good, actually you have two solutions, either you replace this line `export { default as ViewPropTypes } from './exports/ViewPropTypes';` with `export const ViewPropTypes = { style: null };`
or you create a new file like you did.
In my humble opinion, it is easier to just replace the line than creating a new file with the content.
Note: each time that you will either install, remove, or update a package, your modifications will go away. So to avoid that add a script in your package.json.
So follow my comment above but instead of the code displayed there use this script instead.
**Python: fix.py OS-agnostic, use this if possible for your case**
```python
import os
import sys
print("✅ Fixing PropTypes issues")
dir_path = os.path.dirname(os.path.realpath(__file__))
rnw_filename = dir_path + "/../node_modules/react-native-web/dist/index.js"
def remove_line_from_file(file_name, text_to_remove):
with open(file_name, "r+") as f:
d = f.readlines()
f.seek(0)
for i in d:
if i != text_to_remove:
f.write(i)
f.truncate()
def append_new_line(file_name, text_to_append):
"""Append given text as a new line at the end of file"""
if text_to_append in open(file_name).read():
print("⏭️ Skipping...")
else:
with open(file_name, "a+") as file_object:
file_object.write("\n")
file_object.write(text_to_append)
file_object.close()
# Remove the unnecessary line first
remove_line_from_file(rnw_filename, "export { default as ViewPropTypes } from './exports/ViewPropTypes';")
# Fix
append_new_line(rnw_filename, "export const ViewPropTypes = { style: null };")
```
And there you go.
@username_11 you can use it as well it is easier
username_12: 67 | duration: PropTypes.number,
68 | visible: PropTypes.bool,
69 | position: PropTypes.number,
I can't figure out why it recognizes viewproptypes but not the style part
username_4: Hi @username_12, this means that the ViewPropTypes is not defined correctly i.e undefined. This explains why you get the property 'style' of an undefined object.
What is the library that requires **ViewPropTypes**?
username_12: Module.../../../react-native-root-toast/lib/ToastContainer.js
node_modules/react-native-root-toast/lib/ToastContainer.js:66
username_4: @username_12 I just saw it, it looks good. Try to apply the fix manually. Does it work?
If not, please could you explain what you did exactly?
username_12: @username_4, if I do nothing:
Module not found: Can't resolve 'react-native-web/dist/exports/ViewPropTypes' in '/node_modules/react-native-button'
If I use your python postinstall script, it adds "export const ViewPropTypes = { style: null };" to the bottom of index file. I then receive error:
Module not found: Can't resolve 'react-native-web/dist/exports/ViewPropTypes' in '/node_modules/react-native-button'
node_modules/react-native-button/Button.js
Then if I add the alias in my webpack:
'react-native-web/dist/exports/ViewPropTypes': 'react-native-web/dist/index.js',
I get the original style error I described TypeError: Cannot read property 'style' of undefined
username_4: Hi @username_12,
Then the problem is that at the top of your index.js file, you have something like this:
`export { default as ViewPropTypes } from './exports/ViewPropTypes';`.
Check for that line and replace it to this `export const ViewPropTypes = { style: null };`.
```diff
- export { default as ViewPropTypes } from './exports/ViewPropTypes';
+ export const ViewPropTypes = { style: null };
```
Let me know if it works or not.
username_12: I did not have export { default as ViewPropTypes } from './exports/ViewPropTypes'; at the top of the file. However, if I add it to the file and then actually create the folder/file with only export default function ViewPropTypes(props) {} inside. it works.
But after that now i get a similar error for Text.propTypes and I am not sure how to repeat this solution for that propType
username_4: Hi @username_12,
I got a solution for you, I tested it myself.
Go to this file `node_modules\react-native-web\dist\exports\Text\index.js`
Add this line
```diff
...
Text.displayName = 'Text';
+ Text.propTypes = ()=> {};
...
```
It will mock the styles. Now you can integrate it into the Python script too, by simply replacing the `"Text.displayName = 'Text';"` with `"Text.displayName = 'Text';Text.propTypes = ()=> {};"`
Hope it would help.
Franz
username_13: i created script for windows but i guess it's quite easy to use it in mac os or linux.
Just add this script in postinstall section :
`
"scripts": {
"postinstall": "shameForDevsWhoCantAddBackwardCompability.bat",
....
`
shameForDevsWhoCantAddBackwardCompability.bat script:
`
mkdir "node_modules\react-native-web\dist\exports\ViewPropTypes";
echo const ViewPropTypes = { style: null }; > node_modules\react-native-web\dist\exports\ViewPropTypes\index.js
echo export default ViewPropTypes; >> node_modules\react-native-web\dist\exports\ViewPropTypes\index.js
findstr "ViewPropTypes" node_modules\react-native-web\dist\index.js
IF ERRORLEVEL 1 echo export { default as ViewPropTypes } from './exports/ViewPropTypes'; >> node_modules\react-native-web\dist\index.js
`
What it does is simply creates blank ViewPropTypes and exports it. findstr is equivalent for grep.
Hope this will help you as it helped me.
Tested with expo sdk 39 and react-native-web: 0.14.0
username_4: Hi @username_13 did you check my example? https://github.com/username_1/react-native-web/issues/1537#issuecomment-619969653
It is easy and works for Linux, MacOS and Windows
username_13: yeah :) , i just don't want to use any external stuff
username_14: Working solution for "Module not found", tweaked a little bit and inspired from answers above.
1. Create folder && file: ViewPropTypes/index.js in react-native-web/dist/exports
2. Add below code in ViewPropTypes/index.js :
var ViewPropTypes = { style: null }
export default ViewPropTypes;
3. Import above in react-native-web/dist/index.js
export { default as ViewPropTypes } from './exports/ViewPropTypes';
4. run expo r -c
Info: "expo": "^39.0.0"
username_15: @username_4 Script wasn't quite working for me as it wasn't finding the exports. I modified it below and now the errors are gone. The types list can contain all the files with which you are getting errors. I was getting multiple for different dependencies.
```py
import os
dir_path = os.path.dirname(os.path.realpath(__file__))
types = ['ViewPropTypes', 'TextPropTypes', 'ColorPropType', 'EdgeInsetsPropType', 'PointPropType']
for type in types:
BASE_DEST = dir_path+'/node_modules/react-native-web/dist/exports/'+type+'/'
if not os.path.exists(BASE_DEST):
os.mkdir(BASE_DEST)
with open(BASE_DEST+'index.js', "w") as file_object:
file_object.write("module.exports = {}")
file_object.close()
```
username_4: Hi @username_15, That's great. It means that there are many types missing. I see, thanks for sharing your code. It is easier to create folders and files as you can easily check if it doesn't exist already like you did.
username_16: @username_14 your solution works for me!! Thanks for saving my day.
username_17: Building on @username_4 's script (by the way thank you it fixed it), here's my version, for bash:
1. in the root directory, add this as "fix-for-web.sh":
`#!/bin/bash
echo 'Fixing PropTypes issues, for running expo start:web (for web)'
echo "for reference: https://github.com/username_1/react-native-web/issues/1537"
IMPORTS_REACT_NATIVE_WEB=('ViewPropTypes' 'ColorPropType' 'EdgeInsetsPropType' 'PointPropType' '
requireNativeComponent')
for import in "${IMPORTS_REACT_NATIVE_WEB[@]}"
do
echo "Fixing $import ..."
if grep -q "export const $import = { style: null };" ./node_modules/react-native-web/dist/index.js; then
echo "$import fixed already!"
else
echo -e "\nexport const $import = { style: null };">> ./node_modules/react-native-web/dist/index.js
fi
done
`
NOTE: depending on your OS - you might need to run: "chmod +x ./fix-for-web.sh"
2. In package.json:
`
"scripts": {
"start": "expo start",
...
"postinstall": "bash ./fix-for-web.sh"
},
`
username_18: @username_17
Thanks for the solution! And everyone else as well.
This still isn't quite working for me. After running the script (which successfully adds the export lines to `index.js`, I still get the following error:
```
Failed to compile.
./node_modules/react-native-table-component/components/rows.js
Cannot find module: 'react-native-web/dist/exports/ViewPropTypes'. Make sure this package is installed.
You can install this package by running: yarn add react-native-web/dist/exports/ViewPropTypes.
```
I can make the web build compile by adding the following to `webpack.config.js`:
` config.resolve.alias['react-native-web/dist/exports/ViewPropTypes'] = 'react-native-web/dist/index.js';`
However, then I get the errors that `ViewPropTypes` is null, and `.style` is not defined on it.
Anyone have any ideas? Would really appreciate any help. I think it's close, and I need to get this working to get my react-native-web app to compile again.
username_4: Hi @username_18
Did you make sure to always run your script after `npm install` or `yarn install`? If not then you have to add a script.
What solution did you use exactly?
(We ended up bringing many workarounds haha)
username_18: @username_4 Hey,
Yeah I'm testing it before running `npm install`. I understand that it has to be added into `postinstall`, and have verified that the extra lines exist in `./node_modules/react-native-web/dist/index.js` when compiling.
Maybe it's because I'm running `expo build`? That's the way I've been testing. I haven't found a solution yet that works, and there's quite a few affected libraries so I don't want to patch all of them individually.
@username_14 's solution looks like it may work, but I'm not sure how to put it into a script so I can run it on `postinstall`
username_4: Then make sure to export **ViewPropTypes** as an object that includes **style** property.
So it should look like this
```js
export const ViewPropTypes = { style: null };
```
or
```js
export const ViewPropTypes = { style: ()=> null };
```
username_18: Yes, I used the script you provided (thanks for that!), and that exact line is at the bottom of `./node_modules/react-native-web/dist/index.js`.
username_18: Building off @username_14 's solution, I resolved this by adding the following to my `fix-for-web.sh` `post` script:
```
mkdir node_modules/react-native-web/dist/exports/ViewPropTypes
echo "var ViewPropTypes = { style: null }" >> node_modules/react-native-web/dist/exports/ViewPropTypes/index.js
echo "export default ViewPropTypes;" >> node_modules/react-native-web/dist/exports/ViewPropTypes/index.js
echo "export { default as ViewPropTypes } from './exports/ViewPropTypes';" >> node_modules/react-native-web/dist/index.js
```
username_19: use patch-package
first, add script at package.json
(postinstall means after effect installing)
```
+ "postinstall": "patch-package"
```
second, install patch-package
```
npm install --save-dev patch-package
yarn add -D patch-package postinstall-postinstall
```
third
open `react-native-web/dist/index.js` and edit
```
+ export const ViewPropTypes = { style: ()=> null };
```
fourth run patch-package
```
npx patch-package react-native-web
```
check created file at rootDir/patches
good luck
username_20: thx man, it worked! |
dask/dask-kubernetes | 771041714 | Title: Running the tests from master branch fail - "Couldn't gather keys"
Question:
username_0: **What happened**:
make test doesn't work. I keep having the following log which will continuously go on and on for the test "test_cluster_create":
```
dask_kubernetes/tests/test_async.py::test_cluster_create distributed.scheduler - INFO - Clear task state
distributed.scheduler - INFO - Scheduler at: tcp://192.168.2.4:46337
distributed.scheduler - INFO - dashboard at: :8787
distributed.scheduler - INFO - Receive client connection: Client-291dabe6-4152-11eb-869b-9cb6d0f7a683
distributed.core - INFO - Starting established connection
distributed.scheduler - INFO - Register worker <Worker 'tcp://10.244.0.10:36501', name: 0, memory: 0, processing: 1>
distributed.scheduler - INFO - Starting worker compute stream, tcp://10.244.0.10:36501
distributed.core - INFO - Starting established connection
distributed.scheduler - ERROR - Couldn't gather keys {'lambda-abb84b7bea08dbbd080ed10c969c7abc': ['tcp://10.244.0.10:36501']} state: ['memory'] workers: ['tcp://10.244.0.10:36501']
NoneType: None
distributed.scheduler - INFO - Remove worker <Worker 'tcp://10.244.0.10:36501', name: 0, memory: 1, processing: 0>
distributed.core - INFO - Removing comms to tcp://10.244.0.10:36501
distributed.scheduler - INFO - Lost all workers
distributed.scheduler - ERROR - Workers don't have promised key: ['tcp://10.244.0.10:36501'], lambda-abb84b7bea08dbbd080ed10c969c7abc
NoneType: None
distributed.client - WARNING - Couldn't gather 1 keys, rescheduling {'lambda-abb84b7bea08dbbd080ed10c969c7abc': ('tcp://10.244.0.10:36501',)}
distributed.scheduler - INFO - Register worker <Worker 'tcp://10.244.0.10:36501', name: 0, memory: 0, processing: 1>
distributed.scheduler - INFO - Starting worker compute stream, tcp://10.244.0.10:36501
distributed.core - INFO - Starting established connection
^Cdistributed.scheduler - INFO - Scheduler closing...
distributed.scheduler - INFO - Scheduler closing all comms
distributed.scheduler - INFO - Remove worker <Worker 'tcp://10.244.0.10:36501', name: 0, memory: 1, processing: 0>
distributed.core - INFO - Removing comms to tcp://10.244.0.10:36501
distributed.scheduler - INFO - Lost all workers
```
**Anything else we need to know?**:
I tried to keep the environment the same as much as possible, but I'm not sure these minor differences are causing the issue. Worker log:
```
+ '[' '' ']'
+ '[' -e /opt/app/environment.yml ']'
+ echo 'no environment.yml'
+ '[' '' ']'
+ '[' '' ']'
+ exec dask-worker tcp://192.168.2.4:46337 --nthreads 1 --death-timeout 60 --name 0
no environment.yml
distributed.nanny - INFO - Start Nanny at: 'tcp://10.244.0.10:40809'
distributed.worker - INFO - Start worker at: tcp://10.244.0.10:36501
distributed.worker - INFO - Listening to: tcp://10.244.0.10:36501
distributed.worker - INFO - dashboard at: 10.244.0.10:46435
distributed.worker - INFO - Waiting to connect to: tcp://192.168.2.4:46337
distributed.worker - INFO - -------------------------------------------------
distributed.worker - INFO - Threads: 1
distributed.worker - INFO - Memory: 2.08 GB
distributed.worker - INFO - Local Directory: /usr/local/src/dask-worker-space/dask-worker-space/worker-07732v9v
distributed.worker - INFO - -------------------------------------------------
distributed.worker - WARNING - Mismatched versions found
+---------+------------------------+------------------------+------------------------------------------------------+
| Package | This Worker | scheduler | workers |
+---------+------------------------+------------------------+------------------------------------------------------+
| dask | 2020.12.0+13.g76ea3379 | 2020.12.0+15.g30e64df0 | {'2020.12.0+13.g76ea3379', '2020.12.0+15.g30e64df0'} |
| python | 3.8.0.final.0 | 3.8.5.final.0 | {'3.8.5.final.0', '3.8.0.final.0'} |
+---------+------------------------+------------------------+------------------------------------------------------+
distributed.worker - INFO - Registered to: tcp://192.168.2.4:46337
[Truncated]
| dask | 2020.12.0+13.g76ea3379 | 2020.12.0+15.g30e64df0 | {'2020.12.0+13.g76ea3379', '2020.12.0+15.g30e64df0'} |
| python | 3.8.0.final.0 | 3.8.5.final.0 | {'3.8.5.final.0', '3.8.0.final.0'} |
+---------+------------------------+------------------------+------------------------------------------------------+
distributed.worker - INFO - Registered to: tcp://192.168.2.4:46337
distributed.worker - INFO - -------------------------------------------------
distributed.core - INFO - Starting established connection
distributed.worker - INFO - Stopping worker at tcp://10.244.0.10:36501
distributed.nanny - INFO - Worker closed
distributed.dask_worker - INFO - Exiting on signal 15
distributed.nanny - INFO - Closing Nanny at 'tcp://10.244.0.10:40809'
distributed.nanny - INFO - Worker process 18 was killed by signal 15
distributed.dask_worker - INFO - End worker
```
**Environment**:
- Dask version: 2012.12.0
- Python version: 3.8
- Operating System: Arch linux
- Install method (conda, pip, source): clone of master
Answers:
username_0: I did managed to get the tests running if I use `make k8s-make COMMAND=test. Does this mean the worker and the scheduler has to have exactly the same packages and python version including to the minor version?
username_1: It is safest to assume so.
username_0: @username_1 I solved that, and downgraded the python in my conda environment. But it still is in this weird loop.
username_0: Test:
```
dask_kubernetes/tests/test_async.py::test_cluster_create distributed.scheduler - INFO - Clear task state
distributed.scheduler - INFO - Scheduler at: tcp://192.168.2.4:42497
distributed.scheduler - INFO - dashboard at: :8787
distributed.scheduler - INFO - Receive client connection: Client-42527b5f-444a-11eb-b06d-9cb6d0f7a683
distributed.core - INFO - Starting established connection
distributed.scheduler - INFO - Register worker <Worker 'tcp://10.244.0.5:44073', name: 0, memory: 0, processing: 1>
distributed.scheduler - INFO - Starting worker compute stream, tcp://10.244.0.5:44073
distributed.core - INFO - Starting established connection
distributed.scheduler - ERROR - Couldn't gather keys {'lambda-abb84b7bea08dbbd080ed10c969c7abc': ['tcp://10.244.0.5:44073']} state: ['memory'] workers: ['tcp://10.244.0.5:44073']
NoneType: None
distributed.scheduler - INFO - Remove worker <Worker 'tcp://10.244.0.5:44073', name: 0, memory: 1, processing: 0>
distributed.core - INFO - Removing comms to tcp://10.244.0.5:44073
distributed.scheduler - INFO - Lost all workers
distributed.scheduler - ERROR - Workers don't have promised key: ['tcp://10.244.0.5:44073'], lambda-abb84b7bea08dbbd080ed10c969c7abc
NoneType: None
distributed.client - WARNING - Couldn't gather 1 keys, rescheduling {'lambda-abb84b7bea08dbbd080ed10c969c7abc': ('tcp://10.244.0.5:44073',)}
```
Worker:
```
+ exec dask-worker tcp://192.168.2.4:42497 --nthreads 1 --death-timeout 60 --name 0
distributed.nanny - INFO - Start Nanny at: 'tcp://10.244.0.5:46091'
distributed.worker - INFO - Start worker at: tcp://10.244.0.5:44073
distributed.worker - INFO - Listening to: tcp://10.244.0.5:44073
distributed.worker - INFO - dashboard at: 10.244.0.5:37069
distributed.worker - INFO - Waiting to connect to: tcp://192.168.2.4:42497
distributed.worker - INFO - -------------------------------------------------
distributed.worker - INFO - Threads: 1
distributed.worker - INFO - Memory: 2.08 GB
distributed.worker - INFO - Local Directory: /usr/local/src/dask-worker-space/dask-worker-space/worker-q6y8igvo
distributed.worker - INFO - -------------------------------------------------
distributed.worker - INFO - Registered to: tcp://192.168.2.4:42497
distributed.worker - INFO - -------------------------------------------------
distributed.core - INFO - Starting established connection
distributed.worker - INFO - -------------------------------------------------
distributed.worker - INFO - Registered to: tcp://192.168.2.4:42497
distributed.worker - INFO - -------------------------------------------------
distributed.core - INFO - Starting established connection
distributed.worker - INFO - -------------------------------------------------
distributed.worker - INFO - Registered to: tcp://192.168.2.4:42497
distributed.worker - INFO - -------------------------------------------------
distributed.core - INFO - Starting established connection
distributed.worker - INFO - -------------------------------------------------
distributed.worker - INFO - Registered to: tcp://192.168.2.4:42497
distributed.worker - INFO - -------------------------------------------------
```
That just goes on and on.
username_0: I don't see any errors, except that somehow it can't fetch the results.
username_0: This is due to networking. And I'm not sure how this is supposed to work. The client tries to reach out to the workers directly, instead of through the scheduler. This will mean it will need to be able to directly use POD ip's, which is not possible with both kind and minikube, as far as I know. As least not by default.
I read through the docs and my ticket here https://github.com/dask/dask-kubernetes/issues/287
Where it stills sounds like it should be possible to run the client & scheduler outside the cluster (for testing purposes), but I don't see any instructions on how to make this possible?
username_1: I dont think this is true. This should never happen.
username_0: @username_1 I might have described it wrong. The scheduler also runs outside the Kubernetes cluster in this case, unfortunately it doesn't reuse the connection the workers make to the scheduler, but create a new connection from the scheduler to the clients. This fails because the default kind setup doesn't allow. (the function, which is used by the scheduler when asking for results: https://github.com/dask/distributed/blob/658fccc3268bd16226cc5815bf00eed349fe1553/distributed/utils_comm.py#L19 )
Does anyone have this actually working? The docs, and your replies on my other issue made me believe this is actually possible. I'm trying to run dask-kubernetes, the client, and the scheduler outside the cluster, as this makes it possible to test the whole process easily and make adjustments before an actual deployment.
username_1: The scheduler and workers all need to be able to open TCP connections between themselves at any time.
I typically use minikube and add the routing rules as described in the docs. I've not tried the kind instructions for local development, they were added by a contributor. But I expect you would need to set up similar routing rules to ensure you can make connections into the docker network.
username_0: Thanks. I will have another look with minikube instead.
Status: Issue closed
|
Borber/Allendance | 728204565 | Title: 咨询
Question:
username_0: 博主您好,请问您是怎么实现hugo用action做到自动部署?可以请教一下吗
Answers:
username_1: 查看我的 blog 仓库
Status: Issue closed
username_0: 已经解决了 文章参考:[Here](https://www.riotian.com/tech/%E4%BD%BF%E7%94%A8hugo+github-actions%E5%AE%9E%E7%8E%B0%E8%87%AA%E5%8A%A8%E5%8C%96%E9%83%A8%E7%BD%B2%E5%8D%9A%E5%AE%A2/) |
revoxhere/duino-coin | 967664145 | Title: Mining but no DUCO earned
Question:
username_0: Thanks for fixing the mining fix for the AVR, PC, and ESP miner. I finally got them up. There does seem to look like there is one more issue. Been mining for the past few hours now and the balance for DUCO has not moved at all. Not sure what I would call it but I am guessing rewards are not being issued out to miners.
Answers:
username_1: ### I have the same problem as you, I bet everyone does too ###
username_2: Same.
username_2: The users are too much to mining at the same time that might causes the Duino-Coin Main Server overloading and down. Maybe that's the main reason causes all types of miners can't reach the server for mining, even not able to login to the Web Wallet and check the user balance on the Exchange Page.
username_3: Is the issue fixed now?
username_1: not yet
Status: Issue closed
|
krzyzanowskim/CryptoSwift | 751868586 | Title: Crash when using AES CCM
Question:
username_0: **Describe the bug**
I'm trying to decrypt data coming from an API that uses SJCL on the client side and encrypts data with AES CCM, 256 bit key, no padding.
Here is the [demo page](http://bitwiseshiftleft.github.io/sjcl/demo/) to generate test data.
**Reproduce**
Using:
- Xcode 12.2 (12B45b) on macOS 10.15.7
- Pure Swift Package (no app)
- CryptoSwift 1.3.7 from SPM : `.package(url: "https://github.com/krzyzanowskim/CryptoSwift.git", from: "1.3.7")`
Steps to reproduce:
1. Use the [demo page](http://bitwiseshiftleft.github.io/sjcl/demo/) to create data.
2. Decode with AES/CCM (see code below)
3. crash `EXC_BAD_ACCESS (code=1, address=0x7fb2f6f00000)` on `try aes.decrypt(encrypted)`
Code sample:
```Swift
// Data generated from http://bitwiseshiftleft.github.io/sjcl/demo/
// Key generation info used on page:
let password = "<PASSWORD>"
// Salt: 762CB1CB FA62DEDA
let salt: [UInt8] = [
0x76, 0x2C, 0xB1, 0xCB,
0xFA, 0x62, 0xDE, 0xDA]
// Generated key:
// 256 bit key: <KEY>
let rawKey: [UInt8] = [
0x0F, 0xB1, 0x30, 0xBA,
0xCF, 0xBE, 0x98, 0x60,
0x35, 0xC0, 0x94, 0xC7,
0xBC, 0x94, 0xD4, 0x99,
0xE7, 0x9C, 0x89, 0xCD,
0x81, 0x12, 0x55, 0x2C,
0xF6, 0x4D, 0xF6, 0xAE,
0x39, 0xBB, 0x69, 0x12]
// Message:
let message = "Test CryptoSwift decrypt"
// no authenticated additional data
// Encrypted data:
/* JSON transport
{"iv":"dT9+NAjLzdMNvw5x7m2hiw==",
"v":1,
"iter":10000,
"ks":256,
"ts":64,
"mode":"ccm",
"adata":"",
"cipher":"aes",
"salt":"<PASSWORD>=",
"ct":"<KEY>="}
*/
// JSON decoded data:
[Truncated]
// Base64 decoded
let ivDecoded: [UInt8] = Data(base64Encoded: iv)!.bytes
let encrypted: [UInt8] = Data(base64Encoded: ct)!.bytes
let tsInBytes = Int(ts/8)
let ccm = CCM(
iv: ivDecoded,
tagLength: tsInBytes,
messageLength: encrypted.count - tsInBytes)
do {
let aes = try AES(key: rawKey, blockMode: ccm, padding: .noPadding)
let decrypted = try aes.decrypt(encrypted) // <-- crash here
} catch let error {
}
```
Am I doing something wrong here?
Thanks
Answers:
username_0: Here is a repo that reproduces the problem: https://github.com/username_0/test-CryptoSwift
Run the test case: https://github.com/username_0/test-CryptoSwift/blob/main/Tests/MyLibraryTests/CryptoSwiftTests.swift
username_0: Found the problem: wrong size of `iv`
Needs to be the first (15 - lol) bytes, where lol is 2/3/4 if size is <= 1<<16 / <= 1<<24 / > 1<<24.
Status: Issue closed
|
spree/spree_gateway | 430994978 | Title: `The 'AnetApi/xml/v1/schema/AnetApiSchema.xsd:amount' element is invalid`
Question:
username_0: def credit(amount, creditcard, response_code, gateway_options = {})
create_transaction(amount, creditcard, :refund, transaction_options(gateway_options).merge(trans_id: response_code))
end
def void(response_code, creditcard, gateway_options = {})
create_transaction(nil, creditcard, :void, transaction_options(gateway_options).merge(trans_id: response_code))
end
def cancel(response_code)
response = void(response_code, nil)
response = credit(nil, nil, response_code) unless response.success?
response
end
We use cancel method in authorize_net_cim.rb, when code into credit(nil, nil, response_code), it will always return
`=> #<ActiveMerchant::Billing::Response:0x00007fde9655cdd8
@authorization=nil,
@avs_result={"code"=>nil, "message"=>nil, "street_match"=>nil, "postal_match"=>nil},
@cvv_result={"code"=>nil, "message"=>nil},
@emv_authorization=nil,
@error_code="E00003",
@fraud_review=nil,
@message=
"The 'AnetApi/xml/v1/schema/AnetApiSchema.xsd:amount' element is invalid - The value '' is invalid according to its datatype 'Decimal' - The string '' is not a valid Decimal value.",
@params=
{"messages"=>
{"result_code"=>"Error",
"message"=>
{"code"=>"E00003",
"text"=>
"The 'AnetApi/xml/v1/schema/AnetApiSchema.xsd:amount' element is invalid - The value '' is invalid according to its datatype 'Decimal' - The string '' is not a valid Decimal value."}}},
@success=false,
@test=false>`
Answers:
username_1: We are encountering this too. Has there been any progress?
@username_0 If you don't mind, what did you end up doing? |
nicochampda/portetgym | 614690958 | Title: Proposition de contenu
Question:
username_0: Inspiration pour le contenu :
https://www.ctgym.fr/en-savoir-plus/inscriptions-saison-2019-2020-74429
https://www.gymblagnac.com
[PROPOSITION DE CONTENU ET DE SON ORDRE.pdf](https://github.com/nicochampda/portetgym/files/4598724/PROPOSITION.DE.CONTENU.ET.DE.SON.ORDRE.pdf) |
jmiskovic/hexpress | 830953295 | Title: Turning off the reverb on the drums
Question:
username_0: Is there any way to turn off the reverb on the drums?
Love the new record feature btw and this whole project as a whole really.
Answers:
username_0: Oh tilting the phone towards the right works out fine.
Hmm I did notice some crackling noice when using the recording feature but I chalked it up to low processing power on my phone (only 2gb). This also happens only when recording a lot of notes and playing more on top of that at the same time. This is not a problem for me really but maybe it is to your target audience.
Thank you for your blazing fast responses and good luck with the fix.
Status: Issue closed
username_1: I tried to capture some of these tilting features here, but there were some changes since that diagram was created.
 |
milieuinfo/webcomponent-vl-ui-map | 809997731 | Title: [BUG] - [map-layer-switcher] volgende layer kiezen verschuift de kaart
Question:
username_0: **Omschrijf het probleem**
Wanneer je de volgende layer kiest in de layer-switcher, verschuift je kaart telkens tot hij niet meer verder kan.
**Hoe te reproduceren**
Ga naar de demo-pagina en blijf enkele keren klikken op de layer-switcher https://webcomponenten.omgeving.vlaanderen.be/demo/vl-map-overview-map.html
**Gewenst gedrag**
Dat de kaart blijft staan.
**Omgeving (gelieve aan te vullen):**
- Browser Chrome
Status: Issue closed
Answers:
username_1: Duplicatie #181. |
artesaos/site | 144110336 | Title: Desenvolvimento do HTML
Question:
username_0: Galera, vou puxar esse bonde! Quem quer ajudar, cola junto que é nois! :P
Answers:
username_1: Olá, não sou conheço o VUE mas vou me empenhar nesta semana e ajudar no trabalho do HTML.
username_2: Paulo, estamos juntos. Conheço de html e css mas vue e json não sei, quero aprender.
username_1: Vamos deixar registrado aqui os views(screenshots) q foram definidos.
Quem somos -> about_us_v2
username_3: Galera, eu tb não sabia de vue.js e de ontem pra hoje já tive uma bela noção com este cara aqui... é um desconhecido mas ele explica fodasticamente bem.... rs
http://www.vedcasts.com.br/series/vuejs
username_4: Não é desconhecido não @username_3 kkkk
username_3: AHhahahahaha.... @username_4 eu fui irônico.... este cara é pica das galáxia Vedovelli!!!!
username_0: Galera, vou subir um commit add o logo na navbar. Hoje também apresento a minha opnião sobre um novo header/menu.
username_2: Estou envolvido com meu artigo para conclusão do MBA. Não consegui
acompanhar online no slack as conversas. Ontem trabalhei no layout para o
about e contribution, mas o Michael comentou que estão pensando em mudar o
layout.
Queria saber se vai vingar isso mesmo ou se mantemos o que foi acordado no
hangout para concluir esta semana. Não sou vavorável a retrabalho!!! :-)
_______
<<EMAIL>><NAME>, PMP
Administrador, PMP, MBA
[image: ](48) 9914-0400
[image: Visualizar meu perfil no LinkedIn]José Ferrão |
<http://br.linkedin.com/in/josesferrao>[image: @josesferrao] @josesferrao
<https://twitter.com/josesferrao>
username_2: mandando PR
username_5: O que falta para completar essa issue?
username_0: Falta adicionar mais uma seção sobre as contribuições. Vou tentar entregar, se não hoje, na segunda. E falta o footer tb
username_2: Paulo, está faltando no novo layout? Tem algo que eu possa ajudar?
_______
<<EMAIL>><NAME>, PMP
Administrador, PMP, MBA
[image: ](48) 9914-0400
[image: Visualizar meu perfil no LinkedIn]José Ferrão |
<http://br.linkedin.com/in/josesferrao>[image: @josesferrao] @josesferrao
<https://twitter.com/josesferrao>
username_2: Na versão que iniciamos o contributing já está concluído. Ficou bem bacana.
Se quiseres pega apenas a parte da seção e o css para agilizar.
_______
<<EMAIL>><NAME>, PMP
Administrador, PMP, MBA
[image: ](48) 9914-0400
[image: Visualizar meu perfil no LinkedIn]José Ferrão |
<http://br.linkedin.com/in/josesferrao>[image: @josesferrao] @josesferrao
<https://twitter.com/josesferrao>
2016-04-01 18:47 GMT-03:00 <NAME> <<EMAIL>>:
> Paulo, está faltando no novo layout? Tem algo que eu possa ajudar?
>
> _______
>
> <<EMAIL>><NAME>, PMP
> Administrador, PMP, MBA
> [image: ](48) 9914-0400
> [image: Visualizar meu perfil no LinkedIn]<NAME> |
> <http://br.linkedin.com/in/josesferrao>[image: @josesferrao] @josesferrao
> <https://twitter.com/josesferrao>
>
> |
intelliboard/intelliboard | 677317977 | Title: changes for unique class to resolve conflict with other plugins
Question:
username_0: Intelliboard plugin has a page setup.php, over there it's using a class called slider.
This class is conflicting with another community plugin block_slider. The conflict is happening from the CSS used for the intelliboard plugin. It's breaking the dashboard with blocks whoever using that block_slider plugin or any other plugin has got any class named as slider. I believe it is always recommended to use class name specific to the plugin.
I made the fix and pushed and made a pull request #38
The changes I made:
1. Made the change in the setup.mustache from slider to intelliboard-slider.
2. Made changes to the style.css from slider to intelliboard-slider.
Hope it helps. |
ethereum/remix-project | 863128395 | Title: Support Solidity 0.3.6
Question:
username_0: I tried to compile some legacy contracts in Remix and noticed that compiler versions >=0.4.0 worked, but 0.3.6 did not. The loader just kept spinning.
The logs reveal the following:

I think the problem is that since 0.3.6 loads the wasm version it should also use the worker. Probably remix uses the worker for >=0.4.0 compilers, and not for older ones. I suggest to change that rule to >=0.3.6.<issue_closed>
Status: Issue closed |
xanzy/go-gitlab | 1057124137 | Title: Broken build
Question:
username_0: Hi!
This commit breaks build
https://github.com/xanzy/go-gitlab/commit/8f8fc8c529f769832cbfd6c22ce2df4267bf37ca#diff-044b5a63fdba539bf87d25040d91912aec34173e76af7b9e0f40d14da57edebeR670
```
../../xanzy/go-gitlab/gitlab.go:670:26: req.Request.Header.Values undefined (type http.Header has no field or method Values)
```
Answers:
username_1: I think I need some more input as I cannot reproduce that error.
username_0: This reproduces on Go 1.13 which I found out is no longer a supported version ([docs](https://golang.org/doc/devel/release#policy)). Sorry! Works on Go 1.16
Status: Issue closed
|
marcguetg/h5particle | 274982892 | Title: Is meshesPath and particlesPath needed?
Question:
username_0: Right now **meshesPath** and **particlePath** are parameters that are to be set to the relative path to the meshes and particle data respectively. For example: **meshesPath = fields/**
For simplicities sake, I propose eliminating these two parameters and simply standardize the relative path to the mesh be **fields/** and the relative path to the particles be **particles/**
Comments?
Answers:
username_1: This seems reasonable to me ; I am not sure anymore why we decided to have these paths customizable... @username_2: Any idea?
@username_0 Note that, for this discussion and similar ones on the openPMD standard, it is probably more appropriate to use the Issues of the [openPMD-standard repository](https://github.com/openPMD/openPMD-standard). Would that be fine with you?
username_0: @username_1: I'm a little confused here. Are you proposing moving all of the issues in this repository over to openPMD? The original idea was to simplify life so that people who were only interested in developing the version 2 standard here would not have to wade through the other issues in the openPMD issues forum and vice versa.
username_1: Yes, thanks for reminding me this. I now remember that we agreed to opening issues on the present repository.
However, it seems that there are two "version 2" of the standard that are being developed: one on the [wiki of the present repository](https://github.com/marcguetg/h5particle/wiki/openPMD-Standard-V2), and one on the official openPMD repository: https://github.com/openPMD/openPMD-standard/milestone/4. Thus it would probably be good to centralize this, at some point.
That being said, since it is explicitly said in the wiki that it is a first pass as a basis for discussions, I guess that the aim of the v2 in the wiki is to establish a "wishlist" for openPMD in order for it to be used for particle accelerator codes (is that a correct statement?) In that case, I agree you should wait for this first pass to be complete, before merging/reconciling this with the v2 of the official openPMD repository.
username_0: @username_1: In fact I did not realize that there was a V2 being developed at the openPMD repository until you just mentioned it. When I get finished with the draft on this site (I'm hoping within the week) we can take a look and decide the best course of action as how to merge.
The aim of the h5particle wiki here is that I thought it would facilitate the discussion if I presented a concrete draft. Nothing is written in stone. Merging/reconciling can actually begin now if you want. Just be aware that at this point I am in the process of making substantial changes so what you see could look a bit jumbled.
Please be aware that for the purposes of editing, I am keeping everything in one document. My thought was that at the end of finalizing the V2 standard, we can split off anything into an extension as needed.
Another thing is timing. I noticed that there were open issues on the openPMD site that were over 2 years old. I am hoping that the process here can be completed on a time scale of months...
username_2: This seems to be a misunderstanding. "Issues" in the openPMD standard are not "defects" that nobody cares about but ideas where to generalize and move next to. The direction is need and community driven: We always were and are well aware we can not cover all generalized descriptions of particle and mesh formats at once and decided to progress step by step. If you see open issues or missing issues in [the standard](https://github.com/openPMD/openPMD-standard) just jump into the discussion and tell us it is relevant to you (and maybe even how you would address it). I must admit I learned of your repo & plans these days since I saw now issue or report of you on the mainline :)
username_0: @username_2: I don't want to fork the openPMD specification. Ideally my thought was that if we all work together we could come up with an openPMD Version 2 standard that meet the needs of accelerator physicists and that would be official. So please take a look at the draft for the V2 standard on the marcguetg/h5particle wiki and tell me what you think.
username_2: The wiki work looks great! 👍
Can we try together to summarize the changes a bit?
A first view looks to me like we could do the following:
- `Accelerator` openPMD extension:
- let's have the domain-specific naming such as subatomic particles and fixed names of meshes there as well instead of in the base standard (e.g. we have collegues using openPMD for particle based image reconstruction techniques in biology)
- base standard:
- support for lattice coordinates: one could try to map this into a `geometry` instead of only defining it into an extension, but inside the extension it is fine as well with the additional attributes you added!
- additions to `/`: look good so far! we also have some more in the pipe
- `unitSymbol`: we could add labels for how to label things, but this can also go in an extension for now; the idea so far was to by symbol agnostic and just parse `unitDimension` - a domain-specific viz can then just map such (and if necessary also take into account the name of a record) to a symbol with ease
Can you elaborate a bit on the concept behind the particle map? :)
username_0: The idea here was that this does **not** affect any calculations or conversions. It simply is a way to 1) make things clearer to a human viewing the file and 2) Providing a string that a program displaying numbers from the file can also use in the display. Again as an aid for human understanding. The thing is that if the data is originally measured in, say, "miles/hour" there is no way for a program to reconstruct that.
username_2: yes, I understand and like that. Would a general optional "label" as we do it on a few places already do? E.g. a label for the symbol of each record and a label for the unit of each record component?
username_0: "label" would work but "unitsOfMeasurement" I thought was a better descriptor. If you think this is too confusing with "unitDimension" how about something like "unitsDescriptorString"?
username_0: Sorry I missed this question first time around. I'm confused by what you mean by "particle map". The word "map" is not used in the draft I have been working on.
username_2: Ah ok, I was just reading the [`initial thought` section](https://github.com/marcguetg/h5particle/wiki/Initial-Thoughts#particle-map-item) and saw it there
username_2: I think there is a conceptual misunderstanding what base and extension in openPMD is. Just because something is in an extension does not mean it is not standardized or in a domain very important.
The openPMD *base standard* is (scientific) domain agnostic. It describes data and gives enough information to exchange and (dumbly) visualize data without domain knowledge. *Extension* add domain-specific needs, such as namings of records for the domain scientist, methods used in a simulation, interpretations of values in a certain domain's scope, additional attributes, required records, etc.
Extension can in principle also be combined, e.g. one could do a plasma extension and a hydro and a PIC extension and use Plasma + ED-PIC or Plasma + Hydro together with the base standard.
In addition, both the base standard and extensions are not disallowing to add even further records, folders, attributes, etc. if e.g. a specific application needs it but that aspect is not worth standardizing. Examples for that are for example a GPU's random-number matrix generator state or a unique particle-id generator state, etc. that are needed for restarts of a simulation but are not relevant for data exchange or data processing.
username_2: I see your point and initially we intentionally went for "iteration" to avoid confusion with "time steps" which are not necessarily lab time. I migrated this proposal to https://github.com/openPMD/openPMD-standard/issues/148 for discussion.
username_2: Checkpointing (of a simulation) would be dumping and re-storing of its full internal state. In detail this is a very domain- and even application-specific task and not defined how it needs to be done in the openPMD standard. For particle-in-cell codes, we are thinking about [unifying it to some extend](https://github.com/openPMD/openPMD-standard/milestone/3) in future versions of the `ED-PIC` extension. This is relevant for tightly coupled simulations.
username_0: Exactly! That is why, for example, the standard mandates that the names "x", "y", and "z" be used for Cartesian field components. Since the species name is part of the data, it needs to be standardize too. For example, if someone creates a data file using "e+" for the name of electrons, this will not be portable. If the species name where not mandated in the base standard I would agree with you but it is mandated in the base. And ensuring portability is much more important than being agnostic.
username_2: I am a bit confused and can not fully follow why a file that has "base + Accelerator" openPMD markup would not be fully portable. A data reader would check that the extension "Accelerator" is set as a requirement and with that "electrons" as a naming are suddenly known.
username_0: @username_2
Yes the "Accelerator" extension would be portable but the base standard alone is not portable without a species naming convention. I suspect that the only reason why you have not run into trouble before now is that specifying the mass and charge has been good enough. But this is not good enough for a wide range of problems.
My belief is that putting in a species naming convention makes this standard more useful and there is really no drawback in putting it in. Why sacrifice usefulness over ideology?
Since for me personally this will not be an issue (I will always be using the "Accelerator" extension), and you really want agnosticism over portability, I will not argue the point further.
username_3: @username_2 @username_0 How about then having a "Physics" extension - or something generic like that -which would contain the definitions for particles and other data that will be useful across many domains?
username_0: Rereading my draft V2 standard + Extension I realize that nowhere is there a slot to specify the species! This will have to be fixed. The question is this: If the base standard has a species attribute (or mandates that the species is encoded in the directory path like it is in the present version 1), but does not specify what the species names should be, then files created that only use the base standard, but not any extensions, are not portable. And for me portability is of paramount importance.
username_2: I think there is a misunderstanding about the modularity concept in openPMD base standard + extensions that we might be able to better clarify in a VC. Both the base standard and the extensions are part of the openPMD standard, so all aspects are standardized if one writes a new extension for e.g. elementary particle names.
I would suggest we hold this specific sub question, which is also already way out of scope of the original thread question, and discuss this in a follow up. Personally, I would have time for a VC e.g. the week after next week since we have an important lab evaluation during the next week (today was a German holiday for me and Friday you probably all have thanksgiving).
username_3: @username_2 @username_0 I agree that a VC will be an efficient way to clarify things.
username_0: @username_2 @username_3 Sounds good to me. I will be available the week after next. -- Cheers, David
username_2: mass and charge are both only defined in the `ED-PIC` extension and unknown in the base standard.
username_0: Oops! My mistake. Thanks for the correction.
username_1: @username_2 @username_3 @username_0 See your email ; I suggested a date. |
code-for-hamamatsu/covid19 | 589828555 | Title: meta
Question:
username_0: ## 起こっている問題 / The Problem
- xxxx(できるだけ簡潔に/as concise as possible)
## スクリーンショット / Screenshot
<!-- バグであればdeveloper toolからコンソールも合わせて添付 -->
<!-- If it's a bug, attach a screenshot of the developer tool console -->
## 期待する見せ方・挙動 / Expected Behavior
- xxxx(できるだけ簡潔に/as concise as possible)
## 起こっている問題の再現手段 / Steps to Reproduce
1. xxx
2. xxx
3. xxx
## 動作環境・ブラウザ / Environment
- macOS / Windows / Linux / iOS / Android
- Chrome / Safari / Firefox / Edge / Internet Explorer<issue_closed>
Status: Issue closed |
intercom/intercom-php | 178583192 | Title: getUsers with Id
Question:
username_0: ## Version info
- intercom-php version: 3.0
- PHP version: 7.0
## Expected behavior
I try to get the complete user from a hook.
$conversation = $intercomClient->conversations->getConversation($request->get('data')['item']['id']);
dump($conversation);
$user = $intercomClient->users->getUsers(['id' => $conversation->user->id]);
dump($user);
`$user` should contain only one user
## Actual behavior
Even if my userId is good (https://app.intercom.io/a/apps/{myappid}/users/{myUserId}/all-conversations exist) `$user` contains a user.list with all my users. Not only the one I'm looking for
## Logs

Answers:
username_1: Hi @username_0 ,
Can you please try replacing
`$user = $intercomClient->users->getUsers(['id' => $conversation->user->id]);`
with
`$user = $intercomClient->users->getUser($conversation->user->id);`
I hope it will give you your desired result i.e. the details of a single user.
This method `IntercomUsers::getUser()` is not part of any release yet, so, you will have to use latest code from the `master` branch.
username_0: Thanks, I've just saw #163
But the result of getUsers is still not logic right ?
username_1: Is it documented somewhere in the official documentation that you can get details of a single user by id via getUsers() this way? If not, then I think getUsers() is not intended to be used for this purpose. This functionality was simply missing from Intercom PHP SDK before. However, someone from the official team can confirm further :-)
username_0: Ok, in the PHP SDK 1.5 it was working with
$user = $intercomClient->getUser(['id' => $id]);
So same parameters logic of actual `getUsers` but called with `getUser`
So the `getUser` added with #163 will be fine for me. Thanks.
Status: Issue closed
|
pyeve/cerberus | 346921448 | Title: is JSONErrorHandler implemented?
Question:
username_0: ```
And I have asked my fried to try, the same result.
Could anyone please give some advice?
Do i need a higer version of cerberus? I am using 1.2
here is the code in E:\Python\Python36\lib\site-packages\cerberus_collections_init_.py
from cerberus.utils import validator_factory # noqa: F401
from cerberus_collections.error_handlers import * # noqa: F401, F403
from cerberus_collections.versions import __version__ # noqa: F401
VanillaValidator = Validator = cerberus.Validator
Status: Issue closed
Answers:
username_1: you have already opended that issue in the proper repository.
please keep in mind that we do all this in our spare time, raising questions in various places won't help and just costs more time. you'd be better off to take your time to debug the issue. |
cloudant/objective-cloudant | 137079435 | Title: CDTHTTPSessionCookie interceptor should run all other interceptors on it's requests
Question:
username_0: CDTHTTPSessionCookie interceptors should run all the other interceptors on the request's its makes. This would enable logging of HTTP requests made through out library code easier, and it will allow users to add in other functionality they require to those requests.<issue_closed>
Status: Issue closed |
w3ctag/design-reviews | 678690160 | Title: Web page settings to save battery
Question:
username_0: Saluton TAG!
I'm requesting a TAG review of battery-savings.
Saving battery or CPU is important for computing devices that are not connected to an electrical source, or to better share common CPU & GPU resources among running processes. Web sites that are known to have high CPU or battery costs may want to request that the UA optimize for CPU or battery, even if the user has not requested it.
Explainer : https://github.com/username_3/battery-savings/blob/master/explainer.md
Primary contacts (and their relationship to the specification):
- [<NAME>] ([username_3]), [Chromium]
- [<NAME>] ([username_0]), [Chromium]
GitHub repo (if you prefer feedback filed there): https://github.com/username_3/battery-savings/blob/master/explainer.md
Organization/project driving the design: Chromium
External status/issue trackers for this feature (publicly visible, e.g. Chrome Status): https://chromestatus.com/feature/5653874167775232
We'd prefer the TAG provide feedback as (please delete all but the desired option):
🐛 open issues in our GitHub repo for **each point of feedback**
Answers:
username_1: Hi, @username_2 and I took a look at this during the TAG's virtual F2F this week.
It's not clear to me that there's much value in an explicit opt-in from a site for throttling. Most sites won't adopt, so user agents which engage in throttling are going to do so regardless of whether or not the site opts in. Maybe the idea is that, after a site opts in, the user agent could tell the site that it's being throttled in various ways, but it's unclear what the site could then be expected to usefuly do in that scenario. In the `if (throttled) then { ... } else { ... }` case, it seems authors are just as likely to make things worse in the then clause, since there's a lot of misinformation out there about how to optimize sites. So then user agents would stop sending the signal.
username_2: But this is a global thing so it would affect all tasks. It feels like the postTask scheduling proposal might fit this use-cases better.
username_3: Yes, but putting the thread on a "little core" would have a much greater impact on power efficiency, be simpler to implement for browsers and developers, and have more predictable performance.
username_4: Echoing @username_1's thoughts, I also think an opt-in would end up ignored by browsers down the line. Given that the vast majority of sites will not declare anything, browsers need to be able to treat this as ok to throttle. Throttling on low battery is the common case, and declaring that you *don't* want it for some reason is the exception, so the Web platform feature should reflect that to be useful. Have you considered an opt-out instead, where websites can declare that it is important *not* to throttle certain operations?
I'm missing a more extensive list of operations beyond the two used as examples in the explainer (reduced framerate and reduced script speed). Are these the only ones?
The explainer mentions a `reduced-framerate` media query (as `@reduced-framerate`, which I assume is a typo? A media query would be `@media (reduced-framerate)`). This seems a bit too low-level and lacks flexibility. For example, if UAs start conserving battery in different ways in the future, does the MQ stop applying? Are there use cases where you want to handle specific throttling operations differently, or do authors really just need to know that battery savings mode is active? Perhaps a more general `@media (energy-savings)` media query would be more broadly useful?
username_5: @username_0 @username_3 we are just picking this up again in our virtual f2f. We have some additional comments which Ken will write in. Additionally can you let us know the current status on this on your end? There seems to be some privacy & security considerations of this idea that need to be spelled out. Could I ask you to fill out the [privacy & security self check](https://www.w3.org/TR/security-privacy-questionnaire/) and include a privacy & security considerations section in the explainer?
username_2: @username_5, @username_4, @rhiaro and me looked at this in our TAG Virtual F2F breakout
We believe we understand the video conferencing use-case now. Some sites get very quick user interaction and probably won't like to be throttled (A), but other sites like video conf would knows that their users have another hour of meeting left and they are low on battery so they may want to optimize for the battery to last as long as possible and hopefully until the meeting is over (B).
For (A) we think an opt in or opt out won't work, but instead a priority might, like you are going to get throttled so what matters the most to you, high framerate, background tasks, foreground tasks? Then the browser can consider those hints
For (B) an opt in might work, but we think an event that low battery mode is active makes more sense, then the vidconf site can lower video resolution etc and turn off live captions and many other things to save battery.
username_3: There might be others, but I don't know of any right now.
username_3: Consider this situation: User enters a video conferencing site with battery at 100%, and stays on the call for multiple hours. The user would like to extend the length of the call. The best way to do so may be to save battery life along the entire timeline of the call, not just near the end when the battery is getting low.
I don't think the event you suggested would work for this case.
username_3: This TAG review was of the "early review" variety that comes at the Intent to Implement / Intent to Experiment stage of the Blink feature process (maybe we need to revise our process to indicate that more clearly?).
If and when we proceed further with the standards process, I agree there need to be reviews of any privacy and/or security concerns.
If/when we want to proceed further with this proposal, yes it'd be WICG + WebPerfWG I think.
username_5: Ok Chris thanks for that - we are going to close this for now as I think you have our feedback. Please open a new issue when you want to do a full review.
Status: Issue closed
|
CiscoUcs/KUBaM | 1169059496 | Title: Create ISO but not bootable
Question:
username_0: My iso for 5.7 gets created in the /home/%user%/kubam folder but when I mount it to the iso either via vmedia or just through a standard KVM connection it's not bootable. Am I supposed to do something special to make that iso bootable. If I mount it in the OS it looks the same as the normal esx 6.7 iso other than I see the new KS.CFG file I expected to see. |
pytorch/text | 213775113 | Title: error when x is a string instead of a unicode
Question:
username_0: [this line](https://github.com/pytorch/text/blob/a4d597e194e933b384fbce62a13600c4adb6fbd7/torchtext/data.py#L138]), six.text_type.lower requires the input x to be unicode, so when x is a string, this will complains with an error. what about changing this line to be:
```
x = Pipeline(six.text_type.lower)(unicode(x))
```
any ideas
Answers:
username_1: If that works for you and doesn’t break Py3, that sounds good. Sorry that I didn’t actually fix the Python2 problem before.
username_0: Yes, it fixes my problem. And i donot know whether it will break py3.
username_1: Fixed in #21.
Status: Issue closed
username_1: [this line](https://github.com/pytorch/text/blob/a4d597e194e933b384fbce62a13600c4adb6fbd7/torchtext/data.py#L138]), six.text_type.lower requires the input x to be of type unicode, so when x is a string, this will complain with an error. what about changing this line to be:
```
x = Pipeline(six.text_type.lower)(unicode(x))
```
any ideas
Status: Issue closed
|
numery009/SpringBoot-SpringCloud-Docker | 423631441 | Title: Should we save netflix-eureka-naming-server config file in remote repository?
Question:
username_0: I noticed that some microservices project on Github save their config files like `application.properties` on a remote repository and fetch data through **Config Server**, even **Naming-Server**.
So I'm curious about why don't you do that?
Answers:
username_1: I did it on customer-service-mysql micro-services. Fetching the database configuration from the config server.
https://github.com/username_1/Spring-Boot-And-Spring-Cloud/blob/master/customer-service-mysql/src/main/resources/bootstrap.properties
These are the sample micro-services. I have used varieties of Spring-Cloud feature on all the services in various way.
But in the production level we should use Config Server to fetch the properties of our micro-services.
Status: Issue closed
|
jitsi/jitsi-meet | 210965405 | Title: JS Error:Rejecting session-initiate from non-focus user
Question:
username_0: ```
[JitsiMeetJS.js] <Object.getGlobalOnErrorHandler>: UnhandledError: null Script: null Line: null Column: null StackTrace: Error: Rejecting session-initiate from non-focus user: <EMAIL>/focus
[JitsiConference.js] <o.onIncomingCall>: Rejecting session-initiate from non-focus user: <EMAIL>/focus
```
I am using the package from the debian unstable repository, but I get the same error on stable.
Please ask if you need more information.
Answers:
username_1: Did you do any custom changes to the installation? Are you using Nginx / Apache or the builtin server? What browser are you using?
I tried to reproduce it on a freshly installed VM with latest Jitsi Meet unstable, to no avail.
Can you provide full console logs? (use Gist or PasteBin please).
username_0: My setup might not be your usual setup, but I don't know if it's related to this particular error.
`hostA` that has prosody (0.9.x) running `example.com` and `conference.example.com`, on Debian Jessie. `hostB` on Stretch with jitsi related stuff and another prosody (0.9.x) running `meet.example.com`. I also have `focus.meet.example.com` and `videobridge.meet.example.com` that points to hostB.
I have an nginx in front to serve meet (among others), and I emptied videobridge/sip-communicator.properties so jetty doesn't try to bind 443.
I created a local user for focus (<EMAIL>), and added it as admin on conference.example.com.
Here is the config for the whole thing: https://paste.bouah.net/L0z0VEp41-H/
Logs of jicofo: https://paste.bouah.net/QCGlrlexnvh
Logs of jvb: https://paste.bouah.net/1mdiIdplpOH
We both use Chromium version 56.0.2924.87 (64-bit) to test.
username_1: have you adjusted `/etc/jitsi/jicofo/config` to match the focus user you created? Look at `JICOFO_AUTH_DOMAIN`, `JICOFO_AUTH_USER` and `JICOFO_AUTH_PASSWORD` options.
username_0: I created a user `<EMAIL>`, as specified in the config I linked. I can see it connect to the muc on conference.example.com. I didn't create it on example.com because we're using some fancy auth method I need to fix first, but it shouldn't be a problem right?
username_0: I created a user `<EMAIL>`, as specified in the config I linked above. I can see it connect to the muc on conference.example.com. I didn't create it on example.com because we're using some fancy auth method I need to fix first, but it shouldn't be a problem right?
username_1: @username_2 maybe you can help here?
username_2: I don't really get the configuration here, and something is messed up in the configuration I think.
The focus user needs to be admin for its domain: meet.example.com. And from the config it feels like the meet client is trying to connect to different server/domain ... so the meet client needs to connect to the same server where jicofo is connected, and the focus user needs to be admin in order to grand ownership to the first user that enters in a room and will send session initiate to the users connected to that room. |
maximodleon/sabichoso | 345572360 | Title: Add country code mapping
Question:
username_0: Add mapping from 2-code ISO country code to full name. Data --> https://pkgstore.datahub.io/core/country-list/data_json/data/8c458f2d15d9f2119654b29ede6e45b8/data_json.json
Status: Issue closed
Answers:
username_0: :tada: This issue has been resolved in version 1.3.1 :tada:
The release is available on [GitHub release](https://github.com/username_0/sabichoso/releases/tag/v1.3.1)
Your **[semantic-release](https://github.com/semantic-release/semantic-release)** bot :package::rocket: |
HewlettPackard/POSH-HPEOneView | 1165519297 | Title: NEtwork set to Logical Interconnect Group
Question:
username_0: Hi,
Whats the process for adding a new networkset and the contained networks to alogical interconnect group / uplink set?
is there an example
Answers:
username_0: think I've got something that works, not 100% sure on whether its the best way - would be nice if I wasn't rushing to get this lot in and working!
$interconnectGroupName = "ig1"
$networkSetName = "ns1"
```powershell
$lig = Get-OVLogicalInterconnectGroup -Name $interconnectGroupName
$networkSet = get-ovNetworkSet -Name $networkSetName
if ($lig.UplinkSets.networkSetUris -notcontains $networkSet.uri) {
Write-Verbose "NetworkSet not present on lig uplink set, Adding"
$lig | Set-OvUplinkSet -UplinkSetName $uplinkSetName -AddNetworkSet $networkSet | Wait-OVTaskComplete | Out-Null
Get-OVLogicalInterconnect | Where-Object { $_.logicalInterconnectGroupUri -eq $lig.uri } | Update-OVLogicalInterconnect | Wait-OVTaskComplete | Out-Null
} else {
Write-Verbose "Networkset already present on lig uplink set"
}
``` |
evennia/evennia | 200796555 | Title: Feature Request: Move inputfunc splitter from the sessionhandler to the serversession.
Question:
username_0: #### Brief summary of issue / Description of requested feature:
Currently the session.data_in() hook is not used by ServerSession. Instead the inputfunc splitter is located on the SessionHandler. Moving to the ServerSession will provide a single easily modifiable location for handling data in and out. Discussed on IRC:
```
<username_1> As discussed, I think the inputfunc splitter could probably be moved from the sessionhandler to the serversession to give more control for those wanting to operate on all input.
<username_1> The reason I didn't do that originally is because the splitter is such a critical part of the message path that users should not need to modify it. But having one single place to see input coming in is a good idea too.
<username_1> Since one's not likely to override the sessionhandler but instead the serversession.
```<issue_closed>
Status: Issue closed |
amcharts/amcharts4 | 1043586480 | Title: Ready event does not fire when lazy loading the library
Question:
username_0: **Bug description**
I'm lazy loading the library in an Angular app, I noticed that the `ready` event never fires, `inited` event on the other hand works
**Environment (if applicable)**
- amCharts 4.10.22
- Angular 12
Answers:
username_1: Could you elaborate what you mean by lazy-loading?
Do you mean that you import amCharts on demand? Or lazy-load the chart itself using one of the methods provided?
username_0: I mean lazy load the chart library as mentioned here https://www.amcharts.com/docs/v4/tutorials/lazy-loading-amcharts-modules/
username_1: Which of the lazy-loading methods are you using? There are quite a few.
username_1: Thanks. Would you be able to share your chart code so that we can test it?
username_0: Sure!
Chart service
```ts
import { Injectable } from '@angular/core';
import { BehaviorSubject, Observable } from 'rxjs';
import { filter, take } from 'rxjs/operators';
import { AmChartsLib } from './graphs.model';
@Injectable({
providedIn: 'root'
})
export class AmCharts {
// Stream that emits when AmCharts library is loaded and ready to use
private readonly ready = new BehaviorSubject<AmChartsLib>(null);
readonly ready$: Observable<AmChartsLib> = this.ready.asObservable().pipe(
filter((lib: AmChartsLib) => !!lib?.am4core && !!lib?.am4charts),
take(1)
);
constructor() {
Promise.all([
import('@amcharts/amcharts4/core'),
import('@amcharts/amcharts4/charts'),
import('@amcharts/amcharts4/themes/animated'),
import('@amcharts/amcharts4/themes/material')
// tslint:disable-next-line:variable-name no-shadowed-variable
]).then(([am4core, am4charts, am4themes_animated, am4themes_material]: any) => {
// Set up chart themes
am4core.useTheme(am4themes_material.default);
am4core.useTheme(am4themes_animated.default);
// Dispose when routing away
am4core.options.autoDispose = true;
// Emit the libraries as ready to use
this.ready.next({ am4core, am4charts });
}).catch((e: Error) => {
console.error('Error when creating chart', e);
});
}
}
```
Chart component
```ts
import {
Component,
Input,
Output,
EventEmitter,
ViewChild,
ElementRef,
OnInit,
OnChanges,
OnDestroy,
SimpleChanges,
[Truncated]
// Create series
private createSeries({ am4core, am4charts }: AmChartsLib, field: string, name: string, color: string): void {
const series = this.chart.series.push(new am4charts.ColumnSeries());
series.dataFields.valueY = field;
series.dataFields.dateX = 'date';
series.name = name;
series.wheelable = true;
series.columns.template.tooltipText = '{name}: [bold]{valueY}[/]';
series.stacked = true;
series.columns.template.fill = am4core.color(color);
series.columns.template.stroke = am4core.color(color);
series.columns.template.width = am4core.percent(50);
}
private resizeLegend(): void {
this.legendEl.nativeElement.style.height = this.chart.legend.contentHeight + 'px';
}
}
```
username_1: Thanks.
OK, so which ready event does not fire? The one `this.amCharts.ready$.subscribe()` or `this.chart.events.on('ready', ...)`;
username_0: `this.chart.events.on('ready')` is the one not firing, `this.amCharts.ready$` is a service that emits when the library is loaded
username_2: @username_0 I haven't tested it, but I think the problem is that the chart is already ready, so in that case it won't fire the `ready` event. So you should try something like this:
```ts
if (this.chart.isReady()) {
this.ready.emit(this.chart);
} else {
this.chart.events.on('ready', (e: { target: Am4Charts.XYChart }) => this.ready.emit(e.target));
}
```
username_0: hmmm ok, I using `init` instead
Status: Issue closed
|
dosinabox/g2_classic | 527892867 | Title: Руна и Рунический стол
Question:
username_0: если не изготовить руну на руническом столе, то сама руна пропадает.
https://youtu.be/P0vWPcTpFZ0?t=3775
Answers:
username_1: Такое происходит при неудачной попытке изготовления только двух рун: глыбы льда и уничтожения нежити. Исправлю.
username_0: Отлично
Status: Issue closed
username_1: исправлено |
RayBenefield/dev-xp | 764344938 | Title: Reverse the order of transmutation merging
Question:
username_0: ## Expected Behavior
When printing the json for `KLI` the object is printed in reverse, newest keys are at the top of the object. So we can switch the merging in `transmutation` to make new keys show up at the bottom of the object instead.<issue_closed>
Status: Issue closed |
dbthinking/myblog | 681966731 | Title: 《随机漫步的傻瓜》读后感:概率、期望与人生
Question:
username_0: 我们大多数人已经跳出了平均值与中位数的误区,但世界依然有很多陷阱。
最近有好几个朋友咨询我买期房的事情,我正好把所读所思所想聚拢到一起。房价的长远期望大致是横盘或缓慢下跌,从供需角度、从产业支柱等角度分析,各种数据都预测全国房价下跌的必然的,那些掌握长期资源、产业优势的城市风险可能会低一些,虽然政府不敢承担房价崩溃造成动荡与不安,暴跌在理论上是小概率事件,但不代表房价不会因为不可预期因素而崩盘。
地球表面的平均温度是15摄氏度,最低气温是零下89摄氏度,最高气温57摄氏度,极端气温的地方可能不多,但并不是所有的地方都能让人类生存。我们习惯性只谈期望而忽略高度差,也就是预期的最好结果与最坏结果是什么,而我们又能否承受。房价一旦暴跌一半,凭空多出来几十上百万的债务,普通人这一辈子基本就被摧毁了;就算涨一半,赚了几十几百万,对于个体来说,其带来的收获与意义是什么,我们把两种可能都想清楚了吗?
任何在单一决策上 ALL IN,都是不理智的,投资有亏有盈实属正常,人祸易避,天灾难躲,但输光一切而被迫离场,就难有翻身的机会了。一个优秀的投资者,一个成功的赌徒,要禁得起波动与震荡,在长期的博弈游戏中,持续玩下去比偶尔赚大钱更重要。现实媒体总是过度聚焦于稀有事件,不可一世的赢家与一败涂地的输家,让我们过于羡慕成功妄想脱离平凡,不惜拿毁掉自己的人生作为代价。导致很多人不是在做投资,而只是想投机暴富,我们痛惜错过了房地产热、错过了股市暴涨、错过了比特币崛起等诸多机会,等普通人作为投资者进场的时候,基本只能充当接盘侠与韭菜的角色。
那些吸引眼球的成功人士,大概率也是因为运气好,在这几十亿人口的大样本中,在这几千年的时间周期中,超级稀有的偶然事件会发生就不奇怪了,样本高度差也会越来越大,虽然还没有猴子打出莎士比亚的著作,但人类中彩票的事情一直在发生。
世界卫生组织2015年数据(假如这是今年的数据),中国人口预期寿命是76岁,请问你的预期寿命(期望值)是多少,答案会更高一点,因为你已然活了几十年,大样本中的短寿者已经被排除。再到具体的情况上,女性又比男性高,城镇的又比农村高,能续命的又比普通人高。回到成功人士的话题,跟比尔盖茨一样是程序员的有很多,比他水平相似或更好的技术宅也大有人在,但纯技术程序员荣登世界首富的神话基本不会再出现。个体可能拥有与成功人士相似的硬件素质,但时代大背景已经不一样了。人类顶尖的权势梦想,并非努力就能实现。
一个即将到来的婴儿,似乎可以成为他想成为的任何人,似乎可以做到他想做的任何事,似乎可以得到他想要的任何东西。但一旦他被真实孕育,他身处的环境、他遭遇的人们,他经历的事物,只会让期望越来越趋于收敛、逼近现实。这就是有条件期望与无条件期望的区别与无奈。
如果留意此文措辞,可能发现其中充满了不确定的用词与思想。对于未发生的事情,我不知将会如何;对于已发生的事情,我未必理解其根源,如果我希望趋近于真实,用概率表达似乎更恰当。我以前经常困惑于为什么在化学课上,根据观察就能推断出哪是哪些物质发生了哪些化学反应,没人担心可能会存在未知的我们还没想到的情况吗?
观察自身与周边的人们,双重思想是一种常见现象,我当初把网名改为doublethinking,一开始是为了讽刺这种现象,也是自我提醒,但越是琢磨越反而能理性对待人们的矛盾思想,对于没有定论、无法定论的问题,为什么不能允许双重思想(这可能已经脱离奥威尔创造这个词的本意了,维基百科的解释是,同时接受两种相违背的信念的行为)。相比于人类目前离开氧气无法存活的确定性问题,对于那些不确定问题,比如宇宙是将不断膨胀还是终究坍缩,比如人类社会是走向民主还是独裁,在概率上同时接受两种观念,即使是相违背的,也比盲信一种可能更具备科学求真的态度。 |
klausahrenberg/WThermostatBeca | 988480583 | Title: BHT-002-GCLW DLX-LH01 board Unknown MCU commands
Question:
username_0: Hello,
I am reporting Uknown MCU commands for BHT-002-GCLW with DLX-LH01 board:
{"error":"Unknown MCU command: 55 aa 03 07 00 74 11 00 00 70 01 68 00 c8 02 b2 00 96 03 2a 00 96 04 1a 00 c8 01 68 00 c8 02 b2 00 96 03 2a 00 96 04 1a 00 c8 01 68 00 c8 02 b2 00 96 03 2a 00 96 04 1a 00 c8 01 68 00 c8 02 b2 00 96 03 2a 00 96 04 1a 00 c8 01 68 00 c8 02 b2 00 96 03 2a 00 96 04 1a 00 c8 01 68 00 c8 02 b2 00 96 03 2a 00 96 04 1a 00 c8 01 68 00 c8 02 b2 00 96 03 2a"}
{"error":"Unknown MCU command: 55 aa 03 07 00"}
{"error":"Unknown MCU command: 55 aa 03 07 00 08 0c 02"}
They are logged quite frequently almost every second.
Answers:
username_0: release 1.20n
username_1: Same here. Just upgraded to 1.19 and I'm getting a lot of these messages and I'm unable to control it via Home assistant
username_2: I have the same, and on top it seems that the topics used for discovery are not the same as MQTT HA used for auto-discovery. Hence no way to make it work... |
naser44/1 | 104778039 | Title: أبشر ياسوري أفقر دولة بالعالم الاسلامي ستتقاسم معك لقمة الحياة
Question:
username_0: <a href="http://ift.tt/1KN4Ity">أبشر ياسوري أفقر دولة بالعالم الاسلامي ستتقاسم معك لقمة الحياة</a> |
rust-lang/rust | 344554860 | Title: Concatenating bytestrings fails; rustc suggests non-existent `bconcat!()`
Question:
username_0: `concat!()` doesn't work on bytestrings (why?). When trying it, it fails with:
```
1 | const FOO: &[u8] = concat!("foo", "bar");
| ^^^^^^^^^^^^^^^^^^^^^
| |
| expected slice, found str
| help: consider adding a leading `b`: `bconcat!("foo", "bar")`
|
= note: expected type `&'static [u8]`
found type `&'static str`
```
`bconcat!()` sounds interesting, but it isn't documented, and the compiler doesn't accept it:
```
error: cannot find macro `bconcat!` in this scope
--> src/main.rs:1:20
|
1 | const FOO: &[u8] = bconcat!("foo", "bar");
| ^^^^^^^ help: you could try the macro: `concat`
```
I think the real fix here is that `bconcat!()` should exist.
Answers:
username_0: Heh indeed. I originally tried `concat!(b"foo", b"bar")`, but:
```
error: cannot concatenate a byte string literal
--> src/main.rs:1:28
|
1 | const FOO: &[u8] = concat!(b"foo", b"bar");
| ^^^^^^
error: cannot concatenate a byte string literal
--> src/main.rs:1:36
|
1 | const FOO: &[u8] = concat!(b"foo", b"bar");
| ^^^^^^
error: aborting due to 2 previous errors
```
I can't think of a reason why `concat` shouldn't support this though. Is there something I'm missing?
username_1: cc @username_2 #51978 about the `bconcat` diagnostic issue.
username_2: I'll take a look and fix the incorrect suggestion.
The `concat!()` restriction is here:
https://github.com/rust-lang/rust/blob/master/src/libsyntax_ext/concat.rs
It seems to me that the only real restriction that it _should_ have is not mixing byte str literals and str literals when concatenating, and even then it should be possible to convert from one to the other.
Changing the linked code to keep _two_ accumulators, one for `str` and another as a `Vec<u8>` for byte `str`, and returning `base::MacEager::expr(cx.expr_lit(sp, ast::LitKind::ByteStr(rustc_data_structures::sync::Lrc::new(byte_acc))))` would be enough to add the desired functionality.
@rust-lang/lang is there any reason to not change `concat!()` to accept concatenating multiple byte `str`s?
username_2: The diagnostic bug [is fixed in nightly](http://play.rust-lang.org/?gist=b9468573ce64ffd3abcbac8491bb9343&version=nightly&mode=debug&edition=2015) by #51978.
username_2: Closing. There's now an RFC for the feature and the bug has been fixed.
Status: Issue closed
|
jqno/equalsverifier | 354492341 | Title: InaccessibleObjectException: Unable to make field protected boolean javax.swing.AbstractAction.enabled
Question:
username_0: # What steps will reproduce the problem?
Running `org.openstreetmap.josm.data.osm.search.SearchCompilerTest` with JDK 11 and `--illegal-access=deny`
# What is the code that triggers this problem?
```java
@Test
public void testEqualsContract() {
Set<Class<? extends Match>> matchers = TestUtils.getJosmSubtypes(Match.class);
for (Class<?> c : matchers) {
EqualsVerifier.forClass(c).usingGetClass()
.suppress(Warning.NONFINAL_FIELDS, Warning.INHERITED_DIRECTLY_FROM_OBJECT)
.withPrefabValues(TaggingPreset.class, newTaggingPreset("foo"), newTaggingPreset("bar"))
.verify();
}
}
private static TaggingPreset newTaggingPreset(String name) {
TaggingPreset result = new TaggingPreset();
result.name = name;
return result;
}
```
Provide an example of a complete class (equals method, hashCode method, relevant fields) and a call to EqualsVerifier.
```java
public class TaggingPreset extends AbstractAction implements ActiveLayerChangeListener, AdaptableAction, Predicate<IPrimitive> {
....
}
```
# What error message or stack trace does EqualsVerifier give?
```
java.lang.reflect.InaccessibleObjectException: Unable to make field protected boolean javax.swing.AbstractAction.enabled accessible: module java.desktop does not "opens javax.swing" to unnamed module @67065cff
at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:340)
at java.base/java.lang.reflect.AccessibleObject.checkCanSetAccessible(AccessibleObject.java:280)
at java.base/java.lang.reflect.Field.checkCanSetAccessible(Field.java:176)
at java.base/java.lang.reflect.Field.setAccessible(Field.java:170)
at nl.username_1.equalsverifier.internal.reflection.FieldAccessor.modify(FieldAccessor.java:177)
at nl.username_1.equalsverifier.internal.reflection.FieldAccessor.copyTo(FieldAccessor.java:152)
at nl.username_1.equalsverifier.internal.reflection.ObjectAccessor.copyInto(ObjectAccessor.java:116)
at nl.username_1.equalsverifier.internal.reflection.ObjectAccessor.copy(ObjectAccessor.java:81)
at nl.username_1.equalsverifier.EqualsVerifier.withPrefabValues(EqualsVerifier.java:137)
at org.openstreetmap.josm.data.osm.search.SearchCompilerTest.testEqualsContract(SearchCompilerTest.java:721)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:239)
at org.openstreetmap.josm.testutils.JOSMTestRules$TimeoutThread.run(JOSMTestRules.java:684)
```
# What did you expect?
No illegal access :)
# Which version of EqualsVerifier are you using?
2.5.1
# Please provide any additional information below.
Answers:
username_1: Hi!
This is kind of a confusing issue, since I can't simply copy and paste the code into my IDE and run it to see what's wrong.
I see a mention of Swing in the stack trace. Where does that come from in the code?
username_0: The `TaggingPreset` class we use in prefab values extends `javax.swing.AbstractAction`, I have edited the description to make it clearer.
username_0: Standalone example:
```java
public class A {
private B b;
@Override
public int hashCode() {
return 31 + ((b == null) ? 0 : b.hashCode());
}
@Override
public boolean equals(Object obj) {
if (this == obj)
return true;
if (obj == null || getClass() != obj.getClass())
return false;
return java.util.Objects.equals(b, ((A) obj).b);
}
}
public class B extends javax.swing.AbstractAction {
public B(String name) {
super(name);
}
@Override
public void actionPerformed(java.awt.event.ActionEvent e) {
}
}
public class Bug206 {
@org.junit.Test
public void testEqualsContract() {
nl.username_1.equalsverifier.EqualsVerifier.forClass(A.class).usingGetClass()
.suppress(
nl.username_1.equalsverifier.Warning.NONFINAL_FIELDS,
nl.username_1.equalsverifier.Warning.INHERITED_DIRECTLY_FROM_OBJECT)
.withPrefabValues(B.class, new B("foo"), new B("bar"))
.verify();
}
}
```
username_1: Thanks for the update! I see the problem now, and I will look into it.
username_1: Sorry for the radio silence. I've reproduced the issue (which was pretty easy with your update). (See also [this repo](https://github.com/username_1/equalsverifier-issuerepro/tree/issue-206).)
I've been thinking about how to deal with this, but so far I can only conclude that I'm not sure yet 😅. Is this an urgent issue for you?
username_1: OK, thanks! :).
I'm currently working on a large new release, I think I'll look into this one once that release is out. I'll keep you posted.
username_0: Thanks Santa! :santa: :) |
cibernox/ember-power-select | 934194065 | Title: Do not close dropdown on select in multi-select mode
Question:
username_0: When I am using the dropdown in multi-select mode, each time I select an option the dropdown closes and I need to reopen it to select another option.
Is there a way to keep the dropdown "open" while a user selects multiple options?
Answers:
username_1: Yes, pass `@closeOnSelect=false`. Check https://ember-power-select.com/docs/api-reference
Status: Issue closed
|
kylejschulz/mod_0_skills | 723870655 | Title: Assessment Results
Question:
username_0: @username_1
Overall, nice work here! You've performed well enough on this assessment to merit making some changes to achieve being considered `technical ready`. Please make the following updates by EOD Monday:
- [ ] Nice work using naming patterns for your boolean attributes, and coming up with great method ideas for changing those attributes. You can be a bit more accurate in describing what those methods do, however. For example, would method `polish` really toggle `is_polished`, or would it set that attribute to one boolean value every time? Same thing with method 'clean'.
- [ ] In both classes, you have an attribute listed as an array data type. Both of these are singular attributes, and as I see your examples, they really just hold one string at a time. Arrays are more interesting when used as collections, being able to put things in or take them out. Since these ideas are not collections, what data type do you think they should be? Can you think of other ideas for an array attribute for either, or both, classes?
Please let me know if you have any questions!
Answers:
username_1: Mason,
I've made some changes. It didn't make sense to use an array for either example because only one string was listed in each attribute. I added an array in example two. I couldn't think of one for example one that made a lot of sense. I thought i could maybe use an array to list all of the metals that compose sterling silver, but couldn't think of a method that would use it.
I elaborated more on the methods that affected boolean values. I would have included some of that logic in the initial explanation but wasn't sure how thorough you wanted us to be.
Thanks for the helpful input. Please let me know if you'd like any more explanation or edits.
Kyle |
TensorSpeech/TensorFlowTTS | 943401333 | Title: How can I use the pretrain FastSpeech2 model on a new langauge?
Question:
username_0: Hi Community!
Currently, the pretrained TTS models like FastSpeech2 supports ljspeech,kss,baker,libritts & thorsten datasets.
I would like to find out how I can train the model with a Singapore-English language dataset.
Singapore-English language comprises a mixture of words of many different languages - Malay, Tamil, Dialects,English & Chinese.
I have the audio waves & transcript (metadata.csv) files for training.
How can I contribute my datasets to this community? Could anyone recommend me references/guides on how to do so ? I want to specifically train my dataset with FastSpeech2 model :)
Please let me know.
You can also contact me at <EMAIL> to share the knowledge & resources :)
Thank you so much ! |
software-mansion/react-native-reanimated | 695253477 | Title: Question: How to get a Native Module to work with reanimated
Question:
username_0: ## Description
I would like an event from a Native Module to update an `Animated.Node`. Specifically, I want to update a node in response to an event emitter attached to the device orientation, using a module such as https://github.com/cpojer/react-native-heading. I am working with Android and native Java code.
One solution I tried was wrapping the Native Module as a Native UI Component `<DeviceHeading />`, extending a Native View with an `onChange` prop, and trying to use:
```js
const AnimatedDeviceHeading = Animated.createAnimatedComponent(DeviceHeading)
```
You can see the example code https://github.com/username_0/react-native-device-heading/blob/master/example/src/App.tsx
I do not have experience writing native modules and components, but I have read all the docs for both reanimated and react-native for native components, and I cannot find any guidance about how to write a Native Component / Module that works with reanimated.
Ultimately I want to create a view that rotates based on device heading, but using `react-native-reanimated` so that the view rotation happens on the native thread (listening to device heading changes in JS results in a laggy UI).
## Code
https://github.com/username_0/react-native-device-heading/blob/master/example/src/App.tsx
### Package versions
- React: 16.11.0
- React Native: 0.62.2
- React Native Reanimated: 1.13.0
Answers:
username_1: Reanimated should work with any registered native event. I think your problem may be related to not registering events from the JS side, try checking out `UIManager.genericDirectEventTypes` from RN. Closing as this isn't directly related to Reanimated but RN.
Status: Issue closed
|
jfranco1701/STEM-Capstone-2019 | 421295994 | Title: Add option to show/hide event detail information
Question:
username_0: The creator of the event should be able to either show or hide event details like address/date/time.
Answers:
username_1: We are directly showing the event details such as event information, place and date. This ticket can be closed.
Status: Issue closed
|
Baseyoyoyo/Higher-or-Lower | 64515941 | Title: The code base needs test coverage
Question:
username_0: The refactor to classes is looking great, nice work!
Now the code is well abstracted and broken up into small methods, it would be a good move to write some tests for those classes.
Python has a handy testing framework called [unittest](https://docs.python.org/2/library/unittest.html). Let's use that to test some classes, like `PlayerScore` and `Levels`.<issue_closed>
Status: Issue closed |
freechipsproject/rfcs | 582531533 | Title: Register Widths
Question:
username_0: The asymmetry of how `Reg`, `RegNext`, and `RegInit` set their widths can lead to unexpected behavior.
h/t @antonblanchard
<!-- Note: for support questions, you are strongly advised to use [Stack Overflow](https://stackoverflow.com/questions/tagged/chisel). -->
<!--
This repository's issues are reserved for feature requests and bug reports.
Please select the item best describing the issue in each category and delete the other items.
-->
<!-- choose all that apply -->
**Type of issue**: bug report | feature request | other enhancement
<!-- choose one -->
**Impact**: API modification
<!-- choose one -->
**Development Phase**: request
**Other information**
<!-- include detailed explanation, stacktraces, related issues, suggestions how to fix, links for us to have context, eg. Stack Overflow, gitter, etc -->
There's an old PR referencing and documenting some of this behavior: https://github.com/freechipsproject/chisel3/pull/455
**If the current behavior is a bug, please provide the steps to reproduce the problem:**
Consider the following example. While `foo`, `bar`, and `baz` have almost identical construction, `foo` has an unset width, `bar` was a width of 8, and `baz` is a bundle with an internal signal of width `8`.
```scala
package examples
import chisel3._
import chisel3.stage.{ChiselStage, ChiselGeneratorAnnotation}
import chisel3.internal.sourceinfo.UnlocatableSourceInfo
object NoInfo {
implicit val noInfo = UnlocatableSourceInfo
}
import NoInfo._
class DelayedBundle extends Bundle {
val in = UInt(8.W)
}
class Delayed extends MultiIOModule {
val in = IO(Input(UInt(8.W)))
val bundle = IO(Input(new DelayedBundle))
/** The width of foo is unset */
val foo = RegNext(in)
/** The width of bar is set to 8.W */
val bar = Reg(chiselTypeOf(in))
bar := in
/** The width of baz is set to 8.W */
val baz = RegNext(bundle)
Seq(foo, bar, baz).foreach(a => println(a.widthOption))
[Truncated]
To be very concrete, I have implicit classes for doing automatic conversion to bits, words, etc. (i.e., `(Bits) => Vec[UInt]`). These rely on having widths known. Using these methods on `foo` will cause the implicit class methods to error out complaining that they can't be used on something that has an unknown width. However, they work with `bar`. This is unintuitive behavior as it would seem that `foo` and `bar` are identical constructions.
I do realize that `RegNext(0.U)` is probably the reason why this exists the way it does.
**What is the current behavior?**
The width of a `RegNext` is always unset when given a subtype of `Bits`.
**What is the expected behavior?**
The width of a `RegNext` should be set to the type of whatever is next.
**Please tell us about your environment:**
<!-- (examples)
- version: `3.0-SNAPSHOT`
- OS: `Linux knight 4.4.0-92-generic freechipsproject/chisel3#115-Ubuntu SMP Thu Aug 10 09:04:33 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux`
-->
**What is the use case for changing the behavior?**
This clarifies unintuitive behavior and unifies behavior for `RegNext` when given different types. |
cgeo/cgeo | 120885745 | Title: IDEA -> -> Function "own and findet Caches"
Question:
username_0: The great Function: "Own and findet Caches" show on MAIN-Site!!
.....not in properties/cache-details/own and findet Caches/ select/deselect
Answers:
username_1: Why do you need own and found caches on the main screen? AFAIK usually people want to hide these caches, but not to have easy access to them.
username_2: I assume he means the filter. It is available not only from the preferences, but also directly from the map. I couldn't find it in lists, though.
username_2: This issue was mentioned at the OC HQ event.
In the past I was against adding more filter options to the menus before we switch to a more complex filter system, but because this was delayed it might be better to add a checkbox to the list like we did for the live map in #2234.
Status: Issue closed
username_3: Solved with #9081 |
EBI-predocs/research-software | 202508383 | Title: GLIBC Issues
Question:
username_0: A number of packages, when run from a RedHat 6 host crash with errors like:
/lib64/libc.so.6: version `GLIBC_2.14' not found
This is because I built a small number of packages on a RedHat 7 host.
I am currently rebuilding the affected packages, which will resolve this issue.
Answers:
username_0: This should be fixed now, please let me know if you have further issues.
Status: Issue closed
|
sydevs/WeMeditate | 1161904590 | Title: Replace Vimeo with JW Player
Question:
username_0: Most Vimeo support has already been disabled during the content editor overhaul.
We want to switch to JW Player because it will allow us to replace all of these services with 1 service
- Vimeo
- Twitch
- Restream
- Plyr.js (javascript video library)
Currently this ticket is pending on @antontcymbal negotiating an account with the JW Player support. |
jsoverson/grunt-preprocess | 55632265 | Title: Support for grunt targets
Question:
username_0: Hello,
I have the following task described in my gruntfile.js:
```js
preprocess: {
html: {
src: 'src/index.html',
dest: 'dist/index.html'
},
options: {
context: {
DEBUG: false
}
}
}
```
It works great, but if I want to use grunt targets:
```js
preprocess: {
dev: {
html: {
src: 'src/index.html',
dest: 'dist/index.html'
},
options: {
context: {
DEBUG: false
}
}
}
}
```
It does not work anymore. Is this expected behaviour ? Or maybe I did something wrong
Status: Issue closed
Answers:
username_0: Ok, my bad.
I had to remove html into dev/prod to have it working, sorry for the issue. |
reduxjs/redux-toolkit | 622118359 | Title: Missing loading variable at the Documentation examples for createAsyncThunk using rejectWithValue
Question:
username_0: Hi, I was reviewing the documentation for use createAsyncThunk for use it with axios and, in the examples for rejectWithValue there is a loading prop destructured from the selector, but that loading is not used in the code example and is not defined in the initial state

Answers:
username_1: @username_0 Thanks, good catch.
username_0: happy to help
Status: Issue closed
|
google/mediapipe | 855092669 | Title: What is the metric scale of the face landmarks?
Question:
username_0: Hi, I was wondering if there is someway to translate the face landmarks obtained into a metric scale. I would for example need to know the distance from the tip of the nose to the chin.
If I read [the docs](https://google.github.io/mediapipe/solutions/face_mesh.html) then the face_geometry module seems to do the trick. However, I can nowhere find an example of how to do this in python.
Is there a way to get real metric data about face landmarks and if so, is there an example for python?
Answers:
username_1: Hey @username_0,
You are right that the `FaceGeometry` module is currently not available on Python, only on C++. @jiuqiant might be able to navigate you around the timeline there
As for your use-case, I sense that the precision of Face Geometry (in terms of metric units) might not be good enough for you
We have a canonical face 3D model and during runtime we make sure that the scale of detected face is set to the scale of the canonical face model. With perspective cameras, there's a non-trivial challenge of distinguishing between face size and its distance away from a camera: in terms of screen coordinates, a face could seem large either cuz its large on its own OR cuz its closer to the camera. In a way, we sacrifice the scale (by making it a constant) for the sake of better translation estimation (so two faces of roughly the same size could be positioned correctly in respect to the camera)
If that's something you are fine with, then please hang on before MediaPipe has a Python API for FaceGeometry module, otherwise I'd advice you to seek other approaches that try to differentiate between the face size and the distance away from camera. Those approaches will likely be NN-based as they'd take other pieces of information about the face (for example, how other facial features are located relatively to each other) to try making a better guess about its size beyond just assuming it's some constant for every detected face
username_0: Hi There,
Thanks for the elaborate answer. So if I understand correctly media pipe makes every face just as large as the canonical face model. The scale thus remains actually unknown, is that correct?
But if this is the case, what is the use case of the FaceGeometry module? Is it limited to having a slightly more accurate 3D form because it no longer makes use of the weak perspective model?
username_1: One way or another, with a Computer Vision approach you never get the _actual_ value, only some _approximation_. We only approximate each face scale as a scale of some constant abstract face, which is probably the easiest way and lacks any sophistication. If you go a bit further, you'll see that human faces in general have some distribution of sizes so you'd be willing to find some data learning approach that'd try to capture that distribution. However, that'd be a challenging task too as making a camera-invariant, RGB-only face size prediction is generally fairly non-trivial. Apple's ARKit is not RGB-only, they use hardware for accurate depth estimation - that makes their life a bit easier. In general, I think there are probably some good NN-based approaches out there, I just wouldn't expect them go give me the _actual_ face size, only some _approximation_ which is probably better than what Face Geometry module gives you, but will very likely be worse than approaches based on depth sensors
username_2: Thanks for the elaborate answer once again.
**Canonical face model**
I re-read the documentation again and I now realize that the facegeometry module only does a rigid mapping of the canonical face model. This with as goal to find the face pose transformation matrix. This matrix is needed to put virtual models in the correct pose. Furthermore it gives 'a scale' to the model.
Quick questions: Documentation states that triangular topology is 'inherited' from the canonical face model. This is not an actual real time calculation that is performed because always the same 468 vertexes are present so you don't need the canonical face model at runtime to calculate this mesh. Is this correct?
Vertex texture coordinates are inherited from the canonical face model. What's the advantage of this? Why wouldn't you use the metric coordinates of the landmarks? If you use the canonical face model coordinates then you can never have a deforming texture. (As the canonical model is rigid)
**Estimating actual size**
So I understand that it will be very hard to get a good idea of the size. (Maybe with such an NN-based approach but this will take me too long). My current approach would be to get the size by holding an ID-card with known dimensions. I guess this should give me an estimate.
Quick questions: Could I enhance accuracy by taking multiple pictures in multiple poses. Or is Mediapipe really optimized for a frontal picture? I could for example imagine that it is a lot easier to estimate the depth of the nose from a sideview then from a frontal view.
username_0: Thanks for the elaborate answer once again.
Canonical face model
I re-read the documentation again and I now realize that the facegeometry module only does a rigid mapping of the canonical face model. This with as goal to find the face pose transformation matrix. This matrix is needed to put virtual models in the correct pose. Furthermore it gives 'a scale' to the model.
Quick questions: Documentation states that triangular topology is 'inherited' from the canonical face model. This is not an actual real time calculation that is performed because always the same 468 vertexes are present so you don't need the canonical face model at runtime to calculate this mesh. Is this correct?
Vertex texture coordinates are inherited from the canonical face model. What's the advantage of this? Why wouldn't you use the metric coordinates of the landmarks? If you use the canonical face model coordinates then you can never have a deforming texture. (As the canonical model is rigid)
Estimating actual size
So I understand that it will be very hard to get a good idea of the size. (Maybe with such an NN-based approach but this will take me too long). My current approach would be to get the size by holding an ID-card with known dimensions. I guess this should give me an estimate.
Quick questions: Could I enhance accuracy by taking multiple pictures in multiple poses. Or is Mediapipe really optimized for a frontal picture? I could for example imagine that it is a lot easier to estimate the depth of the nose from a sideview then from a frontal view.
username_1: For face tracking, I'd say that MediaPipe doesn't even try doing anything to aggregate predictions over multiple views and come up with a better metric. Under such circumstances, you best shot is to receive the most accurate single short (probably the frontal view, yeah) and to use your ID card trick and try transferring its known size into the face landmark scale
username_3: Hi @username_0, Did you get a chance to go through the above comments. Thanks!
username_4: Wouldn't it be possible to use iris detection to estimate face size in the real world?
username_1: Hey @username_4,
Theoretically, yes, there is an approach worth of exploring (please see #1891 for some additional context). However, there are a few problems with that approach which I'd like to highlight:
- Z derived from left and right eyes might be quite different (please see GIFs in [this comment](https://github.com/google/mediapipe/issues/1642#issuecomment-804362081)), which already suggests that there's significant variation in the depth value signal coming from iris prediction
- You might observe that metric value jumps quite a lot as you are essentially trying to transfer the scale of a small object with a small error (iris) to a larger object (face / head) - this turns the small iris landmark error into a larger face mesh error. This implies using some sort of heavy filtering
- When a user blinks / has eye patch, then you'll not be observing iris such frame, which in turn requires turning this approach into something more complex (maybe extrapolating / hallucinating distance on frames when iris is not visible; or computing some sort of ratio between head pose scale and iris on some initial frames and then only using that ratio + head pose scale on all subsequent frames)
Potentially, I see benefits of using iris landmark prediction to enhance face depth estimation, provided those problems I mentioned are resolved. Unfortunately, me & team didn't have enough time / priority to explore in this direction, so the thoughts I'm sharing in this reply are pretty much as far as we went with the idea. If you'll reach a state of a nice working prototype, then please don't hesitate to share it here! I'll surely help MediaPipe users out & could be later ported into MediaPipe for everyone to share :)
Thanks!
username_3: We are closing this issue as of now, due to lack of activity.
Status: Issue closed
|
chemerisuk/cordova-plugin-firebase-dynamiclinks | 262494368 | Title: Failed to install on iOS
Question:
username_0: Hi Chemerisuk,
I've got this error when updating to 0.11 on iOS.
Failed to install 'cordova-plugin-firebase-dynamiclinks': Error: pod: Command failed with exit code 1.
I have tried pod repo update. Any ideas?
Thanks,
Sam<issue_closed>
Status: Issue closed |
openstack-charmers/zaza | 604637493 | Title: Wishlist: --keep-faulty-model
Question:
username_0: Feature request: most of the time when using "functest-run-suite --keep-model" we're interested in inspecting models with test failures.
It would be useful to have conditional reaping of models, where only successful models would get destroyed, while those with testing failures or errors would be kept for later inspection.
Answers:
username_1: This can be achieved now by using `functest-run-suite --keep-faulty-model`. When an exception occurs during deployment, the model will be kept. Otherwise, models will be destroyed.
Please confirm if this is the expected behavior and close the issue (I can't :))
username_2: Closing due to new feature provided in #458 by @username_1
Status: Issue closed
|
DnAp/Agar.io-Reverse-Engineering | 79460923 | Title: Better ennemy hack calculation
Question:
username_0: Hello,
I really appreciate your project and use it to have more informations while playing Agar :-)
I just notice that the calcul done for mass comparaison is not optimized and is not correct.
Around line 972, you have :
```javascript
if (this.size < minMass * 0.9 / 2) {
this.color = '#57FF07';
} else if (this.size < minMass * 0.9) {
this.color = '#07FFB0';
} else if (this.size * 0.9 > minMass) {
this.color = '#FF3107';
} else {
this.color = '#4106FF';
}
```
You should have that :
```javascript
if (this.size * 0.9 > minMass) {
this.color = '#FF3107';
} else if (this.size < (minMass / 1.414213562) * 0.9) {
this.color = '#57FF07';
} else if (this.size < minMass * 0.9) {
this.color = '#07FFB0';
} else {
this.color = '#4106FF';
}
```
In fact, there is more Red than Green. And there is more Green, you are big and slower.
minMass is the square root of the mass. We must divide by the square root of 2, not 2 directly.
Hope this would help to have better hack :)
Regards.
Answers:
username_1: Thanks, commit it is.
Status: Issue closed
|
CodingGarden/seedling-school-01-frontend-project | 525384091 | Title: MyLaunch Project Submission
Question:
username_0: ## MyLaunch
## Your Name / Title
<NAME>
## Project Description
For those interested in things that happened the day they were born this is perfect! Everyone loves space and so by entering your birthday (or any other date) you can find out if a rocket was launched the day you were born thanks to the [Launch Library API](https://www.launchlibrary.net/docs/1.3/api.html#tippitytop)
* https://github.com/username_0/mylaunch/blob/master/PROJECT_DESCRIPTION.md
## What 3rd Party Web API do you plan to use?
* [Launch Library API](https://www.launchlibrary.net/docs/1.3/api.html#tippitytop)
## Which of the following describes you:
* [x] YouTube Subscriber
* [x] Twitch Follower
* [ ] Patreon Patron
* [x] Superchat Donor
* [ ] Streamlabs Donor
* [x] Coding Garden Moderator
Answers:
username_0: Submitted for project description review.
username_1: ⭐️⭐️⭐️⭐️⭐️
Awesome!! Project description looks great.
Feel free to get started.
username_0: Landing page is complete except for actual links to other pages which don't exist yet. Attached are some screenshots.
<img width="1440" alt="Screen Shot 2019-11-21 at 9 52 47 PM" src="https://user-images.githubusercontent.com/22730186/69394286-9d858100-0ca9-11ea-9ac0-e1508eb4e19a.png">
<img width="351" alt="Screen Shot 2019-11-21 at 9 53 00 PM" src="https://user-images.githubusercontent.com/22730186/69394287-9d858100-0ca9-11ea-9eb2-c0c6435ba228.png">
<img width="351" alt="Screen Shot 2019-11-21 at 9 53 04 PM" src="https://user-images.githubusercontent.com/22730186/69394288-9d858100-0ca9-11ea-9763-24adf8e8b7cc.png">
username_0: I believe it is complete! Woot! It is available [here](http://mylaunch.jajjferris.com/)
username_1: Sweet! Would you like UI feedback or a code review?
username_0: UI feedback would be good. I just chatted with the API maintainers and they gave me a cool idea for an addition to functionality 😄
username_1: Cool!
Some feedback:
* The date picker is OS specific and is kind of hard to enter a birthday because you have to go back many years. Maybe you can have a dedicated year select, month select and day select?
* Could you possibly show launches a few days / weeks before / after a birthday? It's kind of sad to see that there are no launches on .
* Could you possibly show launches that are on the same month / day but in a different year than the user's birth year? Maybe we will see more results.
username_0: * Yeah the date picker is my least favorite feature - thinking about making it just a regular input with a mask instead
* That was the feedback I got from the developers in regards to sad news about your birthday. They also told me they are missing a fair amount of historical data so unless you are pretty young you are out of luck. I am thinking of making multiple card lists as a result - your actual birthday (if there was one), your birthday on every year since. To list all since your birthday could lead to a LOT of results but would be cool.
username_0: I updated the launch date picker UI and functionality on data return.
username_1: I like it! It's a much nicer experience to at least get _some_ result back.
username_0: Yeah, I like what is has become. Actually useful and functional.
username_0: Would love a UI/UX and code review :smile:
username_1: # UI Review
### Things I Like
* Color scheme
* Love the about page. Very succinct.
* Links to video / more info open in a new tab.
* I like the rocket animation during loading
* I can see launches after my birthday if no launch found
### Things to improve
* When clicking submit, remove previous results before loading next results.
* When clicking next for pagination, might scroll to the top
* Might have a separate input box for each: year, month, date; as the placeholder disapears after typing starts and the user might forget the order / format
### Checklist
* [x] Fonts choices are consistent
* [x] Body font is good for reading
* [x] Loading indicators are shown when there is background processing (API requests, calculations etc.)
* [x] Page responds nicely to different screen sizes
* [x] Images resize accordingly
* [x] Images maintain aspect ratio
* [x] Columns break at smaller screen sizes
username_1: # Code Review Checklist
### README
* [x] Readme includes instructions for running the project locally
* Might mention that we need a static file server, or to open index.html in the browser
### HTML
* [x] Consistent Indentation
* [x] Includes meta viewport tag for mobile devices
* `<meta name="viewport" content="width=device-width, initial-scale=1.0">`
* [x] CSS Links and Font Links are in the head of the document
* [x] JavaScript files are linked at the bottom of the page OR at the top with an onload function OR at the top with a defer attribute
* [x] Uses semantic tags where available
* header, time, main, nav, article, section etc.
* See this article for more: https://html.com/semantic-markup/
* [x] Title / Header links back to home page
* [x] No extra elements. Only includes the necessary elements to make things work.
### JavaScript
* [x] Consistent semicolon usage. Either do or do not. There is no inbetween.
* [ ] Consistent quote usage. Either `'` or `"`, don't mix.
* Line 26 uses double quotes
* [ ] Consistent indentation.
* The HTML indentation in the `makeRocketElement` function is a little off
* [ ] Reasonable max line length. Wrap / reformat code when it gets too long.
* `fetchRocketData` has quite a few long lines
* You might try putting the error message in reusable function to reduce line length
* [x] Variable / Function names are clear and concise.
* [x] No abbreviations.
* [x] Variable / Function naming convention is consistent.
* camelCase or PascalCase or snake_case
* [ ] Strings used more than once are in a variable.
* Error messages can be put into variables
* [ ] Functions are as few lines as possible. Code reads like a sentence.
* The `fetchRocketData` could pottentially be split into 1 or 2 functions.
* [ ] Function names dictate intent.
* `fetchRocketData` function name is confusing compared with `makeRocketFetch`, `fetchRocketData` technically doesn't `fetch` it handles all the logic. Might think of a better name. @sk1nni in the chat suggests: `prepareRocketLauch`
* [x] Nested loops avoided where possible.
* [x] Functions take in a reasonable number of parameters. Ideally 3 or less params. Otherwise, use an options object.
* [ ] No extra variables.
* `allBirthdayLaunchesPagination` is unused.
### CSS
* [x] Consistent indentation
* [x] Consistent naming convention
* [x] No duplicate styles. Re-use where applicable.
### Suggested Refactor
A possible way to refactor the fetchRocketData function:
```js
const makeRocketFetch = async (startDate, endDate, offset = 0) => {
const response = await fetch(`${API_URL}${startDate}/${endDate}?limit=${FETCH_LIMIT}&offset=${offset}`);
[Truncated]
const fetchRocketData = async () => {
try {
if (!dateElement.validity.valid) throw new Error(dateElement.validationMessage);
logoElement.classList.toggle('animate', true);
const endDate = getEndDate();
launchResponse = await makeRocketFetch(dateElement.value, getFormattedDate(endDate));
allLaunchesSince.style.display = '';
updatePaginationButtons();
updateDisplayedRockets();
errorElement.classList.toggle('hidden', true);
addSearchResultToPage();
} catch (error) {
birthdayLaunchesContainer.innerHTML = '';
errorElement.classList.toggle('hidden', false);
errorMessageElement.textContent = error.message;
}
logoElement.classList.toggle('animate', false);
};
```
username_0: Thanks for the review! I have addressed the concerns you raised (hopefully) and updated the pagination display so it is easier on the eyes. |
mldbai/mldb | 180329011 | Title: Add missing model evaluation metrics
Question:
username_0: Rationale: Many Kaggle competitions use evaluation metrics that are currently not reported by MLDB. This makes it harder to use to solve competitions.
DOD: Add the following metrics to classifier evaluation:
- Regression
- Mean absolute error: https://www.kaggle.com/wiki/MeanAbsoluteError
- Binary classification
- Log loss: https://www.kaggle.com/wiki/LogarithmicLoss
- Multi-class classification
- Log loss: https://www.kaggle.com/wiki/MultiClassLogLoss
- Hamming Loss: https://www.kaggle.com/wiki/HammingLoss
They should be added in: https://github.com/mldbai/mldb/blob/master/plugins/accuracy.cc
Internal ticket: MLDB-1968
Answers:
username_1: I can work on this starting with Binary Classification Log Loss |
atom/autocomplete-css | 205427350 | Title: Values from another property in hint.
Question:
username_0: Hi, when I write css property between another properties, or at begin of properties, hints from autocomplete-css belong to last written property. When I write new property on end, hint is right.
example:

Atom v 1.13.1
autocomplete-css 0.14.1
Answers:
username_1: Darn, I thought I fixed all the inline bugs :/.
Status: Issue closed
|
Brokemia/TrainerReborn | 684926692 | Title: Can't give item Lightfoot Tabi using correct spelling
Question:
username_0: The Dict file has an incorrect spelling for the Tabi making you have to spell it _LigthfootTabi_
also side note, would be cool to see the reload and save hotkey disabled while on the teleport or spawn screen.
Answers:
username_1: The LigthfootTabi was something that just came over from the old trainer and I hadn't noticed. The thing with hotkeys is something I had noticed and was planning to address at some point. I'll deal with this next time I feel like doing some work on it.
username_1: Working on some changes now. Both of these things will be in the next update |
dalehenrich/metacello-work | 188769859 | Title: Latest Metacello is broken in Pharo6 because of pragma API changes.
Question:
username_0: The pragma API changed in the latest Pharo 6:
In past we had:
- #selector to access to method selector
- #keyword to access to pragma selector
now we have:
- #methodSelector to access to method selector
- #selector to access to pragma selector
I think this broke the latest Metacello for Pharo.
In a Jenkins job I load the latest Metacello for Pharo then I load a project metadataless from GitHub with a baseline.
This is what I see in the stack trace: (Temporary screen here: http://puu.sh/seF7K/88b5bf31f5.png)
…
Loaded -> Metacello-TestsPlatform.squeakCommon-ChristopheDemarey.21 --- github://dalehenrich/metacello-work:master/repository [a2933f5:master] --- cache
Evaluated -> baseline [BaselineOfMetacello] >> reprimeRegistryIssue197
...finished baselineThe method Object>>#ifNotNilDo:ifNil: called from MetacelloProjectRegistration>>#merge: has been deprecated.
Please use #ifNotNil:ifNil: instead
Fetched -> BaselineOfOOnoz-cypress.1 --- github://LeCoinEnBande/OOnoz:development/src [6bc9a1e:development] --- github://LeCoinEnBande/OOnoz:development/src
Loaded -> BaselineOfOOnoz-cypress.1 --- github://LeCoinEnBande/OOnoz:development/src [6bc9a1e:development] --- github://LeCoinEnBande/OOnoz:development/srcMessageNotUnderstood: BaselineOfOOnoz>>baseline
BaselineOfOOnoz(Object)>>doesNotUnderstand: #baseline
[ self configuration perform: pragma selector with: self ] in MetacelloBaselineConstructor(MetacelloAbstractVersionConstructor)>>evaluatePragma: in Block: [ self configuration perform: pragma selector with...etc...
BlockClosure>>ensure:
MetacelloBaselineConstructor(MetacelloAbstractVersionConstructor)>>evaluatePragma:
MetacelloBaselineConstructor>>calculate:project:
MetacelloBaselineConstructor>>on:project:
MetacelloBaselineConstructor class>>on:project:
BaselineOfOOnoz(ConfigurationOf)>>project
BaselineOfOOnoz class(ConfigurationOf class)>>project
MetacelloScriptEngine>>getBaselineProjectUnconditionalLoad:
MetacelloMCBaselineProject>>projectForScriptEngine:unconditionalLoad:
MetacelloMCBaselineProject(MetacelloProject)>>projectForScriptEngine:
MetacelloMCBaselineOfProjectSpec(MetacelloMCProjectSpec)>>versionForScriptEngine:
…
Answers:
username_0: I suppose this changes can't be done in a simple way since Metacello is an inter-smalltalks project? How should this be corrected to not break Metacello on Genstone and Squeak?
Status: Issue closed
username_0: This pragma change will be revert for now. I close. :) |
systemjs/systemjs | 105011605 | Title: Is importing an empty file supported?
Question:
username_0: Hey,
I encountered a rather nasty bug this evening. I attempted to import a file which had only commented out lines within it. Attempting to do so caused my browser to become unresponsive. Looking at the 'Network' tab in Chrome Dev Tools I can see that the last request to successfully complete is the file in question. All XHRs after that point remain pending.
Is this a known issue? Weird edge case? Something corrupt on my end?
Answers:
username_1: Thanks for posting. Can you share an isolated replication or steps to replicate for this?
username_0: https://github.com/username_0/jspmdebug
I built a repo with an example inside of it.
Check it out, run the standard `npm install jspm` `jspm init` and accept all defaults.
I was testing this in a Chrome extension environment, so for me I would then:
- go to chrome://extensions
- check 'developer mode' checkbox
- check 'load unpacked extension' and select parent directory of manifest.json
- navigate to chrome-extension://jbnkffmindojffecdhbbmekbmkkfpmjd/test.html
Observe that page is totally unresponsive. If you open developer tools before navigating to that link then you can see all the network requests go and then it just hangs after trying to load `test.js`
If I delete all the commented out lines (or even just some of them!) from `test.js` then everything runs as expected.
username_1: Thanks for posting this.
I've managed to trace it down that the following regular expression will take a very long time:
```javascript
var a = "//// /collection/\n//import 'test/background/collection/clientErrors.spec';\n//import 'test/background/collection/playlistItems.spec';\n//import 'test/background/collection/playlists.spec';\n//import 'test/background/collection/searchResults.spec';\n//import 'test/background/collection/videos.spec';\n//import 'test/background/collection/streamItems.spec';\n\n//// /model/\n//import 'test/background/model/activePlaylistManager.spec';\n//import 'test/background/model/clientErrorManager.spec';\n//import 'test/background/model/dataSource.spec';\n//import 'test/background/model/playlistItem.spec';\n//import 'test/background/model/playlistItems.spec';\n//import 'test/background/model/relatedVideosManager.spec';\n//import 'test/background/model/signInManager.spec';\n//import 'test/background/model/user.spec';\n//import 'test/background/model/youTubeV3API.spec';"
a.match(/^\s*(\s*\/\/[^\n]*|\s*"[^"]+"\s*;?|\s*'[^']+'\s*;?|\/\*[^\*]*(\*(?!\/)[^\*]*)*\*\/)*\s*System\.register(Dynamic)?\s*\(/)
```
I didn't think there was any bad backtracking going on here, but apparently there must be.
Will have to debug this further to correct the regex.
username_1: NB this should be ported for 0.18 before the 0.19 release.
username_1: Released in 0.19.0.
Status: Issue closed
|
postmanlabs/postman-app-support | 186638161 | Title: Request Font Size
Question:
username_0: <!--
Welcome to the Postman Issue tracker. Any feature requests / bug reports can be posted here.
Any security-related bugs should be reported directly to <EMAIL>
Version/App Information:
-->
1. Postman Version: 4.8.1
2. App (Chrome app or Mac app): Windows Desktop App
3. OS details: Windows 7 64 Bit
4. Is the Interceptor on and enabled in the app: Yes
5. Did you encounter this recently, or has this bug always been there: Always been there
6. Expected behaviour: To be able to increase or decrease the font size for the request the same as the response.
7. Console logs (http://blog.getpostman.com/2014/01/27/enabling-chrome-developer-tools-inside-postman/ for the Chrome App, View->Toggle Dev Tools for the Mac app): N/A
8. Screenshots (if applicable)

<!--
Steps to reproduce the problem:
-->
While I am able to change the font size for the Response for my requests, I am unable to change the font size for the Request, as shown in the Screenshot. This is for the Desktop App for Windows and so Ctrl + Scroll Wheel does not change the font size.
<!--
Some guidelines:
1. Please file Newman-related issues at https://github.com/postmanlabs/newman/issues
2. If it’s a Cloud-related issue, or you want to include personal information like your username / collection names, mail us at <EMAIL>
3. If it’s a question (anything along the lines of “How do I … in Postman”), the answer might lie in our documentation - http://getpostman.com/docs.
-->
Status: Issue closed
Answers:
username_1: @username_0 You can use Cmd/+ to increase the font size. Cmd/- will decrease it, and Cmd/0 will reset it to the default value :) |
avan989/cFE | 496331851 | Title: ES unit test failing on x86_64 due to uint32 memory handle
Question:
username_0: Another case of a `uint32` being used to store a memory pool handle, which fails when running the unit tests on a 64-bit build.
Using the `CFE_ES_MemHandle_t` typedef instead solves the problem.
Answers:
username_0: _Trac comment by jphickey on 2015-10-06 10:28:19:_
Pushed branch `trac-114-es-u32-memhandle-fix`, commit [changeset:f40cfe5] for review
username_0: _Imported from trac issue 114. Created by jphickey on 2015-10-06T10:22:36, last modified: 2019-03-05T14:57:55_
username_0: _Trac comment by glimes on 2015-10-06 12:26:12:_
Recommend accept.
username_0: _Trac comment by sstrege on 2015-10-19 18:02:21:_
approve/accept
username_0: _Trac comment by glimes on 2015-12-22 15:54:31:_
2015-12-22 merged to development
username_0: _Trac comment by sduran on 2015-10-20 12:47:17:_
recommend accept
username_0: _Trac comment by glimes on 2016-02-25 10:17:32:_
these will be fixed in CFE 6.5
username_0: _Trac comment by jhageman on 2019-03-05 14:57:55:_
Milestone renamed
Status: Issue closed
|
NervJS/taro | 498006045 | Title: 引用多端组件,ts无法识别,导致报错
Question:
username_0: **问题描述**
在文件夹内有`fn.h5.js`和`fn.weapp.js`两个文件
在同级目录的其它文件下通过使用`import * as fn from "./fn";`引用模块
ts无法识别,显示`找不到模块“./fn”。ts(2307)`
**复现步骤**
[复现问题的步骤]
1. 创建多个多端组件
2. 通过import语法引入
3. ts报错
**期望行为**
通过ts检查,最好是在`global.d.ts`文件下用某种声明方式通过检查
**报错信息**
`找不到模块“./fn”。ts(2307)`
**系统信息**
Taro CLI 1.3.17 environment info:
System:
OS: macOS 10.14.6
Shell: 5.3 - /bin/zsh
Binaries:
Node: 11.12.0 - /usr/local/bin/node
Yarn: 1.16.0 - ~/.npm-global/lib/node_modules/yarn/bin/yarn
npm: 6.8.0 - ~/.npm-global/bin/npm
npmPackages:
@tarojs/components: 1.3.17 => 1.3.17
@tarojs/mobx: 1.3.17 => 1.3.17
@tarojs/mobx-h5: 1.3.17 => 1.3.17
@tarojs/mobx-rn: 1.3.17 => 1.3.17
@tarojs/plugin-babel: 1.3.17 => 1.3.17
@tarojs/plugin-csso: 1.3.17 => 1.3.17
@tarojs/plugin-sass: 1.3.17 => 1.3.17
@tarojs/plugin-uglifyjs: 1.3.17 => 1.3.17
@tarojs/rn-runner: 1.3.17 => 1.3.17
@tarojs/router: 1.3.17 => 1.3.17
@tarojs/taro: 1.3.17 => 1.3.17
@tarojs/taro-alipay: 1.3.17 => 1.3.17
@tarojs/taro-h5: 1.3.17 => 1.3.17
@tarojs/taro-swan: 1.3.17 => 1.3.17
@tarojs/taro-tt: 1.3.17 => 1.3.17
@tarojs/taro-weapp: 1.3.17 => 1.3.17
@tarojs/webpack-runner: 1.3.17 => 1.3.17
eslint-config-taro: 1.3.17 => 1.3.17
eslint-plugin-taro: 1.3.17 => 1.3.17
nerv-devtools: ^1.4.0 => 1.4.4
nervjs: ^1.4.0 => 1.4.4
stylelint-config-taro-rn: 1.3.17 => 1.3.17
stylelint-taro-rn: 1.3.17 => 1.3.17
npmGlobalPackages:
typescript: 3.6.3
Answers:
username_0: 暂时使用`ts-ignore`跳过检查了
Status: Issue closed
|
aquasecurity/trivy | 595062038 | Title: Support docker registry token
Question:
username_0: Pass Docker Registry v2 Bearer token via `TRIVY_REGISTRY_TOKEN`.
https://docs.docker.com/registry/spec/auth/jwt/
Answers:
username_1: Should this be closed now since https://github.com/aquasecurity/fanal/pull/102 is merged?
username_0: We need to merge this PR.
https://github.com/aquasecurity/trivy/pull/475
username_0: We have to merge https://github.com/aquasecurity/trivy/pull/482.
Status: Issue closed
|
huggingface/tokenizers | 907047293 | Title: special_tokens_mask doesn't include all special tokens
Question:
username_0: Based on the ids this should be [1, 0, 0, 1, 0, 0, 1]
Is this a bug or simply unsupported?
Answers:
username_0: I can see why - it's hardcoded in the post-processor:
https://github.com/huggingface/tokenizers/blob/<KEY>/tokenizers/src/processors/bert.rs#L57
https://github.com/huggingface/tokenizers/blob/c3b3b290398be14ffcfc65450bf094b9b4ed849a/tokenizers/src/processors/roberta.rs#L90
I've created a custom wrapper for my tokenizer anyway so it's not a big deal, I'll do my own post-processing. |
linuxdeepin/developer-center | 419126294 | Title: 多任务预览页面显示bug
Question:
username_0: ## 缺陷描述
只有激活的窗口会显示, 其他的都会以图标的形态出现在屏幕左上角外侧.
## 复现步骤
点击 dock 上的多任务预览图标或者快捷键
## 期望行为
所在桌面上的各个窗口都应该排列在屏幕上
## 截图

### 发行版以及版本
Arch Linux 2019-03-09
Answers:
username_1: I also confirm this one.
Right now there's all kinds of bugs with this feature :(
Most likely it's also connected to https://github.com/linuxdeepin/developer-center/issues/973
username_0: 问题仍然存在.
The problem is still here.
username_2: [1.9.38](https://github.com/linuxdeepin/deepin-wm/pull/50) will release soon
Status: Issue closed
username_0: It's fixed!
撒花! |
Orange-OpenSource/hurl | 1138974877 | Title: Assert fails only if last request int the single hurl file doesn't match
Question:
username_0: Trying to test multiple paths in the same hurl file, but it fails only if last request doesn't match the assertion.
Example:
SUCCEEDS:
```
GET http://{{host}}:8080/tests
GET http://{{host}}:8080/nema
GET http://{{host}}:8080/api/v4/dr/pull
GET http://{{host}}:8080/nema
HTTP/1.1 200
[Asserts]
jsonpath "$.headers.TestBackendName" == "test-1"
```
FAILS:
```
GET http://{{host}}:8080/tests
GET http://{{host}}:8080/nema
GET http://{{host}}:8080/api/v4/dr/pull
HTTP/1.1 200
[Asserts]
jsonpath "$.headers.TestBackendName" == "test-1"
```
`GET http://{{host}}:8080/api/v4/dr/pull` doesn't match assertion but if it's not on the last line in the hurl file - test is successful.
`GET http://{{host}}:8080/nema` matches the assertion and test is successful regardless of the the fact that previous request doesn't .
I think test should fail if assertion is not met for any of the requests in the hurl file.
hurl version: `hurl 1.5.0 libcurl/7.80.0-DEV OpenSSL/1.1.1l zlib/1.2.11`
Is this behavior expected?
Answers:
username_1: Hi @username_0, the assertions sections is optional and test the _associated_ request.
With your first example:
```hurl
GET http://{{host}}:8080/tests
GET http://{{host}}:8080/nema
GET http://{{host}}:8080/api/v4/dr/pull
GET http://{{host}}:8080/nema
HTTP/1.1 200
[Asserts]
jsonpath "$.headers.TestBackendName" == "test-1"
```
Can be summarised:
```
REQUEST-1
REQUEST-2
REQUEST-3
REQUEST-4
RESPONSE-4
```
The second example is
```
REQUEST-1
REQUEST-2
REQUEST-3
RESPONSE-3
```
The Hurl file format describe a list of[ _entry_](https://hurl.dev/docs/entry.html): each entry being a mandatory request and an optional response that can have multiple asserts.
I hope that it is ok with your workflow!
Regards,
username_0: It's not ideal but I can work around it. Thanks for your response!
Status: Issue closed
|
intel/intel-ipsec-mb | 616541238 | Title: redundant assignment of bytes_left
Question:
username_0: Static analysis with Coverity has detected an assignment that is not used In lib/include/snow3g_common.h in function snow3g_8_buffer_ks_32_8():
3280 if (bytes_left >= small_block_size) {
3281 const uint32_t blocks = bytes_left / small_block_size;
3282
3283 length_sub(lengthInBytes, num_lanes, blocks * small_block_size);
CID 98731 (#1 of 1): Unused value (UNUSED_VALUE)
assigned_value: Assigning value from bytes_left - blocks * 8UL to bytes_left here, but that stored value is not used.
3284 bytes_left -= blocks * small_block_size;
3285
Same issue also in lines:
3151 if (bytes_left >= small_block_size) {
3152 const uint32_t blocks = bytes_left / small_block_size;
3153
3154 length_sub(tLenInBytes, num_lanes, blocks * small_block_size);
CID 98724 (#1 of 1): Unused value (UNUSED_VALUE)
assigned_value: Assigning value from bytes_left - blocks * 8UL to bytes_left here, but that stored value is not used.
3155 bytes_left -= blocks * small_block_size;
3156
Perhaps this can be removed?
Answers:
username_1: Sure, thanks for reporting this. Will fix this shortly.
Status: Issue closed
|
mathshistory/mathshistory-site | 525029069 | Title: Add pages for countries
Question:
username_0: The current site has a page for each country, displaying the mathematicians that were born in that country. It would be nice to have similar pages on the new site - perhaps with the new-style map of that country as well?<issue_closed>
Status: Issue closed |
Reisyukaku/loader | 170075180 | Title: Weird memory issues
Question:
username_0: I've tried for too long to figure out what causes these issues that cause good patches to not work. Earlier commits that used the single file 'patches.dat' loaded patches just fine, but as soon as i add in some extra code to load multiple files, things break. I can't use malloc, therefore, i suspect parts of code that need it like [this](https://github.com/username_0/loader/blob/region-change/source/patcher.c#L199) probably are doing undefined behavior. FBI uses calloc for this. |
backend-br/vagas | 668069921 | Title: [Niterói - RJ] Back-end Developer @ Internit
Question:
username_0: ## Descrição da vaga
Saca de PHP e MySQL?
Estamos abrindo mais uma vaga de estágio na Internit, empresa localizada em Niterói que há 23 anos desenvolve sites, lojas virtuais, portais, aplicativos, intranets e outros sistemas web.
## Local
Hoje - Remoto
Pós Covid - Niterói / RJ
## Benefícios
Bolsa + Vale Transporte + Seguro contra acidentes pessoais
#### Diferenciais
Ótimo ambiente de trabalho somado à desafios constantes.
Pós pandemia: Estamos situados em uma rua tranquila com facilidade de estacionamento além de termos um bicicletário na empresa.
## Requisitos
**Obrigatórios:**
Estar na faculdade em cursos da área de TI, conhecimento de PHP, MySQL, Java Script, Html5, Programação orientação a objeto e GIT;
**Desejáveis:**
- conhecimentos em Laravel
**Diferenciais:**
- Profissionais com experiência em Symfony
## Contratação
Estágio
## Nossa empresa
Em 1996 a internet comercial dava os seus primeiros passos. Dois jovens resolveram apostar nesse imenso potencial e abrir uma empresa pioneira. Mas eram tão jovens quanto a Internet e precisaram pedir emancipação aos pais para obter registro de pessoa jurídica. E assim nasceu a Internit, com DNA 100% digital, pronta para crescer junto com a web.
Nossas especialidades são a criação de sites, hotsites, portais, lojas virtuais, blogs, sistemas, aplicativos, CRMs, apresentações em multimídia, intranets e extranets – sempre alinhando bom gosto em design, facilidade de navegação e programação avançada. Foram entregues mais de 1000 projetos
## Como se candidatar
Acesse http://www.internit.com.br/curriculo e cadastre seu currículo.
### Alocação:
- Flexível (Remoto parcial)
### Contratação:
- Estágio
### Experiência
- Estágio |
ilyavoronin/preprint_server | 610262562 | Title: Too verbose logging
Question:
username_0: I believe we should move reference parser log messages to DEBUG level.
For each PDF these messages would be enough:
```
20:32:50.683 [main] INFO com.preprint.server.pdf.PdfHandler - downloading 53: 1307.6490
20:33:00.547 [main] INFO com.preprint.server.pdf.PdfHandler - Success: downloaded
20:33:00.886 [main] INFO com.preprint.server.ref.custom.ReferenceParser - Found 78 references
20:33:00.887 [main] INFO com.preprint.server.ref.CustomReferenceExtractor - Parsed 78 references
20:33:16.451 [main] INFO com.preprint.server.ref.ReferenceExtractor - Validated 64 out of 78
```
It is also good to show 53 out of X, where X is the expected amount of files to download. It will help track download progress.<issue_closed>
Status: Issue closed |
hujiulong/action-qiniu-upload | 754849412 | Title: 上传失败
Question:
username_0: 你好,使用这个 action 上传文件失败了,此处只有一个警告,不知道是否影响运行。
```
(node:3647) [DEP0005] DeprecationWarning: Buffer() is deprecated due to security and usability issues. Please use the Buffer.alloc(), Buffer.allocUnsafe(), or Buffer.from() methods instead.
```
Answers:
username_1: 这个是qiniu sdk内部的警告,对功能没有影响。有别的报错吗
username_0: 没有再试了,上传太慢了,放弃...
Status: Issue closed
|
Triple-T/gradle-play-publisher | 610501599 | Title: Property 'extension.serviceAccountCredentials' does not exist.
Question:
username_0: * Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
BUILD FAILED in 5m 16s
37 actionable tasks: 37 executed
Running after_script
00:01
Uploading artifacts for failed job
```
My gradle looks like
```
buildscript {
ext.kotlin_version = '1.3.61'
repositories {
jcenter()
mavenCentral()
google()
maven{url 'https://oss.sonatype.org/content/repositories/snapshots'}
}
dependencies {
classpath 'com.android.tools.build:gradle:3.6.0'
classpath "org.jetbrains.kotlin:kotlin-gradle-plugin:$kotlin_version"
classpath 'com.google.gms:google-services:4.2.0'
// in the individual module build.gradle files
classpath 'com.github.triplet.gradle:play-publisher:2.0.0-SNAPSHOT'
}
}
allprojects {
repositories {
google()
jcenter()
maven {
url "https://maven.google.com"
}
}
}
task clean(type: Delete) {
delete rootProject.buildDir
}
```
and app.gradle
```
apply plugin: 'com.android.application'
apply plugin: 'kotlin-android'
apply plugin: 'kotlin-kapt'
apply plugin: 'kotlin-android-extensions'
apply plugin: 'com.google.gms.google-services'
apply plugin: 'com.github.triplet.play'
android {
compileSdkVersion 29
[Truncated]
enabled = true
}
buildTypes {
release {
minifyEnabled false
proguardFiles getDefaultProguardFile('proguard-android-optimize.txt'), 'proguard-rules.pro'
signingConfig signingConfigs.release
}
}
}
play {
track = 'alpha' // set track for playstore like 'production','beta','alpha'
serviceAccountEmail = '[email protected]<EMAIL>account.com' //paste your service account email
//jsonFile = file('/PlayJson/play-store-key.json') // copy this line as it is do not modify
serviceAccountCredentials = file("api-7196515771782491829-266529-a514570d0422")
}
```
Please help me
Answers:
username_1: The file doesn't exist 🤷♂️. You maybe meant `serviceAccountCredentials = file('/PlayJson/play-store-key.json')`.
Status: Issue closed
username_1: Remove `serviceAccountEmail`.
username_1: Remove the line with `jsonFile`.
username_0: Thank you, Build Success
username_1: Cool. BTW, I'd highly recommend using the latest version (2.7.5) and not plain 2.0.0.
username_0: @username_1 Okay, i will change it later. Btw, can you help me, what this error mean? it was the result of my pipeline

username_1: It means that you didn't create an app with that package name in the play console. Or that your credentials don't have access.
username_0: Okay, thankyou man |
dadi/cli | 267127960 | Title: Logo rendering incorrectly on ubuntu
Question:
username_0: This is how the logo appears on ubuntu, for me at least.
<img width="231" alt="screen shot 2017-10-20 at 11 08 55" src="https://user-images.githubusercontent.com/1352469/31816227-6d5bce1a-b587-11e7-911e-0a55e3ba31d3.png">
Answers:
username_1: @username_0 what font do you use in the terminal?
username_0: @username_1 it's a monotype. Found the issue. There's a setting in iterm for treating ambiguous spacing characters as double. It's a default in Window command prompt. I'm looking into it.
username_2: I like it, very avant garde
username_0: So the issue is with the character code. U2593 has no fixed width, but U2591 does. This means one will render okay on all clients and the other is kind of a guess.
Here are the two:
Current (Dark shade)
```
▓▓▓▓▓ ▓▓▓▓▓▓▓
▓▓▓▓
▓▓▓▓▓▓▓ ▓▓▓▓
▓▓▓▓
▓▓▓▓▓▓▓
```
Alternative (Mid shade)
```
░░░░░ ░░░░░░░
░░░░
░░░░░░░ ░░░░
░░░░
░░░░░░░
```
username_3: One for @username_2 to decide.
username_3: When you test this, could you share a screenshot of how the logo with the new character looks in the console?
Status: Issue closed
username_3: If this is still an issue, I suggest we move it to https://github.com/dadi/boot, where the logo is being used and pulled into various products. Closing here. |
JuliaLang/julia | 187250893 | Title: k-dimensional (k≥4) array allocation with `Array{type, k}(n₁, n₂, ..., nₖ)`
Question:
username_0: We can allocate memory for a 3-dimensional array by `Array{type,3}(n₁,n₂,n₃)`. For example,
```
julia> A = Array{Int,3}(2,2,2)
2×2×2 Array{Int64,3}:
[:, :, 1] =
4691626752 4691639296
4691626976 4691639408
[:, :, 2] =
4691639856 4691755920
4691639744 0
```
However, the same allocation method does not work for `k`-dimensional array with `k≥4`. For example,
```
julia> B = Array{Int,4}(2,2,2,2)
ERROR: MethodError: no method matching Array{Int64,4}(::Int64, ::Int64, ::Int64, ::Int64)
Closest candidates are:
Array{Int64,4}{T}(::Any) at sysimg.jl:53
```
Is this an expected behavior? I am currently using `B = fill(0,2,2,2,2)` instead as a workaround.
Here is the version info:
```
julia> versioninfo()
Julia Version 0.5.1-pre+4
Commit 887ad05* (2016-10-28 19:39 UTC)
Platform Info:
System: Darwin (x86_64-apple-darwin16.0.0)
CPU: Intel(R) Core(TM)2 Duo CPU T9900 @ 3.06GHz
WORD_SIZE: 64
BLAS: libopenblas (USE64BITINT DYNAMIC_ARCH NO_AFFINITY Penryn)
LAPACK: libopenblas64_
LIBM: libopenlibm
LLVM: libLLVM-3.7.1 (ORCJIT, penryn)
```
Answers:
username_1: Dup of https://github.com/JuliaLang/julia/issues/18459. Fixed on master AFAICT.
It's a new feature and it is possible to work around it so it's unlikely going to be backported.
Status: Issue closed
|
JuliaLang/julia | 523555499 | Title: Distributed test failure: `Future` error propagation broken (testval / put)?
Question:
username_0: Don't think I've seen this one ever before:
```
Distributed (3) | started at 2019-11-15T02:59:31.251
From worker 3: Test Failed at D:\buildbot\worker-tabularasa\tester_win64\build\share\julia\stdlib\v1.4\Distributed\test\distributed_exec.jl:213
From worker 3: Expression: testval == 1
From worker 3: Evaluated: 0 == 1
From worker 3: ERROR: LoadError: There was an error during testing
From worker 3: in expression starting at D:\buildbot\worker-tabularasa\tester_win64\build\share\julia\stdlib\v1.4\Distributed\test\distributed_exec.jl:213
```
https://build.julialang.org/#/builders/32/builds/6108/steps/2/logs/stdio<issue_closed>
Status: Issue closed |
emepyc/project-score | 482587152 | Title: include cancer type column in table
Question:
username_0: @username_1, I don't remember if we agreed on this on not, but I suppose this is now needed for #9 ?
Answers:
username_1: For the table on the gene page, this could be useful - maybe just instead of Tissue. I don't think it's needed for #9, see my comment there
username_0: If unsure I will not implement this for now. We can revisit based on user input
Status: Issue closed
|
OpenCMISS/manage | 136230669 | Title: Include files in Iron project for Visual Studio
Question:
username_0: Not sure if this is a manage issue (i.e., cmake) or an iron issue but the iron visual studio project does not have the include files used in iron visible from visual studio (i.e., we should have an include folder as well as a src folder). Not sure if this means that the iron project does not depend on the include files or not? If not then it probably should as a change to an include file should trigger a re-build.
Answers:
username_1: IDEs like Visual Studio are a bit different from other projects and sometimes you need to add more information into the CMakeLists.txt to make files visible. Without fully looking into it seems like a CMake issue on the Iron repository.
username_2: yap there's more VS-specific options and CMake commands/vars that can be used to make the generated VS projects nicer and more organized. anyone keen to add more structure to it feel free :-)
username_1: This issue was moved to OpenCMISS/iron#50
Status: Issue closed
|
knopp/flutter_reorderable_list | 326415040 | Title: Callback on drag complete
Question:
username_0: Hi Matej,
Thanks for this, been using it for a while now without any problems - more than a proof of concept, it's great! Fluid too!
I want to write the reordered items to a backend. At the moment I am doing it async in the onReorder callback which is not ideal because it's called for each position change. i.e. while dragging. Would prefer to do on drag complete, i.e. when the user releases the drag and the item is in the final position.
An onReorderComplete or onDragComplete callback might be useful.
Answers:
username_1: I definitely like the idea. I'll add the callback as soon as I get back to flutter related work.
username_2: Hi, just adding to the idea, the onReorderComplete could provide as a parameter the items which were affected. For instance, in the example bellow, moving down the item 2 affects only 3 and 4.
```
1 1
2 3 --+
3 4 --+ only these 3 were affected
4 2 --+
5 5
```
BTW, I think it's worth mentioning this feature request flutter/flutter#16763 .
username_1: Callback is there now (`onReorderDone`).
Status: Issue closed
username_3: I'm using your example in a Cupertino app and while onReorder is called, onReorderDone is never called. The dragging overlay item is never removed and the reorder listener can't be triggered again.
username_1: Is there any exception in logs? Looks like something goes wrong during reorder finish.
username_3: No errors in log. I'm going to debug more in depth to see if I can figure out what's happening.
username_3: Result of debugging into end() of _ReorderableListState.
`end(DragEndDetails details) async {
if (_dragging == null) { //_dragging is not null
return;
}
_hapticFeedback();
if (_scrolling) { // _scrolling is false
var prevDragging = _dragging;
_dragging = null;
SchedulerBinding.instance.addPostFrameCallback((Duration timeStamp) {
_dragging = prevDragging;
end(details);
});
return;
}
if (_scheduledRebuild) { // _scheduledRebuild returns false
SchedulerBinding.instance.addPostFrameCallback((Duration timeStamp) {
if (mounted) end(details);
});
return;
}
this._scrollable.position.removeListener(this._scrolled);
var current = _items[_dragging]; // current is null
if (current == null) return; // So we leave method.
final originalOffset = _itemOffset(current);
final dragProxyOffset = _dragProxy.offset;
_dragProxy.updateWidget(current.widget
.childBuilder(current.context, ReorderableItemState.dragProxyFinished));
_finalAnimation = new AnimationController(
vsync: this,
lowerBound: 0.0,
upperBound: 1.0,
value: 0.0,
duration: Duration(milliseconds: 300));
_finalAnimation.addListener(() {
_dragProxy.offset =
lerpDouble(dragProxyOffset, originalOffset, _finalAnimation.value);
_dragProxy.shadowOpacity = 1.0 - _finalAnimation.value;
});
_recognizer?.dispose();
_recognizer = null;
await _finalAnimation.animateTo(1.0, curve: Curves.easeOut);
if (_finalAnimation != null) {
_finalAnimation.dispose();
_finalAnimation = null;
final dragging = _dragging;
_dragging = null;
_dragProxy.hide();
current.update();
_scrollable = null;
if (widget.onReorderDone != null) {
widget.onReorderDone(dragging);
}
}
}`
I think I have further to dig.
username_3: It appears to be something about the way I'm setting keys. If I use UniqueKey() the dragging behavior is normal but never ends. If I use ValueKey() with index as in example the dragging behavior is abnormal (the dragging pushes other items around like a multi-item drag) but does end. If I figure this out I'll update.
username_1: There's definitely something wrong with the keys. It seems like your keys are not stable, i.e. the keys for items change after reorder. What index exactly do you store in the key? Because for each item the example stores the original index. Which doesn't change after reorder. Maybe you're using current index (which changes after each reorder)?
Think of the key as unique identifier of each item. It must not change after reorder. UniqueKey should work, as long as you only create instance for each item once (and then remember it for each item).
username_3: Maybe it's the StreamBuilder I'm using which rebuilds when other elements change (the reorderable list elements are part of a larger structure) and the keys are being reassigned. Also, these reorderable elements can be changed or deleted remotely. This might be a challenge.
username_1: If elements are changed or deleted remotely, you can provide cancellation token to the list and call `cancelDragging` on it. This will interrupt dragging in order to prevent inconsistent results.
As for the rest, you should never need to rebuild the whole list during dragging. From the callback, you should only update your model data. Rebuild of items is triggered automatically.
And for the keys, it really shouldn't be that complicated. Don't your items have any kind of unique identifier?
username_3: It does, and I'm using them. The reorderable list is a copy of the original list which is updated from the copy list when reorder is complete. I added a boolean flag to reduce the chances of unintended initializations of the copy list and this has helped some. onReorderDone now gets called but the animations seem wrong. The dragging item never drops down back into the list after the drag ends. Wondering if that's because in onReorderDone() I fire off a bloc dispatch to update the original data which then rebuilds the whole screen and it doesn't have a chance? |
webtrees/geodata | 299060971 | Title: tools/export: Failed to create flag for England/Cornwall/Isles of Scilly/flag.svg
Question:
username_0: ```bash
Failed to create flag for England/Cornwall/Isles of Scilly/flag.svg
Unable to read image from binary data.
```
No clue what went wrong here, but may be related to the next bug in the output…
Answers:
username_1: It a problem with the `imagemagick` library.
As far as I can tell, the `.svg` file uses a feature that is not supported by the library.
username_1: Originally, webtrees used `.PNG` format because few browsers supported `.SVG`.
This is no longer true.
Maybe we can use the `.SVG` files directly?
username_0: I would think so, yes. |
defunctzombie/node-uuid | 171768089 | Title: Check if id is a valid uuid?
Question:
username_0: Is it possible to use this library to check if a given id is a valid uuid? Or do I need to fall back to regular expressions?
Answers:
username_1: Not currently possible via the library. We have considered adding such functionality via regex tho. If you make a PR we can review it.
username_2: I've provided a simple check function using a regular expression in #21 . This only works for strings right now, let me know if a version that can handle buffers is also needed.
username_0: Great! That's exactly what I wanted. I don't personally have a need for checking on buffers and I would think it would be less common for this use case.
Status: Issue closed
username_1: project has moved here and is being consolidated with node-uuid module https://github.com/kelektiv/node-uuid |
ARGOeu/ARGO | 101991212 | Title: Connectors should verify server certficate
Question:
username_0: Standard Python 2 HTTPS methods expects cafile with all concatenated CA certificates and build manually CA chain. We don't have it since CA chain is built automatically with symlinks in capath. Since we cannot pass it to standard HTTPS methods, we need to empower external library like PyOpenSSL that can figure out CA chain from passed capath.<issue_closed>
Status: Issue closed |
backend-br/vagas | 555181073 | Title: [São Paulo] Back-End Developer Python ou Node na [Pontte]
Question:
username_0: Caso esteja pensando em fazer uma mudança de trabalho ou de propósito, segue uma proposta fresquinha para uma start up bem legal, a Pontte: uma plataforma de empréstimo digital.
Procuramos um Software Engineer Sr. (full stack) - Python ou Node.
**Responsabilidades:**
*Ajudar na evolução da arquitetura e no desenvolvimento da aplicação.
* Monitoramento do sistema.
* Desenvolver e treinar outros engenheiros de software, ajudando-os a avançar em suas carreiras.
* Compartilhe conhecimento de forma eficaz.
* Inspire e conduza a excelência operacional e de engenharia entre as equipes.
* Apoie a equipe de engenharia na obtenção de um alto nível de qualidade técnica, confiabilidade e facilidade de uso.
Qual perfil estamos buscando?
* Desejável 6+ anos de experiência com desenvolvimento de software.
* Conhecimento em arquitetura de software e boas práticas de desenvolvimento.
* Desejável experiência em uma das nossas linguagens utilizadas ( NodeJS/Python).
* Ter conhecimento com HTML, CSS e algum framework em JavaScript, de preferência REACT.
* Capacidade de elaborar testes unitários e de integração.
* Desejável conhecimento em CI/CD.
* Algum conhecimento em arquitetura serverless na AWS.
Contrato CLT, <NAME>, SP- presencial
BENEFÍCIOS:
- VT
- Incentivo aos estudos: R$ 1.500,00 / ano
- Plano de Saúde e Odontológico pago pela Pontte para ele e seus dependentes
- Gympass com até 75% de desconto
- Pipoca e frutas todos os dias
Answers:
username_1: Olá! Tudo bem?
Essa vaga pode ser remota?
Abraços!
http://danielsiqueira.net
username_2: Eu tenho interesse, já estudo/pratico desenvolvimento a mais de 6 anos ( desde a faculdade ), mas trabalhando de fato tenho 3 anos ( já estou na posição Pleno/Sr a mais de 1 ano), trabalhei diretamente com toda stack de vocês(inclusive configuração de CI/CD pela AWS. Caso queria dar uma chance meu email é <EMAIL>.
username_1: <NAME>!
Sim , somente remoto, porque sou do RJ.
Se houver a possibilidade de atuar remotamente CLT, gostaria de me candidatar.
Meu e-mail: <EMAIL>.
Abraços!
username_1: <NAME>!
Sim , somente remoto, porque sou do RJ.
Se houver a possibilidade de atuar remotamente CLT, gostaria de me candidatar.
Meu e-mail: <EMAIL>.
Abraços!
Status: Issue closed
username_3: Issue fechada pelo bot. Motivo: Sem interações em um periodo de 3 meses. |
Subsets and Splits