repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
jsphon/colony | 228411287 | Title: Are the AsyncWorkers, with multiple Threads, a smell
Question:
username_0: In the diagrams, asyncnodes with multiple threads are still displayed as single nodes, which feels misleading.
Also, if there is a target_class instance, then each worker will have its own instance. Which could be confusing.
How could we overcome this? |
nesbox/TIC-80 | 1097289047 | Title: Questionable addition to wiki
Question:
username_0: Do we want this in the Wiki, @nesbox? I know we tell people how to build it from source (and don't do anything to discourage that) but do we want people forking it just to force and distribute the PRO version?
Answers:
username_0: Nevemind, you saw.
Status: Issue closed
|
nattthebear/ff12characterplanner | 850747142 | Title: Do blocked attacks take longer?
Question:
username_0: I saw a stat somewhere that said a blocked attack had a longer animation time to it. If this is true, then does it apply to bows as well? Currently, the simulator doesn't account for this in any way. |
jmix-framework/jmix | 1174459278 | Title: Non-editable DateField does not hide asterisk
Question:
username_0: ### Description
See Russian forum: [topic](https://forum.jmix.ru/t/na-datefield-otobrazhaetsya-otmetka-obyazatelnosti-zapolneniya-dazhe-esli-ekran-v-rezhime-readonly/248).
#### Steps to reproduce
1. Download attached project: [demo.zip](https://github.com/Haulmont/jmix-ui/files/7719682/demo.zip)
2. Open: _Application -> First store entities -> Order_
3. Try to "view" some record.
**ER**
Fields should be non-editable and without asterisk
**AR**
Only DateField has an asterisk |
seek-oss/capsize | 658646144 | Title: rems possible?
Question:
username_0: Given that there can be accessibility issues of setting font sizes by px. Can this hold up with returning rems instead of px values?
More info: https://css-tricks.com/is-it-better-to-use-ems-rems-than-px-for-font-size/
Answers:
username_1: I used rems in tailwind-baseline (https://github.com/username_1/tailwind-baseline/blob/master/src/utils.js) based on the old basekick (https://github.com/username_2/basekick) calculations, but I'm not 100% sure this works correctly in all cases. At least some rounding errors might occur. Interested in this too.
username_2: Just cleaning up ready for publishing and also adding support for fontSize input.
There have been a number of improvements and refinements to the algorithm this week. Once the dust settles, I am curious to investigate em/rem and/or unitless line heights.
Status: Issue closed
|
backstage/backstage | 931516380 | Title: Issues using LDAPS (ActiveDirectory) UNABLE_TO_GET_ISSUER_CERT_LOCALLY
Question:
username_0: Trying to connect Backstage to LDAPS (Active Directory) but I get `UNABLE_TO_GET_ISSUER_CERT_LOCALLY`.
Tried the following
```
yarn config set strict-ssl false
npm config set strict-ssl false
```
No change
Tried to add `cafile` to both yarn and npm containing bort server certificate and Issuing CA in a bundle
```
yarn config set cafile /workspaces/bs/backstage/cert.pem
npm config set cafile /workspaces/bs/backstage/cert.pem
```
No change
Tried to add the certificate to OS (Using devcontainer and `mcr.microsoft.com/vscode/devcontainers/typescript-node:0-14` (Debian) )
```
2021-06-28T10:29:26.231Z catalog warn LDAP client threw an error, UNABLE_TO_GET_ISSUER_CERT Error: unable to get issuer certificate type=plugin
[1] 2021-06-28T10:29:26.232Z catalog warn Processor LdapOrgReaderProcessor threw an error while reading ldap-org:ldaps://ldapserver; caused by LDAP bind failed for CN=LDAP-Bind,OU=Users,OU=MYORG,DC=domain,DC=net, UNABLE_TO_GET_ISSUER_CERT Error: unable to get issuer certificate type=plugin
```
Still issues - Do you have any tip on solution(s) for this?
## Your Environment
`.devcontainer.json`
```
{
"image": "mcr.microsoft.com/vscode/devcontainers/typescript-node:0-14",
"forwardPorts": [3000],
"postCreateCommand": "bash -i -c 'npm install --global yarn'"
}
```
```
$ node -v
v14.17.0
$ npm -v
6.14.13
$ yarn -v
1.22.10
```
Answers:
username_1: Can you make it work when running locally? Probably best to start there, to be sure that the environment isn't putting road blocks in our way.
And this may sound crazy - but are you sure that it's ldaps: rather than just ldap:? Very out-there idea, but let's double check anyway :)
[Here's some info](https://stackoverflow.com/questions/31673587/error-unable-to-verify-the-first-certificate-in-nodejs) about possibilities of adding local cert trust, but it seems like the problem is failure to read the cert at all from remote.
username_0: Thanks @username_1 for your input
I got it working locally (still with devcontainers) by setting this:
`export NODE_TLS_REJECT_UNAUTHORIZED=0` and yes we are only using ldaps.
I guess it has something to do with how NodeJS handles this and we are currently in a PoC of Backstage so no big deal since I now could try out the LDAP integration.
If/when we decide to implement Backstage we will throw some more hard core NodeJS developers on it ;) So I propose to close this issue.
Status: Issue closed
username_2: I have the same issue and i try to add :
`export NODE_TLS_REJECT_UNAUTHORIZED=0`
before executing :
`yarn dev`
I miss somethings ?
Thanks in advance
username_0: We finally did it like:
`package.json`
```
"scripts": {
"dev": "NODE_EXTRA_CA_CERTS=${PWD}/certs/CompanyRootCA.pem concurrently \"yarn start\" \"yarn start-backend\"",
"start": "yarn workspace app start",
```
and when building Docker image:
`Dockerfile`
```
COPY /certs/CompanyRootCA.pem ./
ENV NODE_EXTRA_CA_CERTS=CompanyRootCA.pem
``` |
microsoft/vscode-js-debug | 600614999 | Title: Support for capturing heap and tracing-API-based profiles
Question:
username_0: This issue is a good one for Grace Hopper / Hacktoberfest!
Collecting a heap profile here should be possible with plug-ins to our existing functionality. [CDP docs](https://chromedevtools.github.io/devtools-protocol/v8/HeapProfiler) -- you'll probably want to look at the sampling heap profile initially.
You want to create a new [IProfiler class here](https://github.com/microsoft/vscode-js-debug/blob/48de86f3cc8ec2d629add5e5915be84b1702090e/src/adapter/profiling/index.ts#L86). The interfaces should guide you through what's needed. You'll probably want to abstract and reuse some of the location-massaging code from the CpuProfiler.
Once it's there, the "Start Profiling" command (available when running a Node.js program) will open a picker asking you what kind of profile you want to take.
Additional extra things you could do once you have that working:
- Currently, profiling is duration based, but users could also want to take point-in-time snapshots of their heap. You can add support for this, which will involve some changes in the [ui side of things](https://github.com/microsoft/vscode-js-debug/blob/909e4f1d33860f305bc9ad4fd96c072bf366672c/src/ui/profiling/uiProfileManager.ts#L135-L143) since there's no 'termination condition'
- You can add support for viewing the resulting heap profile files to the visualizer: https://github.com/microsoft/vscode-js-profile-visualizer
General contributing/setup: https://github.com/microsoft/vscode-js-debug/blob/master/CONTRIBUTING.md#development
Answers:
username_1: This is exciting.
I suppose there needs to be an option to `start` and `stop` right?
So we can pick and choose which bits we want to capture
username_0: This issue is a good one for Grace Hopper / Hacktoberfest!
Collecting a heap profile here should be possible with plug-ins to our existing functionality. [CDP docs](https://chromedevtools.github.io/devtools-protocol/v8/HeapProfiler) -- you'll probably want to look at the sampling heap profile initially.
You want to create a new [IProfiler class here](https://github.com/microsoft/vscode-js-debug/blob/48de86f3cc8ec2d629add5e5915be84b1702090e/src/adapter/profiling/index.ts#L86). The interfaces should guide you through what's needed. You'll probably want to abstract and reuse some of the location-massaging code from the CpuProfiler.
Once it's there, the "Start Profiling" command (available when running a Node.js program) will open a picker asking you what kind of profile you want to take.
Additional extra things you could do once you have that working:
- Currently, profiling is duration based, but users could also want to take point-in-time snapshots of their heap. You can add support for this, which will involve some changes in the [ui side of things](https://github.com/microsoft/vscode-js-debug/blob/909e4f1d33860f305bc9ad4fd96c072bf366672c/src/ui/profiling/uiProfileManager.ts#L135-L143) since there's no 'termination condition'
- You can add support for viewing the resulting heap profile files to the visualizer: https://github.com/microsoft/vscode-js-profile-visualizer
General contributing/setup: https://github.com/microsoft/vscode-js-debug/blob/master/CONTRIBUTING.md#development
username_2: Any idea when this might get added? We would love to use this functionality |
nf-core/sarek | 953075352 | Title: [BUG] using with custom genome
Question:
username_0: <!--
# nf-core/sarek bug report
-->
Hi,
I tried to use the Sarek pipe line with costume genome, without dbSNP or known indels. I couldn't make it work with the following, wonder if this pipeline is capable of doing it or did I do it wrong? Thanks!
nextflow run /path/variant_calling/sarek \
-profile docker \
--tools Manta, Strelka \
--input ${input} \
--outdir ${sarek_result} \
--genome "custom" \
--fasta ${ref_fa} \
--bwa $bwa \
--fasta_fai ${fa_fai} \
--email ${email} \
--max_cpus 16 \
--publish_dir_mode "copy" \
-with-report report.html
Answers:
username_1: @username_0
I ran:
```
nextflow run nf-core/sarek -r 2.7.1
-profile docker,test
--genome custom
--fasta https://raw.githubusercontent.com/nf-core/test-datasets/sarek/reference/human_g1k_v37_decoy.small.fasta
--tools manta,strelka
```
without issue, what was your exact command line, and what was your sarek version?
username_0: Thanks. I think the difference from this test case for me is that I am not working with human genome. Not sure if that makes a difference or not.
Here is the command I am using:
nextflow run nf-core/sarek \
-profile docker \
--fasta '/path/Sus_scrofa.Sscrofa11.1.dna.chromosome.MT.fa' \
--input '/path/mtDNA_input.tsv' \
--outdir '/path/mtDNA/results' \
--tools manta, Strelka \
--skip_qc all \
--save_bam_mapped \
--no_strelka_bp \
--igenomes_ignore \
--cpus 16 \
--email <EMAIL>
This is the error message I got:
----------------------------------
Error executing process > 'CreateIntervalBeds (wgs_calling_regions.hg38.bed)'
Caused by:
Process `CreateIntervalBeds (wgs_calling_regions.hg38.bed)` terminated with an error exit status (2)
Command executed:
awk -vFS=" " '{
t = $5 # runtime estimate
if (t == "") {
# no runtime estimate in this row, assume default value
t = ($3 - $2) / 1000
}
if (name == "" || (chunk > 600 && (chunk + t) > longest * 1.05)) {
# start a new chunk
name = sprintf("%s_%d-%d.bed", $1, $2+1, $3)
chunk = 0
longest = 0
}
if (t > longest)
longest = t
chunk += t
print $0 > name
}' wgs_calling_regions.hg38.bed
Command exit status:
2
Command output:
(empty)
Command error:
awk: cannot open wgs_calling_regions.hg38.bed (No such file or directory)
---------------------
It somehow still tried to find some hg38 reference bed file.
Thanks
username_1: @username_0
You hould have both options in your command line:
`--genome custom --igenomes_ignore` |
DIYgod/RSSHub | 739996303 | Title: RSS Proposal: Voicetank
Question:
username_0: <!--
Please ensure the RSS proposal is not listed in [documentation](https://docs.rsshub.app/en) or [issue](https://github.com/DIYgod/RSSHub/issues), website doesn't provide this kind of RSS feed, and provide all the information required by this template.
Otherwise the issue will be closed immediately.
We are flooded with feature requests and short-handed, please try to make it yourself, the [guide](https://docs.rsshub.app/en/joinus) is a good place to start. Submit a pull request when done!
-->
### Website URL
https://www.voicettank.org/
### Website description
新媒体
### What content should be included?
“思想坦克”栏目下所有文章
### Additional description
Answers:
username_1: https://www.voicettank.org/blog-feed.xml
Status: Issue closed
username_2: 建议安装 RSS radar 插件 |
fossasia/pslab-android | 337210475 | Title: Add units to capacitance measurement
Question:
username_0: **Actual Behaviour**
Currently the capacitors values are shown in the form of power of e (scientific form)
**Expected Behaviour**
The capacitors should be shown with units such as pico farad or nano farad etc
**Steps to reproduce it**
Go to MultimeterAcitivity and change it.
**LogCat for the issue**
Not applicable
**Would you like to work on the issue?**
Yes
Answers:
username_1: Also chop off the decimals to two places
username_0: Sure @username_1 :+1:
Status: Issue closed
|
Code4SocialGood/c4sg-services | 214265317 | Title: Remove dependency between Frontend and Backend in local environment
Question:
username_0: _From @username_0 on December 29, 2016 21:49_
Currently, front end developers has to run back end locally to get data feed from Rest API. Setup architecture to enable front end developers to work on front end without running back end application locally.
_Copied from original issue: username_0/C4SG#8_<issue_closed>
Status: Issue closed |
SharePoint/sp-dev-docs | 884458031 | Title: Issue with SPFx 1.12.1 default Teams icon sizes
Question:
username_0: ### Target SharePoint environment
SharePoint Online
### What SharePoint development model, framework, SDK or API is this about?
💥 SharePoint Framework
### Developer environment
Windows
### What browser(s) / client(s) have you tested
- [ ] 💥 Internet Explorer
- [ ] 💥 Microsoft Edge
- [ ] 💥 Google Chrome
- [ ] 💥 FireFox
- [ ] 💥 Safari
- [ ] mobile (iOS/iPadOS)
- [ ] mobile (Android)
- [X] not applicable
- [ ] other (enter in the "Additional environment details" area below)
### Additional environment details
- browser version: N/A
- SPFx version: 1.12.1
- Node.js version: 10.17.0 (although not applicable)
### Describe the bug / error
The PNG images that are put in the teams folder for the color and outline icons that can be used by default are not the correct sizes and do not pass validation with current teams manifests.
Color icon is 92x92 instead of 192x192
Transparent icon is 20x20 instead of 32x32
### Steps to reproduce
1. Generate a hello world web part application using @microsoft/sharepoint yoman template
2. Try and use the image files in the Teams App Studio application and you will get a validation error.
### Expected behavior
The default images should meet the Teams manifest requirements.
Answers:
username_1: @username_0 , thanks to reporting this.
Yes: if you take the icons we scaffold in the Teams folder and use them in a Teams package (.zip file) and side load that directly in Teams rich client schema validation will kick in and will not allow sideloading.
However, "sync to Teams" from SPO App catalog bypass that validation and succeeds.
We took note of the issue and we will provide a fix in the next version of the generator (1.13).
username_2: The fix should be in https://www.npmjs.com/package/@microsoft/generator-sharepoint/v/1.13.0-beta.22 |
kongware/scriptum | 519933451 | Title: Add specialiced trampolines for tail recursion modulo XY
Question:
username_0: - tail call modulo cons
- tail call modulo addition
- tail call modulo multiplication
Status: Issue closed
Answers:
username_0: Instead of utilizing this compiler techniques we should just write the structural fold (i.e. the eleimination rule) for each distinct data type that needs one.
Alternatively CPS transformation can be applied, either manually or with the non-tail recursive trampoline/evaluator that is going to ship with scriptum soon.. |
furas/tensorflow-no-avx | 785323472 | Title: why it doesnt have tf.contrib?
Question:
username_0: `AttributeError: module 'tensorflow' has no attribute 'contrib'`
:(
Answers:
username_0: i use it for NudeNet
[detector.py](https://github.com/notAI-tech/NudeNet/blob/8211bda0fe01c07a9d067252f42adcf27ca1b24d/nudenet/detector.py#L66)
username_1: You have to import it
```
import tensorflow.contrib
```
and later use `tensorflow.contrib` instead of `tf.contrib`
Status: Issue closed
|
emberjs/ember.js | 429538898 | Title: RouterService.urlFor() broken when sub-route has same param name as parent route
Question:
username_0: Given the following route definition:
```javascript
Router.map(function() {
this.route('foo', {path: '/foo/:id'}, function() {
this.route('bar', { path: '/bar/:id' });
});
this.route('baz', {path: '/baz/:bazId'}, function() {
this.route('qux', { path: '/qux/:quxId' });
});
});
```
`RouterService.urlFor()` will return an erroneous URL for the route `foo.bar`, seemingly because their respective parametrized `path` both contain a param with the same name. As a control test, `urlFor()` behaves properly for the route `baz.qux`. See below:
```javascript
Ember.Controller.extend({
router: inject(),
fooBarUrl: computed(function() {
return this.router.urlFor('foo.bar', 123, 456);
// evaluates to "/foo/456/bar/456"
}),
bazQuxUrl: computed(function() {
return this.router.urlFor('baz.qux', 123, 456);
// evaluates to "/baz/123/qux/456"
})
})
```
Reproduction:
- [Twiddle](https://ember-twiddle.com/9476bd9be73ed9de2fd4b532af76b43f?openFiles=controllers.application.js%2C)
- [Gist](https://gist.github.com/username_0/9476bd9be73ed9de2fd4b532af76b43f)
Status: Issue closed
Answers:
username_1: This is working as intended. The way I think about dynamic segments is that there is a single `params` object, so if you have this situation both of the routes are setting the same param and the second will override the value.
Sorry for the confusion!
username_0: @username_1 Interesting. I did not realize this was forbidden as most aspects of the framework work just fine under this condition. Perhaps dups in param names should result in a runtime error (at least on dev builds) so that others don't fall in the same trap. |
DarkPacks/SkyFactory-4 | 502225768 | Title: Placing Block issue
Question:
username_0: ## Bug Report
After using the Building Gadget and placing a block onto between 2 or more of those blocks that were building with the Building Gadget, I get a lag spike. I notice my FPS drops around 50 to 100 FPS when I do this.
## Expected Behaviour
A block to be placed and no FPS drop or game freeze
## Possible Solution
Something to do with the Building Gadget
## Steps to Reproduce (for bugs)
https://youtu.be/u02wIZMDItY?t=14
## Logs
* Client/Server Log: https://gist.github.com/username_0/458ef517618c124e63e15faae32d4d52
## World Information
* Preset: Defualt
* Prestige: No
* Modpack Version world created in: 4.0.8
* Additional Content Installed: None
## Client Information
* Modpack Version: 4.0.8
* Java Version: 14.23.5.2838 ?
* Launcher Used: Twitch
* Memory Allocated: 8 MB
* Server/LAN/Single Player: Single Player
* Optifine Installed: No? Default mod pack
* Shaders Enabled: No? Default mod pack
Let me know if I need to provide more information! Thank you!
Answers:
username_1: We can`t watch your video, it is private
username_0: I fixed that!
username_2: Getting similar issues, but I don't think building gadget is the real cause here.
I get lag spikes when placing/removing blocks, including using pistons. The more blocks getting placed/removed/pistoned/etc, the larger the lag spike. Thus, with building gadgets often placing/removing a lot of blocks at once, the problem becomes a lot more obvious.
For me, the game is beginning to become unplayable as I approach late game and have contraptions that rely on pistons to work.
My guess would be something is doing expensive computations when blocks change.
So far, tried
- breaking all cyclic item collectors
- breaking all feral lanterns
- breaking deep mob learning trial keystone
- shutting down dropper-based lava production (I found a way to stop clogging with a mix of cyclic fluid extraction cable and super-laminar fluiducts)
None of those solved the lag issue, so I think it's unlikely (although not impossible) that they are the cause.
username_2: Update: Currently suspecting InControl as the root cause.
After some experimenting I found that the lag spike happens primarily **when the block change causes a potential change in spawning spaces**.
Example:
<details>
<summary>Flipping this lever causes lag spikes</summary>
<img src="https://user-images.githubusercontent.com/8021147/83800333-27040b80-a65c-11ea-9e95-bf6004767e1b.png"/>
<img src="https://user-images.githubusercontent.com/8021147/83800348-2c615600-a65c-11ea-92e6-37913191dc27.png"/>
</details>
<details>
<summary>To stop the lag spikes, place a block above the bookshelves to prevent spawning spaces from being created/removed</summary>
<img src="https://user-images.githubusercontent.com/8021147/83800653-a1cd2680-a65c-11ea-9f64-c2bf362b8b0a.png"/>
<img src="https://user-images.githubusercontent.com/8021147/83800656-a396ea00-a65c-11ea-8824-051bbd4dfcc5.png"/>
</details>
<details>
<summary>You are likely freeze your your game if you put a stack of blocks in the block placer. Don't do this.</summary>
(that is a block breaker on the opposite side)
<img src="https://user-images.githubusercontent.com/8021147/83800849-fbcdec00-a65c-11ea-8981-c071baef862f.png"/>
</details> |
Codeception/Codeception | 816296808 | Title: Call to undefined method ReflectionUnionType::getName()
Question:
username_0: #### What are you trying to achieve?
Run acceptance test with Webdriver (selenium)
#### What do you get instead?
PHP Fatal error: Uncaught Error: Call to undefined method ReflectionUnionType::getName() in ..... vendor\codeception\codeception\src\Codeception\Lib\Generator\Actions.php:228
### Details
* Codeception version: 4.1.18
* PHP Version: 8.0
* Operating System: Windows
* Installation type: Composer
Answers:
username_1: It seems that you have some helper method with union return type - `public function foo(): int|float`
If you want a quick fix, remove that return type.
username_0: Thank you, it is solved.
Status: Issue closed
username_1: #### What are you trying to achieve?
Run acceptance test with Webdriver (selenium)
#### What do you get instead?
PHP Fatal error: Uncaught Error: Call to undefined method ReflectionUnionType::getName() in ..... vendor\codeception\codeception\src\Codeception\Lib\Generator\Actions.php:228
### Details
* Codeception version: 4.1.18
* PHP Version: 8.0
* Operating System: Windows
* Installation type: Composer
username_1: It must be fixed properly.
Status: Issue closed
|
michael105/minicore | 911661424 | Title: Notes
Question:
username_0: So here is the masterplan for getting the core utils down to 64 kB..
1. Find what is bloating (bello world bloats from 181 to now 241 Bytes.)
2. Enable sort of plugin loading.
Do all work within a function, which is callen with arguments parsed. (E.g. stat(filename,fd))
Have this function with position independent code. Furthermore, the function needs to be located in the binary at a fixed place.
Load, e.g., in ls the binary into memory, and have a wrapper function:
wrap_stat(filename,fd){
if (!addrofstat)
mmap(...);...
asm("jmp address of stat");
}
Funny enough, the real stat's return will return to the place where wrap_stat is invoked.
I'm not really sure whether that's such a great idea 😂 but might work, also performant, and spare the needed bytes. I'll see
Answers:
username_0: Did it. Sort of reinventing shared libraries.
In mini.
Now - it is absolutely possible to get the core tools down to 64kB.
I'm however concerned with my original idea,
having static tools. I guess I have to think about it.
Maybe it is cheating. Maybe I should say, as long the basic functionality is given without loading external "plugins", it's ok.
Maybe I've to sort this out throughly |
Atrejoe/Inky-Calendar-Server | 938812271 | Title: Microsoft.Data.SqlClient.SqlException in /panel/e150487a-bbb4-4009-0912-08d7ba831bfb
Question:
username_0: ## Error in InkyCal
**Microsoft.Data.SqlClient.SqlException** in **/panel/e150487a-bbb4-4009-0912-08d7ba831bfb**
A connection was successfully established with the server, but then an error occurred during the login process. (provider: TCP Provider, error: 0 - No error information)
[View on Bugsnag](https://app.bugsnag.com/noyb-1/inkycal/errors/60e596319c1c360007e122d6?event_id=60e59631007f8f6b08be0000&i=gh&m=ci)
## Stacktrace
:0 - Microsoft.Data.ProviderBase.DbConnectionPool.TryGetConnection(DbConnection owningObject, uint waitForMultipleObjectsTimeout, bool allowCreate, bool onlyOneCheckConnection, DbConnectionOptions userOptions, out DbConnectionInternal connection)
[View full stacktrace](https://app.bugsnag.com/noyb-1/inkycal/errors/60e596319c1c360007e122d6?event_id=60e59631007f8f6b08be0000&i=gh&m=ci)
*Created automatically via Bugsnag* |
watchdogpolska/petycja-php.siecobywatelska.pl | 164610132 | Title: Możliwość wyświetlania konkretnego wpisu
Question:
username_0: Szanowny,
Z poziomu listy #4 powinna być możliwość przejścia do konkretnego wpisu, który wyświetla pełną treść artykułu.
Z wyrazami szacunku
Answers:
username_0: Aby porównać datę utworzenia i modyfikacji najlepiej utworzyć metodę w Modelu `/src/Model/Entity/Post.php`
````php
public function isHasBennModifited(){
// Lte = Less or equal
return $this->created->modify('+15 minutes')->lte($this->modified);
}
````
i potem można odwołać się do tej metody tak samo jak odwołujesz się do tytułu( `$post->title`) to do metody `$post->isHasBennModifited`. a ten kod może być nawet w wyrażenia struktury `if (expression);`
O Entity więcej jest w dokumentacji: http://book.cakephp.org/3.0/en/orm/entities.html
Status: Issue closed
|
jlippold/tweakCompatible | 349812276 | Title: `NoLowPowerAlert` working on iOS 10.2
Question:
username_0: ```
{
"packageId": "com.hiroshit.nolowpoweralert",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "com.hiroshit.nolowpoweralert",
"deviceId": "iPhone6,2",
"url": "http://cydia.saurik.com/package/com.hiroshit.nolowpoweralert/",
"iOSVersion": "10.2",
"packageVersionIndexed": true,
"packageName": "NoLowPowerAlert",
"category": "Tweaks",
"repository": "BigBoss",
"name": "NoLowPowerAlert",
"installed": "0.1-1",
"packageIndexed": true,
"packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.",
"id": "com.hiroshit.nolowpoweralert",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.1.0",
"shortDescription": "Disables low battery alerts",
"latest": "0.1-1",
"author": "hiroshit",
"packageStatus": "Unknown"
},
"base64": "<KEY>
"chosenStatus": "working",
"notes": ""
}
``` |
Draylar/identity | 1107021560 | Title: (Fabric 1.18.1) Piglins spam between hostile and neutral animations/sounds when using an identity that isn't attacked on first sight.
Question:
username_0: **Bug description**
I've seen this happen in the Fabric 1.18.1 version of the mod. Not sure about other versions.
Tested only in **Singleplayer**. _(I'm a loner.)_
If you use an identity that's not supposed to be attacked on first sight by the **Piglins** _(a hostile mob identity)_ but you don't wear any gold armor on yourself, the **Piglins** will start switching back-and-forth between their raised-sword and neutral animations and spamming their anger sounds.
**This does not have any effect on their behavior,** though, and they will remain neutral until they find another excuse to attack.
It might be client-side prediction issues, I've seen a few client-side prediction mess-ups on other games.
I also accidentally created a fork of your mod's code, didn't mean to do that. How do I delete it?
**Game log:**
*Should I really upload a log?*
Answers:
username_1: that happened to me as well, i just morphed back to a player with gold armor |
BabylonJS/Babylon.js | 343714286 | Title: Mesh CullingMode
Question:
username_0: # Feature Request
Add a new property to the Mesh class : cullingMode.
This could be set with static values defining the chosen way to check the culling of the given mesh.
Example of values :
* standard (default) : current culling algo
* bSphereOnly : skip the BBox culling test in the current algo
* optimisticInclusion : quick pre-test of mesh inclusion, then current algo (ex : mesh position inside the rectangle corresponding to the unprojected screen at the mesh position in the camera view system)
* optimisticExclusion : quick pre-test of mesh exclusion, then current algo (ex : bSphere outside the external conical frustum)
Answers:
username_1: Will prepare the work!
username_0: As both the optimistic modes (fast inclusion or exclusion) rely on the camera view matrix, maybe would it be nice to update this matrix once before calling isInFrustum() and not for each mesh ? knowing it's updated anyway before the rendering process ...
username_1: Not sure to get why you say updated for each mesh?
The camera view matrix is updated once before calling the frustum test
username_0: ok
what I was afraid of (not read the full code so far) was that we compute
the cam view matrix on each call to isInFrustum(), so for each mesh.
username_1: Doc: https://github.com/BabylonJS/Documentation/blob/master/content/How_To/scene/Optimizing_your_scene.md#changing-culling-strategy
username_1: Ok first stab done! Feel free to chime in when you can!
username_1: You are right! I'll change that
username_1: @username_0 still interested? :)
username_0: yes... I was wondering if the stuff had started or not
I'll have a look then (I'm on a f..g php code for my "official" work in
the same time)
username_1: It is done and waiting for your updates;)
username_0: Optimistic exclusions are cancelled because they can't really compete the current exclusion tests.
Optimistic Inclusions added in complement on standard and bSphereOnly processes.
#5624
Tests done on my computer with 83 meshes, 81 moving and 2 static (a ground and a column) :
- switching to bSphereOnly processes (with or with optimistic inclusion mode) always gives a good gain vs standard process, whatever none, some or all meshes are in the frustum. It's expected because there aren't any longer the BBox evaluations (so 48 iterations less).
- switching to Optimistic Inclusion processes (with or without bSphereOnly mode) gives a very little gain when all the meshes are visible and no sensible loss when no mesh is visible.
In brief, Optimistic Inclusions work but give very little gain. They keep the same accuracy than the basic mode on what they are applied (standard or bSphereOnly).
bSphereOnly, because it reduces a lot the accuracy, gives a good perf gain. This should not be used with high poly meshes while sending false positives to the GPU has a real rendering cost. This can be very interesting for numerous low poly meshes instead.
Status: Issue closed
|
kokorin/Jaffree | 1021048433 | Title: Image written out at wrong timestamp argument
Question:
username_0: public static void main(String[] args) throws Exception {
Path path = Paths.get("/mnt/Data/software/ffmpeg-git-20210908-amd64-static");
args = new String[] {"/mnt/files/input.mkv", "/mnt/files/bitmap.jpg"};
Path pathToSrc = Paths.get(args[0]);
try (InputStream inputStream =
Files.newInputStream(pathToSrc);
) {
FFmpeg.atPath(path)
.addInput(PipeInput.pumpFrom(inputStream).setPosition(60 * 1000))
.addArguments("-frames:v", "1")
.addArguments("-q:v", "1")
.addOutput(UrlOutput.toUrl("/mnt/Data/downloads/direct.jpeg"))
.execute();
}
}
}
Answers:
username_1: Of course, there is a difference between the 2 approaches. And there are a lot of things that can influence on that: keyframes, size, metainf location and much more.
You can try to do the same with OS pipes ( `|` symbol). I'm sure you will get the same result.
One should understand that InputStreams are non-seekable by their nature, while local files are seekable.
And again: it's not an issue with Jaffree. Closing.
Status: Issue closed
username_1: With `InputStream` it can't be done with 100% reliability. Check an option in ffmpeg for exact positioning.
Read this: http://trac.ffmpeg.org/wiki/Seeking
Use `ChannelInput` if you can, it allows seeking. |
xiaojiaqi/k8seasy_release_page | 906786573 | Title: 免费版?用户名只能aaa?
Question:
username_0: it@it:~$ sudo ./installer --genkey -hostlist=192.168.81.130
Error on cmd: sudo gpasswd -a k8s wheel err: exit status 3 file: /home/aaa/kubs/common/system_init/system_adduser.go 40 data: gpasswd: group 'wheel' does not exist in /etc/group
哪个企业敢用这样的 |
genebean/indiansprings-hugo | 575091286 | Title: Figure out forms
Question:
username_0: There is at least one form on the WordPress site... need to figure out how to migrate that. Netlify has some integration that may help with this.
Answers:
username_0: - https://docs.netlify.com/forms/setup/
- https://harrycresswell.com/articles/forms-with-netlify/
Status: Issue closed
|
gitpod-io/workspace-images | 520482411 | Title: Consider install the latest LTS version of Node.js
Question:
username_0: `nvm install --lts --latest-npm` install latest LTS node.js and npm
Answers:
username_1: @username_2 did https://github.com/gitpod-io/workspace-images/pull/117 fix this?
username_0: Using `nvm install --lts` we won't have to keep updating manually every time.
username_2: I much prefer @username_0' approach, as it removes the need for us to constantly watch https://nodejs.org/ and manually update versions here.
However, with our new "dazzle" build this won't work well, as Dazzle caches each layer separately and forever, unless the `RUN` instruction is modified.
This means that switching to `nvm install --lts` will install the correct LTS once, but then that version will remained cached forever, even when newer versions come out.
username_3: Fixed
username_2: Indeed, thanks @username_3!
Status: Issue closed
|
holistics/dbml | 912201249 | Title: markdown language is not supported in fields' notes
Question:
username_0: markdown language is not supported except for project's notes.
Answers:
username_1: Hi, AFAIK, you can write note with markdown syntax anywhere using by putting it between ''' ''' (3 single quotes). For example:
```
table user{
id integer [primary key, increment]
name varchar (20) [note: 'some note abc']
job job_status [note: '''
# Field's note
**markdown content here**
''']
Note: '''
# Table's note
**markdown content here**
'''
}
```
If there's any case that we overlooked, please let me know.
Status: Issue closed
username_2: Markdown support for the notes of tables & fields

 |
gsoft-inc/sg-orbit | 918795182 | Title: 🐛 Shortcuts to title sections are misplaced
Question:
username_0: ## Describe the bug

When hovering a title the # that represents the direct link to it is not showing totally.
## Steps to reproduce
/?path=/story/borders--page hover on Borders
## Expected results
The # is visible in it's entirety.
## Reproducible demo
Visit any doc page and hover a title<issue_closed>
Status: Issue closed |
Instagram/IGListKit | 268883049 | Title: IGListBatchContext.h files need new header format
Question:
username_0: ⚠️ OSS Workshop issue. Please do not action unless at my workshop.
IGListBatchContext.h files need new header format
```
//
// Copyright (c) 2016-present, Facebook, Inc.
// All rights reserved.
//
// This source code is licensed under the BSD-style license found in the
// LICENSE file in the root directory of this source tree. An additional grant
// of patent rights can be found in the PATENTS file in the same directory.
//
// GitHub:
// https://github.com/Instagram/IGListKit
//
// Documentation:
// https://instagram.github.io/IGListKit/
//
``` |
python-effect/effect | 177927682 | Title: refactoring: raise_ is implemented 3 times
Question:
username_0: These are all the same function:
- https://github.com/python-effect/effect/blob/826cbc5b6917683ba6ddf70cd693d3028ecdf226/effect/_test_utils.py#L52
- https://github.com/python-effect/effect/blob/826cbc5b6917683ba6ddf70cd693d3028ecdf226/effect/_base.py#L183
- https://github.com/python-effect/effect/blob/826cbc5b6917683ba6ddf70cd693d3028ecdf226/effect/test_retry.py#L63
so probably some of them should import the one implementation |
carbon-design-system/carbon | 716916055 | Title: Investigate adding copy functionality by default to copy button functionality
Question:
username_0: Wondering if there is any downside to including something like [react copy to clipboard](https://www.npmjs.com/package/react-copy-to-clipboard) as a dependency so users do not have to add in their own copy functionality.
Answers:
username_1: There is also this react hook...
https://github.com/streamich/react-use/blob/master/docs/useCopyToClipboard.md
username_1: This issue is a duplicate of https://github.com/carbon-design-system/carbon/issues/4448.
Please read the conditions I described in its comments. |
yidun/captcha-ios-demo | 1099885512 | Title: 关闭按钮的问题
Question:
username_0: 
在NSObject层做的封装:https://github.com/yidun/NTESVerifyCode/issues/2
Answers:
username_0: 

username_0: 这个问题,我本地的处理方案:


 |
coolsimulations/SurvivalPlus | 971067862 | Title: Everything is fuel...
Question:
username_0: For whatever your mod is making it so nearly every block burns in a furnace and smelts 6 items. I tested it in a modpack with just survivalplus and jei to make sure.
Answers:
username_1: Before I can comment on this, is this Forge or Fabric? And what Minecraft version is this issue occurring on?
username_1: Fixed with https://github.com/username_1/SurvivalPlus/commit/4b6f993e1fc292df17694decb1b94a0bca08d6fe. Will be released as part of 0.1.5b for 1.14-1.16 and 0.1.6 for 1.17-1.18.
Status: Issue closed
username_2:  |
material-components/material-components-android | 332316682 | Title: error: failed linking references.
Question:
username_0: I follow Get Started. Have maven and lib already but still Error: Failed linking references
Answers:
username_1: Do you have multiple color resources in different folders? I would check those, I got this error because the new gradle version now requires you to have all colors that are in the night / nonnight folder also in the default folder.
username_2: I ran into this issue when upgrading to alpha 3. The solution for me was to change `Widget.MaterialComponents.TextInputLayout.OutlineBox` to `Widget.MaterialComponents.TextInputLayout.OutlinedBox` in my XML layout files.
username_0: Thank you for all your support. @username_2 thank you for letting me know this. I have fixed it.
Status: Issue closed
|
X-Sharp/XSharpPublic | 1021325040 | Title: Some intellisense issues
Question:
username_0: 1. Intellisense shows completion list even for unknown identifiers:

And even after closing the list with ESC, it appears again after typing the next letter
2. Type the following, again an incorrect completion list is shown as in (1), while there should be no completion list for a usual:
```
CLASS Something
EXPORT test AS INT
END CLASS
FUNCTION Start() AS VOID STRICT
LOCAL u := Something{} AS USUAL
? u:
```

3. Start typing in the last line, with the intention to type "? u:test-". After typing the last "t", the completion list keeps showing "ToString", which should not be included, since it does not start with "test":

4. Now type the minus sign, this causes "u:ToString(-" to be autotyped:

As a general comment, I think that when the completion list is about to be shown, but none of the items in the list match the identifier the user has typed, then the list must not be displayed. And if it was displayed before, then it must by closed after a typed character causes the new identifier to not be part of the list anymore. I think this alone will solve many of those problems.
Answers:
username_0: For example, see this code:
```
CLASS Something
EXPORT test AS INT
END CLASS
FUNCTION Start() AS VOID STRICT
LOCAL o AS Something
o:test:=o:
```
when you type the first "o:", then the completion list is correct. But after you type the second "o:" (after the := sign), then the completion list is again wrong, contains all those items of (2). And then, when I try to type
o:test:=o:test-
because of this problem, the editor ends up typing for me instead
o:test:=o:ToString(-1
username_0: Another issue, this one is not critical, but I think it was working before:
4.
LOCAL n AS INT
n:ToString(): // no member completion here
username_0: 10. Wrong paren added when auto-completing events:
Type this:

and commit the selected item (which is an event). Then a bogus opening paren is also typed:
FUNCTION Start() AS INT
System.Console.CancelKeyPress( |
rmeleromira/ansible-zencash-nodes | 400362460 | Title: secure nodes switched to SSL cert check from supernode
Question:
username_0: https://github.com/username_1/ansible-zencash-nodes/blob/bc32cbdcea101d5d4795ca1a6bc4f5a4f79934bd/roles/master/tasks/ufw.yml
Needs to be adjusted, tried adjusting it, but keep failing, help appreciated
Answers:
username_0: {% for securenode in groups['securenodes'] %}
-A PREROUTING -i {{ipv6_interface}} -d {{hostvars[securenodes]. ansible_host}} -p tcp --dport 9033 -j DNAT --to-destination {{hostvars[securenodes]. ansible_host}}:9033
{% endfor %}
username_0: got this, but seems faulty
username_1: Secure nodes do not need ufw rules for the zend port, this is a forward needed for ipv4 support for super nodes. Secure nodes don't actually have ufw installed.
My secure nodes are not experiencing ssl check errors.
username_0: theres a new check tough:

its announced here:
https://blog.zencash.com/enhancing-the-certificate-check-process-for-secure-node-tracking-system/
username_1: I've seen the announcement, but your exception closed after 3 minutes and all my nodes are up, I think that's a server side thing we can't do anything about, the deployment code as is already is in compliance. The real test will come on the 19th when the fall back is discontinued, but we should be good.
A way to test would be to use another host with ipv6 enabled and run
```
curl -6 https://you_node_dns
```
If curl doesn't complain about your cert you should be good.
username_0: it closes yes, but seeying this message 10 times a day, and with alot of nodes thats alot of emails per day..
curl -6 https://secnode2.miningspeed.com:9033
curl: (52) Empty reply from server
Is that normal?
username_0: and are you on v 0.3.1 of the tracker i suppose..
--
username_0: openssl s_client -connect secnode9.miningspeed.com:9033
gethostbyname failure
connect:errno=0
username_1: I haven't had time to look into this. openssl command doesn't have ipv6
implemented for the connect function, that's why I gave you the curl
command. Your certificate is fine
username_2: I see these errors occasionally but I've never missed a payment
username_1: https://blog.zencash.com/node-tracking-system-changes-and-enhancements/
I'll keep this issue open, but there's not much I can do since it's not related to using the playbook; I'll wait for their follow up for this statement. It deploys fine and that's all it's meant to handle. I've left the day 2 operations to the operators, hence this playbook being for expert node operators.
username_1: I haven't gotten a failed ssl check in a few weeks now, last one was Sep 23. Seems like they fixed whatever issue was going on in the tracking servers. Reopen this if you're still seeing the issue.
Status: Issue closed
|
PrivateBin/docker-nginx-fpm-alpine | 676397980 | Title: Unable to create a paste
Question:
username_0: While trying to create a new paste I'm encountering the following error:
`Could not create paste: server error or not responding`
`
2020/08/10 20:26:17 [error] 198#198: *13 FastCGI sent in stderr: "PHP message: PHP Fatal error: Uncaught Exception: unable to write to file /srv/data/salt.php in /srv/lib/Persistence/AbstractPersistence.php:120
Stack trace:
#0 /srv/lib/Persistence/ServerSalt.php(86): PrivateBin\Persistence\AbstractPersistence::_store('salt.php', '<?php # |b129e4...')
#1 /srv/lib/Persistence/TrafficLimiter.php(83): PrivateBin\Persistence\ServerSalt::get()
#2 /srv/lib/Persistence/TrafficLimiter.php(120): PrivateBin\Persistence\TrafficLimiter::getHash('sha256')
#3 /srv/lib/Controller.php(201): PrivateBin\Persistence\TrafficLimiter::canPass()
#4 /srv/lib/Controller.php(125): PrivateBin\Controller->_create()
#5 /var/www/index.php(18): PrivateBin\Controller->__construct()
`
I know its probably a permission error, but I ran the docker image exactly as the documentation says:
`
docker run -d \
--restart="always" \
--read-only \
-p 8080:8080 \
-v /opt/privatebin/code/data:/srv/data \
privatebin/nginx-fpm-alpine
`
Status: Issue closed
Answers:
username_0: OK found this:
https://github.com/PrivateBin/docker-nginx-fpm-alpine/issues/2 |
s-fleck/lgr | 757796834 | Title: Add default extra fields to messages
Question:
username_0: I have a json appender where I would like to automatically add some meta-data without having to specify them in the log function every time.
For example, I have a loop which operates on some entities, I'd like to automatically add the entity ID to each log message emitted in the loop.
The motivation is removal of extra clutter in the logging code, where often I'd have 4-5 extra arguments which don't change inside the entire function/loop body.
In code (non-working):
```R
appender$set_default_meta(a = "b", id = 33)
...
lg$info("Hello")
lg$warn("Warning")
```
Both logs would include the `a` and `id` extra keys without me having to specify them repeatedly.
Answers:
username_1: This is already possible with (abusing) Filters! And since this is actually something I also needed, lgr already includes a constructor for Filters for your exact usecase!
`lg$add_filter(FilterInject$new(a = "b", id = 33), "inject")`
see `?EventFilter` for more details
username_0: Genius! Is this described somewhere in the docs? Maybe a short vignette or a blog post would be nice, I think this is fairly common use-case in data processing.
username_1: Ok great that it works :)
No it's not super well documented yet, only in the help of `?EventFilter` and `with_log_value()` (which is just a wrapper for temporary setting a Filter that injects extra values)
I'm leaving this issue open because I agree it should be better documented (how it works, and also WHY it works). Sadly, refining the documentation is a lot of work :(. In the meanwhile, note that lgr is basically a very close clone of python logging and most of the documentation of python logging applies - with small modifications - also to lgr (in this case [Filter Objects](https://docs.python.org/3/library/logging.html#filter-objects))
username_1: I just saw, this was actually documented in the vignette already, I just forgot about it ;) I gave the section a slightly better heading now.
Status: Issue closed
|
flavorjones/mini_portile | 687625930 | Title: Travis CI is failing. Either fix it or turn it off.
Question:
username_0: We have reliable test coverage of Linux in the Concourse pipelines:
* master https://ci.nokogiri.org/teams/nokogiri-core/pipelines/mini_portile
* PRs: https://ci.nokogiri.org/teams/nokogiri-core/pipelines/mini_portile-pr
We haven't had a green build there in two years and nobody's cared. And the .travis.yml file has gone three years without updates.<issue_closed>
Status: Issue closed |
crimx/ext-saladict | 1094926345 | Title: Safari 版插件无法访问 anki connect
Question:
username_0: <!--
反馈前请阅读
- 使用说明: https://saladict.username_1.com/manual.html
- 常见问题以及答复: https://saladict.username_1.com/q&a.html
- 请先在 issues 页面搜索你的问题,很可能已被解决。
-->
<!-- 这是隐藏的信息 -->
<!-- 👆这样括起来的信息将被隐藏,填写时注意不要写在里面。 -->
<!-- 点击编辑器上方的 preview 可预览效果 -->
<!--
⚠️请_完整_填写以下模板描述问题,否则反馈将会被系统关闭。
⚠️请_完整_填写以下模板描述问题,否则反馈将会被系统关闭。
⚠️请_完整_填写以下模板描述问题,否则反馈将会被系统关闭。
⚠️请_完整_填写以下模板描述问题,否则反馈将会被系统关闭。
⚠️请_完整_填写以下模板描述问题,否则反馈将会被系统关闭。
⚠️请_完整_填写以下模板描述问题,否则反馈将会被系统关闭。
⚠️请_完整_填写以下模板描述问题,否则反馈将会被系统关闭。
⚠️请_完整_填写以下模板描述问题,否则反馈将会被系统关闭。
⚠️请_完整_填写以下模板描述问题,否则反馈将会被系统关闭。
⚠️请_完整_填写以下模板描述问题,否则反馈将会被系统关闭。
(重要事情已经说了十遍😅)
-->
## 设备信息
- 操作系统: [macOS 12.1] <!-- 如 [Win10] -->
- 浏览器版本: [Safari Version 15.2 (17612.3.6.1.6)] <!-- 如 [Chrome77] -->
- 沙拉查词版本: [v7.20.0] <!-- 如 [v7.0.0] (在沙拉查词设置页面左上角查看) -->
<!-- 请在下方 ## 开头行之间的空白处填写,点击编辑器上方的 preview 预览效果 -->
## 描述问题
<!-- 客观描述出现了什么问题 -->
无法连接 Anki Connect
## 复现步骤
<!--
如何重复触发这个不正确的行为,如:
1. 打开某某某......
2. 点击某某某......
3. 滚动到某某某......
4. 问题出现
请提供具体页面和具体操作,而不是「任意页面」「选任一单词」,即便问题确实在多处出现。
-->
沙拉查词设置页面 -> 单词管理 -> Anki Connect
点击 ___检查 Anki Connect___ 提示 ___无法连接 Anki Connect,请确认 Anki 已在运行。___
## 期待的正常行为
<!-- 请描述正常情况下应该出现什么结果 -->
正常连接
## 截图
<!-- 可选,需要情况下,可借助截图描述问题 -->
## 额外信息
<!-- 可选,更多有助于理解问题的描述和资料 -->
在 Safari 的开发工具,查看请求信息,在点击 ___检查 Anki Connect___ 后,请求中可以看到 http://127.0.0.1:8765 的请求结果是 403。(猜测可能是由于插件无权限访问 http://127.0.0.1:8765 导致)
如果手动打开 http://127.0.0.1:8765 则能正常显示 Anki Connect。
Chrome 版本插件可以正常连接。
Answers:
username_1: 感觉你的猜测是对的,Safari 有允许插件访问所有网站么?
username_0: 有的,在所有网页中可以正常查词。
username_1: 那可能是 Safari 不允许了。
username_0: 经查证,是 Anki Connect 对请求的 Headers 中的 Origin 进行了判断,仅允许了 Chrome 和 Firefox 浏览器的插件进行访问。
已经向 Anki Connect 提交 PR 修复次问题([Allow safari-web-extension to access Aniki Connect. #297](https://github.com/FooSoft/anki-connect/pull/297)),等待 Anki Connect 合并发布即可结束此 issue 。
Status: Issue closed
username_0: Aniki Connect 已合并。 |
trufflesuite/drizzle | 408207243 | Title: drizzle object not getting populated post upgrade to React 16.8.1, drizzle 1.3.3
Question:
username_0: i upgraded to the following:
react 16.8.1,
drizzle 1.3.3,
web3 1.0.0-beta.43 (breaking in 1.0.0-beta.37 also)
following this, the drizzle object is not getting populated when the following code is used in index.js :
`const drizzle = new Drizzle(drizzleOptions, drizzleStore);`
both the drizzleOptions and drizzleStore seems to be getting populated properly. the outputted drizzle object is showing 0 'ContractList', 'Contract' property and 'web3' property are both empty. Any help would be greatly appreciated.
Answers:
username_1: Hi @username_0, version 1.3.3 should be pinned at `beta.35`--have you tried using `beta.35` and what were the results? There are known issues related to `web3` integration, which may be affecting you.
Status: Issue closed
|
saltstack/salt | 592909771 | Title: netacl states docs page has dead hyperlink
Question:
username_0: Relevant salt docs page:
- https://docs.saltstack.com/en/master/ref/states/all/salt.states.netacl.html
Relevant salt docs file and error:
```
writing output... [ 65%] ref/states/all/salt.states.netacl
(line 27) broken https://napalm.readthedocs.io/en/latest/installation.html - 404 Client Error: Not Found for url: https://napalm.readthedocs.io/en/latest/installation.html
```
Comments
- manually verified<issue_closed>
Status: Issue closed |
wso2/micro-integrator | 658118659 | Title: ClassNotFoundException: com.hazelcast.core.HazelcastInstance cannot be found by synapse-commons_2.1.7.wso2v176
Question:
username_0: Getting the following exception intermittently when shutting down the server. If we start in osgi mode , the probability is high.
```bash
[2020-07-16 17:05:41,183] ERROR {synapse-commons} - [SCR] Error while attempting to deactivate instance of component Component[
name = throttle.core.services
factory = null
autoenable = true
immediate = true
implementation = org.apache.synapse.commons.throttle.core.internal.ThrottleServiceComponent
state = Disabled
properties =
serviceFactory = false
serviceInterface = null
references = {
Reference[name = hazelcast.instance.service, interface = com.hazelcast.core.HazelcastInstance, policy = dynamic, cardinality = 0..1, target = null, bind = setHazelcastInstance, unbind = unsetHazelcastInstance]
}
located in bundle = synapse-commons_2.1.7.wso2v176 [231]
] java.lang.NoClassDefFoundError: com/hazelcast/core/HazelcastInstance
at java.lang.Class.getDeclaredMethods0(Native Method)
at java.lang.Class.privateGetDeclaredMethods(Class.java:2701)
at java.lang.Class.getDeclaredMethod(Class.java:2128)
at org.eclipse.equinox.internal.ds.model.ServiceComponent.getMethod(ServiceComponent.java:156)
at org.eclipse.equinox.internal.ds.model.ServiceComponent.deactivate(ServiceComponent.java:380)
at org.eclipse.equinox.internal.ds.model.ServiceComponentProp.deactivate(ServiceComponentProp.java:161)
at org.eclipse.equinox.internal.ds.model.ServiceComponentProp.dispose(ServiceComponentProp.java:387)
at org.eclipse.equinox.internal.ds.model.ServiceComponentProp.dispose(ServiceComponentProp.java:102)
at org.eclipse.equinox.internal.ds.InstanceProcess.disposeInstances(InstanceProcess.java:344)
at org.eclipse.equinox.internal.ds.InstanceProcess.disposeInstances(InstanceProcess.java:306)
at org.eclipse.equinox.internal.ds.Resolver.disposeComponentConfigs(Resolver.java:724)
at org.eclipse.equinox.internal.ds.Resolver.disableComponents(Resolver.java:700)
at org.eclipse.equinox.internal.ds.SCRManager.stoppingBundle(SCRManager.java:554)
at org.eclipse.equinox.internal.ds.SCRManager.bundleChanged(SCRManager.java:233)
at org.eclipse.osgi.internal.framework.BundleContextImpl.dispatchEvent(BundleContextImpl.java:973)
at org.eclipse.osgi.framework.eventmgr.EventManager.dispatchEvent(EventManager.java:234)
at org.eclipse.osgi.framework.eventmgr.ListenerQueue.dispatchEventSynchronous(ListenerQueue.java:151)
at org.eclipse.osgi.internal.framework.EquinoxEventPublisher.publishBundleEventPrivileged(EquinoxEventPublisher.java:234)
at org.eclipse.osgi.internal.framework.EquinoxEventPublisher.publishBundleEvent(EquinoxEventPublisher.java:140)
at org.eclipse.osgi.internal.framework.EquinoxEventPublisher.publishBundleEvent(EquinoxEventPublisher.java:132)
at org.eclipse.osgi.internal.framework.EquinoxContainerAdaptor.publishModuleEvent(EquinoxContainerAdaptor.java:231)
at org.eclipse.osgi.container.Module.publishEvent(Module.java:493)
at org.eclipse.osgi.container.Module.doStop(Module.java:651)
at org.eclipse.osgi.container.Module.stop(Module.java:515)
at org.eclipse.osgi.container.ModuleContainer$ContainerStartLevel.decStartLevel(ModuleContainer.java:1861)
at org.eclipse.osgi.container.ModuleContainer$ContainerStartLevel.doContainerStartLevel(ModuleContainer.java:1753)
at org.eclipse.osgi.container.SystemModule.stopWorker(SystemModule.java:275)
at org.eclipse.osgi.internal.framework.EquinoxBundle$SystemBundle$EquinoxSystemModule.stopWorker(EquinoxBundle.java:202)
at org.eclipse.osgi.container.Module.doStop(Module.java:653)
at org.eclipse.osgi.container.Module.stop(Module.java:515)
at org.eclipse.osgi.container.SystemModule.stop(SystemModule.java:207)
at org.eclipse.osgi.internal.framework.EquinoxBundle$SystemBundle$EquinoxSystemModule$1.run(EquinoxBundle.java:220)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.ClassNotFoundException: com.hazelcast.core.HazelcastInstance cannot be found by synapse-commons_2.1.7.wso2v176
at org.eclipse.osgi.internal.loader.BundleLoader.findClassInternal(BundleLoader.java:512)
at org.eclipse.osgi.internal.loader.BundleLoader.findClass(BundleLoader.java:423)
at org.eclipse.osgi.internal.loader.BundleLoader.findClass(BundleLoader.java:415)
at org.eclipse.osgi.internal.loader.ModuleClassLoader.loadClass(ModuleClassLoader.java:155)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
... 33 more
```
Answers:
username_0: If the https://github.com/wso2/wso2-synapse/blob/master/modules/commons/src/main/java/org/apache/synapse/commons/throttle/core/internal/ThrottleServiceComponent.java#L36 get activated, while deactivating https://github.com/wso2/wso2-synapse/blob/master/modules/commons/src/main/java/org/apache/synapse/commons/throttle/core/internal/ThrottleServiceComponent.java#L70 this happens.
Status: Issue closed
|
TabakoffLab/PhenogenCloud | 621910411 | Title: TypeError: e.trackDataTable is undefined
Question:
username_0: View details in Rollbar: [https://rollbar.com/username_0/PhenogenCloud/items/479/](https://rollbar.com/username_0/PhenogenCloud/items/479/)
```
TypeError: e.trackDataTable is undefined
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 634, in TrackMenu/e.generateTrackTable
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 638, in success
File "https://phenogen.org/gene.jsp", line 120, in i
File "https://phenogen.org/gene.jsp", line 120, in fireWith
File "https://phenogen.org/gene.jsp", line 122, in y
File "https://phenogen.org/gene.jsp", line 122, in c
File "https://cdnjs.cloudflare.com/ajax/libs/rollbar.js/2.8.1/rollbar.min.js", line 1, in u.prototype.instrumentNetwork/</<
File "https://phenogen.org/gene.jsp", line 122, in send
File "https://phenogen.org/gene.jsp", line 122, in ajax
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 91, in GenomeSVG/d.getAddMenus
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 135, in GenomeSVG
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 259, in GeneTrack/b.setupDetailedView
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 251, in GeneTrack/b.setSelected
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 299, in GeneTrack/b.draw
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 304, in GeneTrack
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 57, in GenomeSVG/d.addTrack/<
File "https://phenogen.org/gene.jsp", line 1868, in send/<
File "https://phenogen.org/gene.jsp", line 1865, in call
File "https://phenogen.org/gene.jsp", line 1868, in e
File "https://cdnjs.cloudflare.com/ajax/libs/rollbar.js/2.8.1/rollbar.min.js", line 1, in u.prototype.instrumentNetwork/</<
File "https://phenogen.org/gene.jsp", line 122, in send
File "https://phenogen.org/gene.jsp", line 122, in ajax
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 90, in GenomeSVG/d.getAddMenus
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 135, in GenomeSVG
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 259, in GeneTrack/b.setupDetailedView
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 251, in GeneTrack/b.setSelected
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 299, in GeneTrack/b.draw
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 304, in GeneTrack
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 57, in GenomeSVG/d.addTrack/<
File "https://phenogen.org/gene.jsp", line 1868, in send/<
File "https://phenogen.org/gene.jsp", line 1865, in call
File "https://phenogen.org/gene.jsp", line 1868, in e
File "https://cdnjs.cloudflare.com/ajax/libs/rollbar.js/2.8.1/rollbar.min.js", line 1, in u.prototype.instrumentNetwork/</<
File "https://phenogen.org/gene.jsp", line 122, in send
File "https://phenogen.org/gene.jsp", line 122, in ajax
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 90, in GenomeSVG/d.getAddMenus
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 135, in GenomeSVG
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 259, in GeneTrack/b.setupDetailedView
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 251, in GeneTrack/b.setSelected
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 299, in GeneTrack/b.draw
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 304, in GeneTrack
File "https://phenogen.org/javascript/gdb.2.9.2.min.js", line 50, in GenomeSVG/d.addTrack/<
File "https://phenogen.org/gene.jsp", line 1868, in send/<
File "https://phenogen.org/gene.jsp", line 1865, in call
File "https://phenogen.org/gene.jsp", line 1868, in e
``` |
ericknavarro/CompiladoresEjercicio3 | 474902332 | Title: Validación de break en ciclos y mensajes de retorno en funciones
Question:
username_0: - De momento no se valida que el break este contenido en un ciclo ya que si se coloca algo como lo siguiente
function number prueba(){
break;
}
No detecta el error del break, aunque si muestra el siguiente error
**Una función de tipo Void no puede retornar valores, solamente puede retornar vacío.**
- También cuando queremos retornar en una función y solo ponemos return
function number prueba(){
return; // debería arrojar un error mencionando que el retorno necesita un valor
}
Por el momento arroja este error
**Una función de tipo Number no puede retornar un valor que no sea numérico.**
- Por ultimo cuando tenemos varias opciones para salir de una función es posible necesitar mas de un retorno por lo tanto
function number prueba(){
if(3 < 5){
return 3; // si se cumple la condición no habría problema, pero si no cumple no retorna
// por lo tanto seria un error, ya que necesita de otro return fuera del if
}
}
Answers:
username_1: El issue fue corregido con la [propuesta de Pavel](https://github.com/username_1/CompiladoresEjercicio3/pull/28).
Status: Issue closed
|
Liam0205/liam0205.github.io | 398534728 | Title: 自然语言处理的数学原理(一) | 始终
Question:
username_0: https://liam.page/2015/07/25/mathematics-theory-of-natural-language-processing-1/
一个基本的搜索引擎的工作,基本上可以分成以下三个部分: 利用网络爬虫下载网页,分析网页关键词,制成索引备用; 理解用户输入,确定检索关键词; 根据关键词和网页索引,按照相关性排序列出搜索结果。 第一个部分主要涉及网络爬虫技术、图论、自然语言处理等技术;第二个部分主要涉及自然语言处理;第三个部分同样涉及自然语言处理。 自然语言,即是人类用来交流的语言。 由此可见,自然语言处理(NLP, Na |
buxlabs/pure-conditions | 423610062 | Title: update isEmptyObject
Question:
username_0: ```js
function isObjectEmpty (object) {
return Object.keys(object).length === 0 && object.constructor === Object
}
```
Answers:
username_1: Hi @username_0 , I fixed it, check my PR :)
https://github.com/buxlabs/pure-conditions/pull/15
Status: Issue closed
|
FriendsOfSymfony/FOSRestBundle | 345210005 | Title: Return every param validation error.
Question:
username_0: Hello and sorry for having to ask this here. I'm having a problem with the paramFetcher and the param validation.I have this two annotations on my method and I need the paramFetcher to return the user a response specifying BOTH of the invalid params, not just one. If both params do not pass the validation I need to return a response explaining what happened and not just the first one. If I send both of them blank I only get an answer for the first one.
`* @RequestParam(name="nombre", allowBlank=false, strict=true, nullable=false, description="Nombre del archivo.")
* @RequestParam(name="archivo_mime_type", strict=true, nullable=false, allowBlank=false, description="Contiene el mime type del archivo.")`
Does the Bundle provide a way of doing it?
Thank you
Status: Issue closed
Answers:
username_1: @username_0 Did you find a workaround for that? |
benjaminaplin/benjaminaplin.github.io | 95714443 | Title: A Couple Things...
Question:
username_0: Hey Fritz: I have a few things that I could use your help with:
1. I am having some trouble with the '.this' within objects. I am getting an error saying that a player method 'is not a function' when I use '. this'.
2. Again having an issue with scope, where i can get the value of player.arrayPlayerHand only in some parts of the program. I have tried moving the function calls and have tested it in the browse console, but haven't figured it out yet..
Thanks!
Ben
Answers:
username_1: OK. Let's meet in a bit.
Status: Issue closed
|
DataBiosphere/azul | 1033016282 | Title: "Error importing data" for "large" PFB manifest
Question:
username_0: 
From <NAME>: the error XXX not found usually means that the data contains "foreign keys," that is some data entity references another data entity, but the entity being referenced wasn't found. This could be simply because it's not in the PFB, or possibly the PFB is out of order … the entities being referenced need to be earlier in the file than the later ones.
Answers:
username_0: Close this?
username_1: @hannes-ucsc : "Ticket lacks concrete reproduction, therefore fix is hard to verify. Closing.
Status: Issue closed
|
commercialhaskell/stack | 926416720 | Title: Windows msys2: pacman reports PGP signature errors.
Question:
username_0: It suggests installing the icu development package. But when I try to say `stack exec -- pacman -S icu-devel` I get PGP signature errors.
### Steps to reproduce
`stack exec -- pacman -S icu-devel`
### Expected
Latest icu-devel packages installed.
### Actual
```
resolving dependencies...
looking for conflicting packages...
Packages (2) icu-68.2-1 icu-devel-68.2-1
Total Download Size: 1.17 MiB
Total Installed Size: 57.45 MiB
Net Upgrade Size: 25.99 MiB
:: Proceed with installation? [Y/n] y
:: Retrieving packages...
icu-devel-68.2-1-x86_64 1197.3 KiB 2.77 MiB/s 00:00 [###########################################################] 100%
(2/2) checking keys in keyring [###########################################################] 100%
(2/2) checking package integrity [###########################################################] 100%
error: icu: signature from "<NAME> <<EMAIL>>" is unknown trust
:: File /var/cache/pacman/pkg/icu-68.2-1-x86_64.pkg.tar.zst is corrupted (invalid or corrupted package (PGP signature)).
Do you want to delete it? [Y/n] n
error: icu-devel: signature from "<NAME> <<EMAIL>>" is unknown trust
:: File /var/cache/pacman/pkg/icu-devel-68.2-1-x86_64.pkg.tar.zst is corrupted (invalid or corrupted package (PGP signature)).
Do you want to delete it? [Y/n] n
error: failed to commit transaction (invalid or corrupted package (PGP signature))
Errors occurred, no packages were upgraded.
```
### Stack version
```
$ stack --version
Version 2.7.1, Git revision 8afe0c2932716b0441cf4440d6942c59568b6b19 x86_64 hpack-0.34.4
```
### Method of installation
* Official binary, downloaded from stackage.org or fpcomplete's package repository
Answers:
username_1: Workaround: stack exec -- pacman -Sy msys2-keyring
The packaged msys2 is almost a year old now; so is pacman's trustdb.
username_2: Me too:
- https://github.com/agda/agda/issues/5768
and I bothered the wrong people when it is really a `stack` problem:
- https://github.com/actions/virtual-environments/issues/4987 |
haskell-infra/hackage-trustees | 288347838 | Title: microlens-0.4.8.2 has garbage inside
Question:
username_0: It looks like the tarball for microlens-0.4.8.2 didn't get made quite right. In particular, the file src/Lens/Micro/Extra.hs inside has garbage and not Haskell code, causing the build to fail outright.
It would be nice to let cabal's solver know not to use this version if possible.
Answers:
username_0: Oh, and it looks like upstream has been notified here:
https://github.com/aelve/microlens/issues/99
username_1: After confirming that multiple .hs files were in fact binary data I've updated the constraint on this particular version to be unsatisfiable.
username_2: http://hdiff.luite.com/cgit/microlens/diff?id=0.4.8.2&id2=0.4.8.1
https://matrix.hackage.haskell.org/package/microlens Matrix is red atm
username_3: It looks like some random memory buffers from the machine producing the src-dist (containing something that looks like a bash history fragment) got written into the .tar... and judging from those shell fragments, this release could have been tarred up by Stack.
username_0: Looks like other packages may be affected. microlens-th-0.4.1.2 just failed for me. I'm checking in on the rest of the microlens-* family now.
username_0: microlens-th appears (to me) to be the only other package in the family to be affected.
username_1: microlens-th-0.4.1.2 was also affected
username_4: The corrupted versions have been marked as unsatisfiable and I've uploaded new versions.
Thanks guys!
Status: Issue closed
|
DavBfr/dart_pdf | 583197507 | Title: RichText - undeline in the sentence
Question:
username_0: **Describe the bug**
Trying to write a contract with underline in some sentences, makes the underline goes overline.
In other widget, not the pdf, it works normally.
**To Reproduce**
pdf.addPage(MultiPage(
pageFormat:
PdfPageFormat.a4.copyWith(marginBottom: 1.0 * PdfPageFormat.cm),
build: (Context context) => <Widget>[
Container(
alignment: Alignment.center,
margin: const EdgeInsets.only(
top: 1.0 * PdfPageFormat.cm,
bottom: 1.0 * PdfPageFormat.cm),
child: Text("PROCURAÇÃO",
style: TextStyle(
fontWeight: FontWeight.bold,
decoration: TextDecoration.underline))),
RichText(
softWrap: true,
textAlign: TextAlign.justify,
text: TextSpan(
style: TextStyle(),
children: <TextSpan>[
TextSpan(
text: 'PODERES:',
style: TextStyle(
fontWeight: FontWeight.bold,
decoration: TextDecoration.underline,
)),
TextSpan(
text:
" Os mais amplos e ilimitados poderes para representar o(a) Outorgante perante o foro em geral, judicial, extrajudicial ou arbitral, em qualquer instância ou grau de jurisdição, a fim de defenderem seus interesses e proporem ações que julgarem necessárias, conferindo aos Outorgados os poderes das cláusulas "),
TextSpan(
text: "'ad e extra judicia'",
style: TextStyle(fontStyle: FontStyle.italic)),
TextSpan(
style: TextStyle(),
text:
", para representarem-no(a) em Juízo ou fora dele, para o fim de propor e/ou contestar ações cíveis, falimentares, criminais, cautelares, requerer, promover medidas e diligências, inclusive mandado de segurança, intervir, requerer certidões, prestar cauções, habilitar créditos, assinar termos e compromissos, acompanhar feitos até final decisão, com trânsito em julgado, firmar acordos, representa-lo em cumprimento de sentença, e quaisquer outras medidas judicias ou administrativas sobre o fato abaixo narrado, outorgando-lhes ainda os "),
TextSpan( style: TextStyle(
fontWeight: FontWeight.bold,
decoration: TextDecoration.underline,
),
text:
'especiais poderes para confessar, reconvir, variar de ação, acordar, transigir, desistir, receber, requerer alvarás de levantamento, levantar valores e dar quitação ',
),
TextSpan(
text:
"notificar, interpelar, protestar, impetrar mandados de segurança, assinar termos e compromissos, declarar residência e estado de hipossuficiência e substabelecer a quem convier,",
style: TextStyle(fontWeight: FontWeight.bold)),
TextSpan(
text:
" com ou sem reservas de iguais poderes, independente de necessidade de intimação ou ciência do(a) Outorgante, podendo os Outorgados enfim, praticarem todos os atos necessários ao bom e fiel cumprimento do mandato, que é outorgado com fim específico para representar o(a) Outorgante em Ações de Reparação por Dano Ambiental, Dano Moral e Material, e/ou Compensações, extra e/ou judiciais, propostas em razão dos danos morais e materiais por ele(a) sofrido por conta de danos ambientais e/ou poluições marítimas ocorridos que afetaram ou venham a afetar seu exercício profissional, e/ou a sua cultura e modo de ser e viver, especialmente no que se refere ao acidente ambiental a seguir nominado, especialmente por conta do "),
TextSpan(
text: Projeto,
style: TextStyle(
fontStyle: FontStyle.italic,
decoration: TextDecoration.underline,
)),
[Truncated]
],
),
),
]));
**Flutter Doctor**
[✓] Flutter (Channel stable, v1.12.13+hotfix.8, on Mac OS X 10.15.3 19D76, locale pt-BR)
[✓] Android toolchain - develop for Android devices (Android SDK version 29.0.2)
[✓] Xcode - develop for iOS and macOS (Xcode 11.3.1)
[✓] Android Studio (version 3.5)
[✓] VS Code (version 1.43.0)
[✓] Connected device (1 available)
Answers:
username_1: This is fixed in the master branch.
Underline with justified text was not working.
I also improved the decorations with a nicer render.
Status: Issue closed
|
datosgobar/monitoreo-apertura | 340688664 | Title: Migrar indicadores con IDS nulos
Question:
username_0: En la migración realizada en el issue #53 se omitieron algunos indicadores debido a que algunos catálogos cambiaron de título.
Los catálogos cuyos indicadores no se migraron son los siguientes:
Ids de nodos faltantes:
* "aaip" es el title "Datos de la Agencia de Acceso a la Información Pública"
* "cultura" es el title "....cultura..."
* "salud" es el title "....MSAL..." |
chime-experiment/ch_pipeline | 783651028 | Title: Possible bug in handling of cross-pol visibilities in ringmap maker
Question:
username_0: Ue-Li has been comparing ringmaps with GBT maps, and has found that GBT U maps look completely different from U derived from ringmaps. It turns out that the Fourier modes of each map are individually correlated, but with the opposite sign; this points to a possible issue with how cross-pol visibilities are handled in the ringmap maker. (On the other hand, the XX and YY ringmaps are a good match with GBT I and Q maps.)
I don't think Ue-Li tracks issues on here, but I can relay any questions or comments to him.
Status: Issue closed
Answers:
username_0: This was resolved in the beams call: Ue-Li has been using outdated ringmaps, and the issue he highlighted has already been fixed in the ringmap maker |
code-cracker/code-cracker | 49293292 | Title: Create PropertyChangedEventArgs statically
Question:
username_0: Sample Video in https://pbs.twimg.com/tweet_video/B2tTZRNCIAAHlqx.mp4
Reference implementation (VB) in
https://github.com/ljw1004/dotnet-code-analyzer/blob/master/dotnet-code-analyzer/Analyzer_XamlAppTips/UseStaticChangedCS.vb
This is essentially a refactoring, so we are classifying it as a `Hidden` diagnostic.
Diagnostic Id: `CC0106`
This:
````csharp
private string name;
public string Name
{
get
{
return name;
}
set
{
name = value;
PropertyChanged?.Invoke(this, new PropertyChangedEventArgs(nameof(Name));
}
}
````
Becomes:
````csharp
private string name;
public string Name
{
get
{
return name;
}
set
{
name = value;
PropertyChanged?.Invoke(this, nameChanged);
}
}
private static PropertyChangedEventArgs nameChanged = new PropertyChangedEventArgs(nameof(Name);
````
If there is already a field named `nameChanged`, add `1` to it, if there is already on with `1`, add `2`, and so on.
Answers:
username_1: I added the test case and diagnostic id. It is up for grabs.
username_2: @username_1, this one looks interesting. I would like to take care of analyzer/code-fix.
username_1: Go for it.
Status: Issue closed
username_1: Closed on #760. |
JannisJost/QuantumAccelerator | 921076529 | Title: RAM usage after temp file scan is closed
Question:
username_0: **Describe the bug**
The temp file scanner still uses a lot of RAM after beeing closed.
**To Reproduce**
Steps to reproduce the behavior:
1. Run the temp file scanner
2. When the list of folders appears close the window
**Expected behavior**
The temp file scanner should free the RAM once closed because it uses loads amount of memory.
**Desktop (please complete the following information):**
- OS: Windows 10 21H1 (19043.1055)
- Version: 0.1-beta
Answers:
username_0: The memory leak in the search engine now is fixed
Status: Issue closed
|
JanDeDobbeleer/oh-my-posh | 1144689922 | Title: [Command] conditional commands - as described in the documentation - don't work
Question:
username_0: ### Code of Conduct
- [X] I agree to follow this project's Code of Conduct
### What happened?
According to the [documentation](https://ohmyposh.dev/docs/command) the segment below should output a pretty printed date when the current directory is a git repository and a standard date via `date` otherwise. This works as expected in a git repository but not when not in a repository (the git error (`fatal: not a git repository (or any of the parent directories): .git`) is shown instead).
### Theme
directly from the [documentation](https://ohmyposh.dev/docs/command)
```json
{
"$schema": "https://raw.githubusercontent.com/username_1/oh-my-posh/main/themes/schema.json",
"version": 1,
"blocks": [
{
"type": "prompt",
"alignment": "right",
"segments": [
{
"type": "command",
"style": "plain",
"properties": {
"shell": "bash",
"command": "git log --pretty=format:%cr -1 || date +%H:%m:%S"
}
}
]
}
]
}
```
### What OS are you seeing the problem on?
Windows, Linux
### Which shell are you using?
powershell
### Log output
```shell
[1mVersion:[0m 7.19.1
[1mSegments:[0m
ConsoleTitle(false) - 0 ms - ]0;tmp
command(true) - 36 ms - [37m fatal: not a git repository (or any of the parent directories): .git
[0m
[1mRun duration:[0m 36.527554ms
[1mCache path:[0m /home/thorsten/.cache/oh-my-posh
[1mLogs:[0m
2022/02/19 13:26:28 Args duration: 111ns, args:
[Truncated]
2022/02/19 13:26:28 Getenv duration: 4.674µs, args: WSL_DISTRO_NAME
2022/02/19 13:26:28 IsWsl duration: 12.77µs, args:
2022/02/19 13:26:28 debug: Pwd
/home/thorsten/tmp
2022/02/19 13:26:28 Pwd duration: 3.735µs, args:
2022/02/19 13:26:28 PathSeperator duration: 118ns, args:
2022/02/19 13:26:28 PathSeperator duration: 66ns, args:
2022/02/19 13:26:28 PathSeperator duration: 67ns, args:
2022/02/19 13:26:28 PathSeperator duration: 108ns, args:
2022/02/19 13:26:28 PathSeperator duration: 65ns, args:
2022/02/19 13:26:28 User duration: 897ns, args:
2022/02/19 13:26:28 Host duration: 2.951µs, args:
2022/02/19 13:26:28 GOOS duration: 103ns, args:
2022/02/19 13:26:28 TemplateCache duration: 129.925µs, args:
2022/02/19 13:26:28 debug: Getenv
2022/02/19 13:26:28 Getenv duration: 2.284µs, args: XDG_CACHE_HOME
2022/02/19 13:26:28 CachePath duration: 23.612µs, args:
```<issue_closed>
Status: Issue closed |
jbreslin33/footballhome | 650986019 | Title: club profile: add "Club Member" button, Player Interest, Player Lead, Parent Interest, Parent lead etc.
Question:
username_0: also make it clear if you remove as club_member and if you make club_member you cannot be lead etc. also you cannot be lead and prospect etc.
Answers:
username_0: added buttons for leads etc. but they are non functional. need php and db files and code still.
Status: Issue closed
|
DropD/reentry | 305067975 | Title: Add 'group' to EntryPoint instance
Question:
username_0: When retrieving `EntryPoints` they are initially all grouped in an entry point `group`, but the `EntryPoint` itself has no knowledge of this. It would be nice to be able to inquire which group it belongs to
Status: Issue closed
Answers:
username_1: Closed for now, because this would break backwards compatibility with pkg_resources |
ng-lightning/ng-lightning | 178073086 | Title: Angular 2.0 Support
Question:
username_0: Do we need to do anything to make this work with angular 2.0 release? If we do, maybe we should know and make a pull request?
Regards
Status: Issue closed
Answers:
username_1: Hi @username_0, as of version 0.22.0 ng-lightning is fully compatible with Angular 2.0 final release. |
ansilove/ansilove | 507365843 | Title: use of undeclared identifier for ANSILOVE_FONT_*
Question:
username_0: I'm trying to re-install ansilove after switching to a different Mac running High Sierra.
I couldn't get the default installation instructions for to work `libansilove` or `ansilove`, but I had success by changing the cmake command slightly:
```
cmake -S .. -B .
```
After this, `make` worked perfectly for `libansilove`. However, I am running into a problem compiling `ansilove`. Here is the error output:
```
[ 16%] Building C object CMakeFiles/ansilove.dir/src/ansilove.c.o
In file included from /Users/dummy/repos/ansilove/src/ansilove.c:32:
/Users/dummy/repos/ansilove/src/fonts.h:60:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP737'
ANSILOVE_FONT_CP737, /* Greek */
^
/Users/dummy/repos/ansilove/src/fonts.h:61:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP775'
ANSILOVE_FONT_CP775, /* Baltic */
^
/Users/dummy/repos/ansilove/src/fonts.h:62:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP850'
ANSILOVE_FONT_CP850, /* Latin 1 */
^
/Users/dummy/repos/ansilove/src/fonts.h:63:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP852'
ANSILOVE_FONT_CP852, /* Latin 2 */
^
/Users/dummy/repos/ansilove/src/fonts.h:64:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP855'
ANSILOVE_FONT_CP855, /* Cyrillic */
^
/Users/dummy/repos/ansilove/src/fonts.h:65:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP857'
ANSILOVE_FONT_CP857, /* Turkish */
^
/Users/dummy/repos/ansilove/src/fonts.h:66:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP860'
ANSILOVE_FONT_CP860, /* Portuguese */
^
/Users/dummy/repos/ansilove/src/fonts.h:67:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP861'
ANSILOVE_FONT_CP861, /* Icelandic */
^
/Users/dummy/repos/ansilove/src/fonts.h:68:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP862'
ANSILOVE_FONT_CP862, /* Hebrew */
^
/Users/dummy/repos/ansilove/src/fonts.h:69:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP863'
ANSILOVE_FONT_CP863, /* French-canadian */
^
/Users/dummy/repos/ansilove/src/fonts.h:70:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP865'
ANSILOVE_FONT_CP865, /* Nordic */
^
/Users/dummy/repos/ansilove/src/fonts.h:71:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP866'
ANSILOVE_FONT_CP866, /* Russian */
^
/Users/dummy/repos/ansilove/src/fonts.h:72:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP869'
ANSILOVE_FONT_CP869, /* Greek */
^
/Users/dummy/repos/ansilove/src/fonts.h:73:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP437'
ANSILOVE_FONT_CP437, /* IBM PC 80x25 */
^
/Users/dummy/repos/ansilove/src/fonts.h:74:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP437_80x50'
ANSILOVE_FONT_CP437_80x50, /* IBM PC 80x50 */
^
/Users/dummy/repos/ansilove/src/fonts.h:75:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP737'
ANSILOVE_FONT_CP737, /* Greek */
^
/Users/dummy/repos/ansilove/src/fonts.h:76:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP775'
ANSILOVE_FONT_CP775, /* Baltic */
^
/Users/dummy/repos/ansilove/src/fonts.h:77:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP850'
ANSILOVE_FONT_CP850, /* Latin 1 */
^
/Users/dummy/repos/ansilove/src/fonts.h:78:2: error: use of undeclared identifier 'ANSILOVE_FONT_CP852'
ANSILOVE_FONT_CP852, /* Latin 2 */
^
fatal error: too many errors emitted, stopping now [-ferror-limit=]
20 errors generated.
make[2]: *** [CMakeFiles/ansilove.dir/src/ansilove.c.o] Error 1
make[1]: *** [CMakeFiles/ansilove.dir/all] Error 2
make: *** [all] Error 2
```
Answers:
username_0: Sorry, this was my mistake. I pulled changes for `ansilove` but forgot to pull changes to `libansilove`, so I was trying to build incompatible versions. After I got everything to the latest version, it built fine.
Status: Issue closed
|
thuss/standalone-migrations | 169764151 | Title: support for ActiveRecord 5
Question:
username_0: the Gemfile currently has
gem 'activerecord', ENV['AR'] || ["~> 4.2.5", ">= 4.2.5.1"]
gem 'railties', ENV['AR'] || ["~> 4.2.5", ">= 4.2.5.1"]
rails 5 is out, and with it activerecord - would be awesome if we could get an update. thanks :)
Answers:
username_1: We always welcome pull requests!
username_2: Just created a pull request for this issue, please check it out!
username_1: Closing this issue since we now support AR 5.2.0
Status: Issue closed
|
randoop/randoop | 615304927 | Title: Cannot find the Java compiler. Check that classpath includes tools.jar
Question:
username_0: I am trying to use 4.2.3 version.
I have JAVA_HOME installed and hence mvn clean install is working.
I tried running command
java -classpath %RANDOOP_JAR% randoop.main.Main gentests --testclass=sunlife.egistration.controller.RegistrationIdentificationController
Error:
Randoop for Java version 4.2.3.
Cannot find the Java compiler. Check that classpath includes tools.jar.
Classpath:
/C:/Xfunction/randoop-4.2.3/randoop-all-4.2.3.jar
Answers:
username_0: C:\Users\j515>echo %JAVA_HOME%
**C:\eclipse47\jdk180_172**
C:\Users\j515>echo %PATH%
C:\ProgramData\Oracle\Java\javapath;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem;C:\Windows\System32\WindowsPowerShell\v1.0\;C:\Windows\System32\OpenSSH\;C:\Lotus\Notes;C:\Program Files\Git\cmd;C:\Program Files\nodejs\;C:\Program Files (x86)\Personal Communications\;C:\Program Files (x86)\IBM\Trace Facility\;C:\Program Files\Common Files\IBM\MQM\bin64;C:\Program Files\Common Files\IBM\MQM\bin;C:\Program Files\Common Files\IBM\MQM\tools\c\samples\bin64;C:\Program Files\Common Files\IBM\MQM\tools\c\samples\bin;C:\Users\j515\AppData\Local\Microsoft\WindowsApps;C:\Users\j515\Downloads\apache-maven-3.6.2\bin;C:\Users\j515\AppData\Local\Programs\Microsoft VS Code\bin;C:\Users\j515\AppData\Roaming\npm;C:\Program Files\Git\mingw64\bin;C:\Program Files\Git\usr\bin;C:\Program Files\Git\cmd;C:\Xfunction\Zookeper\bin;**C:\eclipse47\jdk180_172\bin;C:\eclipse47\jdk180_172\lib\tools.jar;**
username_1: Your second message shows that you added `tools.jar` to your path (where the operating system looks for executables), but that is different than the classpath (where Java looks for classes).
Status: Issue closed
|
therufa/mdi-vue | 706070843 | Title: Support for Vue 3
Question:
username_0: Does this work with Vue 3? I think Vue 3 has removed support for functional template/component.
Answers:
username_1: Thanks for bringing this up! It seems it does break with vue 3 therefore it should be moved to plain components or something else. Sadly I have limited time during the upcoming weeks, so this might take a little longer to implement.
username_2: @username_1
The template.js has to be converted to support Vue 3 syntax, in Vue 3 `functional` attribute is removed from template
Here is a guide how: https://v3.vuejs.org/guide/migration/functional-components.html#_3-x-syntax
I could try to work on this if you want ?
Status: Issue closed
|
leachjustin18/those-who-serve | 509652399 | Title: Create ability/page to add servants
Question:
username_0: This will include:
* Ability to add there name
* Create user name from name (first initial and last name).
* If name already exists, add number to one.
* Allow suffix, like III, as we do have that
* Add ability to add jobs to a user. |
googleapis/google-cloud-python | 438508900 | Title: Pub/Sub: wrong inherited documentation for StreamingPullFuture
Question:
username_0: The [documentation](https://googleapis.github.io/google-cloud-python/latest/pubsub/subscriber/api/futures.html) for `StreamingPullFuture`'s methods like `result()` and `exception()` doesn't make sense because it describes Publish futures not streaming pull futures.
[`StreamingPullFuture`](https://github.com/googleapis/google-cloud-python/blob/a42b0f1abe83ce981c9573f4c56acc72e4344a2f/pubsub/google/cloud/pubsub_v1/subscriber/futures.py#L20) inherits [`futures.Future`](https://github.com/googleapis/google-cloud-python/blob/a42b0f1abe83ce981c9573f4c56acc72e4344a2f/pubsub/google/cloud/pubsub_v1/futures.py#L24), which is where the documentation got carried over.<issue_closed>
Status: Issue closed |
bioboxes/rfc | 74149157 | Title: Bioboxes should provide a validator for the binning container
Question:
username_0: @username_1 has created an RFC for the binning container. We aim to provide a
validator for each type of biobox as this makes it simpler for biobox
developers to ensure their biobox matches the specification. The binning biobox
is however a more difficult case as it requires large reference databases to
compare against.
Answers:
username_0: @username_1 do you think it might be possible to provide a small bioboxes dummy
database for testing? For example something a few megabytes in size? If I have
understood the RFC correctly, you could specify this as part of the `databases`
section in the biobox.yml?
username_1: I'm afraid that when I make the database smaller a binning bioboxes might not work as expected, so I will concentrate on building a bioboxes service.
If this does not work for whatever reason, I will reconsider providing a smaller database.
username_1: I think if a binning tool developer is using any database(ncbi, taxonomy, refseq, ...) he has to download the database anyway and mount it to his tool. So I created a validator for bioboxes [binning tools](https://github.com/bioboxes/validator-binning). If the developer is using one of the databases he can mount it to the validator by using one of the following parameters:
* -R, --REFSEQ /path/to/your/refseq/database
* -T, --TAXONOMY /path/to/your/taxonomy/database
* -N, --NCBI /path/to/your/ncbi/database
* -C, --COG /path/to/your/cog/database
* -B, --BLASTDB /path/to/your/blastdb/database
This way the validator builds the yaml by using the definitions in the rfc and mounts the provided databases.
Status: Issue closed
username_0: Great Peter. I looked at the validate script in the code - you're cycling through each of these options when you're testing each binning container or each one has a specific arguments?
username_1: This part in the validate script:
https://github.com/bioboxes/validator-binning/blob/master/src/validate#L17-L54
1. Checks which argument is provided to the validate script.
2. Depending on the argument provided to the script it builds a yaml and a docker command used in the cucumber test.
The reason for this is that if you are using one of the databases listed in the [rfc](https://github.com/bioboxes/rfc/blob/master/container/binning/rfc.mkd#databases-definition) you have to mount it to the container and build the yaml used in the cucumber tests. |
vnmakarov/yaep | 186717670 | Title: make install doesn't install allocate.h
Question:
username_0: after `make install`, i try
~$ echo '#include<yaep.h>' >test.c
~$ gcc test.c
In file included from test.c:1:0:
/usr/local/include/yaep.h:47:22: fatal error: allocate.h: No such file or directory
compilation terminated.
Answers:
username_1: Thank you for the report. I've just fixed this issue. |
encode/databases | 512990080 | Title: Transaction rollback in `with transaction` block causes IndexError
Question:
username_0: Should I be able to manually rollback transaction when using transaction comntext?
```python3
import asyncio
from databases import Database
def database():
return Database("postgresql://db",
min_size=1,
max_size=4,
user="tigen",
password="<PASSWORD>",
database="test")
async def run():
async with database() as db:
async with db.transaction() as tx:
await db.fetch_one("select 1")
await tx.rollback()
if __name__ == "__main__":
asyncio.run(run())
```
When run, it causes `IndexError`:
```
Traceback (most recent call last):
File "/usr/local/lib/python3.7/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/usr/local/lib/python3.7/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/app/app/foo.py", line 22, in <module>
asyncio.run(run())
File "/usr/local/lib/python3.7/asyncio/runners.py", line 43, in run
return loop.run_until_complete(main)
File "/usr/local/lib/python3.7/asyncio/base_events.py", line 579, in run_until_complete
return future.result()
File "/app/app/foo.py", line 18, in run
await tx.rollback()
File "/usr/local/lib/python3.7/site-packages/databases/core.py", line 275, in __aexit__
await self.commit()
File "/usr/local/lib/python3.7/site-packages/databases/core.py", line 305, in commit
assert self._connection._transaction_stack[-1] is self
IndexError: list index out of range
```
Using Databases 0.2.5 with Python 3.7.4.<issue_closed>
Status: Issue closed |
sequelize/sequelize | 452279176 | Title: sequelize sync force true changes column type
Question:
username_0: I have a table `EventUserDish`, and it works as expected in dev, and production. But during testing, this table gets modified during `beforeEach` hook, and the column type seems to be getting updated not based on how I define it.
Here is my model definition:
`const eventUserDishFactory = (sequalize: Sequelize) => {
const EventUserDish = <EventUserDishModel>sequalize.define('EventUserDish', {
id: {
allowNull: false,
autoIncrement: true,
primaryKey: true,
type: DataTypes.INTEGER
},
eventId: {
allowNull: false,
type: DataTypes.INTEGER
},
dishId: {
allowNull: false,
type: DataTypes.INTEGER
},
userId: {
allowNull: false,
unique: false,
type: DataTypes.INTEGER
},
count: {
allowNull: true,
type: DataTypes.INTEGER
}
});
EventUserDish.associate = (model: IDB) => {
EventUserDish.belongsTo(model.Dish, {
foreignKey: 'dishId',
as: 'dish'
});
EventUserDish.belongsTo(model.Event, {
foreignKey: 'eventId',
as: 'event'
});
EventUserDish.belongsTo(model.User, {
foreignKey: 'userId',
as: 'user'
});
};
return EventUserDish;
};`
After `beforeEach` hook, `userId` column now is `UNIQUE`. There are other weird behaviors happening like certain tables are not synced, so during testing, these tables will come up as relationship not exits.
I am not sure if this has anything to do with Typescript or is it a sequelize sync issue.<issue_closed>
Status: Issue closed |
adewg/ICAR | 626238280 | Title: ICAR Milking Type Code might be misleading
Question:
username_0: When providing an MilkVisitEventResource, we can supply a milkingType with it. The possible enum values are listed in https://github.com/adewg/ICAR/blob/ADE-1/enums/icarMilkingTypeCode.json
As the only possible values are "Manual" and "Automatic", this might be slightly misleading. With a milking robot, you can still attach a cow manually in some models, and in conventional milking, the milking process is most of the time automatic.
I'd propose (unfortunately as a breaking change for a new version) to either define something like "Robot" a "Conventional" as values or be even more specific like "Robot Auto Attach", "Robot Manual Attach", "Conventional" (and maybe other values).
Answers:
username_0: To add some more complexity: Automatic might even be more ambiguous as it does not describe the type of milking parlor, e.g. it could be an automatic rotor or an automatic robot for free cow traffic.
username_1: This particular enumerationd ates from the original XML specification (developed before I was involved with the group).
However, on investigation, I believe this reflects the differentiation in the ICAR Guidelines between "fully automatic" systems that collect data such as milk weights and any other metrics "without manual intervention" and other systems where manual intervention occur in some way. So it is not trying to define the device (there is a device object for this), but simply which rules are expected to be used.
Section 2 of the Guidelines (milk recording), clause 6.8 covers some of this, and section 11 (milk recording devices) also covers this and refers to the ISO requirements for such devices.
Bearing this in mind, I recommend that this particular field be left as it is (because it reflects the guidelines) and if we need to differentiate the methods of milking and attachment in more detail, then we either identify capabilities in the Device object, or add another field for milking equipment type. |
graphql-java/graphql-java | 635078031 | Title: [v15.0 error] The field type 'Long' is not present when resolving type
Question:
username_0: root.graphql
```
type Query {
}
type Mutation {
}
type Empty {
_: String
}
input PageInput {
pageNum: Int
pageSize: Int
}
type Page {
pageNum: Int
pageSize: Int
startRow: Int
endRow: Int
totalCount: Long
totalPages: Int
firstPage: Boolean
lastPage: Boolean
hasPreviousPage: Boolean
hasNextPage: Boolean
}
```
When I upgrade the graphql-java version from 14.0 to 15.0, an error appears:
```
Caused by: graphql.schema.idl.errors.SchemaProblem: errors=[The field type 'Long' is not present when resolving type 'Page' [@18:1]]
at graphql.schema.idl.SchemaGenerator.makeExecutableSchema(SchemaGenerator.java:251)
at graphql.schema.idl.SchemaGenerator.makeExecutableSchema(SchemaGenerator.java:227)
```
Answers:
username_0: Sorry I found the answer
https://github.com/graphql-java/graphql-java-extended-scalars
Status: Issue closed
|
java-deobfuscator/deobfuscator | 247498203 | Title: Hide Access Deobfuscation (Stringer) failed
Question:
username_0: [Stringer] [HideAccessTransformer] Starting
[Stringer] [HideAccessTransformer] Found 7 decryptors
Deobfuscation failed. Please open a ticket on GitHub
com.javadeobfuscator.deobfuscator.executor.exceptions.ExecutionException: java.lang.UnsupportedOperationException @ a/a/a/a Hi(Ljava/lang/Object;Ljava/lang/Object;Ljava/lang/Object;)Ljava/lang/Object;
at com.javadeobfuscator.deobfuscator.executor.MethodExecutor.execute(MethodExecutor.java:1296)
at com.javadeobfuscator.deobfuscator.executor.MethodExecutor.execute(MethodExecutor.java:85)
at com.javadeobfuscator.deobfuscator.transformers.stringer.HideAccessObfuscationTransformer.lambda$null$5(HideAccessObfuscationTransformer.java:126)
at java.util.stream.ForEachOps$ForEachOp$OfRef.accept(Unknown Source)
at java.util.stream.ReferencePipeline$2$1.accept(Unknown Source)
at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(Unknown Source)
at java.util.stream.AbstractPipeline.copyInto(Unknown Source)
at java.util.stream.AbstractPipeline.wrapAndCopyInto(Unknown Source)
at java.util.stream.ForEachOps$ForEachOp.evaluateSequential(Unknown Source)
at java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(Unknown Source)
at java.util.stream.AbstractPipeline.evaluate(Unknown Source)
at java.util.stream.ReferencePipeline.forEach(Unknown Source)
at com.javadeobfuscator.deobfuscator.transformers.stringer.HideAccessObfuscationTransformer.lambda$transform$6(HideAccessObfuscationTransformer.java:108)
at java.util.stream.ForEachOps$ForEachOp$OfRef.accept(Unknown Source)
at java.util.stream.ReferencePipeline$3$1.accept(Unknown Source)
at java.util.HashMap$ValueSpliterator.forEachRemaining(Unknown Source)
at java.util.stream.AbstractPipeline.copyInto(Unknown Source)
at java.util.stream.AbstractPipeline.wrapAndCopyInto(Unknown Source)
at java.util.stream.ForEachOps$ForEachOp.evaluateSequential(Unknown Source)
at java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(Unknown Source)
at java.util.stream.AbstractPipeline.evaluate(Unknown Source)
at java.util.stream.ReferencePipeline.forEach(Unknown Source)
at com.javadeobfuscator.deobfuscator.transformers.stringer.HideAccessObfuscationTransformer.transform(HideAccessObfuscationTransformer.java:107)
at com.javadeobfuscator.deobfuscator.Deobfuscator.start(Deobfuscator.java:172)
at com.javadeobfuscator.deobfuscator.DeobfuscatorMain.run(DeobfuscatorMain.java:105)
at com.javadeobfuscator.deobfuscator.DeobfuscatorMain.main(DeobfuscatorMain.java:26)
Caused by: java.lang.UnsupportedOperationException
... 26 more
I looked in the code, and it looks like the code uses an "invokedynamic" instruction at MethodExecutor.
Answers:
username_1: https://github.com/java-deobfuscator/deobfuscator/blob/src/main/java/com/javadeobfuscator/deobfuscator/executor/MethodExecutor.java#L1296
There is the error
username_2: InvokeDynamic aren't supported now ._.
username_0: Yeah, but it seems that Retroindy can create INDY_xx equivalents of those methods.
So is it possible to create a program that deobfuscates those statements?
username_3: Retroindy will probably cause stringer obfuscation to fail deobfuscating.
Status: Issue closed
|
oba-lang/oba | 728829970 | Title: A variables type should not change after assignment
Question:
username_0: Right now Oba's documentation claims that the type system is dynamic and strong, but it's really dynamic and weak. The VM should complain when attempting to reassign a value that is already defined with a different type<issue_closed>
Status: Issue closed |
tommy16102/Movie | 1036518461 | Title: issue 2
Question:
username_0: 한 것
1. axios로 현재 상영영화(일별박스오피스), 개봉예정영화(전체영화에서 filter로 개봉예정인거 가져옴), 순위(주간,주중,주말)
해야하는 것
1. px -> %/em/rem
2. 로그인 state -> id/password구분하기
3. 기타 component로 더쪼개기
4. 영화 axios로 가져온거 css하기
5. 기타 등등<issue_closed>
Status: Issue closed |
department-of-veterans-affairs/va.gov-team | 1092845675 | Title: 526: replace `LoadingIndicator` with web component
Question:
username_0: ## Description
All instances of `LoadingIndicator` need to be replaced with `<va-loading-indicator>` web component.
## Tasks
- [ ] Replace React `LoadingIndicator` component with `va-loading-indicator` web component
- [ ] Update unit tests
## Definition of done
- [ ] All tasks complete
- [ ] All tests passing<issue_closed>
Status: Issue closed |
composer/composer | 324165077 | Title: Dependency resolution on local packagist
Question:
username_0: The documentation states in several places that nested respository dependencies will not be resolved. However this is not the case for libraries hosted on packagist.com. When a library hosted on packagist.com specifies required dependent libraries composer will resolve any implied dependencies.
For example when [pear/log](https://packagist.org/packages/pear/log) is installed to an app using a composer.json like;
```
{
"requires" : { "pear/log": ">1.13" }
}
```
composer will install the log library **and also** the required pear/pear_exception library. That is, it resolves the dependency and installs it. I would like to benefit from required dependency resolution.
The [documentation](https://getcomposer.org/doc/05-repositories.md#composer) says of the 'composer' repository type: "This is also the repository type that packagist uses. To reference a composer repository, supply the path before the packages.json file. In the case of packagist, that file is located at /packages.json..."
So I have created a packages.json file for our internal libraries hoping that, then, composer will identify and install implied dependencies. Using the packages.json I am able to install the libraries if all the libraries are specified in an application composer.json like this:
```
{
"repositories": [
{
"type": "composer",
"url": "file:///<directory path to folder containing packages.json>"
}],
"require": {
"lyquidity/xml" : "dev-master",
"lyquidity/utilities" : "dev-master",
"lyquidity/XPath2" : "dev-master",
"lyquidity/XBRL" : "dev-master"
}
}
```
This indicates that the structure of my packages.json is correct and composer can use it. However, what I'd like to do is be able to use a composer.json like this:
```
{
"repositories": [
{
"type": "composer",
"url": "file:///<directory path to folder containing packages.json>"
}],
"require": {
"lyquidity/XBRL" : "dev-master"
}
}
```
The composer.json of the XBRL library and its definition in packages.json both specify the other three libraries (xml, utilities and XPath2) as required dependencies so my hope is that by using the same sort of package definition used by packagist.org I will be able to benefit from dependency resolution but it does not work for me.
Am I barking up the wrong tree or is there something else I should be doing to have composer install dependencies automatically.
Interestingly, although the pear/log library is not explicitly defined in the app composer.json above, composer does resolve this as a dependency *it is required by all our libraries) and installs it from packagist.org.
For completeness, here is my packages.json:
```
{
"packages": {
"lyquidity/utilities": {
"dev-master": {
[Truncated]
"license": "GPL version 3.0",
"authors": [
{
"name": "<NAME>",
"homepage": "https://github.com/orgs/lyquidity/people/username_0"
}
],
"require": {
"php": ">=7.0",
"pear/log": ">1.13",
"lyquidity/utilities" : "dev-master",
"lyquidity/xml" : "dev-master",
"lyquidity/XPath2" : "dev-master"
},
"autoload": { "files": ["./XBRL.php"] }
}
}
}
}
```
Answers:
username_1: Just to make things clear, *dependencies* are always resolved recursively. Only the "repositories" key of dependencies is not loaded recursively.
As for your problem, it seems to stem from the fact you're using dev-master releases, and not allowing a minimum-stability value of `dev`, so it can't find those releases. The root package defines the allowed stability for all packages. By hoisting the dev-master requires to the root composer.json you're effectively whitelisting these packages to be allowed in dev stability as well. That's why it "fixes" it.
Status: Issue closed
|
crystal-lang/crystal | 268976129 | Title: Unexpected behavior of `typeof(self)` restriction
Question:
username_0: This example works for now:
```crystal
class Foo
def self.foo(foo : typeof(self))
pp typeof(self) # => Class
end
end
Foo.foo Foo.new
```
It's unexpected behavior because `typeof(self)` inside method body yields `Class` but `typeof(self)` inside argument restriction seems `Foo`.
And also, this doesn't work:
```crystal
class Foo
def self.foo(foo : typeof(self))
pp typeof(self) # => Class
end
end
Foo.foo Foo
# no overload matches 'Foo.foo' with type Foo:Class
# Overloads are:
# - Foo.foo(foo : typeof(self))
```
This behavior make no sense.
Additionaly, this behavior made me confused in such case:
```crystal
class Foo
def self.foo(foo : typeof(self.bar))
pp foo
end
def self.bar
1
end
def bar
:bar
end
end
Foo.foo 1
# no overload matches 'Foo.foo' with type Int32
# Overloads are:
# - Foo.foo(foo : typeof(self.bar))
```
In this case, restriction `typeof(self.bar)` is `Int32` beacause `self` in instance of `Foo` and so `self.bar` calls `def bar` not `def self.bar`.
- - -
I guess this behavior comes from `self` type restriction:
```crystal
class Foo
def self.foo(foo : self)
pp typeof(self) # => Class
end
end
Foo.foo Foo.new
```
I think this behavior is right because `self` type restriction is special case of using `self`, we should know about this. However, I expect an expression inside `typeof` evaluate as normal, even if restriction.
Answers:
username_0: Note that this behavior is expected in spec.
https://github.com/crystal-lang/crystal/blob/6d1a6a71880bf596ad37484e323b3f87f7817967/spec/compiler/semantic/restrictions_spec.cr#L89-L99
username_1: `typeof` won't work in the future in method signatures. I'll send a PR to remove this soon.
username_2: @username_1 please open an issue first so we can discuss it!
Status: Issue closed
|
jlippold/tweakCompatible | 679679363 | Title: `NoVoiceMail` working on iOS 13.6.1
Question:
username_0: ```
{
"packageId": "net.limneos.novoicemail",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "net.limneos.novoicemail",
"deviceId": "iPhone10,5",
"url": "http://cydia.saurik.com/package/net.limneos.novoicemail/",
"iOSVersion": "13.6.1",
"packageVersionIndexed": true,
"packageName": "NoVoiceMail",
"category": "Tweaks",
"repository": "BigBoss",
"name": "NoVoiceMail",
"installed": "1.0-32",
"packageIndexed": true,
"packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.",
"id": "net.limneos.novoicemail",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.1.5",
"shortDescription": "Removes VoiceMail button from Phone.app",
"latest": "1.0-32",
"author": "<NAME>",
"packageStatus": "Unknown"
},
"base64": "<KEY>",
"chosenStatus": "working",
"notes": ""
}
```
Answers:
username_1: This issue is being closed because your review was accepted into the tweakCompatible website.
Tweak developers do not monitor or fix issues submitted via this repo.
If you have an issue with a tweak, contact the developer via another method.
username_1: This issue is being closed because your review was accepted into the tweakCompatible website.
Tweak developers do not monitor or fix issues submitted via this repo.
If you have an issue with a tweak, contact the developer via another method.
Status: Issue closed
username_1: This issue is being closed because your review was accepted into the tweakCompatible website.
Tweak developers do not monitor or fix issues submitted via this repo.
If you have an issue with a tweak, contact the developer via another method. |
transhumandesign/kag-base | 351936549 | Title: Crates with siege weapons must be picked up while unpacking
Question:
username_0: Very small change, but it should prevent people from unpacking siege weapons by accident when they try to buy something from shops.
Answers:
username_1: going to be a problem sometimes - for example, you couldn't leave a ballista on the ground to unpack and then go protect it from the enemies.
username_0: Usually you make safe place for siege before unpacking it, or get one or two teammates to defend it. Also self-assembling ballistas seem kinda weird to me.
username_1: can't really make safe space or get any teammates to help when you just stole a ballista from the enemy base
username_0: Setting up a spawn point right in enemy base shouldn't be easy.
username_1: people who let enemies steal their ballista that easily should suffer consequences. plus it's not even easy: you need to get to the enemy base first, steal a ballista that will often be protected by a bit of blocks, and protect it enough so that it will unpack and some teammates will spawn on it, and then it's still probably goign to get killed really fast. with this change its just going to be near impossible if there's an enemy near, especially if moving stops the unpacking. I would honestly rather change it so you can't have it picked up if you want to unpack, solves some of the misclick problems (for example using tunnel while holding a crate) + it makes more sense
username_0: Maybe.
username_2: Closing this in that case 👍 Thanks for the discussion.
Status: Issue closed
|
sql-machine-learning/sqlflow | 526596765 | Title: File '/tmp/sqlflow227199809/input.sql' cannot be read
Question:
username_0: ```
**Expected Behavior**
What you expected to happen.
**Screenshots**
**Environment (Please complete the following information):**
- OS:
- Browser:
- Version:
**Additional Notes**
Answers:
username_1: In my case, I've encountered an error:
```sql
%%sqlflow
SELECT *
FROM iris.train
TO TRAIN DNNClassifier
WITH model.n_classes = 3,
model.hidden_units = [10, 10],
train.epoch = 100
COLUMN sepal_length, sepal_width, petal_length, petal_width
LABEL class
INTO sqlflow_models.my_dnn_model;
```
```
_Rendezvous: <_Rendezvous of RPC that terminated with:
status = StatusCode.UNKNOWN
details = "thirdPartyParse failed: ERROR StatusLogger Unrecognized format specifier [d]
ERROR StatusLogger Unrecognized conversion specifier [d] starting at position 16 in conversion pattern.
ERROR StatusLogger Unrecognized format specifier [thread]
ERROR StatusLogger Unrecognized conversion specifier [thread] starting at position 25 in conversion pattern.
ERROR StatusLogger Unrecognized format specifier [level]
ERROR StatusLogger Unrecognized conversion specifier [level] starting at position 35 in conversion pattern.
ERROR StatusLogger Unrecognized format specifier [logger]
ERROR StatusLogger Unrecognized conversion specifier [logger] starting at position 47 in conversion pattern.
ERROR StatusLogger Unrecognized format specifier [msg]
ERROR StatusLogger Unrecognized conversion specifier [msg] starting at position 54 in conversion pattern.
ERROR StatusLogger Unrecognized format specifier [n]
ERROR StatusLogger Unrecognized conversion specifier [n] starting at position 56 in conversion pattern.
ERROR StatusLogger No log4j2 configuration file found. Using default configuration: logging only errors to the console. Set system property 'log4j2.debug' to show Log4j2 internal initialization logging.
ERROR StatusLogger Unrecognized format specifier [d]
ERROR StatusLogger Unrecognized conversion specifier [d] starting at position 16 in conversion pattern.
ERROR StatusLogger Unrecognized format specifier [thread]
ERROR StatusLogger Unrecognized conversion specifier [thread] starting at position 25 in conversion pattern.
ERROR StatusLogger Unrecognized format specifier [level]
ERROR StatusLogger Unrecognized conversion specifier [level] starting at position 35 in conversion pattern.
ERROR StatusLogger Unrecognized format specifier [logger]
ERROR StatusLogger Unrecognized conversion specifier [logger] starting at position 47 in conversion pattern.
ERROR StatusLogger Unrecognized format specifier [msg]
ERROR StatusLogger Unrecognized conversion specifier [msg] starting at position 54 in conversion pattern.
ERROR StatusLogger Unrecognized format specifier [n]
ERROR StatusLogger Unrecognized conversion specifier [n] starting at position 56 in conversion pattern.
NoViableAltException(312@[205:64: ( ( KW_AS )? alias= identifier )?])
at org.antlr.runtime.DFA.noViableAlt(DFA.java:158)
at org.antlr.runtime.DFA.predict(DFA.java:116)
at org.apache.hadoop.hive.ql.parse.HiveParser_FromClauseParser.tableSource(HiveParser_FromClauseParser.java:4171)
at org.apache.hadoop.hive.ql.parse.HiveParser_FromClauseParser.atomjoinSource(HiveParser_FromClauseParser.java:1600)
at org.apache.hadoop.hive.ql.parse.HiveParser_FromClauseParser.joinSource(HiveParser_FromClauseParser.java:1854)
at org.apache.hadoop.hive.ql.parse.HiveParser_FromClauseParser.fromSource(HiveParser_FromClauseParser.java:1527)
at org.apache.hadoop.hive.ql.parse.HiveParser_FromClauseParser.fromClause(HiveParser_FromClauseParser.java:1370)
at org.apache.hadoop.hive.ql.parse.HiveParser.fromClause(HiveParser.java:45198)
at org.apache.hadoop.hive.ql.parse.HiveParser.atomSelectStatement(HiveParser.java:39792)
at org.apache.hadoop.hive.ql.parse.HiveParser.selectStatement(HiveParser.java:40044)
at org.apache.hadoop.hive.ql.parse.HiveParser.regularBody(HiveParser.java:39690)
at org.apache.hadoop.hive.ql.parse.HiveParser.queryStatementExpressionBody(HiveParser.java:38900)
at org.apache.hadoop.hive.ql.parse.HiveParser.queryStatementExpression(HiveParser.java:38788)
at org.apache.hadoop.hive.ql.parse.HiveParser.execStatement(HiveParser.java:2396)
at org.apache.hadoop.hive.ql.parse.HiveParser.statement(HiveParser.java:1420)
at org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:220)
at org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:178)
at org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:173)
at org.sqlflow.parser.HiveQLParserAdaptor.ParseAndSplit(HiveQLParserAdaptor.java:44)
at org.sqlflow.parser.ParserAdaptorCmd.main(ParserAdaptorCmd.java:54)
Exception in thread "main" java.lang.StringIndexOutOfBoundsException: String index out of range: 235
at java.lang.String.substring(String.java:1963)
at org.sqlflow.parser.HiveQLParserAdaptor.ParseAndSplit(HiveQLParserAdaptor.java:71)
at org.sqlflow.parser.ParserAdaptorCmd.main(ParserAdaptorCmd.java:54)
```
Seems like a internal parser bug: `Exception in thread "main" java.lang.StringIndexOutOfBoundsException: String index out of range: 235`
username_2: @username_0 I am not able to reproduce it in `iris-dnn.ipynb`, where I started the Jupyter notebook by `docker run -it -p 8888:8888 sqlflow/sqlflow`.

Could you please check if the file `/tmp/sqlflow227199809/input.sql` exists? If it exists, what is the permission of the file?
username_2: @username_1 Please make sure your Docker image is the latest where it has installed the latest Java parser. I believe this error is fixed by #1202.
username_3: change 755 to 0755 can fix this bug.
username_0: I had checked the existence of the temp file and `input.sql` but the permission of those were not 755 specified in code. @username_3 and me has found this problem was caused by the format of permission. When change 755 into 0755 can solve this problem.
username_0: We can not use public docker image in internal environment. Sad.
Status: Issue closed
username_0: ```
**Expected Behavior**
What you expected to happen.
**Screenshots**
**Environment (Please complete the following information):**
- OS:
- Browser:
- Version:
**Additional Notes**
Status: Issue closed
username_2: @username_3 Thank you so much for spotting this bug. I realized that we need to write `perm so.FileMode` in octal format, which can be done by prepending a `0` in front of the digit. |
fastify/fastify-redis | 276596351 | Title: way not support asyn await ?
Question:
username_0: ```
const { redis} = fastify;
const xxx = await redis.get('xxxx');
```
Answers:
username_1: Hi!
Internally we use the official [redis](https://github.com/NodeRedis/node_redis) client.
If the driver does not support it we can't do it as well.
Status: Issue closed
|
eerimoq/mqttools | 824874898 | Title: feature request: CLI options for external hosts
Question:
username_0: This could really use options for --host and --port to connect to non-local mqtt brokers.
Answers:
username_1: It already does already has those options.
username_0: Apologies. I thought I'd searched through the usage tree, but there it is. :)
Status: Issue closed
|
remisharrock/c-programming-with-linux-MOOC-issues-tracker | 374050202 | Title: Object versus shared library (unit 7-2)
Question:
username_0: Link to the thread: https://courses.edx.org/courses/course-v1:Dartmouth_IMTx+DART.IMT.C.07+2T2018/discussion/forum/ab4624ae490e2b2c7589f718b60f2c26c6a031c8/threads/5bd07e5bef5b3309f5000704
Answers:
username_1: @username_2 @username_3 Thoughts on this? I'm not very knowledgeable about library compilation so not sure of the correct answer in this case.
username_1: @username_2 This is how it's done in the preceding video which seems to work fine, but looking around it seems that the user is right that typically creating a shared library takes a few more steps - I am not sure what the best (or even correct!) option here might be...
username_2: @username_1 @username_3 We really need Rémi's input here. I typically go for simply a .o rather than a .so in my own work, but I am not knowledgable enough to know the advantages/disadvantages of those two. We'd have to change those videos in a future iteration if we wanted to go away from the "shared" part of the library? Is this what the learner is getting at?
username_3: https://randu.org/tutorials/c/libraries.php
Creating Libraries
If you have a bunch of files that contain just functions, you can turn these source files into libraries that can be used statically or dynamically by programs. This is good for program modularity, and code re-use. Write Once, Use Many.
A library is basically just an archive of object files.
Creating Libraries :: Static Library Setup
First thing you must do is create your C source files containing any functions that will be used. Your library can contain multiple object files.
After creating the C source files, compile the files into object files.
To create a library:
ar rcs libmylib.a objfile1.o objfile2.o objfile3.o
This will create a static library called libname.a. Rename the "mylib" portion of the library to whatever you want.
That is all that is required. If you plan on copying the library, remember to use the -p option with cp to preserve permissions.
Creating Libraries :: Static Library Usage
Remember to prototype your library function calls so that you do not get implicit declaration errors.
When linking your program to the libraries, make sure you specify where the library can be found:
gcc -o foo foo.o -L. -lmylib
The -L. piece tells gcc to look in the current directory in addition to the other library directories for finding libmylib.a.
You can easily integrate this into your Makefile (even the Static Library Setup part).
also here
https://medium.com/@Cu7ious/how-to-use-dynamic-libraries-in-c-46a0f9b98270
1. How to create the dynamic library
To demonstrate how to create the dynamic library we will use the codebase of the project form the previous article:
To start we need to create the object files first with command gcc -fPIC -c *.c
As you have noticed, this time we created the object files with the -fPIC flag. This flag stands for Position Independent Code, a characteristic required by shared libraries.
On the next step we will create the library named
gcc -shared -Wl,-soname,libtools.so -o libtools.so *.o
The -shared key tells the compiler to produce a shared object which can then be linked with other objects to form an executable. -Wl flag passes an options to linker with following format -Wl,options, in case of our example it sets the name of library, as it will be passed to the linker.
As the result, we’ve got the library that is ready to be used.
2. How to use a dynamic library
Because of its purpose, our library has to be shared dynamically during the linking with other programs, and to make it happen, we have add a path to the library to the LD_LIBRARY_PATH environment variable:
export LD_LIBRARY_PATH=.:$LD_LIBRARY_PATH
In case of our example, it is the current working directory, and we can use the . to add its path. Now the operating system is aware of where to look if some program will request a functionality from the library. Alternatively to configure dynamic linker run-time bindings, we can use the ldconfig command with the -n and -L flags.
And the last but not least: a program that wants to rely on the functionality of our library must be compiled in the following manner:
gcc our_sources.c -L. -ltools -o resulted_program
where -L flag specifies the path to the library, in our case it is current directory, and -l flag specifies the name of the library to use. Please note that we didn’t provide the lib prefix, as well as the .so extension: they were resolved by the compiler.
As we can see, our program compiles and executes. |
EmilTholin/svelte-routing | 461857511 | Title: How to handle 404 or "no match"
Question:
username_0: How do I try handle a no match ?
Seeing how react routing https://tylermcginnis.com/react-router-handling-404-pages/
I'm curious how might I handle a '404' situation?
Answers:
username_0: ohkay solved it
used something like this
```
<Router url="{url}">
<nav>
<NavLink to="/">home</NavLink>
<NavLink to="about">about</NavLink>
<NavLink to="research">research</NavLink>
<NavLink to="team">team</NavLink>
</nav>
<Route path="about" component="{About}" />
<Route path="research" component="{Research}" />
<Route path="team" component="{Team}" />
<Route path="/" component="{Home}" />
<Route patth="*" component={NoMatch}/>
</Router>
```
username_1: ```html
<Router>
<Route path="blog" component="{Blog}" />
<Route path="about" component="{About}" />
<Route path="/"><Home /></Route>
<Route component={NotFound} />
</Router>
```
Status: Issue closed
username_0: oh thanks yeah that's way simpler!
Sorry I missed it in the docs! |
avan989/cFE | 495991592 | Title: Failure to test should be FAIL
Question:
username_0: The bamboo test plan does not currently complain if it is
unable to stage and run unit tests on a target, for the
simple reason that our test list is currently entirely
driven by parsing results returned by the target.
If the target VM is offline (as it was last weekend),
there are no indications that the test programs did not
run, and the presence of a few test results (the ones from
static analysis) keeps Bamboo happy.
The plan itself, or its top level scripts (same thing), needs
to keep track of the list of test programs, and generate
test failure reports for any such program for which it
does not obtain results.
Answers:
username_0: _Imported from trac issue 71. Created by glimes on 2015-06-22T10:44:28, last modified: 2019-03-05T15:05:53_
username_0: _Trac comment by glimes on 2015-06-22 10:45:09:_
See also OSAL ticket [cfs_osal:#59].
username_0: _Trac comment by glimes on 2015-06-25 09:53:02:_
The Bamboo support scripts for the CFS_CFE build have been
updated to include recent enhancements developed within
the CFS_OSAL test plan context, and now includes logic to
record -- and use -- markers indicating the exit status of
test programs, what signal killed them, or the duration of
a timeout that they exceeded.
Additionally, all current (development) CFE unit tests
write the supporting text for a test, followed by a recognizable
line indicating PASSING or FAILING status; and at the end of
the log file, record a summary of the number of PASSING
and FAILING tests.
The log file parser for CFE now includes logic to take note
of the summary lines, and produce a "ran to completion" test
case. This test case is a PASS if the last lines of the file
contained the summary; if the summary was not present or if
there was unexpected text following it, it is a FAIL.
This successfully reports early termination of CFE_ES,
not just as a report of the exit status, but also with
a failure of this program completion test that includes all
log file text since the previous test result. The complete
report of this failure can be observed (until it expires) at:
https://babelfish.arc.nasa.gov/bamboo/browse/CFS-CFSCFE-JOB2-357/test/case/44118458
Note that development of the test output log parsing code is
still in progress.
Status: Issue closed
|
odalic/odalic-ui | 230739259 | Title: No label for node with no header in the original CSV file
Question:
username_0: If there is no label in the input CSV file for certain column, then relation discovery phase shows empty node:

I would expect to show something, e.g. column 1 or name of the class if classified
Answers:
username_1: Fixed as of https://github.com/odalic/odalic-ui/commit/f5e35e432b567825507ce73b74dc05ab4b01b566 .
Status: Issue closed
|
avan989/cFE | 495988738 | Title: CFE ES makes assumptions about OSAL opaque objects
Question:
username_0: OSAL returns object identifiers which are defined as uint32 values. In the current implementation of OSAL they happen to be zero-based but this should not be a requirement; in fact there are several advantages to making these identifiers non-zero-based.
The primary offender is the ES core application using the task ID from OSAL directly as an array index.
To ensure compatibility the OSAL object IDs should be treated as opaque integers of undefined range.
Answers:
username_0: _Imported from trac issue 10. Created by jphickey on 2014-12-30T20:56:40, last modified: 2019-03-05T14:57:55_
username_0: _Trac comment by glimes on 2015-02-11 11:51:49:_
See [changeset:a220922]
username_0: _Trac comment by acudmore on 2015-04-06 14:30:57:_
recommend accept
Does this impact the ARINC653 OSAL implementation?
username_0: _Trac comment by jphickey on 2015-04-06 11:41:34:_
This is ready for review/merge
Latest is [changeset:738fdbe]. (updated to address issue found after the fact)
username_0: _Trac comment by glimes on 2015-04-07 12:51:11:_
Tested changeset [changeset:738fdbe] as part of the ic-2015-03-10 merge.
username_0: _Trac comment by glimes on 2016-02-16 13:16:45:_
Susie confirmed these tickets have been approved for CFE 6.5
username_0: _Trac comment by glimes on 2015-04-13 15:10:28:_
Part of integration candidate 2015-03-10,
committed to cFS CFE Development branch on 2015-04-10
as part of merge [changeset:7d6f6d0].
username_0: _Trac comment by jhageman on 2019-03-05 14:57:55:_
Milestone renamed
Status: Issue closed
|
w3c/wai-atag-report-tool | 599511570 | Title: inconsistent use of ATAG numbering
Question:
username_0: Example of current numbering:
- Navigation: "A.3"
- Main heading: "UI operable" (with reference to A.3 below that)
- Guideline: "A.3.1: Provide keyboard access to authoring features"
- Success Criteria: "Keyboard Access (Minimum)" (with references to "A.3.1.1 in later links and form labels)
This inconsistent explicit and implicit use of ATAG numbering is confusing. I find myself consistently trying to process what comes from ATAG and what not, and how the different pieces of content relate to each other. In particular, the introduction of new handles (ie. "UI operable") that is not part of ATAG and inconsistently associated with A.3 (see issue #56) increases this confusion.
Suggested numbering:
- Navigation: "A.3"
- Main heading: "A.3: Editing-views are operable"
- Guideline: "A.3.1: Provide keyboard access to authoring features"
- Success Criteria: "A.3.1.1 Keyboard Access (Minimum)"
Answers:
username_1: Agree with all, except…
how would you feel if forward/back button would show: “A.3: Editing-views are operable” rather than just A.3?
username_1: Actually, I tried it with just A.3 and like the simplicity so agree with the suggestion. I had already made the other changes, so that closes this issue.
Status: Issue closed
|
momeemt/Blackvas | 706020228 | Title: プラグインの実装
Question:
username_0: # プラグインの実装
#11 、 #12 のような大型の機能を実装するとき、Canvasのみを触りたいユーザーにとっては無駄にリソースが大きくなり不便になるので、プラグイン として切り分けられると便利。
プラグインを作ること自体はそう難しくないだろうが、Vuex、VueRouterなどが本当に実現するのかという疑問はあり。
でも、shapeやコンポーネントを切り出してプラグイン として読み込めるようになるのは便利。 |
ManageIQ/manageiq-ui-classic | 477383406 | Title: Tagging issue and error while adding/editing Groups
Question:
username_0: While adding/editing tags for a Group, it is not possible to remove selected values of chosen categories. Also error occurs right after choosing a value from the drop down for a selected category, while editing an existing Group.
**Scenario 1:** (adding new Group)
1. _Configuration > Access Control > Groups_ (accordion)
2. _Configuration > Add a new Group_
3. Set some tags for a new group (_My Company Tags_ tab), choose some _Category_ (for _Specific Tags_) and Value
4. Try to remove the existing tag or the value
=> it does not work, nothing happens in the UI, also no error, only adding new values and/or categories work
**Scenario 2:** (editing Group)
1. _Configuration > Access Control > Groups_ (accordion)
2. _Configuration > Edit this Group_
3. In _My Company Tags_ tab, choose some new Category (for _Specific Tags_)
4. Choose some _Value_ from the drop down for selected _Category_
=> error:
```
[----] I, [2019-08-06T15:10:28.905185 #20558:2b1c7a4bce10] INFO -- : Started POST "/ops/rbac_group_field_changed" for 127.0.0.1 at 2019-08-06 15:10:28 +0200
[----] I, [2019-08-06T15:10:28.956162 #20558:2b1c7a4bce10] INFO -- : Processing by OpsController#rbac_group_field_changed as */*
[----] I, [2019-08-06T15:10:28.956346 #20558:2b1c7a4bce10] INFO -- : Parameters: {"id"=>74, "cat"=>106, "val"=>110, "check"=>1, "tree_typ"=>"tags", "op"=>{"id"=>74, "cat"=>106, "val"=>110, "check"=>1, "tree_typ"=>"tags"}}
[----] F, [2019-08-06T15:10:28.968789 #20558:2b1c7a4bce10] FATAL -- : Error caught: [NoMethodError] undefined method `split' for 74:Integer
/home/username_0/manageiq/manageiq-ui-classic/app/controllers/ops_controller/ops_rbac.rb:747:in `rbac_field_changed'
/home/username_0/manageiq/manageiq-ui-classic/app/controllers/ops_controller/ops_rbac.rb:221:in `rbac_group_field_changed'
```
Plus it is not possible to remove selected tags or value(s) of a chosen _Category_, the same as in Scenario 1.

---
**Note:**
I've tested the same scenarios in the latest 5.11 and it works fine.
Answers:
username_0: @miq-bot assign @username_1
username_0: Note that this issue is still present, just different error occurs also for Scenario 1:
```
[----] I, [2019-09-30T11:09:29.961702 #10861:2ac33881df50] INFO -- : Started POST "/ops/rbac_group_field_changed" for ::1 at 2019-09-30 11:09:29 +0200
[----] I, [2019-09-30T11:09:30.003754 #10861:2ac33881df50] INFO -- : Processing by OpsController#rbac_group_field_changed as */*
[----] I, [2019-09-30T11:09:30.003961 #10861:2ac33881df50] INFO -- : Parameters: {"id"=>"39", "cat"=>"119", "check"=>1, "tree_typ"=>"tags", "op"=>{"id"=>"39", "cat"=>"119", "check"=>1, "tree_typ"=>"tags"}}
[----] F, [2019-09-30T11:09:30.019162 #10861:2ac33881df50] FATAL -- : Error caught: [NoMethodError] undefined method `name' for nil:NilClass
/home/username_0/manageiq/manageiq-ui-classic/app/controllers/ops_controller/ops_rbac.rb:1086:in `rbac_group_get_form_vars'
/home/username_0/manageiq/manageiq-ui-classic/app/controllers/ops_controller/ops_rbac.rb:754:in `rbac_field_changed'
/home/username_0/manageiq/manageiq-ui-classic/app/controllers/ops_controller/ops_rbac.rb:221:in `rbac_group_field_changed'
```
username_1: I forgot to link the issue, but it should be fixed by https://github.com/ManageIQ/manageiq-ui-classic/pull/6235
Status: Issue closed
username_0: Closing this issue as it is already fixed. |
inbo/INLA | 787758637 | Title: INLA Shiny app deployment error
Question:
username_0: Hi there,
I've been trying to deploy a Shiny app using R-INLA and it looks like one potential solution would be installing INLA via devtools::install_github() given Shiny doesn't handle non-CRAN and non-Github repos very well. I've tried using this repo and it worked locally of course and it deploys. Once I actually deploy though, I can't perform inla functions - it looks like I run into linux issues:
e.g. a few lines of this
/opt/R/4.0.3/lib/R/library/INLA/bin/linux/64bit/inla.mkl: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.27' not found
I was wondering if you had any thoughts and/or have tried using this INLA repo for Shiny app deployment.
Thanks!
Jonathan<issue_closed>
Status: Issue closed |
spring-projects/spring-data-mongodb | 1014322674 | Title: Dynamic agregation pipeline from method parameter
Question:
username_0: I'm looking for dynamic @Aggregation pipeline affectation.
you know already that we could affect a string array to @Aggregation annotation like this:
@Aggregation({
"{ $unwind : { path : $skills } }",
"{ $group : { _id : $skills, names : { $push : $name } } }",
"{ $project : { _id : 0, skill : $_id, names : 1 } }"
})
Flux<SkilledPeople> getSkilledPeople();
and also a string like this:
@Aggregation("{ $project : { _id : 0, skill : $_id, names : 1 } }")
Flux<SkilledPeople> getSkilledPeople();
peeking the old @query annotation, we could pass the query like this:
@Query("?0")
Page<Person> filter(Document document, Pageable pageable);
### My question is, could we have similar approach link in @Query annotation?
@Aggregation("?0")
Flux<SkilledPeople> getSkilledPeople(String[] pipeline);
or like this?
@Aggregation("?0")
Flux<SkilledPeople> getSkilledPeople(List<String> pipeline);
stack question: https://stackoverflow.com/questions/69423466/dynamic-agregation-pipeline-from-method-parameter |
codefresh-contrib/cfstep-helm | 352651400 | Title: Support multiple Helm Repositories from single Codefresh Pipeline
Question:
username_0: As a Codefresh user I would like to specify multiple Helm Repositories in a single Codefresh pipeline.
Acceptance Criteria:
AC1: Can import shared configuration from any Codefresh Helm Repository
AC2: Can specify the Helm Repository to use during the Helm Pipeline Step<issue_closed>
Status: Issue closed |
OndraZizka/csv-cruncher | 366815603 | Title: Quote all SQL identifiers to retain the case.
Question:
username_0: * all parts of SQL statements are converted to upper case before processing, except identifiers in double quotes and strings in single quotes
* identifiers, both unquoted and double quoted, are then treated as case-sensitive
* most database engines follow the same rule, except, in some respects, MySQL and MS SQLServer.
To retain the case, we would have to quote all uses of columns everywhere.
Answers:
username_0: Looks like this works in commit 0b62ca4 (which will likely become 2.1.0).
Status: Issue closed
username_0: * all parts of SQL statements are converted to upper case before processing, except identifiers in double quotes and strings in single quotes
* identifiers, both unquoted and double quoted, are then treated as case-sensitive
* most database engines follow the same rule, except, in some respects, MySQL and MS SQLServer.
To retain the case, we would have to quote all uses of columns everywhere.
Status: Issue closed
|
salesforce/bazel-eclipse | 717822237 | Title: Implement inline annotations in BUILD files for IDEs
Question:
username_0: We are supporting Project View files, which provide hints and metadata for how IDEs should import slices of Bazel workspaces. We will continue to support them, but...
I prefer an inline approach. I think we should support (and encourage) inline annotations in comments in the BUILD file. This initial story will choose just one use case, and then we build from there.
Some ideas:
# @projectname salad-apple-api
This will be useful when there are multiple packages named apple-api imported in the workspace. By default BEF will name the first it finds 'apple-api', and the second 'apple-api2'. By using the annotation, better names will be used for the projects during import.
# @jdk11
BEF currently reads .bazelrc data to configure the JDK for all projects imported from the workspace. The user can then override it for each project, manually. It is the JDK used by JDT to do code completion and inline compilation. But it would be nicer if this was configured by annotation, then all users would get it automatically. This could be read out of the rules (javacopt) but if there are conflicts within the package it is better for an annotation to tell us what to do.
# @workingset orion
# @workingset fusion
# @workingset troika
In this example, this Bazel package is a component used by three teams: orion, fusion, and troika teams within ACME Inc.
(every big company has had a big project code-named 'fusion' at some point). These working sets would be the equivalent for the Project View 'directories' concept. The user would choose one or more workingsets during import, and all packages annotated with those labels would be imported. Then, using Eclipse Working Sets, the user could toggle which projects were active in the view. |
sisoputnfrba/foro | 443018299 | Title: Problema con el update de la VM
Question:
username_0: Buenas! Como les va?
Hace como mas de 3 o 4 semanas que vengo tratando de updatear la VM (con los updates esos que caen de seguridad y etc) y me salta este archivo de error despues de que se crashea el update inmediatemente arranca:

Y dentro del archivo dice ésto:

Sin embargo tengo como 40gb de espacio libre? Que se rompió?
Answers:
username_1: updatear estamos hablando de todas las actualizaciones que aparecen cuando prendés la VM?
Como cátedra recomendamos no actualizarla, ya que se puede dar el caso que alguna biblioteca relacionada al TP cambie y después el día de la entrega algo no les compile o funcione como deba.
Dicho eso, en cuanto al error no estoy completamente seguro, pero ahí dice que lo que falta es memoria, no disco. Puede que poniendo más RAM ande
otra razón puede ser que Lubuntu de 32 bits perdió soporte hace unos meses. Tal vez se deba por eso también
username_0: Tengo 4gb de ram asignados a la VM, me tiraria a pensar que no le dan los 32 bits para direccionar el tamaño de memoria de manera correcta, cierro el issue nomas, Gracias!
Status: Issue closed
|
NativeScript/theme | 470103410 | Title: demo app not 6.0 compat
Question:
username_0: Is the `/app` dir in this repo not the demo app?
Answers:
username_1: I think they meant this branch:
https://github.com/NativeScript/theme/tree/next
username_2: Indeed. I should merge in master one of these days.
username_0: Yea prob makes sense with the timing of the 6.0 release going GA
username_0: FWIW I can't get `next` to work.
```
~/p/theme (next|✚1…) $ git branch
ls
master
* next
~/p/theme (next|✚1…) $ npm install
...
The current project contains a webpack.config.js file located at /Users/ryan/projects/theme/webpack.config.js that differs from the one in the new version of the nativescript-dev-webpack plugin located at /Users/ryan/projects/theme/node_modules/nativescript-dev-webpack/templates/webpack.javascript.js. Some of the plugin features may not work as expected until you manually update the webpack.config.js file or automatically update it using "./node_modules/.bin/update-ns-webpack --configs" command.
...
tns run ios
Searching for devices...
Cannot find module 'nativescript-dev-sass/lib/before-prepare.js'
```
```
~/p/theme (next|✚1…) $ tns --version
6.0.1
~/p/theme (next|✚1…) $ node --version
v10.16.0
~/p/theme (next|✚1…) $ npm --version
6.9.0
```
Status: Issue closed
|
Subsets and Splits