repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
brianshub/final-project | 206367330 | Title: Feedback on milestone #2
Question:
username_0: Great to see you have your structure to build off of. Look forward to seeing it come to life and I totally acknowledge that in-class I had only asked for HTML structure for this milestone, so this is excellent! |
pyomeca/ezc3d | 704091125 | Title: Error in compiling static library on Ubuntu 20.04
Question:
username_0: Hi @username_1,
I'm trying to compile a statically linked library for Ubuntu 20.04 platform. I can successfully compile a shared library but not static. I'm getting the following error while running `make`:
```bash
Scanning dependencies of target ezc3d
[ 4%] Building CXX object CMakeFiles/ezc3d.dir/src/math/Matrix.cpp.o
[ 9%] Building CXX object CMakeFiles/ezc3d.dir/src/math/Matrix33.cpp.o
[ 13%] Building CXX object CMakeFiles/ezc3d.dir/src/math/Matrix66.cpp.o
[ 18%] Building CXX object CMakeFiles/ezc3d.dir/src/math/Vector3d.cpp.o
[ 22%] Building CXX object CMakeFiles/ezc3d.dir/src/math/Vector6d.cpp.o
[ 27%] Building CXX object CMakeFiles/ezc3d.dir/src/Analogs.cpp.o
[ 31%] Building CXX object CMakeFiles/ezc3d.dir/src/Channel.cpp.o
[ 36%] Building CXX object CMakeFiles/ezc3d.dir/src/Data.cpp.o
[ 40%] Building CXX object CMakeFiles/ezc3d.dir/src/ezc3d.cpp.o
[ 45%] Building CXX object CMakeFiles/ezc3d.dir/src/Frame.cpp.o
[ 50%] Building CXX object CMakeFiles/ezc3d.dir/src/Group.cpp.o
[ 54%] Building CXX object CMakeFiles/ezc3d.dir/src/Header.cpp.o
[ 59%] Building CXX object CMakeFiles/ezc3d.dir/src/Parameter.cpp.o
[ 63%] Building CXX object CMakeFiles/ezc3d.dir/src/Parameters.cpp.o
[ 68%] Building CXX object CMakeFiles/ezc3d.dir/src/Point.cpp.o
[ 72%] Building CXX object CMakeFiles/ezc3d.dir/src/Points.cpp.o
[ 77%] Building CXX object CMakeFiles/ezc3d.dir/src/Subframe.cpp.o
[ 81%] Building CXX object CMakeFiles/ezc3d.dir/src/modules/ForcePlatforms.cpp.o
[ 86%] Linking CXX static library libezc3d.a
[ 86%] Built target ezc3d
Scanning dependencies of target _ezc3d_swig_compilation
[ 90%] Swig compile ezc3d_python.i for python
[ 90%] Built target _ezc3d_swig_compilation
Scanning dependencies of target _ezc3d
[ 95%] Building CXX object binding/python3/CMakeFiles/_ezc3d.dir/CMakeFiles/_ezc3d.dir/ezc3d_pythonPYTHON_wrap.cxx.o
[100%] Linking CXX shared module _ezc3d.so
/usr/bin/ld: ../../libezc3d.a(Matrix.cpp.o): relocation R_X86_64_PC32 against symbol `_ZNK5ezc3d6Matrix6nbColsEv' can not be used when making a shared object; recompile with -fPIC
/usr/bin/ld: final link failed: bad value
collect2: error: ld returned 1 exit status
make[2]: *** [binding/python3/CMakeFiles/_ezc3d.dir/build.make:86: binding/python3/_ezc3d.so] Error 1
make[1]: *** [CMakeFiles/Makefile2:198: binding/python3/CMakeFiles/_ezc3d.dir/all] Error 2
make: *** [Makefile:130: all] Error 2
```
I'm following the same steps that I followed on macOS to successfully compile the static library:
```bash
git clone https://github.com/pyomeca/ezc3d.git
cd ezc3d && mkdir build && cd build
cmake ../ -DBUILD_SHARED_LIBS=FALSE -DCMAKE_BUILD_TYPE=Release -DBUILD_EXAMPLE=FALSE -DBINDER_PYTHON3=ON -DPython3_EXECUTABLE=<path to python3.8 binary> -DSWIG_EXECUTABLE=<path to swig root>/swig -DSWIG_DIR=<path to swig root>/Lib
make -j $(nproc)
```
Do I need to set any additional `cmake` variables on Ubuntu? It seems like the linking is happening for a shared library using libezc3d.a which is compiled for static linking.
Thank you,
Ankur
Answers:
username_1: Hi there!
I am glad it now works :)
I have never compiled it statically on Linux before. It seems that the -fPIC compilation option is not there. I will add this during the day. I always make sense to have this option anyway :)
username_1: #152 Should solve the issue :)
Thanks again!
Status: Issue closed
|
strapi/strapi | 1116160443 | Title: Transaction in strapi
Question:
username_0: Hello,
I'm new on strapi and I'm investigating if I should start using it or not.
Until now, I could replicate most of the functionality I need, however, I was not able to create a MySQL transaction.
I have searched in the documentation and forums but I was not able to find an example of how to create a transaction so I can commit multiple queries and rollback if something has gone wrong.
Is there any solution for this?
Thank you.
Status: Issue closed
Answers:
username_1: Hi thanks for the question. I'm closing this here however since it's not an issue but a question that is perfect for our [Community Forum](https://forum.strapi.io). Once you post it there, I and other community members will be happy to share our ideas. |
dotnet/sdk | 735324713 | Title: ILLink warnings tests should include the version of the runtime libraries in its output
Question:
username_0: As seen in https://github.com/dotnet/installer/pull/8966 it's actually pretty tricky to figure out which version of the runtime/framework assemblies are used when running the [ILLink_verify_analysis_warnings_hello_world_app](https://github.com/dotnet/sdk/blob/0ae5136153da2d364288be1d9cdd4804d8279adf/src/Tests/Microsoft.NET.Publish.Tests/GivenThatWeWantToRunILLink.cs#L288) test. For the test it should be pretty easy to print out the exact version of one of the assemblies from the framework (CoreLib would be best).
Answers:
username_1: While doing this, we should double-check that the test is running against the correct TFM. It requests `net5.0`, but based on https://github.com/dotnet/sdk/pull/14338#issuecomment-725007437 I think it has been running against `net6.0` bits. I would suggest updating the test to run against `net6.0` (intentionally :)).
Status: Issue closed
|
zooniverse/aggregation-for-caesar | 859646526 | Title: ensure dropdown values have there labels exported in look up table
Question:
username_0: Currently, the hash values for dropdown value are not included in the string lookup table output from the workflow config script.
In the dropdown config, each option has the format:
```
{""label"":""T0.selects.0.options.*.0.label"",""value"":""10c4674f90d0d8""}
```
In the string lookup the code gives:
```
T0.selects.0.options.*.0.label: '1'
```
And the aggregation gives:
```
{'10c4674f90d0d8': <count>}
```
This is less than helpful.
Options:
1. keep the string extraction script the same but use `label` rather than `value` in the dropdown extractor.
2. update the string extraction script to include both the `label` and the `value` for dropdowns (would need an `if` statement).<issue_closed>
Status: Issue closed |
gbif/ipt | 263651934 | Title: Fix XSS issues
Question:
username_0: The security scanner run by Smithsonian has found XSS issues to address. Liaise in private with those developers and ensure the issues are addressed.
Answers:
username_0: Fixed with https://github.com/gbif/ipt/commit/bed9f2e461df3c05764e5f661306e61e8e794de0
Status: Issue closed
username_1: The security scanner run by Smithsonian has found XSS issues to address. Liaise in private with those developers and ensure the issues are addressed.
Status: Issue closed
username_2: Passes all xss tests from AppSpider and https://detectify.com/
username_1: Great @username_2, thanks. But the above commit https://github.com/gbif/ipt/commit/bed9f2e461df3c05764e5f661306e61e8e794de0 broke other parts of the IPT's functionality that still require fixing.
username_2: excluded /manage and /admin from firing the filter: https://github.com/gbif/ipt/commit/3ce4adbd9103ce38617695b44d75746e93620235 |
appirio-tech/connect-app | 537273072 | Title: Admin User not able to Edit a project phase when they are not part of the project
Question:
username_0: 1. Login as Admin user <EMAIL> in the test env ( https://connect-test01.topcoder-dev.com/ )
2. For a project where this user is not a member, add a phase
3. Then click on Edit pencil icon of the added phase
4. Check
Actual:
I am not able to edit the phase. Clicking on Edit just makes the pencil icon disappear.


Answers:
username_0: cc: @username_3 @username_2
username_1: @username_2 it seems a valid issue and needed to addressed before the release.
username_2: Fixed via https://github.com/appirio-tech/connect-app/commit/073b0a3fb4acb7e77f3656a33c80466313bedbbd.
Deployed to both TEST and DEV env.
Tested via <EMAIL>:

username_3: Admin is able to Add and Delete Phases. However, Admin user is still not able to Update Phase.
username_0: @username_2 Being an admin user <EMAIL> , I am not able to edit the phase of this project https://connect-test01.topcoder-dev.com/projects/8588
However, there is no issue if I edit the phase of a newly created project. Can you please check why? Thanks.
username_3: I notice the same behavior even for Copilot user for existing projects.
username_2: @username_0 @username_3 I'm not sure if I found the same issues as you, but there was another issue in Webkit based browsers which caused showing: `An invalid form control with is not focusable.` in the browser console. I could reproduce it for any user.
It's now fixed and deployed to both DEV and TEST. Let me know if you still have issues with updating the phases.
username_3: Verified Edits by Admin and Copilot.
Status: Issue closed
|
quarkusio/code.quarkus.io | 966198255 | Title: Create a api guide for external quarkus codegen tooling
Question:
username_0: @username_1 you can already use http://stage.code.quarkus.io/ for this (this is more a pre-prod than a staging). Most of the time the changes gets to code.quarkus.io a few minutes/hours later.
Answers:
username_1: It would be great to have some staging server that we can hit using automatic tests with the latest changes in web APIs.
username_0: @username_1 you can already use http://stage.code.quarkus.io/ for this (this is more a pre-prod than a staging). Most of the time the changes gets to code.quarkus.io a few minutes/hours later. |
pseudomuto/protoc-gen-doc | 406272912 | Title: Using glob for proto files
Question:
username_0: It is possible to use glob sintax to reference the proto files?
`docker run --rm -v /home/work/proto/**/:/protos -v /home/work/a:/out pseudomuto/protoc-gen-doc --doc_opt=markdown,docs.md`
Answers:
username_1: Any luck with this, @username_0? I'm trying to find out how to get the tool to traverse the directory.
username_0: Well, what I ended up doing was to flatten the proto files and after that run the doc generation.
I was on a mac so I did this to flatten the protos:
```
mkdir ./proto_flat
find ./proto -name '*.proto' -exec mv {} ./proto_flat \;
```
And then:
`docker run --rm -v /home/work/proto_flat -v /home/work/doc:/out pseudomuto/protoc-gen-doc --doc_opt=markdown,docs.md` |
solana-labs/solana | 435809805 | Title: Re-evaluate Influx cloud subscription plan
Question:
username_0: #### Problem
We have subscribed to Influx Cloud Pro II plan. Our usage of Influx DB could be optimized to change to a lower plan.
#### Proposed Solution
Analyze our usage of Influx DB. See if we can buffer data in fullnode for longer before sending it to Influx Cloud. Goal is reduce peak to average usage of Influx DB (remove spikes). This might be tricky, as there are multiple testnets sending data randomly.
Answers:
username_1: Dup of #4116
Status: Issue closed
|
nodejh/nodejh.github.io | 208460284 | Title: 使用 SQL *Plus 管理 Oracle 数据库
Question:
username_0: ## 导入导出
数据库的导入导出也是一个很常见的需求。
#### 导出
```
$ exp [username]/[password]@[orcl] file=./database.dmp full=y
```
+ `username` 是数据库用户名
+ `password` 是数据库用户密码
+ `orcl` 是数据库实例名称
+ `file` 后面的参数是导出的数据库文件存放位置及文件名
+ `full` 其值为 `y` 表示全部导出,默认为 `no`。
如果只需导出某几张表,可以指定 `tables` 参数:`tables='(tableName, tableName1)'`。
#### 导入
```
$ imp [username]/[password]@[orcl] file=./database.dmp
```
和导出数据库语法一样,只是关键字不一样。
#### 执行 SQL 文件
执行 SQL 文件的方法有很多种。如下:
**使用 SQL PLUS 命令**
```
$ sqlplus [username]/password@[orcl] @path/file.name
```
或者远程执行:
```
$ sqlplus [username]/password@server_IP/service_name @path/file.name
```
如果sql脚本文件比较复杂,包含了begin end语句,就会不断显示行号,解决办法就是在 sql 脚本的最后用 `/` 符号结尾。
**在 SQL PLUS 中执行**
```
SQL>start file_path
```
```
SQL>@ file_path
```
其中 `file_path` 是文件路径。
---
参考
+ [Oracle表空间(tablespaces)](http://www.cnblogs.com/fnng/archive/2012/08/12/2634485.html)
+ [Oracle数据库,实例,表空间,用户,表之间的关系](http://zyjustin9.iteye.com/blog/2193804)
+ [Oracle 在Sqlplus 执行sql脚本文件](http://nvd11.blog.163.com/blog/static/2000183122012111524636835/) |
kelektiv/node.bcrypt.js | 185127882 | Title: node-gyp rebuild fails on Ubuntu 16.04.1 + nodejs 4.6.1 + bcrypt 0.8.7
Question:
username_0: This is the error/debug log I'm getting:
```
make: Entering directory '/home/koneo/bundle/programs/server/npm/node_modules/bcrypt/build'
CXX(target) Release/obj.target/bcrypt_lib/src/blowfish.o
CXX(target) Release/obj.target/bcrypt_lib/src/bcrypt.o
CXX(target) Release/obj.target/bcrypt_lib/src/bcrypt_node.o
g++: internal compiler error: Killed (program cc1plus)
Please submit a full bug report,
with preprocessed source if appropriate.
See <file:///usr/share/doc/gcc-5/README.Bugs> for instructions.
bcrypt_lib.target.mk:98: recipe for target 'Release/obj.target/bcrypt_lib/src/bcrypt_node.o' failed
make: *** [Release/obj.target/bcrypt_lib/src/bcrypt_node.o] Error 4
make: Leaving directory '/home/koneo/bundle/programs/server/npm/node_modules/bcrypt/build'
gyp ERR! build error
gyp ERR! stack Error: `make` failed with exit code: 2
gyp ERR! stack at ChildProcess.onExit (/usr/lib/node_modules/npm/node_modules/node-gyp/lib/build.js:276:23)
gyp ERR! stack at emitTwo (events.js:87:13)
gyp ERR! stack at ChildProcess.emit (events.js:172:7)
gyp ERR! stack at Process.ChildProcess._handle.onexit (internal/child_process.js:200:12)
gyp ERR! System Linux 4.4.0-36-generic
gyp ERR! command "/usr/bin/nodejs" "/usr/lib/node_modules/npm/node_modules/node-gyp/bin/node-gyp.js" "rebuild"
gyp ERR! cwd /home/koneo/bundle/programs/server/npm/node_modules/bcrypt
gyp ERR! node -v v4.6.1
gyp ERR! node-gyp -v v3.4.0
gyp ERR! not ok
npm ERR! Linux 4.4.0-36-generic
npm ERR! argv "/usr/bin/nodejs" "/usr/bin/npm" "rebuild"
npm ERR! node v4.6.1
npm ERR! npm v2.15.9
npm ERR! code ELIFECYCLE
npm ERR! [email protected] install: `node-gyp rebuild`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the [email protected] install script 'node-gyp rebuild'.
npm ERR! This is most likely a problem with the bcrypt package,
npm ERR! not with npm itself.
npm ERR! Tell the author that this fails on your system:
npm ERR! node-gyp rebuild
```
Status: Issue closed
Answers:
username_1: Maybe a memory issue on your box or some other configuration. There is not enough information in this log to debug it; your compiler just dies (usually a compiler bug or memory issue)
username_0: Okay. Thanks. I'll investigate further. |
Arcaratus/BloodArsenal | 181679182 | Title: [1.10.2] Please add a way to automatically trigger rituals
Question:
username_0: I have a modpack in the works that uses Blood Magic's custom meteors for all ore generation and other stone-type blocks. However, I'd like for players to be able to fully automate resource gathering eventually, and this presents a problem since there's no way in base Blood Magic to automatically trigger the ritual and the Blood Magic devs don't want to allow fake players to do it.
What I'm hoping you might be able to add is some kind of block that when given an activation crystal and a redstone signal will activate the MRS it's attached to for some cost. Perhaps it could drain will from the chunk somehow to incentivise going into that stuff?
Answers:
username_1: Um, dude, it's planned for the base BM mod for there to be a way to reactivate the meteor ritual. Please be patient.
Status: Issue closed
|
fent/node-ytdl-core | 927796064 | Title: Is not working any more
Question:
username_0: this dependecy is not working any more formats object is empty now
Answers:
username_1: can this be a dupe of https://github.com/fent/node-ytdl-core/issues/939 ?
the library is having problems in general atm
username_2: @username_1 @username_0
YouTube has updated their API
which led to ytdl to not work
The only solution is to wait for ytdl to update their code for the new YouTube API
Thanks
username_3: How long does it usually take?
and How do we know that they did update their code for the new YouTube API?
Thanks
username_4: Just subscribe to release, they will do their update here.
username_5: How to do that
username_4: 
username_3: thanks, but still the question remains unanswered, does anyone know if it usually takes hours? weeks? centuries?
username_4: 
username_6: @username_3: I'm not a maintainer, and I don;t mean to come off rude, but this project is totally open-source. It's pretty much instant if you make it, otherwise, you are dependent on unpaid volunteer's free time. Nobody owes you free development, we're all in it together, and it's not helpful to be like "when will it be done?" a bunch.
For people looking into this, here is a simplified function that also fails, but uses similar to current method:
```js
export const getVideoInfo = async (id, options = {}) => {
const url = new URL(`https://www.youtube.com/get_video_info`)
url.searchParams.set('video_id', id)
url.searchParams.set('eurl', `https://youtube.googleapis.com/v/${id}`)
url.searchParams.set('ps', 'default')
url.searchParams.set('gl', options.country || 'US')
url.searchParams.set('hl', options.lang || 'en')
url.searchParams.set('html5', '1')
const u = url.toString()
const r = await fetch(u)
console.log('URL', u)
console.log('REQUEST', r.status, await r.text())
}
```
```
URL https://www.youtube.com/get_video_info?video_id=K-281doxOMc&eurl=https%3A%2F%2Fyoutube.googleapis.com%2Fv%2FK-281doxOMc&ps=default&gl=US&hl=en&html5=1
REQUEST 404
```
In the context of a youtube player page, you can get lots of great info with this (including video URLS in `streamingData.adaptiveFormats`) from `window.ytInitialPlayerResponse`.
Here is a quick function exploiting those things, which currently works:
```js
import fetch from 'node-fetch'
const regex = /var ytInitialPlayerResponse = (.+);<\/script>/gm
export async function getInfo (id) {
const r = await fetch(`https://www.youtube.com/watch?v=${id}`)
const str = await r.text()
const m = regex.exec(str)
return JSON.parse(m[1])
}
getInfo('K-281doxOMc')
.then(console.log)
```
@fent if this seems like an ok direction, I can make a PR.
username_6: On further inspection, the URLs it returns don't seem to actually open.
username_1: @username_6 no reason to excuse - you're not rude
the `https://www.youtube.com/watch?v=xxx` page is one of the pages we do already fetch
feel free to check out https://github.com/fent/node-ytdl-core/pull/938 for the current hotfix
username_6: Ah yes, sorry, I looked closer at the code. It grabs `getWatchHTMLPage`, `getWatchJSONPage`, `getVideoInfoPage`, where`getWatchHTMLPage` is similar to my code, above. I wonder if for my purposes I need the other data (which is the problem, right?) I am trying to just get an array of video formats that work. Is the data from the other functions needed? I like the idea of only making 1 request, instead of a few. I have this running in a little video proxy that redirects to the raw URL, so the fastest path to the raw URL is probably best for my case, if it's not missing anything else I need.
username_1: there is a reason we use multiple sources - just one is not reliable
username_7: any solution?
username_8: a fix has been implemented in #938, now you just gotta wait for the merger *(or like me manually patch the lib)*
username_3: @username_6
Thank you man for explaining
I had no idea about the volunteering and open source nature of it all
Now I'm so much more in love with FireFox and will try to participate myself even though I only know C++ to use in programming chips (I'm an RF EE) and the code you guys are posting looks like Java
Any guidance to a starting point to learning your stuff would be awesome.
No rude in straightening a spoiled little creature like myself
username_3: @username_6
Thank you man for explaining
I had no idea about the volunteering and open source nature of it all
Now I'm so much more in love with FireFox and will try to participate myself even though I only know C++ to use in programming chips (I'm an RF EE) and the code you guys are posting looks like Java
Any guidance to a starting point to learning your stuff would be awesome.
No rude in straightening a spoiled little creature like myself
username_3: how do you "patch the lib"? and who is "lib"? where does he/she hang out?
username_8: lol lib isn't a person, lib == library and i used [`patch-package`](https://www.npmjs.com/package/patch-package) to patch it
username_3: Thanks. That sounds like a good starting point for an ignoramus like me.
BTW, I know "who" lib is, I'm glad you didn't sense my sarcasm and that it gave you a laugh or a few seconds of hi time hahaha
username_9: Hi
i'm new to this how do you implement the "patch package " into app itself
can somone please explain this.
//Rg
username_10: In node_modules/ytdl-core/lib/info.js
Line 254 replace
let jsonStr = body.split("ytInitialPlayerResponse = {")[1].split("}}};")[0] += "}}}";
Not me who found this out but it may be a temp fix until next version comes
username_9: as i said new at this , but where to find "In node_modules/ytdl-core/lib/info.js"
username_10: in your code, find file node_modules then find ytdl-core then find lib then find info.js
username_10: In node_modules/ytdl-core/lib/info.js
Line 254 replace
let jsonStr = body.split("ytInitialPlayerResponse = {")[1].split("}}};")[0] += "}}}";
Not me who found this out but it may be a temp fix until next version comes
Worked for me anyways
username_9: my "code files" don't under stand
username_10: i think I'm able to help , do you have any messaging apps like discord or messenger?
username_8: it's pretty simple, just follow the steps below:
- install [`patch-package`](https://www.npmjs.com/package/patch-package) using yarn/npm
- have a look at #938 and copy the changes that does into your version of `ytdl-core` *(should be located at `node_modules/ytdl-core`)*
- then run the command `npx patch-package ytdl-core`
- add a new script into your `package.json` with the key `postinstall` and value `patch-package`
- run npm/yarn install
this will make it so every time you run a npm/yarn install it automatically patches `ytdl-core` for you *(so it's like a one time hassle, you can forget about it until `ytdl-core` merges #938)*
username_3: "Apps" stands for applications, right? like in phones or android tablets?
Is this forum all about that stuff?
Another thing, do I have to restart or uninstall/install firefox or windows on my computer every couple of hours for any fix in this browser add-on to take effect?
username_6: @username_10 I think my regex above does essentially the same thing, but is maybe a tiny bit more efficient. As far as I understand it, the problem is in another place, though. It's because the lib combines data from several different sources, and one of them is failing, so looking for `ytInitialPlayerResponse` works well in `getWatchHTMLPage` (which is what these 2 snippets are for) but we are getting errors on `getVideoInfoPage`. @username_1 points out that we need all the data-sources, because different video-types need more data.
As for patching, hand-editing your changes, and all that stuff, I personally would recommend against all this. `npm install` can work with git-repos, and as long as they don't need to be built before being published (ytdl-core is good), you can reference the changes directly. Here is an example of how to install the patched version:
```
npm i git://github.com/D3SOX/node-ytdl-core#fix-info-endpoint-404
```
That is where the PR comes from (`user/repo#branch`)
The bonus of this method is it's very easy to go back to the official version, once the changes are published on npm:
```
npm i ytdl-core@latest
```
username_3: I think I get it now. This is a programming homework help session that I somehow entered while investigating why a firefox browser adon called "youtube video and audio downloader (Dev Edt)" stopped showing download options, but instead it shows "info.player_response is undefined" and that is why I am being ignored even tho I'm the family genius and made it to the seventh grade and memorized the whole multiplication table because I couldn't pay for a scientific calculator. I'll go look for a different adon.
username_6: Ok. That addon has nothing to do with this library, other than depending on it. Go take it up with them.
username_8: this has barely anything to do with ytdl-core as @username_6 said. if you're trying to be funny this isn't the right place buddy
username_1: gonna close this as related to https://github.com/fent/node-ytdl-core/issues/939 and fixed with v4.8.3
feel free to open a new issue / write here / reopen this issue when i'm wrong
Status: Issue closed
|
torrentpier/torrentpier | 206217289 | Title: PHP Notice in POST /admin/admin_attach_cp.php
Question:
username_0: ## Error in TorrentPier
**PHP Notice** in **POST /admin/admin_attach_cp.php**
Undefined variable: search_count_greater
[View on Bugsnag](https://app.bugsnag.com/torrentpier/torrentpier/errors/589b2ce5814a72381f72c081?event_id=589b2ce67a824610002b83c3)
## Stacktrace
/var/www/torrentpier/data/www/demo.torrentpier.me/library/attach_mod/includes/functions_admin.php:235 - search_attachments
[View full stacktrace](https://app.bugsnag.com/torrentpier/torrentpier/errors/589b2ce5814a72381f72c081?event_id=589b2ce67a824610002b83c3)
Answers:
username_0: Fixed in #208
Status: Issue closed
|
docker-java/docker-java | 371178913 | Title: Bind.java Parse does not account for Windows file paths
Question:
username_0: The [Bind.parse()](https://github.com/docker-java/docker-java/blob/80c62136df94502c627e9c40dc9cac2f2bfced0d/src/main/java/com/github/dockerjava/api/model/Bind.java#L98) function does not account for Windows paths that use colons to delimit the drive letter at the beginning of the path.
For example, a perfectly valid docker command line volume option (on Windows) would be `-v c:\temp:c:\logs` The Bind.parse() function assumes that the colons are used as delimiters between the host path & container path. Obviously on Windows this is not the case.
Answers:
username_1: Hi,
Unfortunately I have the same problem. There is any update about the issue?
I'm using the **3.2.0-rc1** version.
Here the error that it gives when I Inspect Containers containing Windows Paths for binding:
`com.fasterxml.jackson.databind.JsonMappingException: Error parsing Bind 'C:\Work\BuildDeploy:C:\inetpub\wwwroot:rw' (through reference chain: com.github.dockerjava.api.command.InspectContainerResponse["HostConfig"]->com.github.dockerjava.api.model.HostConfig["Binds"])`
username_2: We're currently applying some major changes to the library. Once we stabilize them, we will focus on bugfixes, including this one.
I hope you will understand it. Thanks!
@username_0 will you consider re-submitting your PR against current master? :) |
AnswerEquals42/HaskellGame | 349849668 | Title: Animate actor motion
Question:
username_0: Would like to see smooth motion from one node to the next as well as smooth rotation when changing direction.
I think the way to go is to add a current and next field for Position and Facing, then animate between with state stepper. So,
data Actor = Actor
{ actorType :: ActorType
, ...
, facing :: Facing
, facing' :: Facing
, position :: Position
, position' :: Position }
Status: Issue closed
Answers:
username_0: Added ActorAnim type to contain a list of frames, where each frame is a (Float, Float) for the translation magnitude and rotation. Whenever an Actor has frames, the transformations in them are applied. Frames are consumed from the left. |
gatsbyjs/gatsby | 263249858 | Title: Importing bulma css causes postcss-custom-properties error
Question:
username_0: All I did was installed bulma
`yarn add bulma`
Then imported bulma css
`import "bulma/css/bulma.css"`
And this the error I get

Answers:
username_1: does bulma rely on css properties?
username_0: @username_1 Yes, they do use custom css properties. Is that a problem ?
username_1: No, just trying to understand the issue a bit. I think the problem is here that the default css setup adds a bunch of cssNext setup along with it, which isn't a great default..
username_2: I just ran into the same issue. I created a local copy of the gatsby postcss sass plugin and ripped out the postcss (since I don't need it).
username_3: Posting my solution to this for others:
* Install `gatsby-plugin-sass` (see [here](https://github.com/gatsbyjs/gatsby/tree/master/examples/using-sass))
* Create `layouts/index.scss` (see [here](https://github.com/jgthms/bulma/issues/1190#issuecomment-356672849))
```sass
$variable-columns: false;
@import "~bulma/bulma";
```
* Import `index.scss` in `layouts/index.js`
username_4: @username_3 had to add one more step to get that to work:
* Add `"gatsby-plugin-sass"` to `plugins:` in `gatsby-config.js`
Status: Issue closed
username_5: Due to the high volume of issues, we're closing out older ones without recent activity. Please open a new issue if you need help! |
jenkinsci/dynatrace-plugin | 211096491 | Title: compatibility with Pipeline
Question:
username_0: Hi! I'd like to be able to use this plugin with pipeline scripts. We can (I believe) simply extend some of the action classes in order to wrap it in, so I'm willing to make the necessary updates and ensure test coverage/passing etc. just need to know where is strategically the best place to fork/branch at this time.
Thanks,
<NAME>
Answers:
username_1: I'd say fork it to your own namespace and have take a stab at it
username_2: Hi there, I am working to extend this plugin for jenkins pipeline support. I think I made a good progress but I am stuck at one issue that does not seem to work. It is failing to set the environment variables during step 1 and step 2. I don't get any exception/error but when I echo all environment variables, it is not showing any of the following. If I use the plugin ui flow (non pipeline), it seems to be working. Could you please take a look at my code and suggest what is wrong? I am very new to the plugin development and in fact it is first time, I am extending any plugin. If you can suggest any improvements after looking at my code, that would also be greatly appreciated.
**Environment variables that are supposed to be set:**
dtMarker=testservice
dtPassword=<PASSWORD>
dtProfile=testprofile
dtServerUrl=https://<someurl>:8021
dtTestrunID=<test id>
dtUsername=<username>
dtVersionBuild=36
dtVersionMajor=testservice
My code is here: [https://github.com/username_2/dynatrace-plugin/tree/jenkins-pipeline-support](https://github.com/username_2/dynatrace-plugin/tree/jenkins-pipeline-support)
I also posted this on stack overflow for help with more details. It contains more details about the issue I am facing: [https://stackoverflow.com/questions/44804903/dynatrace-app-mon-plugin-jenkins-pipeline-compatibility-issue-setting-environm](https://stackoverflow.com/questions/44804903/dynatrace-app-mon-plugin-jenkins-pipeline-compatibility-issue-setting-environm)
username_2: Just to add one more point, I started the development with base code from release 2.0.5 [https://github.com/jenkinsci/dynatrace-plugin/releases/tag/dynatrace-dashboard-2.0.5](https://github.com/jenkinsci/dynatrace-plugin/releases/tag/dynatrace-dashboard-2.0.5). I noticed there was some issue with the code in the master branch and I wasn't able to do testConnection to dynatrace server with that code. But the code from release tag worked fine, so I took that code as my base code.
username_0: my understanding is that the build parameters won't echo as environment variables (presumably because they can be accessed by any node at any point on the pipeline). have you confirmed that your implementation is strictly not working, or is it just not logging as you'd expect?
username_2: Hi @username_0 - The code is able to set the build parameters during pipeline flow. However, it is not able to set them as environment variables. If I use this plugin in a regular free style job, I see both build parameters and environment variables set, which I am able to use in my maven command via %dtTestrunID% (my build node is windows machine). I think my implementation seems to be working since it is able to generate the test id.
If I run 'set' command after step 2 (register test run) in a free style job, I can see all env variables set along with build parameters. However, if I run the same 'set' command after calling this step in the pipeline, it does not set those env varilables, only sets the build parameters.
//dynatrace step-2
dynatraceAppMonRegisterTestRun category: 'unit', platform: ''
Also, I thought of using build parameters in the pipeline script since it is setting those properly but I couldn't find a way to access the build parameters in the pipeline script. I tried several options that I found via google search but none worked. Suppose, access build parameters worked, we still need to find the issue of why it is not setting the env variables.
username_0: 'params.MY_PARAMETER_NAME' should resolve
username_2: Just tried, params.dtTestrunID returns null, however I go to jenkins job -> buildnumber -> parameters, I can see all the build parameters set there already. Is it returning null because job is not build with parameters?
username_0: oh wait is it just 'dtTestrunID'? I was looking at a Jenkinsfile when I referenced that, not a normal pipeline script
username_2: for now, I only need access to dtTestrunID in my pipeline script so it can be passed to the mvn build command. I still think there is something wrong with my code that it is unable to set the environment variables while the existing code is able to. If you have some time, would you mind trying my code out at your side? You can compile it to get the hpi file. Also, just wondering how do you usually debug during plugin development?
username_2: @username_0 @username_1 - Just wondering if you guys have some time to look into my code and suggest what could be the issue? Any help would be highly appreciated.
username_3: @username_0 could you please update the plugin and let us know if it suits your needs now?
If you'd like to adjust it on your own I'd suggest to fork the repo. |
wso2/docs-ei | 560796251 | Title: It does not contain a doc for Slack Connector
Question:
username_0: **Description:**
We have not listed the Slack Connector under [1] and do not have a doc that explains its operations.
[1] - https://docs.wso2.com/display/ESBCONNECTORS/WSO2+ESB+Connectors
Answers:
username_1: For more recent connectors, the docs reside in GitHub repo itself: https://github.com/wso2-extensions/esb-connector-slack
We have added some notes in high level doc pages to point to these [1] [2].
[1] - https://ei.docs.wso2.com/en/latest/micro-integrator/references/connectors/connectors-overview/#example-scenarios-and-operations
[2] - https://ei.docs.wso2.com/en/latest/micro-integrator/use-cases/integration-use-case/connectors/
Status: Issue closed
|
communitybridge/easycla | 596474498 | Title: V2 API (metrics/total-count) - Issues with repositoriesCount when few added repos are disabled in github
Question:
username_0: ## Background
**Prerequisite**
1. Github Organization should be connected to Easy CLA Appilcation and should be configured for All Repositories
2. Valid project should exist
3. All Repos of the github should be added to the Project
4. Trigger the lambda function to get latest metrics and note down **repositoriesCount** count


**Steps to Follow**
1. Launch github organization
2. Navigate to Settings -> Install Github Apps
3. Click Configure for DEV CommunityBridge: EasyCLA
4. Under Repository Access section click on Only select repositories
5. Select only few repositories for e.g 2 repositories
6. Click Save


**_Though in Configure Github Repositories only 2 repositories are shown which is as expected, still all repos are shown in Foundation details page for the project_**

7. Trigger the Lambda function to fetch latest results
10. Observed count is not decremented

Status: Issue closed
Answers:
username_1: @username_0 - current behavior is working as expected. The system detects that the GH repos has been disabled and sends an email to the Project Managers informing of this. No further action is taken by the EasyCLA system - we are waiting on the PM's to fix/resolve the problem rather than make changes to the database. |
pantsbuild/pants | 361415849 | Title: Make incremental compiles work hermetically
Question:
username_0: Currently, Incremental compiles on a sub-target level don't work hermetically. They should.
Answers:
username_1: Closing as this will fall out naturally once compiles are ported to V2, so we don't need a dedicated ticket for it.
Status: Issue closed
|
nvs-vocabs/S11 | 522196060 | Title: Pupa
Question:
username_0: Term name (PrefLabel)
Pupa
Definition
The life stage of some insects undergoing transformation between immature and mature stages. The pupal stage is found only in holometabolous insects, those that undergo a complete metamorphosis. The processes of entering and completing the pupal stage are controlled by the insect's hormones, especially juvenile hormone, prothoracicotropic hormone, and ecdysone.
Sources/references
<NAME>.; DeLong, <NAME>.; Triplehorn, <NAME>. (2004). Introduction to the Study of Insects (Sixth ed.).
Answers:
username_1: @username_0 The following term has been added to S11 and will be published on 16/11/2019 as http://vocab.nerc.ac.uk/collection/S11/current/S1180
Status: Issue closed
|
Lombiq/Orchard-Training-Demo-Module | 585172182 | Title: OrchardCore documentation links broken
Question:
username_0: The links to https://orchardcore.readthedocs.io/ are broken. For example in `YourFirstOrchardCoreController.cs` there is a link to https://orchardcore.readthedocs.io/en/latest/OrchardCore.Modules/OrchardCore.Localization/README/, which should probably be https://orchardcore.readthedocs.io/en/latest/docs/reference/modules/Localize/
Answers:
username_1: Would you mind submitting a PR for the broken links you encounter?
Status: Issue closed
username_1: Fixed them. |
Azure/azure-powershell | 538880568 | Title: Cmdlets review link
Question:
username_0: Cmdlets review link is broken (https://github.com/Azure/azure-powershell-cmdlet-review-pr)
Can you please provide us the exact URL.

Answers:
username_1: @username_0 are you from Microsoft or official partners? It's a private repo and you need to apply access permission.
username_0: Hi ,
I am internal employee and my alias is kakattam. Can you please let me know in email how to apply access permission .
username_1: Resolve it by internal email.
Status: Issue closed
|
logaretm/vee-validate | 219489147 | Title: Support For Internet Explorer
Question:
username_0: It seems validation fails on internet explorer and browsers that do not support promises.
#### Versions:
- VueJs: 2.x
- Vee-Validate: 2.0.0-beta.25
### Description:
It seems validation fails on internet explorer and browsers that do not support promises.
### Steps To Reproduce:
Run on internet explorer < 10
Answers:
username_1: You can polyfill promises using existing polyfills out there, or my favorite: https://polyfill.io/v2/docs/
I try to include as little polyfills as possible to keep the library size from growing too much, especially that promises are likely to be used by another library, so polyfilling them multiple times can be counter intuitive.
Status: Issue closed
username_0: thanks.
username_2: I'd suggest putting at least a small note on the readme mentioning that support for Promises is required via a polyfill or lib for IE support. In my case Promises aren't used by anything else and this was causing errors in production.
username_1: @username_2 noted
username_3: getting an error like below when my code is runned in IE11 but,not getting the error when it is runned in chrome.
err:
SCRIPT1002: Syntax error
jquery-2.1.1.js (2,2536)
when removing vee-validate no error its running fine..how to fix this..? |
giantotter/giantotter_public | 113607051 | Title: Can't leave a game and start a new one.
Question:
username_0: This is not an issue now, since we only have a single game running, however when we allow for multiple games, there is the possibility of hitting this.
The room naming convention is based on the nickname of the player. If player 'A' leaves a game in progress in room '#A', that game doesn't end because there are other people playing in it. When 'A' tries to create a new game, the room bot will attempt to place them in a room named '#A'.
Answers:
username_1: no longer relevant
Status: Issue closed
|
vrbvillor/vrbvillor.github.io | 159761014 | Title: $cookieStore
Question:
username_0: - service in module ngCookies
Provides a key-value (string-object) storage, that is backed by session cookies. Objects put or retrieved from this storage are automatically serialized or deserialized by angular's toJson/fromJson.
Requires the ngCookies module to be installed. |
antlr/grammars-v4 | 228574209 | Title: Synchronize topics, labels and directories
Question:
username_0: I suppose to use the same names in lowercase for unification and beauty.
Rename the following labels:
* `C` -> `c`
* `c#` -> `csharp`
* `t-sql` -> `tsql`
* `Swift` -> `swift`
* `VB` -> `vb`
Answers:
username_1: done
Status: Issue closed
|
mido/mido | 912965085 | Title: Informal port names
Question:
username_0: Given these port names:
```python
'Midi Through:Midi Through Port-0 14:0'
'SH-201:SH-201 15:0'
```
I'd like to be able to open them with:
```python
mido.open_input('sh-201')
mido.open_output('integra')
```
This is pretty easy to add to the `Backend` class so it works the same for all backends. It would first check for an exact match, and if not found search of an case insensitive match anywhere in the string and return the first port that matches.
https://github.com/gbevin/sendmidi which does this exact thing.
The RtMidi backend already has similar functionality for omitting port numbers and client name. |
odin-lang/Odin | 1096732312 | Title: Compiler crash on `distinct` types in `union` variants (Parapoly)
Question:
username_0: ## Context
Odin: dev-2022-01:b02f2953
OS: Arch Linux, Linux 5.15.12-arch1-1
CPU: AMD Ryzen 5 2500U with Radeon Vega Mobile Gfx
RAM: 7652 MiB
## Expected Behavior
`b` is assigned a variable of type `A`
## Current Behavior
the compiler crashes during type checking.
## Failure Information (for bugs)
the compiler outputs the following error message and exits with error code `132`
```
bug.main
lb_emit_conv: src -> dst
Not Identical []int != C
Not Identical []int != union {A, B}
Not Identical 7f00b0036c60 != 7f00b0038920
Not Identical 7f00b0036c60 != 7f00b00389a0
src/llvm_backend_expr.cpp(2113): Panic: Invalid type conversion: '[]int' to 'C' for procedure 'bug.main'
```
### Steps to Reproduce
compile the following code
```odin
package bug
import "core:fmt"
import "core:slice"
A :: distinct []int
B :: distinct []int
C :: union { A, B }
main :: proc() {
a: C = A{}
switch a in a {
case A:
b := A(slice.clone(([]int)(a)))
case B:
case:
}
}
``` |
cloudfoundry/hwc-buildpack | 941540084 | Title: **Release:** hwc-buildpack 3.1.19
Question:
username_0: **Dependency Changes:**
```diff
No dependency changes
```
**New Commits on Develop**:
1505b3a Update libbuildpack
7dbc438 Adds integration suite matrix
6a1c994 Merge pull request #47 from cloudfoundry/automation/github-config/update
db4cda5 Updating github-config
f5a0415 Merge pull request #46 from cloudfoundry/automation/github-config/update
b1440a6 Updating github-config
Refer to [release instructions](https://docs.cloudfoundry.org/buildpacks/releasing_a_new_buildpack_version.html).
Answers:
username_1: No changes to merit a release
Status: Issue closed
|
scunning1975/mixtape | 826073527 | Title: Typo in reference to figure
Question:
username_0: On [this page](https://mixtape.scunning.com/instrumental-variables.html#the-problem-of-weak-instruments), there is a reference to figure 49 whereas all other figures are references as Figure 7.x. Here's the paragraph:
"Figure 49 shows the first stage, and it is really interesting. Look at all those 3s and 4s at the top of the picture. There’s a clear pattern—those with birthdays in the third and fourth quarter have more schooling on average than do those with birthdays in the first and second quarters. That relationship gets weaker as we move into later cohorts, but that is probably because for later cohorts, the price on higher levels of schooling was rising so much that fewer and fewer people were dropping out before finishing their high school degree."
Answers:
username_1: Thank you!
Status: Issue closed
|
Kinto/kinto | 217590217 | Title: Improve configuration of accounts administrators
Question:
username_0: In the accounts API, the setup of administrators requires two steps:
* creating the actual accounts
* adding the principals in the settings
Instead we could have a `kinto createadmin` command for example. |
galasa-dev/projectmanagement | 620781212 | Title: Galasa Ecosystem Documentation
Question:
username_0: Attaching a document with areas to be covered and will create a child issue for each area.[https://app.zenhub.com/files/220212715/fcd6e7a5-5901-4fa8-9f23-1dd0446b86fa/download](https://app.zenhub.com/files/220212715/fcd6e7a5-5901-4fa8-9f23-1dd0446b86fa/download)
Answers:
username_1: I'm assuming this can move to backlog as Caroline has started the work. |
m-labs/nmigen | 547060698 | Title: vendor.xilinx_7series: Vivado TIMING-2 Warning
Question:
username_0: *Vivado* issues TIMING-2 (default severity: CRITICAL) when `create_clock` uses a signal.
Here is a potential solution to please `create_clock`: https://github.com/username_0/nmigen/commit/9e79a30655123c02a2a3025d8f5af0ce8c076b7f .
Answers:
username_1: Can you explain what does the filter expression in your commit does?
username_0: @username_1 this is indeed ugly and should be fixed, it's purpose is to filter negative ports of differential clock inputs, ending with `__n`. So it matches ports ending with `__p` or `__io`. Internal clocks are matched using `^.+/.+$`, for signals like `my_module/clk_out`.
username_1: You did explain why it's ugly, which is helpful (thanks!). Unfortunately, you didn't explain what it *does*, so I can't help you work out a nicer replacement. I don't have the Xilinx flavor of Tcl API memorized.
username_0: Here's an excerpt from [UG906](https://www.xilinx.com/support/documentation/sw_manuals/xilinx2019_2/ug906-vivado-design-analysis.pdf) explaining the TIMING-2 warning and how to use `create_clock` correctly:

The Tcl fragment does exactly that, it starts from the clock signal and determines it's startpoint using `all_fanin -flat -startpoints_only`. The returned could contains `my-signal__p my-signal__n`, therefore `filter -regexp` filters `my-signal__n`, so only a single item remain.
```Tcl
[filter -regexp [all_fanin -flat -startpoints_only \
[get_nets {{signal|hierarchy("/")}}]] {NAME =~{(^.*__p$)|(^.*__io$)|(^.+/.+$)}}]
```
Status: Issue closed
|
AmoDinho/formulaone-graphql-client | 364542463 | Title: Uncontrolled/Controlled Components
Question:
username_0: ```
index.js:2178 Warning: `value` prop on `input` should not be null. Consider using an empty string to clear the component or `undefined` for uncontrolled components.
in input (at CreateDriver.js:85)
in div (at CreateDriver.js:70)
in div (at CreateDriver.js:69)
in CreateDriver (at App.js:9)
in App (at src/index.js:22)
in ApolloProvider (at src/index.js:21)
```
```
Warning: A component is changing an uncontrolled input of type number to be controlled. Input elements should not switch from uncontrolled to controlled (or vice versa). Decide between using a controlled or uncontrolled input element for the lifetime of the component. More info: https://fb.me/react-controlled-components
in input (at CreateDriver.js:85)
in div (at CreateDriver.js:70)
in div (at CreateDriver.js:69)
in CreateDriver (at App.js:9)
in App (at src/index.js:22)
in ApolloProvider (at src/index.js:21)
```
Answers:
username_0: I needed to change the initial state values from `null` to `0`.
Status: Issue closed
|
electron-userland/electron-builder | 1178573531 | Title: Signing app in Mac
Question:
username_0: I'm traying to sign my app in mac, but I alwasy get the same error: **"identity name is specified but no valid identity" with this name in the keychain**
I tried set "identity" attribute in my "electron-builder.config.json" with:
- Apple Development: XXXXX XXXXX (XXXXXX)
- XXXXX XXXXX (XXXXXX)
- (XXXXXX)
- 2352352342363464634634 Apple Development: XXXXX XXXXX (XXXXXX)
... and always get the same error... |
miniclip/gonsul | 814280851 | Title: YAML expansion does not work as expected
Question:
username_0: Version: 0.3.0
Problem
YAML expansion only works for individual values (string, bool, number), but not nested dictionaries. In the case of both individual and nested values in a single YAML file, only individual values are parsed and written to Consul, and nested dictionaries are ignored.
Example.yaml
key1: value1
key2: value2
dict:
key3: value3
sub:
key4: value
Result: Only key1 and key2 end up in Consul, the rest is ignored
Answers:
username_1: Hello @username_0 ,
We've recently bumped the yaml dependency to a newer version, can you confirm if this still happens with the latest release?
Cheers.
username_1: Since we had no reply for a bit, I'm closing this.
Status: Issue closed
|
filecoin-project/devgrants | 1066953805 | Title: Next Step Microgrant: <Pluskit>
Question:
username_0: ### 1. What is your project? (max 100 words)
Pluskit is a Fil+ eco-tools platform dedicated to providing functional, operational and typole-rich tools and plug-ins for eco-participants. It now provides data board of General Verification to help eco-users easily and clearly know the status of DataCap application and allocation.
It uses Python3.0 language and the overall framework uses Django, which conforms to the restful style. The database uses mysql8 and sets indexes in key fields. Due to the huge amount of data, Redis is used as the cache, which will greatly increase the speed of interface return, improve the user experience, and contribute to the development of IPFS.
GitHub:https://github.com/username_0/Pluskit
Website: https://pluskit.io/
### 2. How are you planning to improve this project? (max 200 words)
We have three plans in the next step:
-Develop Large Datasets Board and Automatic Verification Data Board, containing the status of DataCap applications, allocations and storage, to help ecological users understand the reality of verified data storage.
-Launch 1~2 tools of DataCap allocation and recording to help reduce DataCap acquisition-allocation time.
-Carry out a round of promotion to let more people know and use our products, so as to help more users in the ecosystem.
### 3. Will the work be Open Source?
Yes, the work will be Open Source.
### 4. If selected, do you agree to share weekly updates and a grant report upon conclusion?
Yes, Of course.
### 5. Does your proposal comply with our Community Code of Conduct?
Yes
### 6. Links and submissions
<!-- Complete each step, and include the link of the published submission (or "Yes" if there is no URL) -->
* Have you submitted to the Filecoin Community Showcase ([instructions](https://github.com/filecoin-project/community#showcase-your-project))?
Yes
* Have you created a thread introducing your project in the community discussion board's [ Project Showcase category](https://github.com/filecoin-project/community/discussions/categories/project-showcase)?
https://github.com/filecoin-project/community/discussions/390
* If your project began at a hackathon, have you submitted it for the relevant Protocol Labs prizes? Include links here if available:
### Additional questions:
* For each team member(s), please list name, email, Github account, and role in the project.
Product Manager:chic,https://github.com/chicChang
Front-end developer:jiaming,https://github.com/zhf0437
Backend developer:Mr. Zhou,https://github.com/EzrealoveQXY
* If your project was created as part of an event or hackathon:
* What was the name of the event? (e.g. ETHGlobal NFTHack, Cal Hacks hello:world, Chainlink, CivHacks, GameDevJ, ETHGlobal Scaling Ethereum)
* Please link to your hackathon submission
Answers:
username_1: Looking forward to seeing more of this!
username_2: Hi, folks. I would like to approve this microgrant but I need at least one email address. Can you please message back with one or more email addresses for a team member?
username_0: @username_2
username_0: @username_2
Sure, <EMAIL> or <EMAIL>. Thanks!
username_0: Triage notifications on the go with GitHub Mobile for iOS or Android.
username_2: This microgrant has been approved! Look out for an email to submit payment details. Let's use your existing Community Discussion thread for updates.
Status: Issue closed
|
fullcalendar/fullcalendar-vue | 474088721 | Title: eventSource's rendering property ignored
Question:
username_0: In my application i tried to load events from 2 different sources, each of them is an array. The first source contains events an employee has to attend to. The second source contains absences of each employee and needs to be rendered as background events.
To realize that, i created 2 arrays as Vue prop and bound them in an array of eventSources to add other options, like `rendering` or `editable`. Unfortunately the `rendering` property seems to be ignored.
https://codesandbox.io/embed/fullcalendar-vue-fd8rp
Answers:
username_1: The issue is not specific to Vue, this doesn't work in the basic debugging template either:
https://codepen.io/anon/pen/wVJwLx?editors=0010
Only some properties available in the event source object, as descibed here:
https://fullcalendar.io/docs/event-source-object#event-source-options
So, I will move this to the main repo as a feature request. |
material-components/material-components-ios | 404496314 | Title: MDCBottomAppBarView layout quirks
Question:
username_0: I'm not sure of the correct way add the MDCBottomAppBarView to my view. Right now I have a storyboard and I've created a view at the bottom and set its Custom Class to MDCBottomAppBarView.
I don't know what the correct height is for the view. I've pinned it to the bottom and sides and then I want to pin my content view to the top of the bottom app bar. To do this I need to give it a specific height.
It seems if I pick anything under around 90 that the bar is shifted down somewhat and there is a pretty big gap between the bar and the content view.
See here for <90 px height.
<img width="471" alt="screen shot 2019-01-29 at 4 27 33 pm" src="https://user-images.githubusercontent.com/1138573/51941703-38dbfb80-23e3-11e9-8edf-72234b252e2b.png">
If I give it exactly 90 height or more then it lays out correctly but there is still a large gap between the content and the bar:
<img width="425" alt="screen shot 2019-01-29 at 4 32 28 pm" src="https://user-images.githubusercontent.com/1138573/51941813-8193b480-23e3-11e9-8c37-1f96203917d2.png">
What is the correct way to add this as a subview such that lays out correctly and is flush with the content view?
Answers:
username_1: @username_2 Could you please take a look?
username_0: Is there a canonical way of adding this with autolayout? Is this the correct way?
Status: Issue closed
username_0: I'm not sure of the correct way add the MDCBottomAppBarView to my view. Right now I have a storyboard and I've created a view at the bottom and set its Custom Class to MDCBottomAppBarView.
I don't know what the correct height is for the view. I've pinned it to the bottom and sides and then I want to pin my content view to the top of the bottom app bar. To do this I need to give it a specific height.
It seems if I pick anything under around 90 that the bar is shifted down somewhat and there is a pretty big gap between the bar and the content view.
See here for <90 px height.
<img width="471" alt="screen shot 2019-01-29 at 4 27 33 pm" src="https://user-images.githubusercontent.com/1138573/51941703-38dbfb80-23e3-11e9-8edf-72234b252e2b.png">
If I give it exactly 90 height or more then it lays out correctly but there is still a large gap between the content and the bar:
<img width="425" alt="screen shot 2019-01-29 at 4 32 28 pm" src="https://user-images.githubusercontent.com/1138573/51941813-8193b480-23e3-11e9-8c37-1f96203917d2.png">
What is the correct way to add this as a subview such that lays out correctly and is flush with the content view?
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/123642146](http://b/123642146)
username_0: Alright. I've switched to using autolayout in my view that contains the bottom app bar. My constraints are like this:
- pinned to left and right sides of the superview
- pinned to bottom safe area
- height constraint connected to code via IBOutlet
As I'm sure you know the total height of the view differs depending on whether the FAB is showing. So I've added this code to adjust the view height
```swift
func layoutBottomAppBar() {
let size = bottomBar.sizeThatFits(view.bounds.size)
barHeightConstraint.constant = size.height
}
override func viewWillLayoutSubviews() {
super.viewWillLayoutSubviews()
layoutBottomAppBar()
}
#if swift(>=3.2)
@available(iOS 11, *)
override func viewSafeAreaInsetsDidChange() {
super.viewSafeAreaInsetsDidChange()
layoutBottomAppBar()
}
#endif
```
This pins my content view to the _top of the FAB_ not to the top of the bar. I don't want that gap between my content and the bar. Based on setting a breakpoint inside `MDCBottomAppBarView` here;
```objective-c
- (CGPoint)getFloatingButtonCenterPositionForAppBarWidth:(CGFloat)appBarWidth {
CGPoint floatingButtonPoint = CGPointZero;
CGFloat navigationBarTopEdgeYOffset = CGRectGetMinY(self.navBar.frame);
CGFloat midX = appBarWidth / 2;
floatingButtonPoint.y = MAX(0, navigationBarTopEdgeYOffset - self.floatingButtonVerticalOffset);
...
```
I found that the correct offset is 38. Is there a clean way to access this value programmatically so that I can adjust my constraints accordingly? I've tried `floatingButtonVerticalOffset` but that always seems to be `0.0`
username_2: @username_0 Hi, sorry for the late reply. There are two assumptions I want to clarify: (1) the height of the bottom app bar view doesn't differ depending on whether the FAB is showing or not, (2) the safeAreaInset is taken into consideration when calculating the height of the bottom app bar view. So I would suggest try removing the `bottomInset` calculation from `layoutBottomAppBar` function first in this case.
To align the content view, we suggest using a full screen size content view and put bottom app bar above it according to the guidance and examples.
If aligning the bottom of your content view with the top of bottom app bar is something your app specifically want to achieve, you can achieve it by aligning the bottom of content view with (the center of the FAB + floatingButtonVerticalOffset).
username_0: Okay. I can see in one of the BottomAppBar component examples with a tableview it adds a content inset to the tableview to account for the height of the bottom bar. So in the case where the controller that holds my bottom bar can have a child view controller I'll either need to:
1. Know that the child has a scroll view and set the content inset of the scroll view
2. Require that the child conform to some kind of protocol so I can tell it the height of the bottom bar so that it can set its own content inset.
3. For child VCs that don't have a scroll view add some margin to the bottom so they don't have views that are covered by the bottom bar.
Does that sound right?
username_2: @username_0 That sounds right. As you mentioned, one of the BottomAppbar examples sets the contentInset by bottomAppbar's size, which is a good example in this case. |
pact-foundation/pact-specification | 688804297 | Title: Using Matchers.ISO8601_DATETIME_WITH_MILLIS_FORMAT doesn't work as expected.
Question:
username_0: In a pact-node consumer test, I have -
```
describe('edit a policy', () => {
const REQUEST_BODY = {
name: 'account-holder',
version: '2020-07-10T01:22:09.488Z',
};
const RESPONSE_BODY = {
name: Matchers.like('account-holder'),
version: Matchers.ISO8601_DATETIME_WITH_MILLIS_FORMAT,
additionalAttributes: Matchers.like({}),
};
beforeEach(() =>
provider.addInteraction({
state: 'policy account-holder already exist',
uponReceiving: 'edit a policy with body ',
withRequest: {
method: 'PUT',
headers: {
'Content-Type': 'application/json',
},
path: '/policies/account-holder/versions/2020-07-10T01:22:09.488Z',
body: REQUEST_BODY,
},
willRespondWith: {
status: 200,
headers: {
'Content-Type': 'application/json; charset=utf-8',
},
body: RESPONSE_BODY,
},
})
);
it('returns success', () =>
client.updatePolicyByNameAndVersion(REQUEST_BODY).then((policy) => {
expect(policy).toEqual(Matchers.extractPayload(RESPONSE_BODY));
}));
});
```
against fixture -
```
{
name: 'account-holder',
version: new Date().toISOString(),
additionalAttributes: { name: 'some other data object' },
},
```
This fails with
```
Failure/Error: expect(response_body).to match_term expected_response_body, diff_options, example
Actual: {"name":"account-holder","version":"2020-08-17T03:25:28.175Z","additionalAttributes":{"name":"some other data object"}}
Diff
--------------------------------------
Key: - is expected
+ is actual
Matching keys and values are not shown
{
- "version": "^\\d{4}-[01]\\d-[0-3]\\dT[0-2]\\d:[0-5]\\d:[0-5]\\d\\.\\d{3,}([+-][0-2]\\d(:?[0-5]\\d)?|Z)$"
+ "version": "2020-08-17T03:25:28.175Z"
}
Description of differences
--------------------------------------
* Expected "^\\d{4}-[01]\\d-[0-3]\\dT[0-2]\\d:[0-5]\\d:[0-5]\\d\\.\\d{3,}([+-][0-2]\\d(:?[0-5]\\d)?|Z)$" but got "2020-08-17T03:25:28.175Z" at $.version
```
What makes it work is - Thank you @username_2 for the answer.
```
version: Matchers.term({
matcher: Matchers.ISO8601_DATETIME_WITH_MILLIS_FORMAT,
generate: // Some date example
}),
```
But, using the Matchers.ISO8601_DATETIME_WITH_MILLIS_FORMAT, in such a way wasn't obvious. Maybe the doco could reflect this?
Status: Issue closed
Answers:
username_1: Actually neither of those is the correct way of using the matcher, the correct way is to use the `iso8601DateTimeWithMillis()` matcher function.
The [docs](https://github.com/pact-foundation/pact-js/#matching) are very clear on this:

(P.S. thanks for raising, but for Pact JS specific problems please raise them at https://github.com/pact-foundation/pact-js/)
username_2: @username_1 I don't think the documentation is clear. At least, I can't tell how to use it from that table (every time I need a matcher I end up going and reading the source instead - I've been meaning to address this).
I think there are two improvements we could make here:
* I don't think the helper regex functions should be exposed on the `Matchers` object, as this creates an inconsistent API. This is especially a problem now that JS code autocompletion is better.
* We could improve the documentation of the matchers to make it clear what functions are available and how to use them.
username_2: Also, I want to transfer this issue to pact-js, but I don't have the option. Possibly it's because I don't have permissions on `pact-specification`. Are you able to do it? |
jmeter-maven-plugin/jmeter-maven-plugin | 366964214 | Title: Release 2.8.0
Question:
username_0: @username_1 release, tag=`2.8.0`
Answers:
username_1: @username_0 OK, I will release it now. Please check the progress [here](http://www.username_1.com/t/16089-1)
username_1: @username_0 Oops, I failed. You can see the full log [here](http://www.username_1.com/t/16089-1) (spent 3min)
```
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-parameter-documenter/2.0.9/maven-plugin-parameter-documenter-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-artifact-manager/2.0.9/maven-artifact-manager-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/reporting/maven-reporting-api/2.0.9/maven-reporting-api-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-container-default/1.0-alpha-9-stable-1/plexus-container-default-1.0-alpha-9-stable-1.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-core/2.0.9/maven-core-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/doxia/doxia-sink-api/1.0-alpha-10/doxia-sink-api-1.0-alpha-10.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-repository-metadata/2.0.9/maven-repository-metadata-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-registry/2.0.9/maven-plugin-registry-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-error-diagnostics/2.0.9/maven-error-diagnostics-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-parameter-documenter/2.0.9/maven-plugin-parameter-documenter-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/reporting/maven-reporting-api/2.0.9/maven-reporting-api-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/commons-cli/commons-cli/1.0/commons-cli-1.0.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-descriptor/2.0.9/maven-plugin-descriptor-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-repository-metadata/2.0.9/maven-repository-metadata-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/doxia/doxia-sink-api/1.0-alpha-10/doxia-sink-api-1.0-alpha-10.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-monitor/2.0.9/maven-monitor-2.0.9.jar
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-interactivity-api/1.0-alpha-4/plexus-interactivity-api-1.0-alpha-4.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-error-diagnostics/2.0.9/maven-error-diagnostics-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/classworlds/classworlds/1.1/classworlds-1.1.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-descriptor/2.0.9/maven-plugin-descriptor-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/commons-cli/commons-cli/1.0/commons-cli-1.0.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-utils/1.5.8/plexus-utils-1.5.8.jar
Downloading: https://repo.maven.apache.org/maven2/commons-lang/commons-lang/2.3/commons-lang-2.3.jar
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-interactivity-api/1.0-alpha-4/plexus-interactivity-api-1.0-alpha-4.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-monitor/2.0.9/maven-monitor-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-api/1.2/enforcer-api-1.2.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-rules/1.2/enforcer-rules-1.2.jar
Downloaded: https://repo.maven.apache.org/maven2/classworlds/classworlds/1.1/classworlds-1.1.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-common-artifact-filters/1.2/maven-common-artifact-filters-1.2.jar
Downloaded: https://repo.maven.apache.org/maven2/commons-lang/commons-lang/2.3/commons-lang-2.3.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/beanshell/bsh/2.0b4/bsh-2.0b4.jar
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-utils/1.5.8/plexus-utils-1.5.8.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-dependency-tree/2.0/maven-dependency-tree-2.0.jar
3/10 KB
5/10 KB
8/10 KB
3/75 KB 8/10 KB
3/75 KB 10/10 KB
5/75 KB 10/10 KB
8/75 KB 10/10 KB
11/75 KB 10/10 KB
13/75 KB 10/10 KB
16/75 KB 10/10 KB
Downloaded: https://repo.maven.apache.org/maven2/org/beanshell/bsh/2.0b4/bsh-2.0b4.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-component-annotations/1.5.5/plexus-component-annotations-1.5.5.jar
19/75 KB 10/10 KB
19/75 KB 3/31 KB 10/10 KB
21/75 KB 3/31 KB 10/10 KB
21/75 KB 3/31 KB 10/10 KB
21/75 KB 6/31 KB 10/10 KB
21/75 KB 7/31 KB 10/10 KB
21/75 KB 10/31 KB 10/10 KB
21/75 KB 11/31 KB 10/10 KB
24/75 KB 11/31 KB 10/10 KB
24/75 KB 14/31 KB 10/10 KB
24/75 KB 15/31 KB 10/10 KB
27/75 KB 15/31 KB 10/10 KB
27/75 KB 18/31 KB 10/10 KB
27/75 KB 19/31 KB 10/10 KB
29/75 KB 19/31 KB 10/10 KB
29/75 KB 22/31 KB 10/10 KB
29/75 KB 23/31 KB 10/10 KB
29/75 KB 26/31 KB 10/10 KB
32/75 KB 26/31 KB 10/10 KB
32/75 KB 26/31 KB 3/55 KB 10/10 KB
32/75 KB 26/31 KB 5/55 KB 10/10 KB
32/75 KB 26/31 KB 8/55 KB 10/10 KB
32/75 KB 29/31 KB 8/55 KB 10/10 KB
32/75 KB 29/31 KB 11/55 KB 10/10 KB
32/75 KB 31/31 KB 11/55 KB 10/10 KB
32/75 KB 31/31 KB 13/55 KB 10/10 KB
32/75 KB 31/31 KB 16/55 KB 10/10 KB
32/75 KB 31/31 KB 19/55 KB 10/10 KB
32/75 KB 31/31 KB 21/55 KB 10/10 KB
32/75 KB 31/31 KB 24/55 KB 10/10 KB
32/75 KB 31/31 KB 27/55 KB 10/10 KB
32/75 KB 31/31 KB 29/55 KB 10/10 KB
36/75 KB 31/31 KB 29/55 KB 10/10 KB
40/75 KB 31/31 KB 29/55 KB 10/10 KB
44/75 KB 31/31 KB 29/55 KB
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-api/1.2/enforcer-api-1.2.jar (10 KB at 140.2 KB/sec)
48/75 KB 31/31 KB 29/55 KB
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-i18n/1.0-beta-6/plexus-i18n-1.0-beta-6.jar
48/75 KB 31/31 KB 32/55 KB
48/75 KB 31/31 KB 36/55 KB
48/75 KB 31/31 KB 40/55 KB
48/75 KB 31/31 KB 44/55 KB
48/75 KB 31/31 KB 48/55 KB
52/75 KB 31/31 KB 48/55 KB
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-component-annotations/1.5.5/plexus-component-annotations-1.5.5.jar (0 B at 0.0 KB/sec)
56/75 KB 31/31 KB 48/55 KB
56/75 KB 31/31 KB 52/55 KB
60/75 KB 31/31 KB 52/55 KB
60/75 KB 31/31 KB 55/55 KB
64/75 KB 31/31 KB 55/55 KB
68/75 KB 31/31 KB 55/55 KB
72/75 KB 31/31 KB 55/55 KB
75/75 KB 31/31 KB 55/55 KB
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-common-artifact-filters/1.2/maven-common-artifact-filters-1.2.jar (31 KB at 406.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-i18n/1.0-beta-6/plexus-i18n-1.0-beta-6.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-dependency-tree/2.0/maven-dependency-tree-2.0.jar (55 KB at 659.3 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-rules/1.2/enforcer-rules-1.2.jar (75 KB at 886.7 KB/sec)
[WARNING] Rule 0: org.apache.maven.plugins.enforcer.RequireMavenVersion failed with message:
Detected Maven Version: 3.3.9 is not in the allowed range 3.5.0.
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 49.421 s
[INFO] Finished at: 2018-10-04T20:50:38+00:00
[INFO] Final Memory: 21M/456M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-enforcer-plugin:1.2:enforce (enforce-maven-version) on project jmeter-maven-plugin: Some Enforcer rules have failed. Look above for specific messages explaining why the rule failed. -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
container 2dd2bceb5c0831fbb71d253fd30ef875d084741068889c678f7f6aafd838028e is dead
Thu Oct 4 22:51:47 CEST 2018
```
username_0: @username_1 release, tag=2.8.0
username_1: @username_0 OK, I will release it now. Please check the progress [here](http://www.username_1.com/t/16089-427168367)
username_1: @username_0 Oops, I failed. You can see the full log [here](http://www.username_1.com/t/16089-427168367) (spent 3min)
```
1b7e9809d33b: Pull complete
b16fb2bbe4f5: Pull complete
Digest: sha256:e5fc92de44cf2af5446bc023205f6ab144ae1ca1050d6983f09c03e9edd9b7c5
Status: Downloaded newer image for maven:3.5-jdk-8
+ docker ps --filter=status=exited
+ grep --quiet '\sjmeter-maven-plugin_jmeter-maven-plugin_286\s*$'
+ ls -al .
total 364
drwx--x--x 3 username_1 username_1 4096 Oct 4 23:06 .
drwxrwxrwt 11 root root 20480 Oct 4 23:06 ..
-rw-rw-r-- 1 username_1 username_1 444 Oct 4 23:06 end.sh
-rwxrwxr-x 1 username_1 username_1 338 Oct 4 23:06 entry.sh
-rw-rw-r-- 1 username_1 username_1 6 Oct 4 23:06 pid
-rw-rw-r-- 1 username_1 username_1 287323 Oct 4 23:06 pubring.gpg
drwxrwxr-x 4 username_1 username_1 4096 Oct 4 23:06 repo
-rwxrwxr-x 1 username_1 username_1 7027 Oct 4 23:06 run.sh
-rw-rw-r-- 1 username_1 username_1 665 Oct 4 23:06 script.sh
-rw-rw-r-- 1 username_1 username_1 10050 Oct 4 23:06 secring.gpg
-rw-rw-r-- 1 username_1 username_1 569 Oct 4 23:06 settings.xml
-rw-rw-r-- 1 username_1 username_1 10012 Oct 4 23:06 stdout
++ pwd
++ pwd
+ docker run -t --rm -v /tmp/username_1-hAy0:/main --env=JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 --env=head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git --env=tag=2.8.0 '--env=scripts=( '\''export '\''\'\'''\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''tag=2.8.0'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''head_branch=master'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''author=username_0'\''\'\'''\'''\'' '\'';'\'' '\''mkdir /home/r/gpg'\'' '\'';'\'' '\''mv /home/r/pubring.gpg /home/r/gpg/pubring.gpg'\'' '\'';'\'' '\''mv /home/r/secring.gpg /home/r/gpg/secring.gpg'\'' '\'';'\'' '\''sudo chmod -R 700 /home/r/gpg'\'' '\'';'\'' '\''mvn versions:set "-DnewVersion=${tag}"'\'' '\'';'\'' '\''export PATH=$JAVA_HOME/bin:$PATH'\'' '\'';'\'' '\''set'\'' '\'';'\'' '\''javac -version'\'' '\'';'\'' '\''git commit -am "${tag}"'\'' '\'';'\'' '\''mvn clean deploy -Prelease --settings /home/r/settings.xml'\'' '\'';'\'' )' --env=head_branch=master --env=author=username_0 --hostname=docker --privileged --memory=6g --memory-swap=16g --oom-kill-disable --cidfile=/tmp/username_1-hAy0/cid -w=/main -v /var/run/docker.sock:/var/run/docker.sock --name=jmeter-maven-plugin_jmeter-maven-plugin_286 maven:3.5-jdk-8 /main/entry.sh
WARNING: Your kernel does not support swap limit capabilities or the cgroup is not mounted. Memory limited without swap.
+ set -e
+ set -o pipefail
+ shopt -s dotglob
+ useradd -m -G sudo r
+ echo '%sudo ALL=(ALL) NOPASSWD:ALL'
+ cp -R /root/.bashrc /root/.m2 /root/.profile /home/r
+ cp -R ./cid ./end.sh ./entry.sh ./pid ./pubring.gpg ./repo ./run.sh ./script.sh ./secring.gpg ./settings.xml ./stdout /home/r
+ rm -rf repo
+ chown -R r:r /home/r
+ chmod a+x /home/r/script.sh
+ su --login r --command /home/r/script.sh
mesg: cannot open /dev/pts/0: Permission denied
+ set -e
+ set -o pipefail
+ shopt -s expand_aliases
+ alias 'sudo=sudo -i'
+ export HOME=/home/r
+ HOME=/home/r
+ cd /home/r/repo
+ export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64
+ export head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git
+ head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git
+ export tag=2.8.0
+ tag=2.8.0
+ export head_branch=master
+ head_branch=master
+ export author=username_0
+ author=username_0
+ mkdir /home/r/gpg
+ mv /home/r/pubring.gpg /home/r/gpg/pubring.gpg
+ mv /home/r/secring.gpg /home/r/gpg/secring.gpg
+ sudo -i chmod -R 700 /home/r/gpg
/home/r/script.sh: line 9: sudo: command not found
container 35e2fae0d97b72d4a77b9655037ffa9d091dffa2e5bea895b06a2e0f8f2a14c0 is dead
Thu Oct 4 23:08:39 CEST 2018
```
username_0: @username_1 release, tag=`2.8.0`
username_1: @username_0 OK, I will release it now. Please check the progress [here](http://www.username_1.com/t/16089-427304501)
username_0: @username_1 release, tag=`2.8.0`
username_1: @username_0 Oops, I failed. You can see the full log [here](http://www.username_1.com/t/16089-427304501) (spent 3min)
```
gpg: gpg-agent is not available in this session
gpg: encrypted with 1 passphrase
gpg: original file name='pubring.gpg'
gpg: WARNING: message was not integrity protected
+ rm -rf pubring.gpg.enc
++ pwd
++ pwd
+ gpg --verbose --keyring=/tmp/username_1-9Dcz/.gpg/pubring.gpg --secret-keyring=/tmp/username_1-9Dcz/.gpg/secring.gpg --decrypt repo/secring.gpg.asc
gpg: armor header: Version: GnuPG/MacGPG2 v2
gpg: armor header: Comment: GPGTools - https://gpgtools.org
gpg: public key is 9A839C3A
gpg: using subkey 9A839C3A instead of primary key 9AF0FA4C
gpg: using subkey 9A839C3A instead of primary key 9AF0FA4C
gpg: encrypted with 2048-bit RSA key, ID 9A839C3A, created 2014-08-17
"Rultor.com <<EMAIL>>"
gpg: AES256 encrypted data
gpg: original file name='tmp.OVMNksGv'
+ gpg --no-tty --batch --verbose --decrypt --passphrase username_1-key:jmeter-maven-plugin/jmeter-maven-plugin secring.gpg.enc
gpg: armor header: Version: GnuPG/MacGPG2 v2
gpg: armor header: Comment: GPGTools - https://gpgtools.org
gpg: CAST5 encrypted data
gpg: gpg-agent is not available in this session
gpg: encrypted with 1 passphrase
gpg: original file name='secring.gpg'
gpg: WARNING: message was not integrity protected
+ rm -rf secring.gpg.enc
+ rm -rf .gpg
+ '[' -z 2.8.0 ']'
+ [[ 2.8.0 =~ ^[a-zA-Z0-9\.\-]+$ ]]
+ echo 'tag name is valid: "2.8.0"'
tag name is valid: "2.8.0"
+ cd repo
++ git tag -l 2.8.0
+ '[' ']'
+ export BRANCH_NAME=__username_1
+ BRANCH_NAME=__username_1
++ git show-branch __username_1
++ wc -l
+ '[' 0 -gt 0 ']'
+ git checkout -b __username_1
Switched to a new branch '__username_1'
+ docker_when_possible
+ true
++ uptime
++ sed 's/ /\n/g'
++ tail -n 1
+ load=2.20
++ echo 2.20 '>' 30
++ bc
+ '[' 0 -eq 1 ']'
+ echo 'load average is 2.20, low enough to run a new Docker container'
load average is 2.20, low enough to run a new Docker container
+ break
+ cd ..
+ '[' -n '' ']'
+ use_image=username_0/username_1maven
+ docker pull username_0/username_1maven
Using default tag: latest
Error response from daemon: manifest for username_0/username_1maven:latest not found
'cid' file is absent, container wasn't started correctly
```
username_0: @username_1 release, tag=`2.8.0`
username_1: @username_0 OK, I will release it now. Please check the progress [here](http://www.username_1.com/t/16089-427306823)
username_1: @username_0 Oops, I failed. You can see the full log [here](http://www.username_1.com/t/16089-427306823) (spent 3min)
```
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-model/2.0.9/maven-model-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-parameter-documenter/2.0.9/maven-plugin-parameter-documenter-2.0.9.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/reporting/maven-reporting-api/2.0.9/maven-reporting-api-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-registry/2.0.9/maven-plugin-registry-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-core/2.0.9/maven-core-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-container-default/1.0-alpha-9-stable-1/plexus-container-default-1.0-alpha-9-stable-1.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-repository-metadata/2.0.9/maven-repository-metadata-2.0.9.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-error-diagnostics/2.0.9/maven-error-diagnostics-2.0.9.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/doxia/doxia-sink-api/1.0-alpha-10/doxia-sink-api-1.0-alpha-10.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/reporting/maven-reporting-api/2.0.9/maven-reporting-api-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-error-diagnostics/2.0.9/maven-error-diagnostics-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-repository-metadata/2.0.9/maven-repository-metadata-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-parameter-documenter/2.0.9/maven-plugin-parameter-documenter-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-monitor/2.0.9/maven-monitor-2.0.9.jar
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-interactivity-api/1.0-alpha-4/plexus-interactivity-api-1.0-alpha-4.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-descriptor/2.0.9/maven-plugin-descriptor-2.0.9.jar
Downloading: https://repo.maven.apache.org/maven2/commons-cli/commons-cli/1.0/commons-cli-1.0.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/doxia/doxia-sink-api/1.0-alpha-10/doxia-sink-api-1.0-alpha-10.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/classworlds/classworlds/1.1/classworlds-1.1.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-descriptor/2.0.9/maven-plugin-descriptor-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-interactivity-api/1.0-alpha-4/plexus-interactivity-api-1.0-alpha-4.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/commons-cli/commons-cli/1.0/commons-cli-1.0.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-utils/1.5.8/plexus-utils-1.5.8.jar
Downloading: https://repo.maven.apache.org/maven2/commons-lang/commons-lang/2.3/commons-lang-2.3.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-api/1.2/enforcer-api-1.2.jar
Downloaded: https://repo.maven.apache.org/maven2/classworlds/classworlds/1.1/classworlds-1.1.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-monitor/2.0.9/maven-monitor-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-rules/1.2/enforcer-rules-1.2.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-common-artifact-filters/1.2/maven-common-artifact-filters-1.2.jar
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-utils/1.5.8/plexus-utils-1.5.8.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/commons-lang/commons-lang/2.3/commons-lang-2.3.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/beanshell/bsh/2.0b4/bsh-2.0b4.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-dependency-tree/2.0/maven-dependency-tree-2.0.jar
3/10 KB
5/10 KB
5/10 KB 3/31 KB
5/10 KB 5/31 KB
8/10 KB 5/31 KB
8/10 KB 8/31 KB
10/10 KB 8/31 KB
10/10 KB 11/31 KB
10/10 KB 11/31 KB 3/75 KB
10/10 KB 13/31 KB 3/75 KB
10/10 KB 13/31 KB 5/75 KB
10/10 KB 16/31 KB 5/75 KB
10/10 KB 19/31 KB 5/75 KB
10/10 KB 19/31 KB 8/75 KB
10/10 KB 21/31 KB 8/75 KB
10/10 KB 21/31 KB 11/75 KB
10/10 KB 24/31 KB 11/75 KB
10/10 KB 27/31 KB 11/75 KB
10/10 KB 27/31 KB 13/75 KB
10/10 KB 29/31 KB 13/75 KB
10/10 KB 31/31 KB 13/75 KB
10/10 KB 31/31 KB 16/75 KB
10/10 KB 31/31 KB 19/75 KB
10/10 KB 31/31 KB 21/75 KB
10/10 KB 31/31 KB 24/75 KB
10/10 KB 31/31 KB 27/75 KB
Downloaded: https://repo.maven.apache.org/maven2/org/beanshell/bsh/2.0b4/bsh-2.0b4.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-component-annotations/1.5.5/plexus-component-annotations-1.5.5.jar
10/10 KB 31/31 KB 29/75 KB
3/55 KB 10/10 KB 31/31 KB 29/75 KB
5/55 KB 10/10 KB 31/31 KB 29/75 KB
5/55 KB 10/10 KB 31/31 KB 32/75 KB
8/55 KB 10/10 KB 31/31 KB 32/75 KB
11/55 KB 10/10 KB 31/31 KB 32/75 KB
13/55 KB 10/10 KB 31/31 KB 32/75 KB
16/55 KB 10/10 KB 31/31 KB 32/75 KB
19/55 KB 10/10 KB 31/31 KB 32/75 KB
21/55 KB 10/10 KB 31/31 KB 32/75 KB
24/55 KB 10/10 KB 31/31 KB 32/75 KB
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-api/1.2/enforcer-api-1.2.jar (10 KB at 153.7 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-i18n/1.0-beta-6/plexus-i18n-1.0-beta-6.jar
27/55 KB 31/31 KB 32/75 KB
29/55 KB 31/31 KB 32/75 KB
32/55 KB 31/31 KB 32/75 KB
32/55 KB 31/31 KB 36/75 KB
32/55 KB 40/75 KB
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-common-artifact-filters/1.2/maven-common-artifact-filters-1.2.jar (31 KB at 482.2 KB/sec)
32/55 KB 44/75 KB
32/55 KB 48/75 KB
36/55 KB 48/75 KB
40/55 KB 48/75 KB
44/55 KB 48/75 KB
48/55 KB 48/75 KB
52/55 KB 48/75 KB
55/55 KB 48/75 KB
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-component-annotations/1.5.5/plexus-component-annotations-1.5.5.jar (0 B at 0.0 KB/sec)
55/55 KB 52/75 KB
55/55 KB 56/75 KB
55/55 KB 60/75 KB
55/55 KB 64/75 KB
55/55 KB 68/75 KB
55/55 KB 72/75 KB
55/55 KB 75/75 KB
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-i18n/1.0-beta-6/plexus-i18n-1.0-beta-6.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-dependency-tree/2.0/maven-dependency-tree-2.0.jar (55 KB at 701.5 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-rules/1.2/enforcer-rules-1.2.jar (75 KB at 775.8 KB/sec)
[WARNING] Rule 0: org.apache.maven.plugins.enforcer.RequireMavenVersion failed with message:
Detected Maven Version: 3.3.9 is not in the allowed range 3.5.0.
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 59.419 s
[INFO] Finished at: 2018-10-05T09:48:38+00:00
[INFO] Final Memory: 21M/427M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-enforcer-plugin:1.2:enforce (enforce-maven-version) on project jmeter-maven-plugin: Some Enforcer rules have failed. Look above for specific messages explaining why the rule failed. -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
container eeca8535cff91d34b6222f2e0cf15d6ab17bed814cadf09515676efcbfc0261f is dead
Fri Oct 5 11:49:35 CEST 2018
```
username_0: @username_1 release, tag=`2.8.0`
username_1: @username_0 OK, I will release it now. Please check the progress [here](http://www.username_1.com/t/16089-427310498)
username_1: @username_0 Oops, I failed. You can see the full log [here](http://www.username_1.com/t/16089-427310498) (spent 3min)
```
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-registry/2.0.9/maven-plugin-registry-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-parameter-documenter/2.0.9/maven-plugin-parameter-documenter-2.0.9.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/reporting/maven-reporting-api/2.0.9/maven-reporting-api-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-container-default/1.0-alpha-9-stable-1/plexus-container-default-1.0-alpha-9-stable-1.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-core/2.0.9/maven-core-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/doxia/doxia-sink-api/1.0-alpha-10/doxia-sink-api-1.0-alpha-10.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-repository-metadata/2.0.9/maven-repository-metadata-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-artifact-manager/2.0.9/maven-artifact-manager-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-error-diagnostics/2.0.9/maven-error-diagnostics-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/reporting/maven-reporting-api/2.0.9/maven-reporting-api-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-repository-metadata/2.0.9/maven-repository-metadata-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/commons-cli/commons-cli/1.0/commons-cli-1.0.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-descriptor/2.0.9/maven-plugin-descriptor-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-parameter-documenter/2.0.9/maven-plugin-parameter-documenter-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/doxia/doxia-sink-api/1.0-alpha-10/doxia-sink-api-1.0-alpha-10.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-interactivity-api/1.0-alpha-4/plexus-interactivity-api-1.0-alpha-4.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-monitor/2.0.9/maven-monitor-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-error-diagnostics/2.0.9/maven-error-diagnostics-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/classworlds/classworlds/1.1/classworlds-1.1.jar
Downloaded: https://repo.maven.apache.org/maven2/commons-cli/commons-cli/1.0/commons-cli-1.0.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-descriptor/2.0.9/maven-plugin-descriptor-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-interactivity-api/1.0-alpha-4/plexus-interactivity-api-1.0-alpha-4.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/commons-lang/commons-lang/2.3/commons-lang-2.3.jar
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-utils/1.5.8/plexus-utils-1.5.8.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-api/1.2/enforcer-api-1.2.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-monitor/2.0.9/maven-monitor-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/classworlds/classworlds/1.1/classworlds-1.1.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-rules/1.2/enforcer-rules-1.2.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-common-artifact-filters/1.2/maven-common-artifact-filters-1.2.jar
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-utils/1.5.8/plexus-utils-1.5.8.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/commons-lang/commons-lang/2.3/commons-lang-2.3.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-dependency-tree/2.0/maven-dependency-tree-2.0.jar
Downloading: https://repo.maven.apache.org/maven2/org/beanshell/bsh/2.0b4/bsh-2.0b4.jar
3/10 KB
5/10 KB
8/10 KB
10/10 KB
10/10 KB 3/31 KB
10/10 KB 5/31 KB
3/75 KB 10/10 KB 5/31 KB
5/75 KB 10/10 KB 5/31 KB
5/75 KB 10/10 KB 8/31 KB
8/75 KB 10/10 KB 8/31 KB
8/75 KB 10/10 KB 11/31 KB
11/75 KB 10/10 KB 11/31 KB
11/75 KB 10/10 KB 13/31 KB
13/75 KB 10/10 KB 13/31 KB
16/75 KB 10/10 KB 13/31 KB
16/75 KB 10/10 KB 16/31 KB
19/75 KB 10/10 KB 16/31 KB
21/75 KB 10/10 KB 16/31 KB
21/75 KB 10/10 KB 19/31 KB
24/75 KB 10/10 KB 19/31 KB
24/75 KB 10/10 KB 21/31 KB
24/75 KB 10/10 KB 24/31 KB
27/75 KB 10/10 KB 24/31 KB
Downloaded: https://repo.maven.apache.org/maven2/org/beanshell/bsh/2.0b4/bsh-2.0b4.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-component-annotations/1.5.5/plexus-component-annotations-1.5.5.jar
29/75 KB 10/10 KB 24/31 KB
32/75 KB 10/10 KB 24/31 KB
3/55 KB 32/75 KB 10/10 KB 24/31 KB
3/55 KB 32/75 KB 10/10 KB 27/31 KB
5/55 KB 32/75 KB 10/10 KB 27/31 KB
8/55 KB 32/75 KB 10/10 KB 27/31 KB
11/55 KB 32/75 KB 10/10 KB 27/31 KB
11/55 KB 32/75 KB 10/10 KB 29/31 KB
11/55 KB 32/75 KB 10/10 KB 31/31 KB
13/55 KB 32/75 KB 10/10 KB 31/31 KB
16/55 KB 32/75 KB 10/10 KB 31/31 KB
16/55 KB 36/75 KB 10/10 KB 31/31 KB
16/55 KB 40/75 KB 10/10 KB 31/31 KB
16/55 KB 44/75 KB 10/10 KB 31/31 KB
16/55 KB 48/75 KB 10/10 KB 31/31 KB
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-api/1.2/enforcer-api-1.2.jar (10 KB at 140.2 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-i18n/1.0-beta-6/plexus-i18n-1.0-beta-6.jar
19/55 KB 48/75 KB 31/31 KB
21/55 KB 48/75 KB 31/31 KB
24/55 KB 48/75 KB 31/31 KB
27/55 KB 48/75 KB 31/31 KB
29/55 KB 48/75 KB 31/31 KB
29/55 KB 52/75 KB 31/31 KB
29/55 KB 56/75 KB 31/31 KB
29/55 KB 60/75 KB 31/31 KB
29/55 KB 64/75 KB 31/31 KB
32/55 KB 64/75 KB 31/31 KB
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-component-annotations/1.5.5/plexus-component-annotations-1.5.5.jar (0 B at 0.0 KB/sec)
32/55 KB 68/75 KB 31/31 KB
32/55 KB 72/75 KB 31/31 KB
32/55 KB 75/75 KB 31/31 KB
36/55 KB 75/75 KB 31/31 KB
40/55 KB 75/75 KB 31/31 KB
44/55 KB 75/75 KB 31/31 KB
48/55 KB 75/75 KB 31/31 KB
52/55 KB 75/75 KB 31/31 KB
55/55 KB 75/75 KB 31/31 KB
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-common-artifact-filters/1.2/maven-common-artifact-filters-1.2.jar (31 KB at 406.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-i18n/1.0-beta-6/plexus-i18n-1.0-beta-6.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-rules/1.2/enforcer-rules-1.2.jar (75 KB at 886.7 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-dependency-tree/2.0/maven-dependency-tree-2.0.jar (55 KB at 651.4 KB/sec)
[WARNING] Rule 0: org.apache.maven.plugins.enforcer.RequireMavenVersion failed with message:
Detected Maven Version: 3.3.9 is not in the allowed range 3.5.0.
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 01:01 min
[INFO] Finished at: 2018-10-05T10:03:34+00:00
[INFO] Final Memory: 21M/431M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-enforcer-plugin:1.2:enforce (enforce-maven-version) on project jmeter-maven-plugin: Some Enforcer rules have failed. Look above for specific messages explaining why the rule failed. -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
container 797263bb98904e19cdc5e8f5d036c885832a9fd6b31aa8ed74af8a4cb39deae8 is dead
Fri Oct 5 12:04:48 CEST 2018
```
username_0: @username_1 release, tag=`2.8.0`
username_1: @username_0 OK, I will release it now. Please check the progress [here](http://www.username_1.com/t/16089-427549317)
username_1: @username_0 Oops, I failed. You can see the full log [here](http://www.username_1.com/t/16089-427549317) (spent 19s)
```
+ set -e
+ set -o pipefail
++ dirname ./run.sh
+ cd .
+ echo 5476
+ echo '1.68 838a77339'
1.68 838a77339
+ date
Sat Oct 6 08:06:08 CEST 2018
+ uptime
08:06:08 up 90 days, 16:50, 0 users, load average: 2.07, 2.11, 2.15
+ head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git
+ image=username_0/username_1maven:1.0.0
+ head_branch=master
+ author=username_0
+ tag=2.8.0
+ vars=('--env=JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64' '--env=head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git' '--env=tag=2.8.0' '--env=scripts=( '\''export '\''\'\'''\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''tag=2.8.0'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''head_branch=master'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''author=username_0'\''\'\'''\'''\'' '\'';'\'' '\''mkdir /home/r/gpg'\'' '\'';'\'' '\''mv /home/r/pubring.gpg /home/r/gpg/pubring.gpg'\'' '\'';'\'' '\''mv /home/r/secring.gpg /home/r/gpg/secring.gpg'\'' '\'';'\'' '\''sudo chmod -R 700 /home/r/gpg'\'' '\'';'\'' '\''mvn versions:set "-DnewVersion=${tag}"'\'' '\'';'\'' '\''export PATH=$JAVA_HOME/bin:$PATH'\'' '\'';'\'' '\''set'\'' '\'';'\'' '\''javac -version'\'' '\'';'\'' '\''git commit -am "${tag}"'\'' '\'';'\'' '\''mvn clean deploy -Prelease --settings /home/r/settings.xml'\'' '\'';'\'' )' '--env=head_branch=master' '--env=author=username_0')
+ scripts=('export '\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''' ';' 'export '\''head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git'\''' ';' 'export '\''tag=2.8.0'\''' ';' 'export '\''head_branch=master'\''' ';' 'export '\''author=username_0'\''' ';' 'mkdir /home/r/gpg' ';' 'mv /home/r/pubring.gpg /home/r/gpg/pubring.gpg' ';' 'mv /home/r/secring.gpg /home/r/gpg/secring.gpg' ';' 'sudo chmod -R 700 /home/r/gpg' ';' 'mvn versions:set "-DnewVersion=${tag}"' ';' 'export PATH=$JAVA_HOME/bin:$PATH' ';' 'set' ';' 'javac -version' ';' 'git commit -am "${tag}"' ';' 'mvn clean deploy -Prelease --settings /home/r/settings.xml' ';')
+ directory=
+ container=jmeter-maven-plugin_jmeter-maven-plugin_286
+ as_root=false
+ mkdir -p /home/username_1/.ssh
+ echo -e 'Host github.com\n\tStrictHostKeyChecking no\n'
+ chmod 600 /home/username_1/.ssh/config
+ git clone <EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git repo
Cloning into 'repo'...
+ cd repo
+ git config user.email <EMAIL>
+ git config user.name username_1
+ '[' -z 'export '\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''' ']'
+ cd ..
+ cat
+ '[' false = true ']'
+ cat
+ chmod a+x entry.sh
+ cat
+ echo 'export '\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''' ';' 'export '\''head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git'\''' ';' 'export '\''tag=2.8.0'\''' ';' 'export '\''head_branch=master'\''' ';' 'export '\''author=username_0'\''' ';' 'mkdir /home/r/gpg' ';' 'mv /home/r/pubring.gpg /home/r/gpg/pubring.gpg' ';' 'mv /home/r/secring.gpg /home/r/gpg/secring.gpg' ';' 'sudo chmod -R 700 /home/r/gpg' ';' 'mvn versions:set "-DnewVersion=${tag}"' ';' 'export PATH=$JAVA_HOME/bin:$PATH' ';' set ';' 'javac -version' ';' 'git commit -am "${tag}"' ';' 'mvn clean deploy -Prelease --settings /home/r/settings.xml' ';'
+ sensitive=()
+ gpgconf --reload gpg-agent
+ gpg --keyserver hkp://pool.sks-keyservers.net --verbose --recv-keys 9AF0FA4C
gpg: requesting key 9AF0FA4C from hkp server pool.sks-keyservers.net
gpgkeys: key 9AF0FA4C can't be retrieved
gpg: no valid OpenPGP data found.
gpg: Total number processed: 0
gpg: keyserver communications error: keyserver helper general error
gpg: keyserver communications error: unknown pubkey algorithm
gpg: keyserver receive failed: unknown pubkey algorithm
'cid' file is absent, container wasn't started correctly
```
username_0: @username_1 release, tag=`2.8.0`
username_1: @username_0 OK, I will release it now. Please check the progress [here](http://www.username_1.com/t/16089-427549732)
username_1: @username_0 Oops, I failed. You can see the full log [here](http://www.username_1.com/t/16089-427549732) (spent 20s)
```
+ set -e
+ set -o pipefail
++ dirname ./run.sh
+ cd .
+ echo 11291
+ echo '1.68 838a77339'
1.68 838a77339
+ date
Sat Oct 6 08:15:47 CEST 2018
+ uptime
08:15:47 up 90 days, 16:59, 0 users, load average: 2.09, 2.08, 2.10
+ head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git
+ image=username_0/username_1maven:1.0.0
+ head_branch=master
+ author=username_0
+ tag=2.8.0
+ vars=('--env=JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64' '--env=head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git' '--env=tag=2.8.0' '--env=scripts=( '\''export '\''\'\'''\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''tag=2.8.0'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''head_branch=master'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''author=username_0'\''\'\'''\'''\'' '\'';'\'' '\''mkdir /home/r/gpg'\'' '\'';'\'' '\''mv /home/r/pubring.gpg /home/r/gpg/pubring.gpg'\'' '\'';'\'' '\''mv /home/r/secring.gpg /home/r/gpg/secring.gpg'\'' '\'';'\'' '\''sudo chmod -R 700 /home/r/gpg'\'' '\'';'\'' '\''mvn versions:set "-DnewVersion=${tag}"'\'' '\'';'\'' '\''export PATH=$JAVA_HOME/bin:$PATH'\'' '\'';'\'' '\''set'\'' '\'';'\'' '\''javac -version'\'' '\'';'\'' '\''git commit -am "${tag}"'\'' '\'';'\'' '\''mvn clean deploy -Prelease --settings /home/r/settings.xml'\'' '\'';'\'' )' '--env=head_branch=master' '--env=author=username_0')
+ scripts=('export '\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''' ';' 'export '\''head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git'\''' ';' 'export '\''tag=2.8.0'\''' ';' 'export '\''head_branch=master'\''' ';' 'export '\''author=username_0'\''' ';' 'mkdir /home/r/gpg' ';' 'mv /home/r/pubring.gpg /home/r/gpg/pubring.gpg' ';' 'mv /home/r/secring.gpg /home/r/gpg/secring.gpg' ';' 'sudo chmod -R 700 /home/r/gpg' ';' 'mvn versions:set "-DnewVersion=${tag}"' ';' 'export PATH=$JAVA_HOME/bin:$PATH' ';' 'set' ';' 'javac -version' ';' 'git commit -am "${tag}"' ';' 'mvn clean deploy -Prelease --settings /home/r/settings.xml' ';')
+ directory=
+ container=jmeter-maven-plugin_jmeter-maven-plugin_286
+ as_root=false
+ mkdir -p /home/username_1/.ssh
+ echo -e 'Host github.com\n\tStrictHostKeyChecking no\n'
+ chmod 600 /home/username_1/.ssh/config
+ git clone [email protected]:jmeter-maven-plugin/jmeter-maven-plugin.git repo
Cloning into 'repo'...
+ cd repo
+ git config user.email <EMAIL>
+ git config user.name username_1
+ '[' -z 'export '\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''' ']'
+ cd ..
+ cat
+ '[' false = true ']'
+ cat
+ chmod a+x entry.sh
+ cat
+ echo 'export '\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''' ';' 'export '\''head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git'\''' ';' 'export '\''tag=2.8.0'\''' ';' 'export '\''head_branch=master'\''' ';' 'export '\''author=username_0'\''' ';' 'mkdir /home/r/gpg' ';' 'mv /home/r/pubring.gpg /home/r/gpg/pubring.gpg' ';' 'mv /home/r/secring.gpg /home/r/gpg/secring.gpg' ';' 'sudo chmod -R 700 /home/r/gpg' ';' 'mvn versions:set "-DnewVersion=${tag}"' ';' 'export PATH=$JAVA_HOME/bin:$PATH' ';' set ';' 'javac -version' ';' 'git commit -am "${tag}"' ';' 'mvn clean deploy -Prelease --settings /home/r/settings.xml' ';'
+ sensitive=()
+ gpgconf --reload gpg-agent
+ gpg --keyserver hkp://pool.sks-keyservers.net --verbose --recv-keys 9AF0FA4C
gpg: requesting key 9AF0FA4C from hkp server pool.sks-keyservers.net
gpgkeys: key 9AF0FA4C can't be retrieved
gpg: no valid OpenPGP data found.
gpg: Total number processed: 0
gpg: keyserver communications error: keyserver helper general error
gpg: keyserver communications error: unknown pubkey algorithm
gpg: keyserver receive failed: unknown pubkey algorithm
'cid' file is absent, container wasn't started correctly
```
username_0: @username_1 release, tag=`2.8.0`
username_1: @username_0 OK, I will release it now. Please check the progress [here](http://www.username_1.com/t/16089-427558573)
username_1: @username_0 Oops, I failed. You can see the full log [here](http://www.username_1.com/t/16089-427558573) (spent 3min)
```
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-model/2.0.9/maven-model-2.0.9.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-artifact-manager/2.0.9/maven-artifact-manager-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/reporting/maven-reporting-api/2.0.9/maven-reporting-api-2.0.9.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/doxia/doxia-sink-api/1.0-alpha-10/doxia-sink-api-1.0-alpha-10.jar
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-parameter-documenter/2.0.9/maven-plugin-parameter-documenter-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-container-default/1.0-alpha-9-stable-1/plexus-container-default-1.0-alpha-9-stable-1.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-repository-metadata/2.0.9/maven-repository-metadata-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-core/2.0.9/maven-core-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-error-diagnostics/2.0.9/maven-error-diagnostics-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/reporting/maven-reporting-api/2.0.9/maven-reporting-api-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/commons-cli/commons-cli/1.0/commons-cli-1.0.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-parameter-documenter/2.0.9/maven-plugin-parameter-documenter-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-descriptor/2.0.9/maven-plugin-descriptor-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/doxia/doxia-sink-api/1.0-alpha-10/doxia-sink-api-1.0-alpha-10.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-interactivity-api/1.0-alpha-4/plexus-interactivity-api-1.0-alpha-4.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-repository-metadata/2.0.9/maven-repository-metadata-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/maven-monitor/2.0.9/maven-monitor-2.0.9.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-error-diagnostics/2.0.9/maven-error-diagnostics-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/classworlds/classworlds/1.1/classworlds-1.1.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-plugin-descriptor/2.0.9/maven-plugin-descriptor-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-utils/1.5.8/plexus-utils-1.5.8.jar
Downloaded: https://repo.maven.apache.org/maven2/commons-cli/commons-cli/1.0/commons-cli-1.0.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/commons-lang/commons-lang/2.3/commons-lang-2.3.jar
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-interactivity-api/1.0-alpha-4/plexus-interactivity-api-1.0-alpha-4.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-api/1.2/enforcer-api-1.2.jar
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/maven-monitor/2.0.9/maven-monitor-2.0.9.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-rules/1.2/enforcer-rules-1.2.jar
Downloaded: https://repo.maven.apache.org/maven2/classworlds/classworlds/1.1/classworlds-1.1.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-common-artifact-filters/1.2/maven-common-artifact-filters-1.2.jar
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-utils/1.5.8/plexus-utils-1.5.8.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/beanshell/bsh/2.0b4/bsh-2.0b4.jar
3/10 KB
5/10 KB
8/10 KB
10/10 KB
Downloaded: https://repo.maven.apache.org/maven2/commons-lang/commons-lang/2.3/commons-lang-2.3.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-dependency-tree/2.0/maven-dependency-tree-2.0.jar
10/10 KB 2/75 KB
10/10 KB 5/75 KB
10/10 KB 6/75 KB
10/10 KB 9/75 KB
10/10 KB 10/75 KB
10/10 KB 13/75 KB
10/10 KB 14/75 KB
10/10 KB 17/75 KB
10/10 KB 18/75 KB
10/10 KB 21/75 KB
10/10 KB 22/75 KB
10/10 KB 22/75 KB 3/31 KB
10/10 KB 25/75 KB 3/31 KB
10/10 KB 25/75 KB 5/31 KB
10/10 KB 26/75 KB 5/31 KB
10/10 KB 29/75 KB 8/31 KB
10/10 KB 29/75 KB 5/31 KB
Downloaded: https://repo.maven.apache.org/maven2/org/beanshell/bsh/2.0b4/bsh-2.0b4.jar (0 B at 0.0 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-component-annotations/1.5.5/plexus-component-annotations-1.5.5.jar
10/10 KB 29/75 KB 11/31 KB
10/10 KB 30/75 KB 11/31 KB
10/10 KB 30/75 KB 13/31 KB
10/10 KB 33/75 KB 13/31 KB
10/10 KB 34/75 KB 13/31 KB
10/10 KB 34/75 KB 16/31 KB
10/10 KB 37/75 KB 16/31 KB
10/10 KB 38/75 KB 16/31 KB
10/10 KB 41/75 KB 16/31 KB
10/10 KB 42/75 KB 16/31 KB
10/10 KB 42/75 KB 19/31 KB
10/10 KB 45/75 KB 19/31 KB
10/10 KB 46/75 KB 19/31 KB
10/10 KB 46/75 KB 21/31 KB
10/10 KB 46/75 KB 24/31 KB
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-api/1.2/enforcer-api-1.2.jar (10 KB at 151.3 KB/sec)
Downloading: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-i18n/1.0-beta-6/plexus-i18n-1.0-beta-6.jar
49/75 KB 24/31 KB
49/75 KB 27/31 KB
50/75 KB 27/31 KB
53/75 KB 27/31 KB
53/75 KB 29/31 KB
54/75 KB 29/31 KB
3/55 KB 54/75 KB 29/31 KB
5/55 KB 54/75 KB 29/31 KB
5/55 KB 57/75 KB 29/31 KB
5/55 KB 58/75 KB 29/31 KB
5/55 KB 58/75 KB 31/31 KB
8/55 KB 58/75 KB 31/31 KB
8/55 KB 61/75 KB 31/31 KB
11/55 KB 61/75 KB 31/31 KB
11/55 KB 62/75 KB 31/31 KB
13/55 KB 62/75 KB 31/31 KB
16/55 KB 62/75 KB 31/31 KB
19/55 KB 62/75 KB 31/31 KB
19/55 KB 65/75 KB 31/31 KB
19/55 KB 66/75 KB 31/31 KB
21/55 KB 66/75 KB 31/31 KB
24/55 KB 66/75 KB 31/31 KB
24/55 KB 69/75 KB 31/31 KB
24/55 KB 70/75 KB 31/31 KB
27/55 KB 70/75 KB 31/31 KB
27/55 KB 73/75 KB 31/31 KB
29/55 KB 73/75 KB 31/31 KB
29/55 KB 75/75 KB 31/31 KB
32/55 KB 75/75 KB 31/31 KB
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-component-annotations/1.5.5/plexus-component-annotations-1.5.5.jar (0 B at 0.0 KB/sec)
36/55 KB 75/75 KB 31/31 KB
40/55 KB 75/75 KB 31/31 KB
44/55 KB 75/75 KB 31/31 KB
48/55 KB 75/75 KB 31/31 KB
52/55 KB 75/75 KB 31/31 KB
55/55 KB 75/75 KB 31/31 KB
Downloaded: https://repo.maven.apache.org/maven2/org/codehaus/plexus/plexus-i18n/1.0-beta-6/plexus-i18n-1.0-beta-6.jar (0 B at 0.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-common-artifact-filters/1.2/maven-common-artifact-filters-1.2.jar (31 KB at 400.8 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/enforcer/enforcer-rules/1.2/enforcer-rules-1.2.jar (75 KB at 931.0 KB/sec)
Downloaded: https://repo.maven.apache.org/maven2/org/apache/maven/shared/maven-dependency-tree/2.0/maven-dependency-tree-2.0.jar (55 KB at 659.3 KB/sec)
[WARNING] Rule 0: org.apache.maven.plugins.enforcer.RequireMavenVersion failed with message:
Detected Maven Version: 3.3.9 is not in the allowed range 3.5.0.
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 59.731 s
[INFO] Finished at: 2018-10-06T09:08:23+00:00
[INFO] Final Memory: 22M/455M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-enforcer-plugin:1.2:enforce (enforce-maven-version) on project jmeter-maven-plugin: Some Enforcer rules have failed. Look above for specific messages explaining why the rule failed. -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
container af81cbd863409e6625c9fc9f01fe63acbad367b7503b2412081e5803eb9f4772 is dead
Sat Oct 6 11:09:34 CEST 2018
```
username_0: @username_1 release, tag=`2.8.0`
username_1: @username_0 OK, I will release it now. Please check the progress [here](http://www.username_1.com/t/16089-427560977)
username_1: @username_0 Oops, I failed. You can see the full log [here](http://www.username_1.com/t/16089-427560977) (spent 3min)
```
+ set -e
+ set -o pipefail
++ dirname ./run.sh
+ cd .
+ echo 30067
+ echo '1.68 838a77339'
1.68 838a77339
+ date
Sat Oct 6 11:52:15 CEST 2018
+ uptime
11:52:15 up 90 days, 20:36, 0 users, load average: 2.08, 2.09, 2.13
+ head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git
+ image=username_0/username_1maven:1.0.0
+ head_branch=master
+ author=username_0
+ tag=2.8.0
+ vars=('--env=JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64' '--env=head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git' '--env=tag=2.8.0' '--env=scripts=( '\''export '\''\'\'''\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''tag=2.8.0'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''head_branch=master'\''\'\'''\'''\'' '\'';'\'' '\''export '\''\'\'''\''author=username_0'\''\'\'''\'''\'' '\'';'\'' '\''mkdir /home/r/gpg'\'' '\'';'\'' '\''mv /home/r/pubring.gpg /home/r/gpg/pubring.gpg'\'' '\'';'\'' '\''mv /home/r/secring.gpg /home/r/gpg/secring.gpg'\'' '\'';'\'' '\''sudo chmod -R 700 /home/r/gpg'\'' '\'';'\'' '\''mvn versions:set "-DnewVersion=${tag}"'\'' '\'';'\'' '\''export PATH=$JAVA_HOME/bin:$PATH'\'' '\'';'\'' '\''set'\'' '\'';'\'' '\''javac -version'\'' '\'';'\'' '\''mvn -v'\'' '\'';'\'' '\''git commit -am "${tag}"'\'' '\'';'\'' '\''mvn clean deploy -Prelease --settings /home/r/settings.xml'\'' '\'';'\'' )' '--env=head_branch=master' '--env=author=username_0')
+ scripts=('export '\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''' ';' 'export '\''head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git'\''' ';' 'export '\''tag=2.8.0'\''' ';' 'export '\''head_branch=master'\''' ';' 'export '\''author=username_0'\''' ';' 'mkdir /home/r/gpg' ';' 'mv /home/r/pubring.gpg /home/r/gpg/pubring.gpg' ';' 'mv /home/r/secring.gpg /home/r/gpg/secring.gpg' ';' 'sudo chmod -R 700 /home/r/gpg' ';' 'mvn versions:set "-DnewVersion=${tag}"' ';' 'export PATH=$JAVA_HOME/bin:$PATH' ';' 'set' ';' 'javac -version' ';' 'mvn -v' ';' 'git commit -am "${tag}"' ';' 'mvn clean deploy -Prelease --settings /home/r/settings.xml' ';')
+ directory=
+ container=jmeter-maven-plugin_jmeter-maven-plugin_286
+ as_root=false
+ mkdir -p /home/username_1/.ssh
+ echo -e 'Host github.com\n\tStrictHostKeyChecking no\n'
+ chmod 600 /home/username_1/.ssh/config
+ git clone <EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git repo
Cloning into 'repo'...
+ cd repo
+ git config user.email <EMAIL>
+ git config user.name username_1
+ '[' -z 'export '\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''' ']'
+ cd ..
+ cat
+ '[' false = true ']'
+ cat
+ chmod a+x entry.sh
+ cat
+ echo 'export '\''JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64'\''' ';' 'export '\''head=<EMAIL>:jmeter-maven-plugin/jmeter-maven-plugin.git'\''' ';' 'export '\''tag=2.8.0'\''' ';' 'export '\''head_branch=master'\''' ';' 'export '\''author=username_0'\''' ';' 'mkdir /home/r/gpg' ';' 'mv /home/r/pubring.gpg /home/r/gpg/pubring.gpg' ';' 'mv /home/r/secring.gpg /home/r/gpg/secring.gpg' ';' 'sudo chmod -R 700 /home/r/gpg' ';' 'mvn versions:set "-DnewVersion=${tag}"' ';' 'export PATH=$JAVA_HOME/bin:$PATH' ';' set ';' 'javac -version' ';' 'mvn -v' ';' 'git commit -am "${tag}"' ';' 'mvn clean deploy -Prelease --settings /home/r/settings.xml' ';'
+ sensitive=()
+ gpgconf --reload gpg-agent
+ gpg --keyserver hkp://pool.sks-keyservers.net --verbose --recv-keys 9AF0FA4C
gpg: requesting key 9AF0FA4C from hkp server pool.sks-keyservers.net
gpgkeys: key 9AF0FA4C can't be retrieved
gpg: no valid OpenPGP data found.
gpg: Total number processed: 0
gpg: keyserver communications error: keyserver helper general error
gpg: keyserver communications error: unknown pubkey algorithm
gpg: keyserver receive failed: unknown pubkey algorithm
'cid' file is absent, container wasn't started correctly
```
username_0: @username_1 release, tag=`2.8.0`
username_1: @username_0 OK, I will release it now. Please check the progress [here](http://www.username_1.com/t/16089-427570346)
username_1: @username_0 Done! FYI, the full log is [here](http://www.username_1.com/t/16089-427570346) (took me 5min)
Status: Issue closed
|
MicrosoftDocs/windows-itpro-docs | 821383442 | Title: Key Trust or Cert Trust
Question:
username_0: Hello , please specify if the feature is available for WHFB in Key Trust or Cert Trust architecture.
It seems it works only in the Cert Trust WHFB model .
Thanks in advance
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f50a8599-301a-5993-1a86-e722f66ffcf2
* Version Independent ID: fffd0fc9-97f6-cfa9-70fb-7f1308804de0
* Content: [Dual Enrollment - Microsoft 365 Security](https://docs.microsoft.com/en-us/windows/security/identity-protection/hello-for-business/hello-feature-dual-enrollment)
* Content Source: [windows/security/identity-protection/hello-for-business/hello-feature-dual-enrollment.md](https://github.com/MicrosoftDocs/windows-itpro-docs/blob/public/windows/security/identity-protection/hello-for-business/hello-feature-dual-enrollment.md)
* Product: **w10**
* GitHub Login: @mapalko
* Microsoft Alias: **mapalko**
Answers:
username_1: Closing via #9277
Thank you
Status: Issue closed
|
Torann/laravel-geoip | 556375672 | Title: Maxmind GeoLite2-City database too large for Laravel Vapor
Question:
username_0: A requirement of Laravel Vapor is that the compressed project size is < 45mb. While the GeoLite2-City database is simply too large for this on its own being ~69mb uncompressed, the GeoLite2-Country database works just fine sitting at just 4mb uncompressed.
Obviously, this solution would break compatibility for other services, as well as the public API now returning either a partial Location response object or a new response object entirely. However, if this were a configurable option for the `maxmind_database` service it would be an intentional, visible configuration option that would allow support for Laravel Vapor.
The code changes necessary in the `MaxMindDatabaseService` would be minimal. Instead of calling the `city` method on the `Reader` object if the configuration option is `country` we would call `country`.
Given the response from the country database really only consists of `iso_code`, I would think it would be fine to opt for a partially complete `Location` object consisting of only that. Again, this would be an intentional configuration change available to users who do not need/cannot use the entire city database in instances such as Laravel Vapor.
This is more of an idea gathering to see if it would be supported/accepted or if there might be a workaround already out there. I'm more than willing to PR the changes if this approach is accepted.
Answers:
username_1: I think it makes sense to be able to switch between the two depending on use-case.
I've noticed that using `separate-vendor: true` allows the deployment to succeed but I'm not convinced the database is available and working, always seems to return the fallback location. I wonder if perhaps the `storage/app` directory isn't made available to a deployment. May need to be moved to `resources` or something.
Did you have any success getting it working?
username_1: Yeah, I have found that downloading the database to the resources directory seems to make it work.
```
'maxmind_database' => [
'class' => \Torann\GeoIP\Services\MaxMindDatabase::class,
'database_path' => resource_path('geoip.mmdb'),
'update_url' => sprintf('https://download.maxmind.com/app/geoip_download?edition_id=GeoLite2-City&license_key=%s&suffix=tar.gz', env('MAXMIND_LICENSE_KEY')),
'locales' => ['en'],
],
```
And call `geoip:update` as a build step.
```
build:
- "composer install --no-dev"
- "php artisan optimize"
- "php artisan geoip:update"
- "yarn && yarn run prod && rm -rf node_modules"
```
username_0: I use the database path, but that would also work. The important part is including it in your build step. I just recently revisited my GitHub actions workflow and, for either you or anyone else looking to use this workaround, let me save you some time I lost working through this...
Assuming the `vapor.yml` file has been configured with the build step mentioned before
```yml
- php artisan geoip:update
```
Your deployment step should look something similar to this (removed a few lines specific to my workflow)
```yml
- name: Deploy to production
env:
VAPOR_API_TOKEN: ${{ secrets.VAPOR_API_TOKEN }}
CACHE_DRIVER: array
MAXMIND_LICENSE_KEY: ${{ secrets.MAXMIND_LICENSE_KEY }}
run: ./vendor/bin/vapor deploy production
```
This would also apply if you're running any integration tests against, or for whatever reason need this package to function properly during your test suite. You'll need to specify the array cache driver (as the file driver does not support tagging) and be sure to have your maxmind license key set as an environment variable so that `geoip:update` is able to fetch the mmdb.
username_1: Worth noting here too - if you're only interested in the request country and using Cloudflare (or can use Cloudflare) [they can expose the request country as a request header](https://support.cloudflare.com/hc/en-us/articles/200168236-Configuring-Cloudflare-IP-Geolocation#12345682).
username_2: Does that keep working when you go over the 1000 concurrent connections (the default) and Lambda scales up? I would think the database would then be gone again.
username_3: Tried including the database in the deploy steps but due to the lambda size limit could not follow through with it. Certain that it would work constantly if paired with the docker run time.
@username_2 currently exploring using the EFS to store the maxmind file database. It seems to be working fine essentially have turned my config into the following.
```
'maxmind_database' => [
'class' => \Torann\GeoIP\Services\MaxMindDatabase::class,
'database_path' => env('MAXMIND_STORAGE_PATH', storage_path('app/geoip.mmdb')),
'update_url' => sprintf('https://download.maxmind.com/app/geoip_download?edition_id=GeoLite2-City&license_key=%s&suffix=tar.gz', env('MAXMIND_LICENSE_KEY')),
'locales' => ['en'],
]
```
So I can still work with it on local, and the path on vapor to something the likes of `/mnt/app/geoip.mmdb` . Still I'm cautious to call it a solution though due to the nature of the size of payload that a lambda function can handle aside from the fact that I am no expert on streams which I feel could be a solution if that were to be an issue.
Also as a side note(mainly because I didn't know) if you use `request()->server('REMOTE_ADDR')` it will not work in vapor, you have to call it through the proxy `request()->ip()`
username_2: Completely forgot to update this.
I went for a dual Vapor and EC2 with EFS solution. My project leans heavily on Laravel jobs and they were 2 to 3 times slower on Vapor than on an EC2 server. I run the same code on EC2 as Vapor and use the EC2 server just for (scheduled) jobs. Which also means the EC2 updates the MaxMind DB and stores it on EFS.
For Vapor the `database_path` in the config is set to `"/mnt/local/geoip.mmdb"`
EC2 `database_path` is set to `"/mnt/efs/geoip.mmdb"` (where I mounted the EFS storage)
username_1: You can now get city, latitude and longitude using dynamic Cloudflare headers (not through workers, so it's free). If this is roughly all you need it's a great alternative to using GeoIP in your Laravel app. [You can see the available fields here](https://developers.cloudflare.com/ruleset-engine/rules-language/fields/) and [how to add a HTTP request header modification rule here](https://developers.cloudflare.com/rules/transform/request-header-modification/create-dashboard/). |
phanxgames/AetherStory | 855104031 | Title: Characters missing on character screen
Question:
username_0: **Describe the bug**
Sometimes when logging out, your characters do not appear on the character selection screen.
**To Reproduce**
Steps to reproduce the behavior:
1. Unknown
**Expected behavior**
Characters should always appear on the character select screen.
**Screenshots**
(need to screenshot)
**Context:**
- OS: Windows
- Patch Version: 0.9.4
- Game Build: 484 |
Esri/solution.js | 634958252 | Title: Deployed ArcGIS Pro Packages Cannot be Opened Directly from AGOL
Question:
username_0: 1) Deploy a solution with and ArcGIS Pro Package (Recreation Outreach or Sewer Data Management)
2) Navigate to package's Item Details Page
3) Click "Open in ArcGIS Pro"
BUG: ArcGIS Pro will launch, but the item will not download and open the package.
If you click download, and then open the ppkx, everything works correctly. This seems to be an issue with how the pitemx is referencing the package.
Answers:
username_1: @username_0, did you find a workaround for this? If not, seems like it should be critical for the June release.
username_0: @username_1
There are 2 workarounds which can be used.
1) Download the package instead of clicking "Open In ArcGIS Pro". I think this will be what most people do if the open button doesn't work the first time.
2) We ship the project as a zip. This would force people to download.
I'll bump priority and we can make the work arounds be a last resort. If we can't get this addressed.
username_1: @username_0, this worked for me. At first it didn't work, but the problem for me was that I wasn't signed into the org in Pro that I had deployed to and the item was private, so it couldn't reach the item to download and extract. After I signed into the portal in Pro, closed Pro and then tried the Open in Pro it worked fine. I am at 2.6 Beta. Could this be the same issue for you?
We don't create or clone the pitemx file, this is generated dynamically by the item so I am not sure how we could cause an issue here.
username_0: @username_1
At first I would see the issue regardless of the org I was signed into.
I cracked open the pitemx file and noticed that one of the item id's being referenced was the original item from the org used to create the solution item. I changed this to match the new item id and everything then worked fine.
As a test, I then changed the item id back to the original value (from source org, not target), and now the item will open correctly if logged into the target org.
I'm on a daily build of 2.6, so that could be the issue. However, other folks are seeing this behavior with 2.5. I've asked to re-verify what org they are signed into when clicking "Open in ArcGIS Pro"
Status: Issue closed
username_0: 1) Deploy a solution with and ArcGIS Pro Package (Recreation Outreach or Sewer Data Management)
2) Navigate to package's Item Details Page
3) Click "Open in ArcGIS Pro"
BUG: ArcGIS Pro will launch, but the item will not download and open the package.
If you click download, and then open the ppkx, everything works correctly. This seems to be an issue with how the pitemx is referencing the package.
username_1: If you could get an item that produces the pitemx pointing back to the original item id I would be interested in checking it out because we aren't include it as a resource with the item, so would be good to figure out why it continues to point to it.
username_1: @username_0, tested this on QA today with Transit Outreach and the pro project worked both ways (Opening from Pro, or Opening from item details). This is the last issue in Solution.js. Wondering if we can close and wait to see if we get any other reports of an issue?
Status: Issue closed
|
FRC-Utilities/QDriverStation | 269417855 | Title: Robot code not detected
Question:
username_0: I'm running 17.05, downloaded as a .deb on Ubuntu 16.04.
When starting, comms are established with the rio, but I get the following in the logs:
00:00.6 DEBUG DS Events: Robot code status set to false
I'm able to deploy the java code from eclipse to the rio just fine. So, am not sure what might be going on. Also, as far as I can see, if my code what throwing an exception, there's no log that tells me that - or maybe it's just not getting that far.
Also - since I'm connected to the rio via USB networking (IP address 172.22.11.2), I'm unable to edit the network addresses from qdriverstation. I worked around this by putting an entry in /etc/hosts that resolves roboRIO-xxxx-FRC.local to 172.22.11.2. Not sure if this is a known issue or not.
Answers:
username_1: Thanks for the feedback, it may be possible that you have selected the wrong DS protocol version from the settings tab. To clarify the cause of your issue, could you please upload the full log? (You can find them in ~./qdriverstation, or by clicking the "Logs..." button in the console section).
username_0: Log file as requested:
Log created on: Oct 29 2017 - 13:26:07 PM
LibDS version: 2.2
Operating System: Ubuntu 16.04.3 LTS
Application name: QDriverStation
Application version: 17.05
------------------------------------------------------------------------
ELAPSED TIME ERROR LEVEL MESSAGE
------------------------------------------------------------------------
00:00.3 DEBUG Scale factor set to: 1
00:00.5 DEBUG DS Client: Changing team number to 1289
00:00.5 DEBUG DS Events: Team number set to 1289
00:00.5 DEBUG DS Client: Using new FMS address "0.0.0.0"
00:00.5 DEBUG DS Client: Using new radio address "0.0.0.0"
00:00.5 DEBUG DS Client: Using new robot address "0.0.0.0"
00:00.5 DEBUG DS Client: Using new FMS address ""
00:00.5 DEBUG DS Client: Using new radio address ""
00:00.5 DEBUG DS Client: Using new robot address ""
00:00.5 DEBUG DS Client: Switched to FRC 2016 Protocol
00:00.5 DEBUG DS Client: Resetting joysticks...
00:00.5 DEBUG DS Client: Using new FMS address ""
00:00.5 DEBUG DS Client: Using new radio address ""
00:00.5 DEBUG DS Client: Using new robot address ""
00:00.5 DEBUG DS Client: Resetting joysticks...
00:00.5 DEBUG DS Client: Setting position to 0
00:00.5 DEBUG DS Client: Setting alliance to 0
00:00.5 DEBUG Initialized in 585 milliseconds
00:00.6 DEBUG DS Events: Robot code status set to false
00:00.6 DEBUG DS Events: ESTOP set to false
00:00.6 DEBUG DS Events: Robot communications set to true
00:01.5 DEBUG DS Events: FMS communications set to false
username_1: Thanks! Unfortunately it seems that the issue has something to do with QDriverStation, I will check the protocol code later this week. I have to finish some freelance work these days.
username_0: Thank you - sorry for the delay in responding. - I've been dealing with the effects of a wind storm that took out electricity for a few days.
Further data point - with 2016 protocol, I do get robot comms. With 2015 and 2014, there is no robot comms. You probably know that (given the log file), but just wanted to let you know.
username_0: Found myself at work today with some extra time, so I took a dive in. Here's a long shot guess:
frc_2015.c line 71, uint8_t cRobotHasCode = 0x20
Since cRobotHasCode is a uint8_t, this will truncate the supplied 16bit value, resulting in cRobotHasCode getting set to zero.
In read_robot_packet(), line 642, this zero value gets AND'd to the status bits, resulting in the robot's code status getting set to zero.
I'm guessing this was a typo and cRobotHasCode should be set to 0x02
Since I installed from a .deb; I'll try to set up an actual build/test environment this weekend & see if I'm right.
username_1: Thanks for the information, I will check again the 2015 protocol implementation (I did not check that code for almost a year :laughing:).
Status: Issue closed
|
Yardanico/random-stuff | 659899199 | Title: important packages new
Question:
username_0: arraymancer - nimlapack, but won't work on ARC (for now) anyway
cello - shallowCopy
combparser - closure sink inference
docopt - closure sink inference
hts-nim - finalizers to destructors (Error: cannot bind another '=destroy' to: Variant:ObjectType; previous declaration was constructed here implicitly: /home/dian/Things/Nim/pkgstemp/hts/src/hts/vcf.nim(180, 16)) - https://gist.github.com/username_0/5a9f7e9842101443ebf662c58340f406
iterutils - deepCopy
karax - undeclared PNimType
markdown - ORC crash
kdtree - config issue
macroutils - config issue
jstin - config issue
INim - config issue
fidget - config issue
gara - config issue
ggplotnim - config issue
compactdict - config issue
comprehension - config issue
DelaunayNim - config issue
chroma failed - config issue
nim-chronos failed - config issue
ast-patter-matching - config issue
Answers:
username_0: ```
FAIL: https://github.com/status-im/nim-chronicles C
Test "https://github.com/status-im/nim-chronicles" in category "nimble-packages-1"
Failure: reBuildFailed
package test failed
$ nim c --gc:orc -o:chr -r chronicles.nim
Hint: used config file '/home/dian/Things/Nim/config/nim.cfg' [Conf]
Hint: used config file '/home/dian/Things/Nim/config/config.nims' [Conf]
Hint: used config file '/home/dian/.config/nim/nim.cfg' [Conf]
Hint: used config file '/home/dian/.config/nim/config.nims' [Conf]
Hint: used config file '/home/dian/Things/Nim/pkgstemp/chronicles/nim.cfg' [Conf]
..........................................................
/home/dian/.nimble/pkgs/chronos-2.5.1/chronos/transports/common.nim(516, 4) Error: expression cannot be cast to ptr SeqHeader
```
```
FAIL: https://github.com/status-im/nim-chronos C
Test "https://github.com/status-im/nim-chronos" in category "nimble-packages-1"
Failure: reBuildFailed
package test failed
$ nim c --gc:orc -r -d:release tests/testall
Hint: used config file '/home/dian/Things/Nim/config/nim.cfg' [Conf]
Hint: used config file '/home/dian/Things/Nim/config/config.nims' [Conf]
Hint: used config file '/home/dian/.config/nim/nim.cfg' [Conf]
Hint: used config file '/home/dian/.config/nim/config.nims' [Conf]
Hint: used config file '/home/dian/Things/Nim/pkgstemp/chronos/tests/config.nims' [Conf]
.............................................
/home/dian/Things/Nim/pkgstemp/chronos/chronos/asyncloop.nim(567, 13) template/generic instantiation of `withData` from here
/home/dian/Things/Nim/lib/pure/ioselects/ioselectors_epoll.nim(526, 23) template/generic instantiation of `getData` from here
/home/dian/Things/Nim/lib/pure/ioselects/ioselectors_epoll.nim(499, 3) Warning: Cannot prove that 'result' is initialized. This will become a compile time error in the future. [ProveInit]
..
/home/dian/Things/Nim/pkgstemp/chronos/chronos/asyncloop.nim(936, 37) Hint: could not turn 'cbproc' to a sink parameter [Performance]
.
/home/dian/Things/Nim/pkgstemp/chronos/chronos/asyncsync.nim(96, 34) template/generic instantiation of `async` from here
/home/dian/Things/Nim/pkgstemp/chronos/chronos/asyncmacro2.nim(47, 32) Warning: See corresponding Defect; AssertionError is deprecated [Deprecated]
....
/home/dian/Things/Nim/pkgstemp/chronos/chronos/transports/common.nim(516, 4) Error: expression cannot be cast to ptr SeqHeader
```
```
FAIL: https://github.com/unicredit/cello C
Test "https://github.com/unicredit/cello" in category "nimble-packages-1"
[Suite] rotated strings
/home/dian/Things/Nim/pkgstemp/cello/tests/rotated_string.nim(45, 15): Check failed: x[5] == 'f'
x[5] was ,
[FAILED] underlying strings are shared
```
username_0: FAIL: https://github.com/Vindaar/ggplotnim C - depends on arraymancer
username_0: ```
FAIL: https://github.com/LemonBoy/jstin C
/home/dian/Things/Nim/lib/system/assertions.nim(110) test1
/home/dian/Things/Nim/lib/system/assertions.nim(23) raiseAssert
/home/dian/Things/Nim/lib/system/fatal.nim(49) sysFatal
Unhandled exception: JstinDeserializeError wasn't raised by:
let t3 = fromJson[array[5, int]](toJson([1, 2, 3, 4])) [AssertionDefect]
[FAILED] Handling of array & seq
Error: execution of an external program failed: '/home/dian/Things/Nim/pkgstemp/jstin/tests/test1 '
Tip: 2 messages have been suppressed, use --verbose to show them.
Error: Execution failed with exit code 1
... Command: "/home/dian/Things/Nim/bin/nim" c --noNimblePath -d:NimblePkgVersion=0.1.0 "-r" "--path:." "/home/dian/Things/Nim/pkgstemp/jstin/tests/test1"
```
username_0: ```
FAIL: https://github.com/PMunch/macroutils C
Test "https://github.com/PMunch/macroutils" in category "nimble-packages-1"
Failure: reBuildFailed
package test failed
$ nimble test
Executing task test in /home/dian/Things/Nim/pkgstemp/macroutils/macroutils.nimble
Verifying dependencies for [email protected]
Compiling /home/dian/Things/Nim/pkgstemp/macroutils/tests/test1 (from package macroutils) using c backend
Hint: used config file '/home/dian/Things/Nim/config/nim.cfg' [Conf]
Hint: used config file '/home/dian/Things/Nim/config/config.nims' [Conf]
Hint: used config file '/home/dian/.config/nim/nim.cfg' [Conf]
Hint: used config file '/home/dian/.config/nim/config.nims' [Conf]
Hint: used config file '/home/dian/Things/Nim/pkgstemp/macroutils/tests/config.nims' [Conf]
................................
/home/dian/Things/Nim/pkgstemp/macroutils/src/macroutils.nim(1198, 10) Error: type mismatch: got <NimNode, NimNodeKind, proc (n: NimNode, y: sink seq[int]): NimNode{.closure, gcsafe, locks: <unknown>.}>
but expected one of:
```
- closure sink inference
username_0: - combparser/docopt - can't compile because of closure sink inference, need to manually annotate some code with {.nosinks.}
- hts-nim - https://gist.github.com/username_0/5a9f7e9842101443ebf662c58340f406 - need to reorder finalizers so they can be converted to destructors
- iterutils - need deepCopy
- karax - undeclared PNimType
- markdown - Crash in runtime in the cycle collector (!), will try to reproduce that
- chronos and chronicles (depend on chronos) - they cast seqs to the internal representation which is different for seqs_v2, so they need to be adapted (it isn't hard, I already did it manually before)
- cello - fails because shallowCopy is "fake" on arc
- jstin - some test failure
- ggplotnim - depends on arraymancer so will fail
- macroutils - closure sink inference again for some reason
username_0: Nimble packages 2
- nake - closure sink inference
- nimfp - depends on classy, which fails due to closure sink inference
- nimgame2 - same as hts-nim, need to fix order of finalizers for ARC
- nimly - SIGSEGV in allocation at runtime
- nimwc - closure sink inference
- norm - deepCopy
- numericalnim - SIGSEGV at runtime
- parsetoml - deepCopy |
uclibs/application_portfolio | 639801262 | Title: Dashboard: Add URL
Question:
username_0: Add URL (and make linkable) in the My Development and My Production modules of the dashboard. It should be before the View, Edit, Delete. We may even want to reduce the View, Edit, and Delete to icons. They are taking up a lot of real estate in our module.
Answers:
username_0: OR.... If we don't have the space, we can make the title link to the URL I kind of like this idea.
Status: Issue closed
|
sphinx-doc/sphinx | 459242336 | Title: Apidoc doesn't pass the user_template_dir when it recurses
Question:
username_0: **Describe the bug**
Tried to replace module template for apidoc; which seems to have no effect, it works for package and toc template. To note is that --separate is passed as an argument.
**To Reproduce**
Create the template module.rst_t, then run sphinx-apidoc --separate -t=./_templates ...
In the generated rst files you can see that the template was not applied.
**Expected behavior**
Expected that the module template in separate mode was used, which was not.
**Your project**
I have attached only the docs folder since the issue can be observed in it, can't attach the source due to it being proprietary. An easy exemple to check will be the rst file app/app.cleanup, the generated module file doesn't respect the template(ie the module word is still appended after the module name).
[docs.zip](https://github.com/sphinx-doc/sphinx/files/3314963/docs.zip)
**Environment info**
- OS: Win 10
- Python version: 3.6.8
- Sphinx version: 3.0.0.dev20190621
- Sphinx extensions: 'sphinx.ext.apidoc', 'sphinx.ext.autodoc', 'sphinx.ext.coverage', 'sphinx_autodoc_typehints', 'sphinx.ext.inheritance_diagram', 'sphinx_paramlinks', 'sphinx.ext.autosummary', 'sphinxcontrib.httpdomain'
**Additional context**
I already have tried checking the files used by apidoc and I have seen that in sphinx.ext.apidoc, in the function create_package_file, on line 160, when it goes through the submodules, the user_template_dir parameter is not passed. if the parameter is added, the templates are used correctly.
Answers:
username_1: @username_0 Could you try #6519? You can try it with this installation: `pip install https://github.com/username_1/sphinx/archive/6517_passthrough_templatedir_to_separateoption.zip`.
username_0: I have tried it and it works, thanks!
username_1: Thank you for confirming. Merged now.
Status: Issue closed
|
meething/meething | 607859720 | Title: Old videos/participants not removing from UI
Question:
username_0: **Describe the bug**
NB; Tests were done on same LAN.
**To Reproduce**
Steps to reproduce the behavior:
1. Open multiple instances using Chrome and Firefox (2+ instances EACH) with half on Chrome and half on Firefox.
2. Close some of the instances
**Expected behavior**
When an instance/user leaves, user ID and video frame should disappear on other instances.
**Screenshots**

**Desktop (please complete the following information):**
- OS: Win 10
- Browser: chrome to firefox
- Version: latest
**Additional context**
- Link you used? https://meething.hepic.tel/
- Branch you used? (Such as fix_audio) Master/hepic
Add any other context about the problem here.
**Metric**
- Perfomance / Benchmarking
- Latency
- [ ] Cross site scripting
Answers:
username_1: @username_0 thanks for the report!
Those objects appear to have no height, suggesting the video was never connected to them, rather than the user leaving - or perhaps they rejoined? We need a step by step procedure we can replicate OR as you reproduce this issue on your LAN please paste the browser console logs around the time and before the issue is manifested, to see what action has failed.
username_1: This should be fixed. Please reopen if you notice the same issue. Thanks!
Status: Issue closed
|
timyates/mod-jdbc-persistor | 162631530 | Title: JDBC Result set from HANA procedure part 2
Question:
username_0: This is a continuation of 'JDBC Result set from HANA procedure #33'
I have now (at last!) managed to get the result set back from my HANA procedure with the following code using a regular JDBC connection
String sql = "CALL "XX_SCHEMANAME\".mySprocName(?,?,?,?)";
CallableStatement callableStatement = connection.prepareCall(sql);
callableStatement.setString(1, "val1");
callableStatement.setString(2, "val2");
callableStatement.setString(3, "val3");
callableStatement.setInt(4, 123);
boolean b = callableStatement.execute();
resultSet = callableStatement.getResultSet();
--------------------------------
Please will you let me know whether vertx is using callable statements any where? I can't see that it does.
The following dirty hack of JdbcProcessor.doSelect gets me back the result set from the HANA procedure
private void doSelect( final Message<JsonObject> message,
Connection connection,
TransactionalHandler transaction ) throws SQLException {
new BatchHandler( connection, message, transaction ) {
public JsonObject process() throws SQLException {
JsonObject reply = new JsonObject() ;
ArrayList<Map<String,Object>> result = new ArrayList<Map<String,Object>>() ;
// processing
while( ( resultSet != null || valueIterator.hasNext() ) &&
( batchSize == -1 || result.size() < batchSize ) ) {
LimitedMapListHandler handler = new LimitedMapListHandler( batchSize == -1 ? -1 : batchSize - result.size() ) ;
if( resultSet == null ) {
List<Object> params = valueIterator.next() ;
boolean callable = params.size() > 1;// <-----------Flag value is hacked here - needs passing from client properly.
if (callable) {
String sql = message.body().getString( "stmt" ) ;
statement = connection.prepareCall(sql);
int i = 1;
for (Object param: params) {
statement.setObject(i++, param);
}
boolean isResultSet = statement.execute();
if (isResultSet) {
resultSet = statement.getResultSet();
}
}
else {
statementFiller.fill( statement, params ) ;
resultSet = statement.executeQuery() ;
}
}
store( result, handler ) ;
}
reply.putArray( "result", JsonUtils.listOfMapsToJsonArray( result ) ) ;
return reply ;
}
}.handle( message ) ;
} |
strengejacke/sjstats | 396129249 | Title: feature request: adding support for `"svyolr"` objects in `sjstats::p_value`
Question:
username_0: ``` r
# setup
set.seed(123)
library(survey)
#> Loading required package: grid
#> Loading required package: Matrix
#> Loading required package: survival
#>
#> Attaching package: 'survey'
#> The following object is masked from 'package:graphics':
#>
#> dotchart
data(api)
# preparing data
dclus1 <- svydesign(
id = ~ dnum,
weights = ~ pw,
data = apiclus1,
fpc = ~ fpc
)
dclus1 <- update(dclus1, mealcat = cut(meals, c(0, 25, 50, 75, 100)))
# model
m <- svyolr(mealcat ~ avg.ed + mobility + stype, design = dclus1)
# class
class(m)
#> [1] "svyolr"
# summary
summary(m)
#> Call:
#> svyolr(mealcat ~ avg.ed + mobility + stype, design = dclus1)
#>
#> Coefficients:
#> Value Std. Error t value
#> avg.ed -2.6999217 1.13422434 -2.380412
#> mobility 0.0325042 0.02065336 1.573797
#> stypeH -1.7574715 0.69953393 -2.512346
#> stypeM -0.6191463 0.30964416 -1.999541
#>
#> Intercepts:
#> Value Std. Error t value
#> (0,25]|(25,50] -8.8579 3.6862 -2.4030
#> (25,50]|(50,75] -6.5865 3.1089 -2.1186
#> (50,75]|(75,100] -4.9249 2.8634 -1.7200
#> (26 observations deleted due to missingness)
# tidy output
broom::tidy(m)
#> # A tibble: 7 x 5
#> term estimate std.error statistic coefficient_type
#> <chr> <dbl> <dbl> <dbl> <chr>
#> 1 avg.ed -2.70 1.13 -2.38 coefficient
#> 2 mobility 0.0325 0.0207 1.57 coefficient
#> 3 stypeH -1.76 0.700 -2.51 coefficient
#> 4 stypeM -0.619 0.310 -2.00 coefficient
#> 5 (0,25]|(25,50] -8.86 3.69 -2.40 zeta
#> 6 (25,50]|(50,75] -6.59 3.11 -2.12 zeta
#> 7 (50,75]|(75,100] -4.92 2.86 -1.72 zeta
# extracting p-values using `sjstats`
sjstats::p_value(m)
#> Error in stats::coef(summary(fit))[, 4]: subscript out of bounds
```
<sup>Created on 2019-01-05 by the [reprex package](https://reprex.tidyverse.org) (v0.2.1)</sup><issue_closed>
Status: Issue closed |
dapr/dapr | 1035609730 | Title: Actor Reminder Partition No Longer Preview Feature
Question:
username_0: The purpose of this issue is to gather supporting evidence of Actor Reminder Partitioning no longer needing to be a preview feature.
Release version: 1.3
Supported SDKs: Java, Python, PHP, .NET
Missing SDKs: JS
Answers:
username_0: @artursouza - Can you triage this so it doesn't go stale?
username_1: Notice I triaged to 1.6. |
michaelsevilla/noitatressid | 314793519 | Title: relatedwork: scoop up new file systems
Question:
username_0: - octopus
- data lakes at microsoft
- FSMAC: nvram in localFS
- hopfs
- colossusFS
Answers:
username_0: "[...] the conventional wisdom has been that it is too expensive [...] to store hierarchical file system metadata fully normalized in a distributed database".
- Normalized: don't store full path name with metadata |
b0xCH/Traveli | 215252776 | Title: Dashboard: Sorting does not work
Question:
username_0: # Expected behavior
Table data gets sorted.
# Actual behavior
Nothing.
# Steps to reproduce
Go to dashboard and click on a table header column to sort (e.g. for trainrides).
Status: Issue closed
Answers:
username_1: Table sorting on the Dashboard is disabled, as it should show the last trainrides/flights chronologically. |
NewSpring/Apollos | 296839796 | Title: Profile birthday incorrect
Question:
username_0: ## Acceptance Criteria
- [ ] It should the users birthday if available.
- [ ] It should have an appropriate blank/unfilled state.
## Context
Currently the a users birthday is displayed as the current date in Profile>Personal Details
## UI Screenshot or Wireframe (optional)

Answers:
username_0: @username_1 what's the progress on this? Anything I can do to help?
Status: Issue closed
|
urbanairship/ios-library | 92652430 | Title: [UAInboxDBManager saveContext] crash
Question:
username_0: Please fix this crash

Status: Issue closed
Answers:
username_1: Thanks for the report! We request that all crash or bug reports be filed through <EMAIL> as our team doesn't actively monitor GitHub issues.
When you send the email, please include as much information as you can. In this case, make sure to include the Urban Airship SDK version(s) associated with the crash in your application. Since you are using Crashlytics, you can also share a copy of the full crash report with us. There is a "Share" button above the "Devices" graph (it's just cut off on the screenshot above). Clicking that reveals a link that you can send to our support team. Then just include any other info that would be useful, such a steps to reproduce if you've seen it during development.
Thanks again. |
baomidou/mybatis-plus | 394586150 | Title: Oracle中,如何写查询时间范围语句
Question:
username_0: ### 当前使用版本(必须填写清楚,否则不予处理)
3.0.6
### 该问题是怎么引起的?**([最新版](https://search.maven.org/search?q=mybatis-plus)上已修复的会直接**close**掉)**
Oracle 11G中,字段设置为timestamp,此时,传入的字符串时间必须使用Oracle的to_date函数进行格式化才能进行查询,查询语句如下
`SELECT * FROM PROJ_TEACH_RESEARCH_SETUP WHERE PROJ_CREATE_TIME >= TO_DATE('2018-12-26 15:03:33','yyyy-MM-dd HH24:MI:SS')`
### 重现步骤
尝试使用Mybatis-plus的QueryWrapper进行条件查询
`projTeachResearchSetupMapper
.selectList(new QueryWrapper<ProjTeachResearchSetup>().apply("PROJ_CREATE_TIME = TO_DATE('{0}','yyyy-MM-dd HH24:MI:SS')", "2018-12-26 15:03:33"));`
### 报错信息
Could not set parameters for mapping: ParameterMapping{property='ew.paramNameValuePairs.MPGENVAL3',
mode=IN, javaType=class java.lang.Object, jdbcType=null, numericScale=null, resultMapId='null', jdbcTypeName='null', expression='null'}.
Cause: org.apache.ibatis.type.TypeException: Error setting non null for parameter #1 with JdbcType null .
Try setting a different JdbcType for this parameter or a different configuration property. Cause: org.apache.ibatis.type.TypeException:
Error setting non null for parameter #1 with JdbcType null .
Try setting a different JdbcType for this parameter or a different configuration property. Cause: java.sql.SQLException: 无效的列索引
### 补充
这里的关键是我用的参数绑定的写法,如果我放弃参数绑定的写法,直接传入sql,查询能通过。如果想用参数绑定的方式,这里就会这样
Answers:
username_1: 建议使用Date对象当参数,不要使用特定的函数
```
projTeachResearchSetupMapper .selectList(new QueryWrapper<ProjTeachResearchSetup>().eq("PROJ_CREATE_TIME", DateUtils.parseDate("2018-12-26 15:03:33","yyyy-MM-dd HH:mm:ss"));
```
username_0: 确认有效,谢谢,请关闭问题
Status: Issue closed
|
element-plus/element-plus | 846431714 | Title: [Question] Date : Tue Mar 02 2021 23:59:59 is showing as the next day
Question:
username_0: [https://codepen.io/username_0/pen/yLggBgr](https://codepen.io/username_0/pen/yLggBgr)
This date is showing as next day but it's not, it's still the same day.
How i can make it as it is ? without round it to next day ?
02 2021 23:59:59
when i insert this date it's showing me:
03 2021<issue_closed>
Status: Issue closed |
stellar/stellar-core | 107812213 | Title: add a flag to make "auth" flags read-only
Question:
username_0: This allows an issuer to be explicit when it comes to "immutability" of the contract that they provide (non revocable auth in particular).
Such accounts become non-mergeable (as merging would allow to indirectly clear flags).
This also allows to remove an index on trust lines (performance)<issue_closed>
Status: Issue closed |
rwynn/monstache | 188784220 | Title: Question on duplication
Question:
username_0: Just started testing with monstache and so far very impressed. I have a quick question regarding duplicates; specifically how are MongoDB `_id` properties mapped to ElasticSearch, and how when a document in MongoDB is updated is the correct ElasticSearch document updated?
Answers:
username_1: The default mapping is as follows:
```
mongodb database -> elasticsearch index
mongodb collection -> elasticsearch type
mongodb document _id -> elasticsearch document id
```
Any operation on a mongo document (insert, update, delete) will be recorded in the oplog with it's _id and that will eventually be propagated into a bulk index command to ES where the id is specified as the _id in mongodb.
So I think you may be correct above the potential for collisions since mongo _ids are unique at the collection level (not guaranteed across the database) while ES ids must be unique at the index level.
According to [link](http://stackoverflow.com/questions/5303869/mongodb-are-mongoids-unique-across-collections), duplicate _ids across the database are unlikely with autogenerated _ids.
Regardless, I should change the default mapping such that
```
elasticsearch index <- mongodb database . mongodb collection
```
Thanks for giving monstache a try and reporting this issue.
Status: Issue closed
username_0: Awesome, thanks. :+1: |
spring-cloud/spring-cloud-consul | 747496357 | Title: Only ConsulDiscoveryProperties private constructor invoked - NPE on hostInfo property
Question:
username_0: **Describe the bug**
**Stack:**
- Spring Boot: 2.4.0
- Spring Cloud Dependencies: 2020.0.0-M5
- Spring Cloud Consul: 3.0.0-M5
- Spring Cloud Config: 3.0.0-M5
```
<dependency>
<groupId>org.springframework.boot</groupId>
<artifactId>spring-boot-starter-web</artifactId>
</dependency>
<dependency>
<groupId>org.springframework.cloud</groupId>
<artifactId>spring-cloud-starter-consul-discovery</artifactId>
</dependency>
<dependency>
<groupId>org.springframework.cloud</groupId>
<artifactId>spring-cloud-starter-config</artifactId>
</dependency>
```
**Description**
When using `spring-cloud-consul-discovery` and spring-cloud-config, the private constructor of `ConsulDiscoveryProperties` is invoked and not the public one with `InetUtils` as parameter. This causes in a not configured `hostInfo` property. When calling the setter of `hostname` property of `ConsulDiscoveryProperties`, `this.hostInfo.override = true;`is invoked and a NPE is thrown.
**Stack Trace**
```
org.springframework.boot.context.properties.bind.BindException: Failed to bind properties under 'spring.cloud.consul.discovery' to org.springframework.cloud.consul.discovery.ConsulDiscoveryProperties
at org.springframework.boot.context.properties.bind.Binder.handleBindError(Binder.java:363)
at org.springframework.boot.context.properties.bind.Binder.bind(Binder.java:323)
at org.springframework.boot.context.properties.bind.Binder.bind(Binder.java:308)
at org.springframework.boot.context.properties.bind.Binder.bind(Binder.java:238)
at org.springframework.boot.context.properties.bind.Binder.bind(Binder.java:199)
at org.springframework.boot.context.properties.bind.Binder.bind(Binder.java:186)
at org.springframework.cloud.consul.discovery.configclient.ConsulConfigServerBootstrapper.lambda$intitialize$3(ConsulConfigServerBootstrapper.java:57)
at org.springframework.cloud.consul.discovery.configclient.ConsulConfigServerBootstrapper$$Lambda$327/000000000000000000.get(Unknown Source)
at org.springframework.boot.DefaultBootstrapContext.getInstance(DefaultBootstrapContext.java:119)
at org.springframework.boot.DefaultBootstrapContext.getOrElseThrow(DefaultBootstrapContext.java:111)
at org.springframework.boot.DefaultBootstrapContext.get(DefaultBootstrapContext.java:88)
at org.springframework.cloud.consul.discovery.configclient.ConsulConfigServerBootstrapper.lambda$intitialize$4(ConsulConfigServerBootstrapper.java:63)
at org.springframework.cloud.consul.discovery.configclient.ConsulConfigServerBootstrapper$$Lambda$328/000000000000000000.onApplicationEvent(Unknown Source)
at org.springframework.context.event.SimpleApplicationEventMulticaster.doInvokeListener(SimpleApplicationEventMulticaster.java:203)
at org.springframework.context.event.SimpleApplicationEventMulticaster.invokeListener(SimpleApplicationEventMulticaster.java:196)
at org.springframework.context.event.SimpleApplicationEventMulticaster.multicastEvent(SimpleApplicationEventMulticaster.java:170)
at org.springframework.context.event.SimpleApplicationEventMulticaster.multicastEvent(SimpleApplicationEventMulticaster.java:148)
at org.springframework.boot.DefaultBootstrapContext.close(DefaultBootstrapContext.java:131)
at org.springframework.boot.SpringApplication.prepareContext(SpringApplication.java:392)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:325)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:1309)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:1298)
at org.example.TestApp.main(TestApp.java:14)
Caused by: java.lang.IllegalStateException: Unable to set value for property hostname
at org.springframework.boot.context.properties.bind.JavaBeanBinder$BeanProperty.setValue(JavaBeanBinder.java:357)
at org.springframework.boot.context.properties.bind.JavaBeanBinder.bind(JavaBeanBinder.java:98)
[Truncated]
... 37 common frames omitted
14:29:08.130 [main] ERROR o.s.b.d.LoggingFailureAnalysisReporter -
```
**Sample**
Just create a Spring Boot application with the given dependencies above and set the property
```
spring:
cloud:
consul:
discovery:
hostname: some-value
```
Thanks,
Marcel
Answers:
username_1: Is this still a problem with 3.0.0?
username_0: Hi @username_1,
the problem still exists with the described scenario
Status: Issue closed
|
graphql-java-kickstart/graphql-java-tools | 436248688 | Title: Nested Input type object need a direct usage to be registered correctly
Question:
username_0: with version 5.5.1
this doesn't work ( error during graphql service registration )
```gql
input A {
v :Int
}
input B {
a : A
}
type Query {
q1 (b: B ) : Int
}
```
seems that A is not usabe as nested input object if hasn't been called directly by some method in the graphql interface
this **WORKS** instead.. ( the `dumb` ) method seems to be required to register `A` as input type which is required by B... seems a bug in the library to me.. , the previously schema didn't worked neither with `A` as normal `type`
```gql
input A {
v :Int
}
input B {
a : A
}
type Query {
q1 (b: B ) : Int
dumb(a:A):Int
}
```
Answers:
username_1: Related to #261
username_2: This is a major bug, is anyone looking into this or is there an ETA? |
webdriverio/webdriverio | 592891581 | Title: Browser target is not set correctly in capabilities
Answers:
username_1: Multiremote has a different way of capabilities:
```js
capabilities: {
SomeBrowserName: {
capabilities: { ... }
}
}
```
You should only use multiremote if you want to have these different capabilities interact with each other.
You are using `chromeOptions` here which is incorrect. With browsers moving towards WebDriver it should be `goog:chromeOptions` as stated in the docs.
Status: Issue closed
username_1: Closing due to inactivity. |
xamarin/Essentials | 803264757 | Title: [Bug] Xamarin.Essentials WebAuthenticator on iOS cannot complete the registration
Question:
username_0: <!-- Bug report best practices: https://github.com/xamarin/Essentials/wiki -->
### Description
When I try to login on my iPhone (13.4) with the Webautenticator (not the native AppleSignInAuthenticator) I get the dialog to login, but after a click on login, I get an error “Registration not completed”. On the Azure, the service seems to work OK. Google works fine. The AppleSignInAuthenticator works fine.
My app is based on the sample in in the Xamarin.Essentials samples. I tried to test the behavior of the WebAuthenticator when the iOS version is less than 13. My phone has the version 13.4.
The Log from the Azure App Service:
Application:2021-02-07 22:31:09.555 +00:00 [Information] Microsoft.AspNetCore.Hosting.Diagnostics: Request starting HTTP/1.1 GET https://shoporderapi.azurewebsites.net/api/mobileauthshop/Apple - -
Application:2021-02-07 22:31:10.041 +00:00 [Information] Microsoft.AspNetCore.Routing.EndpointMiddleware: Executing endpoint 'ShopOrderWeb.Controllers.MobileAuthShopController.Get (ShopOrderWeb)'
Application:2021-02-07 22:31:10.144 +00:00 [Information] Microsoft.AspNetCore.Mvc.Infrastructure.ControllerActionInvoker: Route matched with {action = "Get", controller = "MobileAuthShop"}. Executing controller action with signature System.Threading.Tasks.Task Get(System.String) on controller ShopOrderWeb.Controllers.MobileAuthShopController (ShopOrderWeb).
Application:2021-02-07 22:31:10.182 +00:00 [Information] AspNet.Security.OAuth.Apple.AppleAuthenticationHandler: Apple was not authenticated. Failure message: Not authenticated
Application:2021-02-07 22:31:10.194 +00:00 [Information] AspNet.Security.OAuth.Apple.AppleAuthenticationHandler: AuthenticationScheme: Apple was challenged.
Application:2021-02-07 22:31:10.202 +00:00 [Information] Microsoft.AspNetCore.Mvc.Infrastructure.ControllerActionInvoker: Executed action ShopOrderWeb.Controllers.MobileAuthShopController.Get (ShopOrderWeb) in 49.5103ms
Application:2021-02-07 22:31:10.203 +00:00 [Information] Microsoft.AspNetCore.Routing.EndpointMiddleware: Executed endpoint 'ShopOrderWeb.Controllers.MobileAuthShopController.Get (ShopOrderWeb)'
Application:2021-02-07 22:31:10.262 +00:00 [Information] Microsoft.AspNetCore.Hosting.Diagnostics: Request finished HTTP/1.1 GET https://shoporderapi.azurewebsites.net/api/mobileauthshop/Apple - - - 302 - - 706.4353ms
### Expected Behavior
The WebAuthenticator can login on iPhone and delivers the accesstoken
### Actual Behavior
The call to the
r = await WebAuthenticator.AuthenticateAsync(authUrl, callbackUrl);
shows the login dialog, but ends with the message “Registration not completed" and I get an exception in the app a the end.
### Basic Information
- IDE: Visual Studio 2019 16.8.4
- Platform Target Frameworks: Xamarin.Forms 5.0.0.1931, Xamaein.Essentials 1.6.1
- iOS: The version of the iOS SDK you are compiling against, 11.0
- Affected Devices: iPhoneX iOS 13.4
Answers:
username_0: Hi, I solved the problem. The configuration for oAuth2 for Apple is more complicated as described here:
https://docs.microsoft.com/en-us/xamarin/ios/platform/ios13/sign-in
It works only for iOS >= 13 . For iOS < 13 and for other platforms, which use iOS oAuth2, it is necessary to configure more in iOS Certificates.
1. Create an identifier for your app explicit
2. Create a Key for the app and download the certificate for Azure
3. Create a new ServiceId -> the id is your clientId in Azure
Here Domain is like the url in azure XXXXXXXX.azurewebsites.net
The redirect url is: https:// XXXXXXXX.azurewebsites.net/signin-apple
The API must work in Azure, because Apple doesn’t accept the localhost for test.
You can read more about the configuration here:
https://medium.com/@sirajul.anik/sign-in-with-apple-verify-mobile-app-payload-under-5-minutes-for-backend-developers-d69c2217ddec
It works well in iOS. In my UWP and Android App I get the token, but the email and the name is empty. I coudn’t find any option to get it. Without the email it is useless to use the iOS oAuth2 in other systems, apart from iOS. If you find a possibility to get the email on Android or UWP, let me know. |
openshift/openshift-ansible | 160125277 | Title: docker upgrade on atomic hosts: how to handle
Question:
username_0: RHEL and CentOS Atomic hosts use the docker-current/docker-latest split. To move to docker-latest (which is recommended, right?), afaict, one has to perform the following steps.
* disable/stop docker systemd service unit,
* enable/start docker-latest systemd service unit,
* set `DOCKERBINARY=/usr/bin/docker-latest` in `/etc/sysconfig/docker`.
If that's desirable, I can change the docker-upgrade playbook as well as the docker role to do that.
Answers:
username_1: We intend to only support `docker` it's my understanding that `docker-latest` should actually conflict with our services however I don't know for sure how that's implemented, or even if it is implemented.
That said, I think https://github.com/openshift/origin/pull/9046 should make this work when you've elected to use docker-latest. We definitely won't be testing with anything other than docker-current.
username_1: Essentially, docker aka docker-current exists to fulfill the requirements of kubernetes and openshift. Whereas docker-latest is there to meet the needs of users who require newer versions of docker than those that have been thoroughly vetted against kubernetes and openshift.
Status: Issue closed
|
Hagbuck/imagetools | 278513848 | Title: [GUI] Choix de la techno
Question:
username_0: Il faut choisir la bibliothèque à utiliser pour développer l'interface graphique.
Il faut prendre en compte, la technologie en elle même et la compatibilité de la licence avec nos objectifs
Answers:
username_0: La **SDL2**, bonne candidate pour ce projet est sous la liscence **zlib** compatible avec notre licence MTI, car la seule restriction est de ne pas se revendiquer auteur de cette dernière
Status: Issue closed
|
google/material-design-icons | 149400158 | Title: Optimized PNG files - make pull request?
Question:
username_0: I have run all of the material-design-icons through a png optimizer, and have the results here: https://github.com/username_0/material-design-icons
It has shaved 1541587 bytes off the png files. Let me know if I should make a pull request.
Answers:
username_1: Hi @username_0,
Thanks for the offer. We've been doing similar work ourselves, and the results will be seen in our next release.
Status: Issue closed
|
josezy/ikaro | 641798364 | Title: Home position
Question:
username_0: * [ ] get home position (maybe MAV_CMD_GET_HOME_POSITION?)
* [ ] display distance to home
tip:
```
COMMAND_LONG {target_system : 1, target_component : 1, command : 511, confirmation : 0, param1 : 242.0, param2 : 1000000.0, param3 : 0.0, param4 : 0.0, param5 : 0.0, param6 : 0.0, param7 : 0.0}
```
Answers:
username_0: in favor of #47
Status: Issue closed
|
ory/k8s | 529272977 | Title: Installation fails with helm3 on minikube
Question:
username_0: **Describe the bug**
The `ory/hydra` chart fails to install under `helm3`
**To Reproduce**
Steps to reproduce the behavior:
1. install [helm3](https://github.com/helm/helm/releases/tag/v3.0.0)
2. `helm repo add ory https://k8s.ory.sh/helm/charts`
3. `helm repo update`
4. try to run `helm install` as per the [official guide](https://k8s.ory.sh/helm/hydra.html)
5. the installation fails with `Error: unable to build kubernetes objects from release manifest: error validating "": error validating data: ValidationError(RoleBinding.roleRef): unknown field "namespace" in io.k8s.api.rbac.v1beta1.RoleRef`
**Expected behavior**
The helm chart is installed successfully
**Environment**
* Version: v1.0.0-rc.14_oryOS.12
* Environment: helm v3.0.0, minikube v1.5.2 on Ubuntu
Answers:
username_1: Helm3 is very new, please use Helm2 for the time being when working with these charts.
username_0: thanks for the tip @username_1, although I actually already did that :-) Still thought it was good to track this one down with its own issue
username_2: This can be worked around by disabling the use of the maester. Use the following flag when installing the chart
```
--set 'maester.enabled=false'
```
username_0: good to know, thanks @username_2
Status: Issue closed
|
zhouzhouhhh/sqlstudy | 394184879 | Title: tests
Question:
username_0: from django.db import models
import time
# Create your models here.
class Department(models.Model):
id = models.CharField(max_length=20, primary_key=True, default='')
name = models.CharField(null=True, max_length=20)
def __str__(self):
return self.id
class Purchase(models.Model):
id = models.CharField(max_length=20, primary_key=True, default='')
count = models.IntegerField(null=True)
single_price = models.FloatField(null=True)
total_price = models.FloatField(null=True)
state = models.IntegerField(null=True, default='0')
create_time = models.DateField(null=True, auto_now_add=True)
def __str__(self):
return self.id
class User(models.Model):
id = models.CharField(max_length=20, primary_key=True, default='')
password = models.CharField(null=True, max_length=50, default='')
username = models.CharField(null=True, max_length=20)
age = models.IntegerField(null=True, blank=True)
state = models.IntegerField(null=True, blank=False, default='0')
identity = models.IntegerField(null=True, blank=False, default='0')
department = models.ForeignKey(Department, to_field='id', on_delete=models.CASCADE, null=True, blank=True)
def __str__(self):
return self.id
class Room(models.Model):
id = models.CharField(max_length=20, primary_key=True, default='')
count = models.IntegerField(null=True)
pos = models.TextField(null=True)
user = models.ForeignKey(User, to_field='id', on_delete=models.CASCADE, null=True, blank=True)
def __str__(self):
return self.id
class fire_extinguisher(models.Model):
id = models.CharField(max_length=20, primary_key=True, default='');
state = models.IntegerField(default='2', null=True)
keep_time = models.IntegerField(null=True) ##以天计算
create_time = models.DateTimeField(auto_now=True, null=True)
description = models.TextField(blank=True, null=True)
kind = models.IntegerField(null=True, blank=False, default='0')
purchase = models.ForeignKey(Purchase, to_field='id', on_delete=models.CASCADE, null=True, blank=True)
room = models.ForeignKey(Room, to_field='id', on_delete=models.CASCADE, null=True, blank=True)
def __str__(self):
return self.id
class Application(models.Model):
id = models.CharField(max_length=20, primary_key=True, default='')
create_time = models.DateField(null=True, auto_now_add=True)
state = models.IntegerField(null=True, default='0')
uid = models.CharField(null=True, max_length=20)
kind = models.IntegerField(null=True, default='0')
text = models.TextField(null=True, default='')
room = models.ForeignKey(Room, to_field='id', on_delete=models.CASCADE, null=True, blank=True)
def __str__(self):
return self.id |
esp-rs/esp32-hal | 1091817652 | Title: Documentation of "esp32-hal" vs "esp-idf-hal"
Question:
username_0: I'm fairly new to embedded development, native development, and rust all at once. I'm choosing this platform combo to learn - maybe there are others like me as an audience? Maybe include some minimum-need-to-know definitions of like IDF and SVD?
Thanks!
Answers:
username_1: I second you on that. I found out that the the esp-idf-hal is the one to use if you want to have access to the std library. But other than that I am as lost as you.
Additionally, there are some definitions about the gpio in both libraries. However, I found now clear example how to set the pins in the esp-idf-hal. Hopefully, someone could share some light.
Status: Issue closed
username_2: Please take a look at the [book section](https://esp-rs.github.io/book/overview.html) on this topic. If this doesn't cover the missing information, please open an issue with specifics on the book repo :)
username_3: Sorry but that does not clarify it for me... what documentation is to use when I want to use gpio without direct use of the registers? esp-idf-hal seams to have no documentation... and there seams to be no tutorial online were I can clarify that for me...
username_1: @username_2 Thx for pointing to the book. Still, as a newbie is difficult to know what the [books is referring](https://esp-rs.github.io/book/overview.html#comparing-std-and-no_std). Just to make sure. Does it mean that _no_std_ is esp32-hal and _std_ is esp32-idf-hal ?
username_2: Apologies if it's not clear! Turns out writing documentation for a project you've been working on for years is hard, as it's easy to gloss over important info. You are correct though. If you think there is a better way to write this section, or some gaps to fill please file an issue on the book repo (or PR if you're feeling adventurous :D).
username_1: @username_2 Thank you for the explanation. I [submitted a PR ](https://github.com/esp-rs/book/pull/19) in an attempt to clarify it. |
strawberry-graphql/strawberry | 989530081 | Title: Add support for the graphql-transport-ws GraphQL over WebSocket Protocol
Question:
username_0: Just for clarification: the legacy `graphql-ws` websocket subprotocols library is called "subscriptions-transport-ws", while the newer `graphql-transport-ws`websocket subprotocols library is called "graphql-ws". Yep, slightly confusing.
According to NPM download stats the [legacy library](https://www.npmjs.com/package/subscriptions-transport-ws) is still downloaded more often than the [new library](https://www.npmjs.com/package/graphql-ws). Since both protocols are incompatible I propose we support both for now and give users a choice to disable them individually.
I already implemented the new protocol for our AIOHTTP integration, which went very well since the new protocol is way better documented than the old one. I'm currently working on tests and a good abstraction that allows us to reuse the code for the ASGI integration. I'll finish work on this next week.
Answers:
username_1: we use `graphql-transport-ws` in combination with starlette & django - works really nice. I'm not sure if supporting the old library/spec makes sense, because the backend code is so much different and there is not much in common.
I _think_ many downloads on the old library are coming from the very first tutorials with subscriptions (which ar quite popular).
A good subscription example with the new protocol in the strawberry documentation would make much more sense IMO.
username_2: @username_1 out of curiosity why are you using starlette instead of Django channels? I'm assuming you're just using starlette for subscriptions only?
Status: Issue closed
username_1: @username_2 sorry for the late reply:
I use Django3 async stuff with uvicorn and an starlette app
Subscriptions just works ootb with starlette, there is no more need for django channels with async support
username_2: Interesting thanks @username_1 . So you're not using Django views at all?
username_1: well maybe this was missleading:
My stack looks simelar to this:
Uvicorn -> starlette -> Handle some routes directly with starlette (e.g. fast async views which don't need the whole django boilerplate middleware, orm stuff) and yes also delivering the websockets (handling `django.setup` myself and mount some websocket middlewares)
Everything not matching a route in starlette is forwarded into the django app (fully async) because django will handle sync/async requests itself with the `django.core.handlers.asgi import ASGIHandler` - obviously you have to handle ratelimits or your django server will be easy to ddos :)
I think this is going out of scope of this issue - if you are interested in this setup we can talk in discord?
username_1: I should also mention with starlette you already get an working graphql [example + endpoint](
https://www.starlette.io/graphql/) with starlette
username_2: Thats a really interesting setup, thanks @username_1 |
probot/create-probot-app | 490029886 | Title: create-probot-app installs probot ^7.2.0
Question:
username_0: I've tested this out with both basic and checks:
`npx create-probot-app` creates a probot where its `package.json` specifies version `^7.2.0`
I noticed the problem because my target repo wasn't reading my `.github` repo's `.github/config.yml` file. Once I updated the version to `9.3.3`, it worked fine.
I've gone through the create-probot-app repo to see if I can find the problem, but the templates look fine: `"probot": "^9.0.1"` is what I see in all of the `package.json` templates. So the problem may be occurring elsewhere.
Answers:
username_1: The version for `probot` in all templates is `^9.0.1` on the `master` branch. However, the latest release - [3.1.0](b516e3c0e84dda2d490d17076f6ac988c13132a8) - is using older dependencies.
A new release needs to be deployed.
username_2: I've published 3.1.1
Status: Issue closed
|
elastic/security-docs | 1164062621 | Title: [DOCS] Update Elastic Agent installation to match new Fleet UI in 8.1
Question:
username_0: ## Description
The Fleet / Add agent flyout UI changed in 8.1.0, so we need to update the instructions for installing Elastic Agent for Endpoint Security:
https://www.elastic.co/guide/en/security/current/install-endpoint.html#enroll-security-agent
In particular, steps 1 and 2 have been swapped, and the agent policy step has been renamed.
We need to update some screenshots as well as text.
## Screenshots
**PREVIOUS UI:**

**8.1 UI:**
<img width="1269" alt="image" src="https://user-images.githubusercontent.com/87339667/157468423-1a9ee856-4542-4495-bad3-2bbe790d1212.png"><issue_closed>
Status: Issue closed |
catamphetamine/captchan | 470097250 | Title: Enhancement: 8chan/lynxchan board search on board baR
Question:
username_0: iIf the black bar on the left allows for direct search instead of needing to click "Show All Boards" that would be convenient.
Answers:
username_1: Hello.
Yeah, I guess that would be a convenient feature.
Will be implemented, maybe by September or so.
username_0: Okay, so maybe we can brainstorm some new feautures while we have the time
username_1: I have a looot of features planned.
But I only write code in free time when I have the mood and when I'm not writing my other open-source projects, so this app is planned for finish somewhere by the end of this year.
username_1: The TODO list FYI.
Mostly in Russian.
https://github.com/username_1/captchan/blob/master/TODO.md
The next features will be:
* Adding threads to the list of tracked ones.
* Adding message posting.
I guess those two will ships somewhere in Autumn.
username_0: Okay so lynxchan is the most famous in https://endchan.xyz and https://waifuist.pro
username_1: https://endchan.xyz seems like funny one, mostly Russian for some reason, but it's inactive too. [Added it](https://username_1.github.io/captchan/?chan=endchan). Just the basic message syntax: seems that they even allow [rainbow-colored text](https://endchan.xyz/.static/posting.html). Runs a [really old](https://endchan.xyz/index.json) version of `lynxchan` though: `1.7.5`. But seems to work (only briefly checked it).
https://waifuist.pro looks like a creepy/perv one, lol.
username_0: @username_1 I could have thrown in bbw-chan.nl but that is way too gross to be included.
username_1: @username_0 Oh shoot, they even got that )
Status: Issue closed
username_1: Added board search in sidebar ("Boards" -> "..." -> type board id or title).

username_0: Thanks! |
AssemblyScript/assemblyscript | 953655840 | Title: AssemblyScript will be deprecated?
Question:
username_0: `AssemblyScript will be severely impacted as the tools it has developed must eventually be deprecated due to unresolvable correctness and security problems
Answers:
username_1: No, UTF-16/UTF-8 issues https://news.ycombinator.com/item?id=27946421 |
guyellis/learn | 470346505 | Title: Update dependency css-loader
Question:
username_0: Update dependency `css-loader` to latest version.
Answers:
username_1: Hi @username_0 ,
As you created a new issue to update many other dependencies, I think this issue may be closed, right?
Anyway, I'm going to update **css-loader** in other issue too.
username_0: Because this wasn't just a simple change-the-version-number of the dependency I split it out into its own issue to isolate it and make it easier to upgrade. However that's fine if it gets updated along with the other issue.
username_1: Ok @username_0 ! No problem!
username_0: Thanks for fixing this @username_1
Tip: If you include `Fixes #802` in the PR request title/subject then when the PR is merged it will automatically close the issue.
Status: Issue closed
|
jeanmarcgp/xtsanalytics | 342068851 | Title: create roll_jdplot function
Question:
username_0: This function would be similar to jdplot, but uses a rolling window when computing its quantiles.
Then, each point is assigned to its proper quantile subset at the time the rolling window quantile is applied.
Run this over the promising jdplots to see how the clusters evolve over time. |
seleniumbase/SeleniumBase | 873716244 | Title: Task: Add a method for asserting the attribute of an element. (backdated)
Question:
username_0: Task: Add a method for asserting the attribute of an element.
Status: Issue closed
Answers:
username_0: (Resolved by https://github.com/seleniumbase/SeleniumBase/pull/889 and released with https://github.com/seleniumbase/SeleniumBase/releases/tag/v1.63.0)
Two methods were added that mostly do the same thing. When using ``--demo`` mode, the element will be highlighted if visible.
```python
self.wait_for_attribute(selector, attribute, value=None,
by=By.CSS_SELECTOR, timeout=None)
"""
Raises an exception if the element attribute/value is not found.
If the value is not specified, the attribute only needs to exist.
Returns the element that contains the attribute if successful.
Default timeout = LARGE_TIMEOUT.
"""
self.assert_attribute(selector, attribute, value=None,
by=By.CSS_SELECTOR, timeout=None)
"""
Raises an exception if the element attribute/value is not found.
If the value is not specified, the attribute only needs to exist.
Returns True if successful.
Default timeout = SMALL_TIMEOUT.
"""
``` |
envoyproxy/envoy | 916791872 | Title: AWS SigV4 signer does not canonicalize URI and query string properly
Question:
username_0: *Description*:
When calculating the String to Sign for an incoming request, Envoy does not properly canonicalize the path and query parameters. This causes AWS to reject certain requests with the error: ```The request signature we calculated does not match the signature you provided. Check your AWS Secret Access Key and signing method. Consult the service documentation for details.```
I've found the following three error cases so far, but there might be others I'm missing. In all cases below, the root problem seems to be a mismatch between the canonical request expected by AWS (first box) and the one calculated by Envoy (second box), either in the URI or query string lines (first two lines following `GET`)
1. Query parameters that are not alphabetically sorted
According to [SigV4 docs (Step 3)](https://docs.aws.amazon.com/general/latest/gr/sigv4-create-canonical-request.html), query parameters should be sorted when building the canonical query string. Envoy seems to use the entire query string intact, without parsing or sorting the params. Requests succeed if the query params sent by the client happen to be correctly sorted, but fail with the signature error below otherwise:
```
$ curl "localhost:10000/v1/?a=test&b=value"
{"statusCode": 200, "body": "{}"}
$ curl "localhost:10000/v1/?b=value&a=test" (edited for clarity)
{"message":"The request signature we calculated does not match the signature you provided.
Check your AWS Secret Access Key and signing method. Consult the service documentation for
details. The Canonical String for this request should have been
GET
/v1/
a=test&b=value
accept:*/*
host:<redacted>.execute-api.us-west-2.amazonaws.com
...
```
```
[2021-06-09 16:21:57.845][252264][debug][http] [external/envoy/source/extensions/common/aws/signer_impl.cc:65] Canonical request:
GET
/v1/
b=value&a=test
accept:*/*
host:<redacted>.execute-api.us-west-2.amazonaws.com
```
2. Query parameters with no value
Per [Step 3](https://docs.aws.amazon.com/general/latest/gr/sigv4-create-canonical-request.html), parameters with no value specified (e.g. `https://my.site/?foo`) should be canonicalized as `foo=`, with a trailing equals sign. Because Envoy uses the original query string (`foo`) instead, these requests always fail.
```
$ curl "localhost:10000/v1/?a" (edited for clarity)
{"message":"The request signature we calculated does not match the signature you provided.
Check your AWS Secret Access Key and signing method. Consult the service documentation for
details. The Canonical String for this request should have been
GET
/v1/
a=
accept:*/*
host:<redacted>.execute-api.us-west-2.amazonaws.com
...
```
```
[2021-06-09 16:22:40.948][252264][debug][http] [external/envoy/source/extensions/common/aws/signer_impl.cc:65] Canonical request:
GET
/v1/
a
[Truncated]
region: us-west-2
host_rewrite: <redacted>.execute-api.us-west-2.amazonaws.com
- name: envoy.filters.http.router
clusters:
- name: some_service
connect_timeout: 5s
type: STRICT_DNS
lb_policy: ROUND_ROBIN
load_assignment:
cluster_name: some_service
endpoints:
- lb_endpoints:
- endpoint:
address:
socket_address:
address: <redacted>.execute-api.us-west-2.amazonaws.com
port_value: 443
transport_socket:
name: envoy.transport_sockets.tls
```
Answers:
username_1: cc: @marcomagdy @username_3
username_2: I can pick this up
username_3: Not stale
username_0: Fixed in #17137. Thanks @username_2!
Status: Issue closed
|
alucardu/TooManyBones | 610681917 | Title: Implement dexterity tracker
Question:
username_0: **Current behavior**
- Currently players have to use a die or their memory to track their dexterity usage.
**Expected behavior**
- Players can use a dexterity tracker located near their player board to track their dexterity usage.
- When they end their turn the dexterity meter is reset to 0.<issue_closed>
Status: Issue closed |
thinktandem/tandem | 465418886 | Title: Add benefit for MyersBriggs and/or CliftonStrengths finder reports
Question:
username_0: @username_1 the thinking here is to incentivize people to do this, eg make it an explicit (although still voluntary) part of onboarding and listed as a benefit. I think we want to do this because its valuable data for both Tandem and the person themselves.
if we draw from some other stipend then people are going to need to make a tradeoff between doing this or something else and that obviously lowers the incentive. We are only talking about probably less than $100 a person here.
Answers:
username_1: These should fall within our existing continuing education stipend.
username_0: @username_1 the thinking here is to incentivize people to do this, eg make it an explicit (although still voluntary) part of onboarding and listed as a benefit. I think we want to do this because its valuable data for both Tandem and the person themselves.
if we draw from some other stipend then people are going to need to make a tradeoff between doing this or something else and that obviously lowers the incentive. We are only talking about probably less than $100 a person here.
username_2: My thoughts:
## Privacy:
On a personal note, I was horrified at the thought of a psych test when Dustin suggested it. The Clifton site was interesting though, so I took it out of my own curiosity. Once I went through the result, it was clear the focus was on professional related strengths, and I was OK with sharing it.
Despite how professionally focused any test like this is, and despite we're recognizing this as an opt-in from the beginning, we should emphasize 1) it's first and foremost for personal curiosity/self improvement, and 2) there's absolutely no expectation to share the results.
## Accuracy:
I'm suspect I could have ended up with different results depending on my mood at the time I took the Clifton Strengths assessment. I also missed a few questions based on the ambiguity of the questions. That said, I think there was value in the results.
## Tandem's Usage of the Data
It makes sense that we might be able to use the data, but are we really going to? If so, we would need to dedicate time to figuring out how to interpret Clifton results and how they would map to preferential assignments with Tandem.
Status: Issue closed
|
asriz7777/learning_testsuites | 375327248 | Title: Preprodtesting : ApiV1ProjectsIdProjectChecksumsGetQueryParamPagesizeNegativeNumber
Question:
username_0: Project : Preprodtesting
Job : UAT
Env : UAT
Region : US_WEST_3
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Tue, 30 Oct 2018 05:30:20 GMT]}
Endpoint : http://172.16.31.10/api/v1/api/v1/projects/0UQbKW8X/project-checksums?pageSize=-1
Request :
Response :
{
"timestamp" : "2018-10-30T05:30:20.287+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/projects/0UQbKW8X/project-checksums"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot ---
Status: Issue closed
Answers:
username_0: Project : Preprodtesting
Job : UAT
Env : UAT
Region : US_WEST_3
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Tue, 30 Oct 2018 05:30:20 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/projects/0UQbKW8X/project-checksums?pageSize=-1
Request :
Response :
{
"timestamp" : "2018-10-30T05:30:20.287+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/projects/0UQbKW8X/project-checksums"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot --- |
jonasman/TeslaSwift | 1155565525 | Title: API Token Reuse Help
Question:
username_0: Hey there,
I seem to be misunderstanding the token reuse process. Logging in via `api.authenticateWeb` successfully logs the user in and keeps them logged in for a few days, but the token always expires and they have to consistently re-login. I am currently saving the returned token after login like such:
`UserDefaults.standard.set(token.jsonString!, forKey: "tesla.token")`
And then immediately calling 'api.reuse(token: token)'.
Then, under my app's initialization process, I run the following:
`if let jsonString = UserDefaults.standard.object(forKey: "tesla.token") as? String,
let token: AuthToken = jsonString.decodeJSON(){
api.reuse(token: token)
print("Reusing")
}`
Is there a better way to handle this, or am I missing the mark altogether? Any help is greatly appreciated, thank you!
Answers:
username_1: that should be enough, do you have the latest version? there were some bugs related to token refresh
username_0: I can try to uninstall and re-install. Newest would be 7.1.0 correct?
username_0: Updated to the newest version. I'll post another issue if it persists. |
Chicago/predicting-e-coli-concentrations | 289001218 | Title: Publish pre-print to Chi Hack Night
Question:
username_0: After talking with @derekeder, he would be interested in posting the final document on Chi Hack Night (presumably [here](https://chihacknight.org/blog/index.html)). Because we're submitting to a journal (and because we don't have funding to publish it as open access), we would need to publish a pre-print version, such as the one from #23.
Rmarkdown makes it easy to output the pre-print as an HTML version, which can be uploaded as a blog entry to the [Chi Hack Night blog repo](https://github.com/chihacknight/chihacknight.org/tree/master/_posts/blog). We can simply take the pre-print version and switch from a PDF output to an HTML output.
We should conclude the entry with a link to the PDF version of the pre-print so people know the proper citation. |
remeda/remeda | 708228380 | Title: Promise helper?
Question:
username_0: Can we handle Promise inside `pipe` function? Ramda has such functionalities.
https://ramdajs.com/docs/#pipeP
Answers:
username_0: It would be great if we can resolve promises like this.
```ts
const res = R.pipe(
fetch('https://jsonplaceholder.typicode.com/todos/1'),
R.resolve(),
R.pick(['id', 'title']),
)
console.log(res) // => { id: 1, title: "delectus aut autem" }
```
where R.resolve is
```ts
function Resolve<T> = <R>(args: Promise<T>, fn: (value: T) => R): Promise<R>
function Resolve<T> = <R>(fn: (value: T) => R): Promise<R>
```
Is it technically possible and acceptable? If so, I would try to implement it.
Thanks
username_0: I've end up to create my own `mapPromise` function.
```ts
export function mapPromise<T, R>(
fn: (a: T extends Promise<infer S> ? S : never) => R,
): (a: T) => Promise<R> {
return async (a: T) => {
return (fn as any)(await a)
}
}
```
```
const res = await R.pipe(
fetch('https://jsonplaceholder.typicode.com/todos/1'),
mapPromise(R.pick(['id', 'title'])),
)
console.log(res) // => { id: 1, title: "delectus aut autem" }
``` |
savonrb/savon | 45874149 | Title: [version 3] Empty body for element containing 'choice' elements
Question:
username_0: This example uses the same wsdl and xsd as ticket #632
``` ruby
require 'bundler/setup'
require 'savon'
require 'pp'
require 'logging'
logger = Logging.logger['root']
logger.add_appenders(Logging.appenders.stdout)
logger.level = :debug
client = Savon.new("test.wsdl")
pp client.services
pp client.operations(:alaService, :alaEndpoint)
order = client.operation(:alaService, :alaEndpoint, :order)
p order.example_body
order.body = {
orderRequest: {
message: {
messageId: "b68b2cf4-475e-11e1-a92e-fb2ff6467c99",
correlationId: "bac72cfa-475e-11e1-a92e-fb2ff6467c99",
sentAt: "2012-01-20T18:30:43Z",
},
orderReferences: {
buyerOrderReference: "REF1234567",
},
buyer: {
buyerIdentifier: "foo",
},
seller: {
sellerIdentifier: "bar",
},
alaNewInstall: {
location: {
addressKey: "123456789012",
},
serviceItem: {
product: {
productName: "adsl24",
},
},
},
},
}
puts order.build
#resp = order.call
#p resp
```
Gives the following output:
```
INFO Savon::Importer : Resolving WSDL document "test.wsdl".
[Truncated]
<xs:choice maxOccurs="unbounded">
<xs:element maxOccurs="1" name="message" type="ala:message" />
<xs:element maxOccurs="1" name="orderReferences" type="ala:orderReferences" />
<xs:element maxOccurs="1" name="buyer" type="ala:buyer" />
<xs:element maxOccurs="1" name="buyerContactDetails" type="ala:buyerContactDetails" />
<xs:element maxOccurs="1" name="seller" type="ala:seller" />
<xs:choice maxOccurs="1">
<xs:element name="alaMigration" type="ala:alaMigration" />
<xs:element name="alaNewInstall" type="ala:alaNewInstall" />
<xs:element name="alaModify" type="ala:alaModify" />
<xs:element name="alaCease" type="ala:alaCease" />
</xs:choice>
<xs:element maxOccurs="1" name="notes" type="ala:notes" />
<xs:any namespace="##other" processContents="lax" minOccurs="0" maxOccurs="unbounded" />
</xs:choice>
</xs:complexType>
</xs:element>
```
I didn't write this xsd, and I'm not sure why it's wrapped in a 'choice' rather than using `minOccurs=0` on each element. |
emirotin/WpPygments | 3710988 | Title: highlighting doesn't work
Question:
username_0: Hi,
I just installed this plugin for wordpress and I noticed that the highlighting doesn't work.
Please advice me what to do.
Answers:
username_0: Again we have the (more or less) the same issue, getting 404
username_1: Sorry to say that, but I'm afraid the plugin is not supported anymore
username_0: When did it happen? since last week or so ?
username_1: I don't know, didn't touch it for a couple of years.
On Mon, Aug 3, 2015 at 11:01 PM <NAME> <<EMAIL>>
wrote:
> When did it happen? since last week or so ?
>
> |
opentracing-contrib/java-spring-rabbitmq | 399104923 | Title: Support for multiple RabbitTemplate beans in the ApplicationContext
Question:
username_0: Currently the Spring Boot auto configuration will blow up on applicaiton startup if there is more than one bean of type RabbitTemplate. It would seem useful to support a collection of RabbitTemplate beans instead for applications that have multiple RabbitTemplate beans.
https://github.com/opentracing-contrib/java-spring-rabbitmq/blob/904bd38465c790a362740ceb8cf4f7f532d35ada/opentracing-spring-rabbitmq-starter/src/main/java/io/opentracing/contrib/spring/rabbitmq/RabbitMqTracingAutoConfiguration.java#L43
Answers:
username_1: Do you want to submit a PR for the next release?
username_0: I've since disabled Spring Boot's RabbitMQ AutoConfiguration for other reasons and so don't have a pressing need for this. If I have time next week I can take a look into solving this.
username_2: Looking forward to this feature.
username_1: I just did some tests with several rabbit templates defined and it works now. Feel free to reopen this if needed.
Status: Issue closed
username_3: @username_1 Hi, i have this trouble right now. In io.opentracing.contrib.spring.rabbitmq.RabbitMqTracingAutoConfiguration bean with name rabbitMqSendTracingAspect required a single bean of rabbitTemplate.
Could you check this for now, please?
username_3: Also, i got the idea to fix this - simply use ((RabbitTemplate) pjp.getTarget()).getMessageConverter and so on in aspect.
username_1: Currently the Spring Boot auto configuration will blow up on applicaiton startup if there is more than one bean of type RabbitTemplate. It would seem useful to support a collection of RabbitTemplate beans instead for applications that have multiple RabbitTemplate beans.
https://github.com/opentracing-contrib/java-spring-rabbitmq/blob/904bd38465c790a362740ceb8cf4f7f532d35ada/opentracing-spring-rabbitmq-starter/src/main/java/io/opentracing/contrib/spring/rabbitmq/RabbitMqTracingAutoConfiguration.java#L43
username_3: @username_1 i want to try my best in my pull-request with changes.
Currently, i have some work done with aspect + unit test, but have no idea how to write IT test with multiple RabbitTemplate beans. Can you help me with that?
In my mind, i should create configuration with some rabbitTemplate beans in it, and then try to publish/publishAndRecieve messages, using both rabbit templates, and assert that both of them trace message. How can i organize my work on it, using already existing test utils?
username_1: @username_3 Thank you for raising this PR!
I'll try to find some time this week to have a look at it.
Perhaps @username_0 could also double check?
username_4: anything update? |
soolar/sl_ascend | 408544823 | Title: Consider Powerlevelling in the Haunted Gallery
Question:
username_0: I've been using your script for some time now as I ascend as Ed. But, every time I need to powerlevel from level 10 to level 11, the script always tries to adventure at the Oil Peak, even though it has other places to choose from, such as the Top Floor of the Castle in the Sky and the Hole in the Sky (which can be unlocked from an NC in the Top Floor), which give more XP than the Oil Peak.
My runs are done with no IoTMs as I have none. So, I think the script should try to powerlevel in the Haunted Gallery if it gives more XP per turn as opposed to the Oil Peak.
I'm not sure why it wants to adventure at the Oil Peak, maybe I'm just a noob. But hopefully, you can look into this!
Answers:
username_1: While I do support a revamp of the power-leveling logic (esp. for shiny-less HCEd runs), I'm not sure that the Gallery is the best place for Myst. farming. IIRC, Gallery = Muscle, Ballroom = Moxie, and Bathroom = Mysticality. I could be completely wrong here, but I've seemed to have better luck by acquiring the antique hand mirror with a few additional adventures in the Bedroom, then switching to the Bathroom with the Scribe servant and some -combat (Shelter of Shed and whatever else) to get the NC with the Myst stats option.
I've tried this method, and I've also tried just letting the script do its thing at Oil Peak (and the Hospital in later levels), and the Bathroom *seems* to be the quickest for Myst gains and power-leveling.
username_0: Huh. I totally forgot there was a way to increase Myst in the Bathroom. I'll try that next time I ascend and see how it goes. But if adventuring in the Bathroom is a faster way than the Gallery for shiny-less runs then perhaps consider that?
I'll change the title to reflect this.
username_0: It seems that the bathroom provides 1.2x Myst (max 300) with the mirror while the gallery provides 3x+1 Myst (max 200), where x is your Myst mainstat. Judging from these formulae, the area that provides more XP for the NC is the gallery.
From my calculations, to be able to get the full 300 substats from the bathroom, you'll need to have at least 250 mainstat, while you only need 67 mainstat to get the full 200 substats from the gallery.
username_1: I think you're right for the lower levels, didn't realize the gallery also had a Myst option. The only other variables that I can think of that might impact this would be 1) The 75% combat chance in the Bathroom vs 80% combat chance in the Gallery, and 2) the combat stats gained from one location vs the other. Though I think Gallery would still win for lower levels, even after factoring those differences. |
angular/clutz | 121366911 | Title: Extending XMLHttpRequest produces wrong TS
Question:
username_0: Typescript differentiates between instance members of function type and instance methods. see [1]
Consuming jquery externs fails, because of the mismatch between:
1) https://github.com/cloud9ide/typescript/blob/master/typings/lib.d.ts#L6297
2) https://github.com/google/closure-compiler/blob/master/contrib/externs/jquery-1.9.js#L1063
We can special case this, because most often closure code extends closure classes and not classed from `lib.d.ts`.
[1]http://www.typescriptlang.org/Playground#src=class%20A%20%7B%0A%09foo%3A%20()%20%3D%3E%20void%3B%0A%7D%0A%0Aclass%20B%20extends%20A%20%7B%0A%09foo()%3A%20void%20%7B%7D%3B%0A%7D
Answers:
username_0: Fixed with https://github.com/angular/clutz/commit/7b087dfd2d6da90740a4786700e7bcf2b621db16
Status: Issue closed
|
steelbrain/linter | 868713906 | Title: Feature request: lint on save
Question:
username_0: "Lint on change" is pretty annoying, especially when you're typing a big chunk of code, and it throws warnings and errors in your face. "Lint on save" would be a much better alternative.
Answers:
username_0: OK, my bad, didn't knew it was doing it by default. Change the request to "explicitly tell user that it lints on save as a default behaviour". |
aws-amplify/amplify-cli | 360093236 | Title: No way to pass in metadata headers on website S3 object.
Question:
username_0: Sometimes you need to add metadata on the website S3 objects in order to insert Cache-Control or Expires headers. Those headers then tell Cloudfront how long to cache that particular object. There needs to be some way to instruct 'amplify publish' to add these headers to a particular file when needed.
This is how you add them manually....
To add a Cache-Control or Expires header field to Amazon S3 objects using the Amazon S3 console
Sign in to the AWS Management Console and open the Amazon S3 console at https://console.aws.amazon.com/s3.
In the Amazon S3 console, in the buckets pane, click the name of the bucket that contains the files.
In the list of objects, select the first object to which you want to add a header field.
Click Actions and click Properties.
In the right pane, expand Metadata.
Click Add More Metadata.
In the Key list, click Cache-Control or Expires, as applicable.
In the Value field, enter the applicable value:
For a Cache-Control field, enter: max-age=number of seconds that you want objects to stay in a CloudFront edge cache
For an Expires field, enter a date and time in HTML format.
Click Save.
Answers:
username_1: @username_0 You can update the hosting Cloudformation template with this information you've mentioned above before you do a publish.
username_0: I see that I could edit './#current-cloud-backend/hosting/S3AndCloudFront/template.json', but is that file safe to edit? Are any of the tools going to overwrite it?
username_1: You can go ahead and edit it.
It would be overwritten only if you go ahead and explicitly run the `amplify hosting configure` command to change some of the configurations tied to the hosting category.
username_2: You can run
`amplify hosting configure`
and select to configure CloudFront when prompt.
The CLI will lead you to specify the default cache behavior.
Status: Issue closed
username_1: @username_0 Let me know if the above-mentioned solutions worked for you. Feel free to re-open if the issue still persists.
username_0: The easiest way I found to solve this was to modify the publish command to addon an aws cli copy of the files with the appropriate cache control. Like this:
aws s3 cp --cache-control='no-cache' build/sw.js s3://`jq -r .WebBucket ../config/stack.json` |
drone/drone-ui | 232623418 | Title: Account page not accessible
Question:
username_0: With drone 0.7 installed, I am logged in to the UI through my GitHub account. I can see all repositories as expected, everything is building correctly. However, navigating to the Account section, all I can see is Loading... both in the sidebar and the main page.
The culprit seems to be the `https://drone.c8y.tech/api/user/repos?all=true` call that errs 500 with "Error fetching repository list. sql: Scan error on column index 21: converting driver.Value type <nil> ("<nil>") to a int: invalid syntax".
What can I do to fix this?
Answers:
username_1: please upgrade to 0.7.1
there was a migration issue in 0.7.0 that was fixed with patch https://github.com/drone/drone/commit/ee54698c62495852538c2fae729e821c1e70ecca
Status: Issue closed
username_0: That worked nicely, thanks! |
kymatio/kymatio | 554391578 | Title: 1D CDGMM
Question:
username_0: Currently we broadcast the last dimension of each filter, and do a elementwise multiplication. This may not work for complex filters.
https://github.com/kymatio/kymatio/pull/477/files#r358764794
Answers:
username_1: I'm ok with it until we add some complex filters
Status: Issue closed
|
livgust/macovidvaccines.com | 810388107 | Title: Display the full menu options on Desktop / add Contribute to it so it docks on mobile/small screens
Question:
username_0: Might be nice to see all the dropdown options from the top left menu on Desktop. Something like:
About | Resources | Give Feedback
Then add Contribute to it so it looks like:
About | Resources | Give Feedback | Contribute
Then make it so on small screens it docks into a button to display the menu (like it currently does all the time). I do something similar on my personal website if you'd like an example of what I'm talking about (https://downing.io/).
This way, as you add menu items they'll be visible to everyone on desktop & when you switch to a smaller screen everything appears on the same dropdown.
Answers:
username_0: Starting working on this - so far I moved the menu items into the navbar itself. Need to figure out how to make it collapsable on mobile devices, since it doesn't look that pretty on screens with <600px width. But here's what it looks like at the moment:
 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.