repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
zyedidia/micro | 278896680 | Title: prebuilt binary does not run on Debian
Question:
username_0: ## Description of the problem or steps to reproduce
The prebuilt binary doesn't seem to work in Debian 8 and Debian 9, at least not over SSH. It complains about an error in the executable's format:
戭獡㩨⼠潨敭爯湹潮⽮楢⽮業牣㩯䬠湡楤楂摲瑡楥渠捩瑨愠獵ﱦ牨湥›敆汨牥椠潆浲瑡搠牥倠潲牧浡摭瑡楥
for some reason it gets converted to an Asian script when I paste it; output from other commands pastes just fine. typing it out, the message actually reads this (sorry, it's in German):
bash: ./micro: Kann die Binärdatei nicht ausführen: Fehler im Format der Programmdatei
## Specifications
Commit hash: 5fc8f84
OS: Debian 8, Debian 9
Terminal: urxvt (SSH-session)
Answers:
username_1: Have you tried building from source? I'm not seeing this issue or the other one you posted about (#943) when running on Ubuntu.
username_0: For clarification, did you see *this* issue when running debian? And did you see the other issue on ubuntu 16.04, as opposed to ubuntu 17.10?
I would try compiling it myself, but I don't know the dependencies.
username_2: Which architecture are you running on? Did you use the correct binary?
I’m using Debian testing (x86_64) and I didn’t encounter any similar issue.
username_0: I'm an idiot, I did indeed forget that my Debian machine is 32 bit. The linux32 binary works.
Status: Issue closed
|
pytest-dev/pytest-django | 173273565 | Title: Allow db access to determine length of parameterized fixture list
Question:
username_0: I have a periodically-increasing set of regression test-cases in my test database, but I don't see a documented way to look in the django db use that data to set the length of my parameterized 'regression_test_data' fixture. As a hack, before 3.0, inside 'pytest_generate_tests' I would check if regression tests are being run and, if so, hit the database using the django db manager using to determine the number of regression tests, but even this hack fails during collection on 3.0 with 'Failed: Database access not allowed, use the "django_db" mark to enable it.'
Ideally test generation referenced from a django_db test could be enhanced to automatically allowed to access db. As an added bonus it would be great if any collection of a django_db test could be completely skipped if the command-line is run to ignore tests marked django_db. If this were implemented in version X.Y, then I could do something like:
```
# test_my_regression.py
@pytest.mark.django_db
def test_my_regression_for_my_function(my_regression_fixture): ...
# conftest.py
def pytest_generate_tests(metafunc):
if 'my_regression_fixture' in metafunc.fixturenames:
# This code won't get called during fast tests, because none of
# my fast tests use my_regression_fixture, and pytest X.Y is smart
# enough to not even call pytest_generate_tests on those skipped tests.
# pytest X.Y checks for django_db tag on test_my_regression_for_my_function,
# and says 'OK, you can use db', rather than raise error
regressions = Regression.objects.filter(...)
metafunc.parametrize('my_regression_fixture', regressions)
```
-Rolf
Answers:
username_1: Not really clear to me (and I don't think pytest-django can skip `pytest_generate_tests`), but this seems to be related: https://github.com/pytest-dev/pytest-django/issues/368.
username_0: Fair enough. In that case, I would request a supported way that I can use the django object manager from inside pytest_generate_tests in 3.0 without the 'Failed: Database access not allowed, use the "django_db" mark to enable it' error.
username_2: This is a bit tricky -- pytest collects test first, then executes them. When the first test that needs the database is executed (the db fixture is requested), then the database will be constructed. Fixtures execute with the tests, not during test collection.
Do you use pytest-djangos database setup? If so, how is the Regression model objects created? By using default pytest-django `Regression.objects.all()` would just be empty anyways?
Storing test cases in the same database that is also used for testing (being rollbacked, flushed etc between tests) seems error prone.
Can you store your regression data outside of the actual test database? Then you could access it from `pytest_generate_tests` without any problems.
username_0: It's already stored outside the test database, but uses the same schema (using django's [manual database selection](https://docs.djangoproject.com/en/1.10/topics/db/multi-db/#manually-selecting-a-database)) If I can't use the convenient django ORM, I'll just do a raw query. Is django.db.connections['regression'] supported at pytest_generate_tests time, or should I not even take that for granted and use the vanilla Python Database API?
username_3: I believe I'm hitting this same issue. Have there been any updates regarding this situation?
I can see @username_2 's point regarding trickiness, but perhaps there would be a way to either explicitly pre-provision the DB? Is there a way to directly access the fixture instantiation logic? |
itchio/itch | 203885643 | Title: Add option to limit download speed
Question:
username_0: So you can make sure it doesn't eat the whole house's capacity. Always wished steam had this.
Answers:
username_1: Actually I think it does! As for itch, it's planned as well - I think another issue mentioned this already.
username_2: I'd like to see this in itch.
username_1: A quick update: itch v25 has code for throttling the overall bandwidth usage, but it's not accessible anywhere in the UI yet.
username_3: Is it accessible in a config file or something?
username_1: @username_3 nope, no setting is accessible anywhere right now, not even via config files.
username_4: @username_1 any update on this? Also what about better bundle support than the website (issue 2401)
Also forgive me for off-topic question: for outside-of-client itch games, is it as easy as moving files to add them to the client? |
armory3d/armory | 263641443 | Title: Gamepad input: button started/released/down not working properly
Question:
username_0: All buttons receives the started/released/down event signals when pressing/releasing button 0. Buttons 1, 2, 3, does not receive any event signals.
```haxe
var gamepad = Input.getGamepad(0);
if(gamepad.released("0")){
trace("button 0");
}
if(gamepad.released("1")){
trace("button 1");
}
if(gamepad.released("2")){
trace("button 2");
}
if(gamepad.released("3")){
trace("button 3");
}
```
**Log ouput from pressing button 0**
```
GameManager.hx:197: button 0
GameManager.hx:201: button 1
GameManager.hx:205: button 2
GameManager.hx:209: button 3
GameManager.hx:197: button 0
GameManager.hx:201: button 1
GameManager.hx:205: button 2
GameManager.hx:209: button 3
GameManager.hx:197: button 0
GameManager.hx:201: button 1
GameManager.hx:205: button 2
GameManager.hx:209: button 3
GameManager.hx:197: button 0
GameManager.hx:201: button 1
GameManager.hx:205: button 2
GameManager.hx:209: button 3
```
All buttons work fine and receives the correct signals in the minimal Kha test project ([attached](https://github.com/armory3d/armory/files/1364643/input_test.zip)).
Answers:
username_1: Little progress, exposed a way to use kha key codes directly:
```hx
import iron.system.Input;
//...
if (gamepad.released(Gamepad.keyCode(0)) {
trace("button 0"); // Cross / A button
}
if (keyboard.released(Keyboard.keyCode(kha.input.KeyCode.Up)) {
trace("up");
}
```
username_0: Hi @username_1, seems like Gamepad.keyCode(0) is not implemented? Gamepad.buttons[0] works though.
username_0: Yes, works now!
Status: Issue closed
|
YaLTeR/MouseTweaks | 941610410 | Title: Loom and Stonecutter wheel tweak compatibility
Question:
username_0: Minecraft 1.14 introduced Loom and Stonecutter, but wheel tweak isn't working with it.
Kinda showcase of the issue [here](https://www.youtube.com/watch?v=R_LZOWfCX74).
Answers:
username_1: They don't work because those inventories have slots with a scrollbar, which takes all wheel input for itself, just like in the creative menu. I'm not sure if this is something I want to override because it's the behavior people will expect coming from vanilla.
username_0: Then I'll probably close this issue, considering it exhausted.
Status: Issue closed
|
Altinn/altinn-studio | 406779457 | Title: expand and minimize file explorer when working on service files
Question:
username_0: *Related to issue #954*
**Functional architect/designer:** @-mention
**Technical architect:** @-mention
**Description**
As a service developer I want to be able to minimize the file explorer when i'm working on service files to get more space for the code editor.
This functionality can possibly be used in other areas of Altinn studio and should be consistant with the rest of the solution.
**Sketch (if relevant)**
// Sketches needed
**Acceptance criterea**
- What is allowed/not allowed
- Validations
- Error messages and warnings
- ...
**Tasks**
- [ ] Sketches
- [ ] Documentation
- [ ] developer tasks
Answers:
username_1: Close this. Need to work on concept for studio designer to see what we need
Status: Issue closed
|
raiden-network/raiden-services | 401285170 | Title: Include reward info to MS update
Question:
username_0: ## Description
**As Josh I want to be able to**
include a reward in the update to the MS
**So that I can**
get a higher chance of the MS monitoring my channel.
## Acceptance Criteria
- [ ]
## Related story / issue
-<issue_closed>
Status: Issue closed |
symfony/symfony | 186122966 | Title: [Form] Ease filter forms with GET
Question:
username_0: Imagine a filter form that is submitted using simple GET.
We've built a special form type with mostly choice types (ie. filter on property X with value Y).
The problem comes with rendering; choice A needed to be simply rendered as a native form select (ie. `{{ form_widget(form.choiceFieldA) }} ``, choice B required text links (ask the designer why ;-)).
I eventually ended up foreaching choices, checking current state, rendering hidden inputs (ie. keep state when choice A was _submitted_), avoiding `form_rest` to be called, building the proper urls/query strings, dealing with multiple options. It was horrible.
However the result was worth it. basically something like this https://getbootstrap.com/components/#nav-pills
Imo. it would be really cool if i got this result with simply `{{ form_links(form.choiceFieldB) }}`.
Any thoughts?
Answers:
username_0: To clarify, i want my radio choice type to be work like this

Status: Issue closed
|
naver/egjs-infinitegrid | 579195412 | Title: Position not ok when one column on vue-infinitegrid
Question:
username_0: ## Description
When I add items to my array when I have one column (mobile) the items are added to it but I cannot see them unless I resize the screen, I inspected the code with dev tools and the position of the elements is not correct:
`style="position: absolute; left: -100000px; top: -100000px;"`
It changes to the correct one after I resize the screen but it should appear correct when I load the new elements. This behaviour does not happen when I have two or more columns on my grid.
Any ideas on why this is happening?
Answers:
username_1: @username_0
Is there any demo I can test?
username_2: I'm having the similar issue while using react version(react-infinitegrid) of lib. Please guide me if any one know to fix it.

username_1: @username_2
Could you show me how to use InfiniteGrid JSX?
username_2: {childElements}
</GridLayout>
</div>
);
}
componentWillUnmount() {}
componentDidMount() {
}
loadNextPage = (groupKey) => {
// Execute call from Parent components
this.props.loadPage(groupKey);
};
}
GalleryGridLayout.propTypes = {};
GalleryGridLayout.defaultProps = {
className: "",
};
const mapStateToProps = (state) => ({});
const mapDispatchToProps = (dispatch) => ({});
export default connect(
mapStateToProps,
mapDispatchToProps
)(InfiniteGalleryGridLayout);
```
username_1: @username_2
In my opinion, the key of group 1 and group 2 of childElements is duplicated.
username_2: Ok Thanks, I've checked it out just. But, its having unique key and group key.

username_2: @username_1 ,
Upto 2 page load, it works properly. But when I try to load/scroll up 3rd time, it messed the position of the second loaded group.
Video link: https://drive.google.com/file/d/1F2A16hpMlCIa7c9DJwjznI1fEOzm2_AQ/view?usp=sharing
username_1: @username_2
Does the key change each time it is rendered? Does the key change each time it is rendered?
username_2: Hi @username_1,
I've just checked it out. It remains the same for each renders for the same item/image.
username_1: @username_2
I made an example. Have you ever tried to copy and paste your code as much as possible. Is there a difference?
https://codesandbox.io/s/confident-jackson-rz7lo?file=/src/App.js:321-331
username_2: Only difference is that I'm passing childelements from the parent component. In your code, childelements are getting generated in the same component.
username_1: @username_2
Can I know the jsx of childElements?
username_2: <ImageItem
key={`${image.id}_${Date.now()}`}
url={image.urls.thumb}
alt={image.alt_description}
width={image.width}
height={image.height}
/>
</div>
```
I'll try to share the codesandbox link with you today or may be tomorrow.
username_1: @username_2
key={`${image.id}_${Date.now()}`}
is't it true that the key doesn't change when rendering?

It seems that the key changes every time it is rendered.
Aren't you writing this way in the parent component?
```js
const childElements = images.map(image => ....)
```
username_2: @username_1,
Oh! My bad. Yeah, Key will differ on render. I understood group key.
Thanks a lot for the torubleshooting.
Status: Issue closed
|
bitcoin/bitcoin | 355642928 | Title: ibd memory usage up in 0.17
Question:
username_0: Just to track the memory usage issue caused by the leveldb subtree change in #13925
See also https://bitcoinperf.com/timeline/?exe=3%2C4%2C2%2C1&base=1%2B23&ben=reindex.522000.dbcache%3D2048.mem-usage&env=1&revs=50&equid=on&quarts=on&extr=on
Answers:
username_1: This was discussed in the meeting of 2018-08-30,
```
21:04 < gmaxwell> I no longer think #14109 is blocking, it appears to be a measurement artifact. pages in the cache in read only mmaps show up in res.
```
so I'm untagging this from 0.17.0
Status: Issue closed
|
xws-bench/battles | 135246181 | Title: Computer:176 Computer:24
Question:
username_0: Syndicate_Thug*Twin_Laser_Turret.Syndicate_Thug*Twin_Laser_Turret.Serissu*Swarm_Tactics*Stealth_Device.Torkil_Mux*Twin_Laser_Turret*Bossk.VSOuter_Rim_Smuggler*Intelligence_Agent*Navigator*Anti-Pursuit_Lasers.Outer_Rim_Smuggler*Intelligence_Agent*Navigator*Anti-Pursuit_Lasers.Prototype_Pilot*Chardaan_Refit*Autothrusters.Prototype_Pilot*Chardaan_Refit*Autothrusters.<br>
http://bit.ly/1oCyQQI<br> |
facebookresearch/hydra | 578209059 | Title: Frequency
Question:
username_0: # 🚀 Feature Request
<!-- A clear and concise description of the feature you are requesting -->
## Motivation
**Is your feature request related to a problem? Please describe.**
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
<!-- Please link to any relevant issues or other PRs! -->
## Pitch
**Describe the solution you'd like**
<!-- A clear and concise description of what you want to happen. -->
**Describe alternatives you've considered**
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
**Are you willing to open a pull request?** (See [CONTRIBUTING](../CONTRIBUTING.md))
## Additional context
Add any other context or screenshots about the feature request here.
Status: Issue closed
Answers:
username_1: Closing. feel free to open a new one and provide actual details if you want anything. |
pimutils/todoman | 753367285 | Title: Usage page on readthedocs shows stack trace
Question:
username_0: The page https://todoman.readthedocs.io/en/stable/usage.html shows a stack trace instead of the help output of todoman.
Answers:
username_1: This is fixed in `master`, a new documentation will be built with the next release.
username_1: Fixed.
Status: Issue closed
|
bbc/psammead | 553691514 | Title: Make MostRead a non-alpha component
Question:
username_0: **Is your feature request related to a problem? Please describe.**
Once all accessibility and UX issues have been addressed, move MostRead out of alpha.
**Describe the solution you'd like**
~Just do it~ Confirm and finish all accessibility requirements, move MostRead out of alpha and begin https://github.com/bbc/simorgh/issues/4986
**Describe alternatives you've considered**
N/A
**Testing notes**
[Tester to complete]
Dev insight: Will there be any potential regression? etc
- [ ] This feature is expected to need manual testing.
**Additional context**
Add any other context or screenshots about the feature request here.<issue_closed>
Status: Issue closed |
DBMS-Consulting/CQT2 | 266363774 | Title: Release 2 : Priority 1 : Pick list should be type-ahead everywhere.
Question:
username_0: All Modules:
Designee, Designee 2, Designee 3, Drug Program and Drug Protocol.
Answers:
username_0: Fixed. Tested on 21-Oct-2017.

------

------

-------

Status: Issue closed
username_0: Fixed |
farmerbb/Taskbar | 393892857 | Title: Make taskbar not overlay apps
Question:
username_0: I'm not sure if this is actually possible or not, but would it be possible to make Taskbar show up underneath all apps rather than an overlay, like how the navbar shows up?
Answers:
username_1: I also do not like overlay.
Maybe taskbar could show on press "home" button. And show the current launcher (home screen) on long press "home" button.
Maybe it can popup like a keyboard. This way, when visible, it will be docked down.
(Off topic 1 - I have 10" tablet and it will be useful to set custom/bigger size of open/close (and other) taskbar buttons.
Off topic 2 - Recent applications can work without "usage access" - showing only applications started from taskbar.) |
CoinAlpha/hummingbot | 1001270690 | Title: Document curl script and end-to-end testing instructions for QA
Question:
username_0: ## Why
This was discussed w/ James this morning. We need to start involving QA into testing gateway v2. The best way to start is to provide them with some simple instructions on how to use the curl test scripts, and also how to start doing end-to-end testing with Hummingbot.
## What
* Write documentations for using curl test scripts
* Write documentations for doing end-to-end tests from Hummingbot |
myzhan/boomer | 238635371 | Title: Could not connect to locust master using example scenario
Question:
username_0: ```shell
➜ loc go run main.go
2017/06/26 14:42:17 Boomer is built without zeromq support. We recommend you to install the goczmq package, and build Boomer with zeromq when running in distributed mode.
2017/06/26 14:42:17 Boomer is connected to master(127.0.0.1:5557) press Ctrl+c to quit.
2017/06/26 14:42:17 read tcp 127.0.0.1:58016->127.0.0.1:5557: read: invalid argument
exit status 1
➜ loc sudo lsof -i :5557
COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME
python2.7 94417 **** 16u IPv4 0x95313e37ccf82897 0t0 TCP *:5557 (LISTEN)
```
```shell
[INFO] [root._exec:219] Cmd : locust --host private-network-host -c 1 -r 1.0 -n 1 -f ../dummy.py --master --master-bind-host=0.0.0.0 --master-bind-port=5557
[2017-06-23 13:21:20,804]. /INFO/stdout: GRAFANA_TAG=generic
[2017-06-23 13:21:20,804]. /INFO/stdout:
[2017-06-23 13:21:20,820]. /INFO/locust.main: Starting web monitor at *:8089
[2017-06-23 13:21:20,821]. /INFO/locust.main: Starting Locust 0.7.5
```
Answers:
username_1: Look at this https://github.com/username_1/boomer/issues/6 for details :D
Status: Issue closed
|
song940/node-escpos | 210730660 | Title: BASE64 image data url support?
Question:
username_0: I am trying to use base64 image data url but am getting error
`TypeError: url.indexOf is not a function`
Is there any way around or am doing something wrong here?
`let device = new Escpos.USB();`
`let printer = new Escpos.Printer(device);`
`Escpos.Image.load(dataUrl, function(image) {`
`device.open(function() {`
`printer`
`.align('ct')`
`.raster(image)`
`.cut()`
`.close();`
`});`
`});`
Answers:
username_1: `url` is the path to the file. It can be a relative path, an http url, a data url, or an in-memory Buffer.
`TypeError: url.indexOf is not a function` means your url is not a `string` . Is your `url` correct?
username_0: i am using Data Url `data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAaoAAAJFCAYAAACMdbM8AAAg.......`
Status: Issue closed
username_0: Solved it, data passed to `Escpos.Image.load` wasn't in perfect shape, a little bit cleaning got the job done.
Thanks for response.
username_2: How did you fixed this ??
username_3: "data passed to Escpos.Image.load wasn't in perfect shape, a little bit cleaning got the job done"
username_4: The problem is back, Here are details, I am using BASE64 image data,
`Escpos.Image.load(data.image,function(image) {
device.open(function() {
printer
.align('ct')
.image(image)
.cut()
.close();
});
});`
That's how my code look's like , result with `.image(image)` is as below
https://ibb.co/kUvsrz
Any suggestions ?
username_4: I am trying to use base64 image data url but am getting error
`TypeError: url.indexOf is not a function`
Is there any way around or am doing something wrong here?
```js
let device = new Escpos.USB();
let printer = new Escpos.Printer(device);
Escpos.Image.load(dataUrl, function(image) {
device.open(function() {
printer
.align('ct')
.raster(image)
.cut()
.close();
});
});
```
Status: Issue closed
|
spraakbanken/korp-frontend | 478937030 | Title: Trend diagram x-axis labels are incorrect for 1900 and earlier
Question:
username_0: The label 1900 is missing completely, and it its place it says 1890. For 1890 it says 1880, and so on. Everything before the year 1910 is offset by 10 years.
Answers:
username_1: This bug was again noticed by a user. Even though it’s not fatal, it’s a bit annoying.
Do you have plans to tackle this bug yourselves? The most recent commit to [Rickshaw](https://github.com/shutterstock/rickshaw) seems to be one year old, so its development seems to have stalled or at least slowed down, which doesn’t feel very promising as regards waiting for the bug to be fixed upstream anytime soon. @jroxendal wrote in an email in June 2018 that if the Rickshaw developers don’t fix this bug, you’d have to hack it yourselves.
username_2: Yes, we will have to do something ourselves. Maybe the best solution is to switch to a library that is better maintained. Unfortunately I don't know when there will be time for this.
username_1: (I’m sorry about the clutter above resulting from pushing a number of rebased commits to our fork with the commit message referring to this issue; I didn’t realize that they’d be mentioned here.)
A better-maintained graph library would certainly be nice, but I suppose that modifying the code for another library might not be straightforward unless the new library is almost a drop-in replacement.
I took for the first time a look at the code I found relevant to this issue in both Korp and Rickshaw, and it appeared to me rather easy to fix, unless of course I overlooked something. I noticed that you (@jroxendal, to be precise) had already implemented a fix for decades in 2014 ([commit b917cc39](https://github.com/spraakbanken/korp-frontend/blob/b917cc3987b3d03e6057a32d039cf80c43947154/app/scripts/results.coffee#L1481-L1494)) but it appeared not to be in use. I amended the fix slightly to work also with decades 1800 and earlier and created pull request #176 of it.
As far as I can see, the labels are now correct; see for example https://korp.csc.fi/korp-test/jn9/?mode=swedish#?corpus=klk_sv_1909,klk_sv_1908,klk_sv_1907,klk_sv_1906,klk_sv_1905,klk_sv_1904,klk_sv_1903,klk_sv_1902,klk_sv_1901,klk_sv_1900,klk_sv_1899,klk_sv_1898,klk_sv_1897,klk_sv_1896,klk_sv_1895,klk_sv_1894,klk_sv_1893,klk_sv_1892,klk_sv_1891,klk_sv_1890,klk_sv_1889,klk_sv_1888,klk_sv_1887,klk_sv_1886,klk_sv_1885,klk_sv_1884,klk_sv_1883,klk_sv_1882,klk_sv_1881,klk_sv_1880&cqp=%5B%5D&lang=sv&search=lemgram%7Ckorp%5C.%5C.nn%5C.1&page=0&result_tab=2 and the trend diagram from there:
However, I’m wondering if the reason why you didn’t use the existing fix also applies to this fix, too. I haven’t noticed any unwanted side-effects, though.
username_2: Really cool! The reason why we didn't use the fix is unfortunately unknown. I will test this on Monday and merge the pull request if everything looks fine.
Status: Issue closed
|
iBaa/PlexConnect | 56800596 | Title: PlexConnect fails with custom scanner S-Babs or BABS
Question:
username_0: Hi.
My anime series are using custom scanner S-Babs and it has stopped working with Plexconnect only.
When I try to enter the Anime section i get an 404 error stating
http://localhost/PMS(PMSServeraddress:32400)/library/sections/10?PlexConnect=S_-_BABS
Can't connect to localhost
The Plexconnect Log does not state anything when trying to enter the Anime setion.
When entering a working section like Cartoons it says
12:15:55 XMLConverter: Section scanner found, updating command.
I am running latest commit
Apple-TV:/Applications/PlexConnect/.git root# git log
commit d16d437e7814a50d6d607436c16fcf8f05ab8004
Author: username_1 <<EMAIL>>
Date: Thu Feb 5 14:27:23 2015 +0000
1.0 (mono) badge
I have an older version of Plexconnect where my Anime still works. With a commit dated Oct 22.
Apple-TV:/Applications/PlexConnect_old root# git log
commit bd96af4d2c6ef338ce617c0ac23d1128d8db7f95
Author: username_1 <<EMAIL>>
Date: Wed Oct 22 20:10:22 2014 +0100
Keep Baa happy ;)
BR
<NAME>in
Answers:
username_1: Please web browser to the following address, replacing PMS_IP:PORT with you PMS IP address and port:-
http://PMS_IP:PORT/library/sections
and post the resulting xml.
username_0: This XML file does not appear to have any style information associated with it. The document tree is shown below.
<MediaContainer size="12" allowSync="0" identifier="com.plexapp.plugins.library" mediaTagPrefix="/system/bundle/media/flags/" mediaTagVersion="1420915278" title1="Plex Library">
<Directory allowSync="0" art="/:/resources/show-fanart.jpg" composite="/library/sections/10/composite/1423255816" filters="1" refreshing="0" thumb="/:/resources/show.png" key="10" type="show" title="Anime" agent="com.plexapp.agents.thetvdb" scanner="S - BABS" language="en" uuid="a4c1f896-05ae-4c1a-b3a3-9c8537e649c2" updatedAt="1423255816" createdAt="1410187822">
<Location id="16" path="/home/yakumo/Archive/Video/Anime/Downloaded"/>
<Location id="37" path="/home/yakumo/Archive/Video/Anime/OwnRips"/>
</Directory>
<Directory allowSync="0" art="/:/resources/show-fanart.jpg" composite="/library/sections/12/composite/1423255818" filters="1" refreshing="0" thumb="/:/resources/show.png" key="12" type="show" title="Cartoons" agent="com.plexapp.agents.thetvdb" scanner="Plex Series Scanner" language="en" uuid="760bfa09-a101-4e6f-a726-ed4dd4b53abf" updatedAt="1423255818" createdAt="1410199651">
<Location id="24" path="/home/yakumo/Archive/Video/Cartoons/Downloaded"/>
<Location id="25" path="/home/yakumo/Archive/Video/Cartoons/OwnRips"/>
</Directory>
<Directory allowSync="0" art="/:/resources/movie-fanart.jpg" composite="/library/sections/11/composite/1423255819" filters="1" refreshing="0" thumb="/:/resources/movie.png" key="11" type="movie" title="Filmer" agent="com.plexapp.agents.themoviedb" scanner="Plex Movie Scanner" language="en" uuid="d21cbea8-02f8-4fb6-af28-b08ecc72607b" updatedAt="1423255819" createdAt="1410199561">
<Location id="17" path="/home/yakumo/Archive/Video/Movies/Downloaded/EngSub"/>
<Location id="18" path="/home/yakumo/Archive/Video/Movies/Downloaded/NoSub"/>
<Location id="19" path="/home/yakumo/Archive/Video/Movies/Downloaded/SweSub"/>
<Location id="20" path="/home/yakumo/Archive/Video/Movies/Downloaded/Kolla igenom"/>
<Location id="21" path="/home/yakumo/Archive/Video/Movies/Ownrips/EngSub"/>
<Location id="22" path="/home/yakumo/Archive/Video/Movies/Ownrips/NoSub"/>
<Location id="23" path="/home/yakumo/Archive/Video/Movies/Ownrips/SweSub"/>
<Location id="39" path="/home/yakumo/Archive/Video/Movies/Downloaded/couch"/>
</Directory>
<Directory allowSync="0" art="/:/resources/movie-fanart.jpg" composite="/library/sections/21/composite/1423256247" filters="1" refreshing="0" thumb="/:/resources/video.png" key="21" type="movie" title="Home Videos" agent="com.plexapp.agents.none" scanner="Plex Video Files Scanner" language="xn" uuid="99f94fc1-45c7-42e6-a64f-e558264d60ff" updatedAt="1423256247" createdAt="1416492517">
<Location id="38" path="/home/yakumo/Archive/Video/Phantom"/>
</Directory>
<Directory allowSync="0" art="/:/resources/movie-fanart.jpg" composite="/library/sections/13/composite/1423256247" filters="1" refreshing="0" thumb="/:/resources/movie.png" key="13" type="movie" title="Music" agent="com.plexapp.agents.themoviedb" scanner="Plex Movie Scanner" language="en" uuid="309764ef-111d-465e-8c18-69ab87bffa59" updatedAt="1423256247" createdAt="1410199689">
<Location id="26" path="/home/yakumo/Archive/Video/Music"/>
</Directory>
<Directory allowSync="0" art="/:/resources/show-fanart.jpg" composite="/library/sections/9/composite/1423249021" filters="1" refreshing="0" thumb="/:/resources/show.png" key="9" type="show" title="Serier" agent="com.plexapp.agents.thetvdb" scanner="Plex Series Scanner" language="sv" uuid="7484e149-36c4-4b23-a153-f6b1439a97a1" updatedAt="1423249021" createdAt="1410186335">
<Location id="14" path="/home/yakumo/Archive/Video/Series/Downloaded"/>
<Location id="15" path="/home/yakumo/Archive/Video/Series/Own Rips"/>
</Directory>
<Directory allowSync="0" art="/:/resources/movie-fanart.jpg" composite="/library/sections/19/composite/1423249033" filters="1" refreshing="0" thumb="/:/resources/movie.png" key="19" type="movie" title="Serietidnings Filmer" agent="com.plexapp.agents.themoviedb" scanner="Plex Movie Scanner" language="en" uuid="3200c934-e423-44f0-8600-f6e913315977" updatedAt="1423249033" createdAt="1410203684">
<Location id="34" path="/home/yakumo/Archive/Video/Movies/Comic Movies/Downloaded"/>
<Location id="35" path="/home/yakumo/Archive/Video/Movies/Comic Movies/Ownrips"/>
</Directory>
<Directory allowSync="0" art="/:/resources/movie-fanart.jpg" composite="/library/sections/14/composite/1423250216" filters="1" refreshing="0" thumb="/:/resources/movie.png" key="14" type="movie" title="Standup" agent="com.plexapp.agents.themoviedb" scanner="Plex Movie Scanner" language="sv" uuid="041ba1ec-a738-4de1-87c8-605583845aa0" updatedAt="1423250216" createdAt="1410199747">
<Location id="27" path="/home/yakumo/Archive/Video/Movies/Standup/OwnRips"/>
<Location id="28" path="/home/yakumo/Archive/Video/Movies/Standup/Downloaded"/>
</Directory>
<Directory allowSync="0" art="/:/resources/movie-fanart.jpg" composite="/library/sections/7/composite/1423251008" filters="1" refreshing="0" thumb="/:/resources/movie.png" key="7" type="movie" title="Svenska Filmer" agent="com.plexapp.agents.themoviedb" scanner="Plex Movie Scanner" language="sv" uuid="e6c80da7-fbb0-44e5-9c08-65b805bd990e" updatedAt="1423251008" createdAt="1410179143">
<Location id="11" path="/home/yakumo/Archive/Video/Movies/Downloaded/Swedish"/>
<Location id="12" path="/home/yakumo/Archive/Video/Movies/Ownrips/Swedish"/>
</Directory>
<Directory allowSync="0" art="/:/resources/movie-fanart.jpg" composite="/library/sections/16/composite/1423251408" filters="1" refreshing="0" thumb="/:/resources/movie.png" key="16" type="movie" title="Tecknat" agent="com.plexapp.agents.themoviedb" scanner="Plex Movie Scanner" language="sv" uuid="4fe49333-b29a-4cf8-849c-26f458d28ee0" updatedAt="1423251408" createdAt="1410199877">
<Location id="30" path="/home/yakumo/Archive/Video/Movies/Cartoons/Ownrips"/>
<Location id="31" path="/home/yakumo/Archive/Video/Movies/Cartoons/Downloaded"/>
</Directory>
<Directory allowSync="0" art="/:/resources/movie-fanart.jpg" composite="/library/sections/18/composite/1423250215" filters="1" refreshing="0" thumb="/:/resources/video.png" key="18" type="movie" title="Träning" agent="com.plexapp.agents.none" scanner="Plex Video Files Scanner" language="xn" uuid="39adea65-4f83-47f9-94e1-e98d0f2fc3b1" updatedAt="1423250215" createdAt="1410199965">
<Location id="33" path="/home/yakumo/Archive/Video/Training"/>
</Directory>
<Directory allowSync="0" art="/:/resources/movie-fanart.jpg" composite="/library/sections/20/composite/1423250216" filters="1" refreshing="0" thumb="/:/resources/video.png" key="20" type="movie" title="Wrestling" agent="com.plexapp.agents.none" scanner="Plex Video Files Scanner" language="xn" uuid="e5847e34-8536-4d34-ba3b-5807d0dbe2f4" updatedAt="1423250216" createdAt="1410205259">
<Location id="36" path="/home/yakumo/Archive/Video/Wrestling"/>
</Directory>
</MediaContainer>
username_1: Ok, I don't have S-BABS scanner to test with but try the following XMLConverter.py:-
http://www26.zippyshare.com/v/ju6VltIz/file.html
Download it and replace the original XMLConverter.py file in your PlexConnect install, you'll need to shutdown PlexConnect before replacing the file. Let me know if it fixes the problem and if it does I'll commit the changes.
username_0: I would be happy to say it works now but I no =(
All sections dissapered and here is what the log is spiting out.
http://pastebin.com/qEEFBiGD
username_1: Ok, that error is nothing to do with the scanner, that's because your version of PlexConnect is too old for the new XMLConverter.py file to work with.
The XMLConverter.py file I posted will only work with the very latest version of PlexConnect. update and try again.
username_0: Did a new git pull and replaced the XMLConverter and we are back to square one.
http://localhost/PMS(PMSServeraddress:32400)/library/sections/10?PlexConnect=S_-_BABS
Can't connect to localhost
username_1: Ok, where can I get this "S-BABS" scanner so I can do some testing myself?
username_2: There u go ;
https://forums.plex.tv/index.php/topic/109441-series-better-absolute-scanner-s-babs/
username_1: Umm... that just thread just gives you a single py file and tells you "To run this, you will need to follow the regular instruction for setting up a scanner" ????? I've got no idea how to do that, isn't there just a complete plug-in package?
Besides that I'm not even sure what the point of this scanner is?
username_0: Place S - BABS.py in
On Linux.
/var/lib/plexmediaserver/Library/Application Support/Plex Media Server/Scanners/Series/
This is a scanner used mainly for Anime that uses absolute ordering ie Name.123.mkv Instead of Season and episode number and then by putting an tvdb.id file next to your anime The scanner then transforms your anime episodeds to thetvdb season and ep releases in plex so it looks good =)
username_1: Why would you not just name your Anime with the correct Season/Episode numbering to start with?
username_1: Ok, grab the latest commit and it should work now :D
username_0: Becuase almost no anime is released that way.
Example [HorribleSubs] Naruto Shippuuden - 397 [720p].mkv
Examlple2 [yibis]_One_Piece_661_[720p][AE41095E].mkv
Almost always an anime series is just epnumber to end no seasons. And if there are "season" it is almost always a new series when they are released in its original form from Japan.
Example: Code Geass: Lelouch of the Rebellion (In us season1) , Code Geass: Lelouch of the Rebellion R2 (In us season2)
What Babs and S-Babs does is that they match the released absolute numbering with the info that is located on thetvdb that are sorted by US television standards more or less.
BTW when setting up the scanner you have select in the PMS web insterface for the whole "Anime Section"
A Anime Demo Archive containing some dummy files for testing
http://yakumo.se/PMS/Anime.zip
username_0: It works with latest commit.
Thank you !! =)
Status: Issue closed
|
tuxBurner/play-akkjobs | 273236645 | Title: How to use it in Scala 2.12.2?
Question:
username_0: Hello, I am a beginner at Play!
And I want to use a task lib in my project(2.6.6 with Scala 2.12.2).
Then I try those way:
write "libraryDependencies += "com.github.username_1" % "play-akkajobs_2.11" % "1.0.1"" at build.sbt;
update Scala version from 2.12.2 to 2.11;
download lib and move it under src/libs;
All of those were not work,now,I don't know what should I do...
Answers:
username_1: Hi @username_0 i have to lift it to play 2.6
Status: Issue closed
username_1: Hi @username_0 i lifted the library to version 2.6.0 so it is now compatible with play 2.6.x.
username_0: Thank you ~!
:) |
BorgWarner-MAI/BW-DMX-Scanner | 181383004 | Title: MainActivity.java line 259
Question:
username_0: #### in danieljuric.borgwarner.MainActivity.checkNormCodeLengh
* Number of crashes: 1
* Impacted devices: 1
There's a lot more information about this crash on crashlytics.com:
[https://fabric.io/borgwarner-turbo-systems-gmbh/android/apps/danieljuric.borgwarner/issues/57f62c3f0aeb16625b32976c](https://fabric.io/borgwarner-turbo-systems-gmbh/android/apps/danieljuric.borgwarner/issues/57f62c3f0aeb16625b32976c)<issue_closed>
Status: Issue closed |
rickychan0611/bwa-u3-jammming-prj | 312008506 | Title: you may want to remove your console.logs
Question:
username_0: https://github.com/rickychan0611/bwa-u3-jammming-prj/blob/master/src/util/Spotify.js#L1-L93
It seems you have a lot of console.logs in this file, which is fine when you're debugging/testing your code but usually you want to remove your logs when you finish your project or are pushing production code. |
rfjakob/earlyoom | 437970344 | Title: make: [Makefile:35: install] Error 1 (ignored)
Question:
username_0: ```
$ sudo make install
[sudo] password for user:
sed "s|:TARGET:|/usr/local/bin|g;s|:SYSCONFDIR:|/etc|g" earlyoom.service.in > earlyoom.service
install -d /usr/local/bin/
install -m 755 earlyoom /usr/local/bin/
pandoc is not installed, skipping earlyoom.1 manpage generation
install -d /etc/default/
install -m 644 earlyoom.default /etc/default/earlyoom
install -d /etc/systemd/system
install -m 644 earlyoom.service /etc/systemd/system
chcon -t systemd_unit_file_t /etc/systemd/system/earlyoom.service
chcon: can't apply partial context to unlabeled file '/etc/systemd/system/earlyoom.service'
make: [Makefile:35: install] Error 1 (ignored)
systemctl enable earlyoom
Created symlink /etc/systemd/system/multi-user.target.wants/earlyoom.service → /etc/systemd/system/earlyoom.service.
```
Answers:
username_0: (Ubuntu 19.04, Debian 9)
username_1: On Fedora 29, systemd unit files must have the right SELinux context:
```
$ ls -lZ /etc/systemd/system
total 68
drwxr-xr-x. 2 root root system_u:object_r:systemd_unit_file_t:s0 4096 Dec 31 20:56 basic.target.wants
drwxr-xr-x. 2 root root system_u:object_r:systemd_unit_file_t:s0 4096 Nov 5 2017 bluetooth.target.wants
lrwxrwxrwx. 1 root root system_u:object_r:systemd_unit_file_t:s0 41 Nov 5 2017 dbus-org.bluez.service -> /usr/lib/systemd/system/bluetooth.service
[...]
```
I imagine Debian 10 will do something similar. But in any case, an error in setting the context is ignored because of the "-" in front of the line in the Makefile.
username_0: ```
$ ls -lZ /etc/systemd/system
итого 52
lrwxrwxrwx 1 root root ? 9 май 17 2018 apache2.service -> /dev/null
lrwxrwxrwx 1 root root ? 9 май 17 2018 apache-htcacheclean.service -> /dev/null
```
By the way, Debian/Ubuntu/Arch/Gentoo and most distros don't use SELinux.
username_1: Yeah, Debian is a bit slow, but looks like Debian 10 at least gets secure boot:
https://www.phoronix.com/scan.php?page=news_item&px=Debian-10-Testing-Secure-Boot
Status: Issue closed
username_0: What does this have to do with SELinux context?
username_2: Am I right in interpreting this to means the installation is successful, since the error is ignored?
I feel this might be worth putting in the README as a warning to users on these systems.
username_0: @username_2 PR are welcome
username_3: I couldn't see that comment on https://github.com/username_1/earlyoom - and consequently I installed and the removed and then reinstalled earlyoom, because I thought there was a problem. |
aikin/thinking-listening-reading | 655430682 | Title: 抽象概念 -> 快速落地实践 -> 系统性学习 -> 巩固
Question:
username_0: 从高效学习到深度学习
高效学习,带来的是快感,关注点是 学以致用。 该如何开始?
先找大牛的分享或者书,了解要学习的技术,包含哪些概念和概念的逻辑关系。 每一个抽象的概念,背后都是一个**模型**支撑,抽象的是对问题的解决。
每个概念都有自己的抽象层次,每一层抽象概念,都由一组逻辑关系关联。
找 Demo 直接动手操作,核心点:Demo 要简单,这样可以快速提供反馈。然后调整复杂度,再验证,再了解反馈。
系统性学习:找一本教科书,因为教科书都是从多维度去写的,而且内容很有层次,覆盖面广。
写在最后:
在所有的过程中,不断巩固内容,How to? 要产生输出,比如:分享,写文章。 |
tensorflow/tensorflow | 639030242 | Title: Support for use of tensorflow_probability.distributions.Distribution instances in model.fit(...)
Question:
username_0: <em>Please make sure that this is a feature request. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:feature_template</em>
**System information**
- TensorFlow version (you are using): 2.1.0
- Are you willing to contribute it (Yes/No): No (have yet to work out the best way to implement)
**Describe the feature and the current behavior/state.**
I have a model that I would like to train using the KL divergence between the model output and a tfp.distributions.Distribution instance. (Note: this model is using tensorflow_probability's probabilistic layers). As of right now, Distributions cannot be plumbed through the model.fit(...) pipeline due to incompatibilities with the DataAdapter classes, as:
1) They are not recognised by TensorLikeDataAdapter as a Tensor-like object (i.e. they fail the _is_tensor() check).
2) tf.keras.utils.Sequence instances that return distributions fail in convert_for_inspection (calling np.array(distribution,dtype='float64') fails as np sees distributions as sequences).
**Will this change the current api? How?**
Change to support this should have no effect on the API, but would expand functionality to users of tf and tfp.
**Who will benefit with this feature?**
Users who use both tf and tfp, and need to use distributions to train models.
**Any Other info.** |
bdwilliamson/vimp | 584719833 | Title: plot_vim may return the wrong y-axis labels
Question:
username_0: In some cases, plot_vim returns incorrect y-axis labels. This can be fixed, for now, by using a different method of plotting (e.g., ggplot).
Status: Issue closed
Answers:
username_0: Fixed in v2.0.1 (because we no longer support plot_vim; instead, we recommend using [`ggplot2`](https://ggplot2.tidyverse.org/) and [`cowplot`](https://github.com/wilkelab/cowplot) [for an example, see the [vignette](https://github.com/username_0/vimp/blob/master/vignettes/introduction_to_vimp.Rmd)]) |
ghi-electronics/TinyCLR-Drivers | 855437710 | Title: V2.1.0 Preview 4: Command 'Winc15x0Interface.GetFirmwareVersion();' before initialization crashes makes board unresponsive when
Question:
username_0: I'm working with a FEZ Feather TinyCLR V2.1.0 Preview 4.
Mistakenly I used the command 'Winc15x0Interface.GetFirmwareVersion();' before the module was initialized. This crashes application and makes the board unresponsive over USB. Firmware had to be erased in LDR mode and firmware had to be updated.
After initialization of the WiFi module the command worked as expected.
Not a big thing but should perhaps be handled more gracefully.
Answers:
username_1: Just tried it. Higher than Preview4, and could not reproduce.
Below is code and output:
```cs
var version = Winc15x0Interface.GetFirmwareVersion();
Debug.WriteLine("version : " + version);
networkController.Enable();
```
```
version : 19.5.4.15567
Mac: 70:b3:d5:fa:c0:0f
ip address :192.168.86.37
subnetmask :255.255.255.0
gate way :192.168.86.1
dns[0] :172.16.58.3
dns[1] :172.16.58.3
Phy staus True
```
This is tested on rc1-internal. Not on preview4.
Status: Issue closed
username_0: Hi, I think you misunderstood my concern. The command only crashes the application and makes the board unresponsive over USB when the command is issued **before** the WiFi module is initialzed. It is clear that it can not work before 'talking to the board' is established. However execution of the command should only be performed by the firmware after the SPI connection is successfully established. After successful initialization the command works, but this was not my concern. |
leonardfischer/wunderground-api | 989678433 | Title: Does this still work?
Question:
username_0: I've tried with my api key and everythng just seems to error out. Extracting the generated URLs seem to lead to dead api end points.
Can anyone confirm that this still works?
Answers:
username_1: Hi @username_0
that's a good question... Weather Underground changed it's API some time ago and (to my knowledge) has no free access anymore, so I can not really validate the functionality right now :( Or is it still possible to get access to their API? In that case I could update the library, like I updated https://github.com/username_1/open-weather-map-api
username_2: this is outdated and do not work.
Status: Issue closed
username_1: Hey @username_2 thanks for the update - I will archive this repo :) |
HomepointXI/Issues | 1099103167 | Title: Cannot start Celaeno
Question:
username_0: <!--
Please do not remove or change any of the template data below. Instead, simply fill out the following information like you are taking a survey or test. When filling out this information, please DO NOT use the `@` symbol. This will trigger unwanted mentions to users that may not be part of this project.
-->
**Date & Time**:
2022/01/11
**Client Version (use `/ver` in game):**
30191204_1
**Character Name:**
Fragile
**Discord Name (if different from in game):**
**Nation:**
sandy
**Job(level)/Sub Job(level):**
WHM/RDM
**NPC or Monster or item Name:**
Celaeno
**Zone name:**
Dangruf Wadi (J-12) (J-8)
**Coordinates (use `!where`):**
x:-155 y:3 z:-164 (96)
**Item tool link (for items issues only):**
**Multi-boxing? (multiple clients on same connection):**
yes
**Steps to reproduce:**
1. Click on Planar Ritt
2. Choose start void watch
3. "Indigo Stratum Abyssite IV is responding..."
4. Nothing happens
Clicking on Planar Ritt doesn't even bring up the menu.
**Any additional information:**
Answers:
username_1: @username_0 it says that the abyssite IS or ISN'T responding?
username_0: it says "the abyssite IS responding"
username_1: And you are at the rift at H-4?
username_0: I tried J-12 and J-8.
username_1: Go to the rift at H-4, the VWNM will only spawn at the one specified on the voidwatch page.
http://wiki.homepointxi.com/Category:Voidwatch#Battle_Procedure
username_0: Can you change the location to J-12? It's too far away.
Status: Issue closed
|
MatthewGrim/Lunar_SPS | 386375985 | Title: Re-design transmitter radius analysis in code to give each orbit its optimum transmitter radius
Question:
username_0: As #45 showed, the optimum transmitter radius varies from orbit to orbit. At the moment, it has not been shown that our optimisation technique finds the global minimum because it does not find the optimum transmitter size for every orbit at the same time. This has been shown in #44 to not be difficult to do, so we can add it to the code for robustness, and to be sure that we are achieving the optimal solution.
This issue documents this re-factor of the optimisation routine.
Answers:
username_0: In this issue, I have talked about the re-design of the satellite transmitter optimisation. This involves the following steps:
- Change the optimize_link_efficiency to choose a transmitter radius optimised for the maximum range of the satellite from the target.
- If the transmitter radius at this point is too small to meet the pointing requirement, increase the beam size until it is just able to.
This produces a set of designs at every altitude that should be able to pass the pointing requirement, and the optimum beam aperture for every orbit. **It does not solve the problem for if the transmitted beam can fit inside the target.**
username_0: For comparison - these should not change or improve:
**AMALIA**
Iteration: 0, Max radius: 1.12
------------------------------- SATELLITE ORBIT -------------------------------
Number of SPS: 1
Optimal orbit altitudes --> Perigee: 1200.0 km, Apogee: 1400.0 km
Orbital period --> 250.28 minutes
Total active time (blackout reduction) --> 6.04 %
Total blackout time --> 43.72 %
Max active period duration --> 0.94 hours
Max blackout period duration --> 13.31 hours
Min range to target --> 2183.26 km
Max range to target --> 2531.95 km
------------------------------ LASER TRANSMITTER ------------------------------
Minimum allowable transmitter power --> 3.44 kW
Transmitter aperture radius: 88.44 cm
Mean link efficiency --> 7.44277 %
Mean power delivered --> 110.78 W
Steady state temperature --> 70.17 Celsius
--------------------------- BATTERY CHARACTERISTICS ---------------------------
Battery capacity --> 11590.17 Whr
Battery mass --> 82.79 kg
Battery cycles --> 21000.4
Battery charge time --> 2.09 hr
--------------------------- RECEIVER CHARACTERISTICS ---------------------------
Receiver Area --> 0.365764447695684 $m^2$
Mean flux at receiver --> 0.44 AM0
Maximum flux at receiver --> 0.47 AM0
Minimum flux at receiver --> 0.4 AM0
Mean heat load on receiver --> 110.78 W
Total energy transferred --> 7099.24 MJ per year
**Sorato**
Iteration: 0, Max radius: 1.12
------------------------------- SATELLITE ORBIT -------------------------------
Number of SPS: 1
Optimal orbit altitudes --> Perigee: 2300.0 km, Apogee: 2300.0 km
Orbital period --> 383.57 minutes
Total active time (blackout reduction) --> 10.42 %
Total blackout time --> 39.34 %
Max active period duration --> 1.91 hours
Max blackout period duration --> 5.5 hours
Min range to target --> 3065.15 km
Max range to target --> 3644.01 km
------------------------------ LASER TRANSMITTER ------------------------------
Minimum allowable transmitter power --> 4.96 kW
Transmitter aperture radius: 105.62 cm
Mean link efficiency --> 1.12192 %
Mean power delivered --> 24.06 W
Steady state temperature --> 67.63 Celsius
--------------------------- BATTERY CHARACTERISTICS ---------------------------
Battery capacity --> 26875.83 Whr
Battery mass --> 191.97 kg
Battery cycles --> 13702.72
Battery charge time --> 3.2 hr
--------------------------- RECEIVER CHARACTERISTICS ---------------------------
Receiver Area --> 0.07863935625457205 $m^2$
Mean flux at receiver --> 0.45 AM0
Maximum flux at receiver --> 0.48 AM0
Minimum flux at receiver --> 0.4 AM0
Mean heat load on receiver --> 24.06 W
Total energy transferred --> 1845.36 MJ per year
username_0: Current results:
**AMALIA**
------------------------------- SATELLITE ORBIT -------------------------------
Number of SPS: 1
Optimal orbit altitudes --> Perigee: 1300.0 km, Apogee: 1300.0 km
Orbital period --> 250.28 minutes
Total active time (blackout reduction) --> 6.12 %
Total blackout time --> 43.64 %
Max active period duration --> 0.87 hours
Max blackout period duration --> 4.99 hours
Min range to target --> 2186.05 km
Max range to target --> 2490.95 km
------------------------------ LASER TRANSMITTER ------------------------------
Minimum allowable transmitter power --> 3.37 kW
Transmitter aperture radius: 92.11 cm
Min link efficiency --> 6.86157 %
Min power delivered --> 100.0 W
Mean link efficiency --> 7.41458 %
Mean power delivered --> 108.06 W
Steady state temperature --> 67.63 Celsius
--------------------------- BATTERY CHARACTERISTICS ---------------------------
Battery capacity --> 8237.55 Whr
Battery mass --> 58.84 kg
Battery cycles --> 21000.4
Battery charge time --> 2.09 hr
--------------------------- RECEIVER CHARACTERISTICS ---------------------------
Receiver Area --> 0.365764447695684 $m^2$
Mean flux at receiver --> 0.43 AM0
Maximum flux at receiver --> 0.45 AM0
Minimum flux at receiver --> 0.4 AM0
Mean heat load on receiver --> 108.06 W
Total energy transferred --> 7164.36 MJ per year
**Sorato**
------------------------------- SATELLITE ORBIT -------------------------------
Number of SPS: 1
Optimal orbit altitudes --> Perigee: 2300.0 km, Apogee: 2300.0 km
Orbital period --> 383.57 minutes
Total active time (blackout reduction) --> 10.42 %
Total blackout time --> 39.34 %
Max active period duration --> 1.91 hours
Max blackout period duration --> 5.5 hours
Min range to target --> 3065.15 km
Max range to target --> 3644.01 km
------------------------------ LASER TRANSMITTER ------------------------------
Minimum allowable transmitter power --> 4.93 kW
Transmitter aperture radius: 111.41 cm
Min link efficiency --> 1.00843 %
Min power delivered --> 21.5 W
Mean link efficiency --> 1.11557 %
Mean power delivered --> 23.78 W
Steady state temperature --> 67.63 Celsius
--------------------------- BATTERY CHARACTERISTICS ---------------------------
Battery capacity --> 26723.73 Whr
Battery mass --> 190.88 kg
Battery cycles --> 13702.72
Battery charge time --> 3.2 hr
--------------------------- RECEIVER CHARACTERISTICS ---------------------------
Receiver Area --> 0.07863935625457205 $m^2$
Mean flux at receiver --> 0.44 AM0
Maximum flux at receiver --> 0.47 AM0
Minimum flux at receiver --> 0.4 AM0
Mean heat load on receiver --> 23.78 W
Total energy transferred --> 1834.91 MJ per year
username_0: These receivers, eye balling it, are in the same orbits-ish, and have slightly better efficiencies. This is a good sign that the code has worked, but I should probably do a double check tomorrow with meld when I am more awake. I will commit this change in any case - it can always be reverted.
username_0: This new implementation is not the default implementation.
Status: Issue closed
|
PyThaiNLP/pythainlp | 467979741 | Title: Alternative syllable tokenizer
Question:
username_0: Hi,
@ponrawee has made a syllable tokeniser using CRF, codenamed `ssg`. Its performance is comparable to PyThaiNLP's algorithm, but faster.
## [Performance Comparison][result]
PyThaiNLP's algorithm is used ground truth.
<img width="951" alt="image" src="https://user-images.githubusercontent.com/1214890/61200948-3826ef80-a70d-11e9-87a9-be6247295be3.png">
## [Speed Comparison][notebook]
<img width="523" alt="image" src="https://user-images.githubusercontent.com/1214890/61200759-b8008a00-a70c-11e9-8c63-f38dcccc19b8.png">
Nevertheless, we suspect that ssg's featurizer is not optimal, causing its speed on long text is much lower than PyThaiNLP.
Repository: https://github.com/ponrawee/ssg
[notebook]: https://colab.research.google.com/drive/1ssqHG5EZ5hHL6ZjlxOuvbNlwC7ZxDe7x#scrollTo=Df2OPVVEFpbc
[result]: https://pythainlp.github.io/tokenization-benchmark-visualization/?experiment-name=Syllable-Tokenisation-PyThaiNLP-vs-CRF&q=%E0%B8%88%E0%B8%A3%E0%B8%A3%E0%B8%A2%E0%B8%B2
Answers:
username_0: hi,
`ssg` is now available via `pip` with CI-integrated and unit-tests added.
```
$ pip install ssg
```
username_1: @username_0 You can add `ssg` like deepcut in `subword_tokenize`, `syllable_tokenize`.
username_0: ^ cc: @ponrawee
Status: Issue closed
|
ACRA/acra | 13124715 | Title: Creating Error Reporter
Question:
username_0: I folowed the advanced Usage tutorial in creating my own error Reporter, but when I try and run it i recieve this error: java.lang.IllegalStateException: Cannot access ErrorReporter before ACRA#init. My code is below for the application class:
@Override
public void onCreate(){
super.onCreate();
SaveCrash crash = new SaveCrash(this);
ACRA.init(this);
ACRA.getErrorReporter().setReportSender(crash);
}
I apologize if it is something easy i am missing, but i followed the instructions on Advanced Usage page to the tee.
Answers:
username_1: I had the same problem. But my problem was with ReportsCrashes. I don't use this annotation. Also I saw it case in sources. When I call init I receive
"ACRA#init called but no ReportsCrashes annotation on Application ", and ErrorReporter is not initialized. Then I receive - throw new IllegalStateException("Cannot access ErrorReporter before ACRA#init");
username_2: @username_1 You cannot use `ACRA.init(Application)` if your Application class does not have annotations.
username_1: Yes, now I see it. But I thought that I could do it when I use a custom sender :)
username_3: Hi,
I have this error too:
with dependency: `compile 'ch.acra:acra:4.9.0'`
`--------- beginning of crash
06-14 10:44:07.045 24277-24277/com.empire E/AndroidRuntime: FATAL EXCEPTION: main
Process: com.empire, PID: 24277
java.lang.RuntimeException: Unable to start activity ComponentInfo{com.empire/com.empire.HomeActivity}: java.lang.IllegalStateException: Cannot access ErrorReporter before ACRA#init
at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2416)
at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2476)
at android.app.ActivityThread.-wrap11(ActivityThread.java)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1344)
at android.os.Handler.dispatchMessage(Handler.java:102)
at android.os.Looper.loop(Looper.java:148)
at android.app.ActivityThread.main(ActivityThread.java:5417)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:726)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:616)
Caused by: java.lang.IllegalStateException: Cannot access ErrorReporter before ACRA#init
at org.acra.ACRA.getErrorReporter(ACRA.java:331)
at com.empire.HomeActivity.setupOCR(HomeActivity.java:132)
at com.empire.HomeActivity.onCreate(HomeActivity.java:260)
at android.app.Activity.performCreate(Activity.java:6251)
at android.app.Instrumentation.callActivityOnCreate(Instrumentation.java:1107)
at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2369)
at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2476)
at android.app.ActivityThread.-wrap11(ActivityThread.java)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1344)
at android.os.Handler.dispatchMessage(Handler.java:102)
at android.os.Looper.loop(Looper.java:148)
at android.app.ActivityThread.main(ActivityThread.java:5417)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:726)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:616) `
`import org.acra.ACRA;
import org.acra.ReportField;
import org.acra.ReportingInteractionMode;
import org.acra.annotation.ReportsCrashes;
import org.acra.sender.HttpSender;
@ReportsCrashes(
formUri = "https://xxxxxx.cloudant.com/acra-xxxxx/_design/acra-storage/_update/report",
reportType = HttpSender.Type.JSON,
httpMethod = HttpSender.Method.POST,
formUriBasicAuthLogin = "-------",
formUriBasicAuthPassword = "------------------",
//formKey = "", // This is required for backward compatibility but not used
customReportContent = {
ReportField.APP_VERSION_CODE,
ReportField.APP_VERSION_NAME,
ReportField.ANDROID_VERSION,
ReportField.PACKAGE_NAME,
ReportField.REPORT_ID,
ReportField.BUILD,
ReportField.STACK_TRACE
},
mode = ReportingInteractionMode.TOAST,
resToastText = R.string.crash_toast_text
)`
The config :
public class HomeActivity extends AppCompatActivity {
@Override
protected void onCreate(Bundle savedInstanceState) {
super.onCreate(savedInstanceState);
ACRA.init(getApplication());
try {
throw new NumberFormatException();
} catch (NumberFormatException e){
ACRA.getErrorReporter().handleException(e);
}
}
thanks! |
michaelrambeau/bestofjs | 252139299 | Title: Add Redux-observable
Question:
username_0: `https://redux-observable.js.org`
`https://github.com/redux-observable/redux-observable`
I think it could be tagged as _redux_ and _functional programming_ .
Answers:
username_1: Hello @username_0
Thank you for the suggestion, `redux-observable` will be available on bestof.js.org very soon.
username_1: Hello @username_0
`redux-observable` project is available on bestof.js.org.
Check here: https://bestof.js.org/tags/redux
Status: Issue closed
|
reactive-commons/reactive-commons-java | 804849942 | Title: Graceful shutdown
Question:
username_0: The connection with the broker and the message handling should be gracefully stopped. Today, When a spring application is shutdown, reactive commons keeps receiving messages and failing because other beans were destroyed. |
zazuko/fso-lod | 115509461 | Title: SPARQL not working
Question:
username_0: The "SPARQL Endpoint" linked in the README is:
A) Not a SPARQL Enpoint but just a YASGUI sparql query editor
B) It is not working as it sends the SPARQL Query to `http://localhost:8080/sparql`
Answers:
username_1: Not sure what you are talking about, on data.admin.ch it says the web interface is at /sparql while the endpoint is /query. YASGUI indeed doesn't have the correct config yet, need to figure out how to properly do that with the npm install of YASGUI.
Status: Issue closed
|
swdotcom/swdc-vscode | 555717878 | Title: Test bug report
Question:
username_0: **Describe the bug**
A clear and concise description of what the bug is.
**Steps to reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Desktop (please complete the following information):**
- OS: [e.g. macOS]
- Plugin version [e.g. Code Time v1.2.1]
**Additional context**
Add any other context about the problem here.<issue_closed>
Status: Issue closed |
stellar/laboratory | 267338986 | Title: invalid base fee in transaction builder triggers a cryptic error message
Question:
username_0: Base fee triggers an error `- "value" argument is out of bounds ` instead of saying that the fee is invalid.
See https://www.stellar.org/laboratory/#txbuilder?params=<KEY>&network=public<issue_closed>
Status: Issue closed |
frappe/frappe | 654107394 | Title: Filters translation bug for Prepared Report
Question:
username_0: <!--
Welcome to the Frappe Framework issue tracker! Before creating an issue, please heed the following:
1. This tracker should only be used to report bugs and request features / enhancements to Frappe
- For questions and general support, use https://stackoverflow.com/questions/tagged/frappe
- For documentation issues, refer to https://frappe.io/docs/user/en or the developer cheetsheet https://github.com/frappe/frappe/wiki/Developer-Cheatsheet
2. Use the search function before creating a new issue. Duplicates will be closed and directed to
the original discussion.
3. When making a bug report, make sure you provide all required information. The easier it is for
maintainers to reproduce, the faster it'll be fixed.
4. If you think you know what the reason for the bug is, share it with us. Maybe put in a PR 😉
-->
## Description of the issue
Using Prepared Report with language different from English, the translated filters parameters was transfered directly to report. It not happen with normal direct report
## Context information (for bug reports)
**Output of `bench version`**
```
(paste here)
```
## Steps to reproduce the issue
For example, General Ledger report in ERPNext
1. Change language of user to other languages like Vietnamese
2. Access General Ledger report in Accounting module
3. Choose Group by: Group by Voucher (Consolidated)
4. Press Generated Report (only happen when have many GL Entry and system switch to Prepared Report)
5. The result of report is false, even we try another group by, the result is the same.
My suspsect for the reason of this bug:
- If the system run report directly, this bug will not happen because the system will translate the right language
- However, if run in Prepared Report, the system put the report running to enqueue, however, enqueue using the default language system is English, not recoginize the original language of user, that why the parameter transfer to report python is wrong.
### Observed result
### Expected result
### Stacktrace / full error message
```
(paste here)
```
## Additional information
OS version / distribution, `Frappe` install method, etc.
Answers:
username_1: this still exist tell today
username_1: i will try to make a pr to fix it |
sul-dlss/sul-requests | 578170473 | Title: Duplicate headings on messages dashboard
Question:
username_0: <img width="527" alt="Screen Shot 2020-03-09 at 1 26 16 PM" src="https://user-images.githubusercontent.com/96776/76254403-a1fa8900-6209-11ea-92ce-2f6d28022784.png">
Both `Add message` buttons go to the same place in this case. Not sure if this is intentional or not (one of them will turn to an `Edit message` button when there is a message), but it can be a bit confusing why there are two things that do the same thing in this case.
Status: Issue closed
Answers:
username_0: Closed by #936 |
siddarth/zoo | 1158997797 | Title: Tests run against legacy Goji
Question:
username_0: The latest version of Goji has it's own GitHub subdomain (https://github.com/goji/goji), but I believe this still runs against https://github.com/zenazn/goji . Would you find any value in either getting a PR against this repo to update to the new repo, or me hosting a slightly altered version of this code at my home against the new repo?<issue_closed>
Status: Issue closed |
threeML/astromodels | 713996308 | Title: Issue with LogL values changing for different parameter values using JointLikelihood
Question:
username_0: I tried to develop a morphology lookup table very similar to `template_model.py`. In my case, I adapted the template model to read in FITS files to build a data frame of flux values that is multi-index by morphology parameters and information of energy, RA, and Dec. I first interpolate over the morphology parameters and then over energy, RA, and dec. The full program is here:
My problem comes to running the fit for my model. When I try to make a fit using `JointLikelihood`, the parameters seem to be sent to the fitter, but the `logL` values do not change
The first interpolation is done in this function
```
def _prepare_interpolators(self):
#Figure out the shape of the data matrices
#para_shape = [x.shape[0] for x in list(self._parameters_grids.values())]
para_shape = map(lambda x: len(x), list(self._parameters_grids.values()))
#interpolate over the parameters
self._interpolators = []
for e in self._E:
for b in self._B:
for l in self._L:
this_data = np.array(self._data_frame[tuple([e, l, b])].values.reshape(*para_shape), dtype=float)
#ti = GridInterpolator(self._parameters_grids.values(), this_data, bounds_error=False, fill_value=0.0 )
ti = GridInterpolator(tuple(map(lambda x: np.array(x), list(self._parameters_grids.values()))),
this_data, bounds_error=False, fill_value=0.0)
self._interpolators.append(ti)
```
the parameter values are to evaluated to give an interpolated map and be used for interpolation over energy, RA, and dec
```
def _interpolate(self, parameter_values):
#create an interpolated map
self._interpolated_map = np.array(map(lambda j: self._interpolators[j](np.atleast_1d(parameter_values)),
range(len(self._interpolators))))
#figure out the shape of the map
map_shape = map(lambda x: len(x), self._map_parameters_grids.values())
#interpolate over map energy, ra, and dec
self._interpolator = GridInterpolator((self._E, self._L, self._B),
self._interpolated_map.reshape(*map_shape), bounds_error=False, fill_value=0.0)
```
the function _interpolate gets called inside the evaluate function, and the self._interpolator should evalute the different values of energy, ra, and dec
```
def evaluate(self, x, y, z, K, hash, lon0, lat0, *args):
lon = x
lat = y
angsep = angular_distance_fast(lon0, lat0, lon, lat)
#transform energy from keV to MeV
#galprop likes MeV, 3ML likes keV
convert_val = np.log10((u.MeV.to('keV')/u.keV).value)
#if only one energy is passed, make sure we can iterate just once
if not isinstance(z, np.ndarray):
[Truncated]
trial values: 3.5187e+26 -> logL = -11079569737.757
trial values: 3.3807e+26 -> logL = -11079569737.757
trial values: 4.0881e+26 -> logL = -11079569737.757
trial values: 2.7351e+26 -> logL = -11079569737.757
trial values: 4.3404e+26 -> logL = -11079569737.757
trial values: 2.395e+26 -> logL = -11079569737.757
trial values: 3.45e+26 -> logL = -11079569737.757
trial values: 3.45e+26 -> logL = -11079569737.757
trial values: 3.45e+26 -> logL = -11079569737.757
trial values: 3.45e+26 -> logL = -11079569737.757
trial values: 3.45e+26 -> logL = -11079569737.757
trial values: 3.45e+26 -> logL = -11079569737.757
trial values: 3.4501e+26 -> logL = -11079569737.757
trial values: 3.45e+26 -> logL = -11079569737.757
trial values: 3.4503e+26 -> logL = -11079569737.757
trial values: 3.4497e+26 -> logL = -11079569737.757
trial values: 3.453e+26 -> logL = -11079569737.757
trial values: 3.4471e+26 -> logL = -11079569737.757
trial values: 3.4798e+26 -> logL = -11079569737.757
```
Answers:
username_1: Hi Ramiro, I think it would be helpful if you included the full function definition here.
username_1: Btw, were you able to verify in the mean time that your function does what you want it to do? I would start with that before trying to fit anything
First step could be to just verify the function values (define a source with this morphology, set the parameters, print the values at a few different energies/positions, change the parameter(s), print again etc).
Next step could be to make model maps for a few different values of your parameter(s). |
banzaicloud/bank-vaults | 934964874 | Title: [vault-env] Inline mutation doesn't use cache
Question:
username_0: **Describe the bug**:
[vault-env] Inline mutation doesn't use cache
- name: DATABASE
value: "postgres://${vault:database/production/creds/user#username}:${vault:database/production/creds/user#password}@${some_ip}:5432/database"
Having configuration like this vault-env does not use cache and makes two reads two from vault. In this case process receives wrong credentials
**Expected behaviour**:
Vault-env uses cache and does one read from vault.
**Steps to reproduce the bug**:
- name: DATABASE
value: "postgres://${vault:database/production/creds/user#username}:${vault:database/production/creds/user#password}@${some_ip}:5432/database"
**Additional context**:
Add any other context about the problem here.
/kind bug
Status: Issue closed
Answers:
username_1: hello,
i have a same bug. please how did you fixed it ? |
go-gitea/gitea | 560891116 | Title: support Github-like CODEOWNERS?
Question:
username_0: We are a team of ~15 engineers using gitea for daily development.
What is the attitude of this project towards supporting the github-like CODEOWNERS system or something like it?
To be honest, I'd find this functionality very useful.
Answers:
username_1: I think this is nice feature but it would require Gitea to have request review functionality first for this
username_2: Doesn't `request review` feature depends on this one oppositely? When a pull request submitted, we need find the code owners and send review requests.
username_3: Do you have any update or plan when to implement this feature?
We want to use [renovate](https://github.com/renovatebot/renovate) to make dependency updates, but Gitea doesn't support to add reviewers for PR through API at the moment (or I didn't find it, but renovate doesn't implemenet this feature yet).
username_4: Still doing, after #12039 and #11355 finished, I will try this feature, add some useful links about this issue.
github: https://docs.github.com/en/github/creating-cloning-and-archiving-repositories/about-code-owners
gitlab: https://gitlab.com/help/user/project/code_owners
username_5: Alright, seems that both PRs are merged. Any news on that? This feature would very nice!
username_6: Added bounty! [](https://www.bountysource.com/issues/88138532-support-github-like-codeowners) |
Norwegian-marine-datacentre/mareano | 21370187 | Title: rekkefølge tegnforklaring
Question:
username_0: Rekkefølgen på tegnforklaring stemmer ikke med rekkefølgen på kartlag på kartbildet.
Answers:
username_0: Lagt til at legend grafikk blir tegnet i samme rekkefølge som for kartbilder. Både når man slår på kartbilder og når man laster siden første gangen (mareano-stasjoner). Viss man etterpå forandrer rekkefølgen under "kartlag" i utvidet kartpanel så vil ikke legends grafikken forandre rekkefølge.
Status: Issue closed
|
facelessuser/pymdown-extensions | 419221407 | Title: snippets not evaluating markdown extra data
Question:
username_0: I'm using mkdocs-markdownextradata-plugin==0.0.5 to reference extra values in my markdown. When I use snippets to include another file I'm noticing that the evaluation of my extra data stopped working as intended. Any ideas on how to workaround?
Answers:
username_0: Further analysis: Other types of extensions are being evaluated, such as superfences and details.
username_1: Do you have an example for reproduction?
username_1: Also, I don't know what mkdocs-markdownextradata-plugin is.
username_0: [repro.zip](https://github.com/username_1/pymdown-extensions/files/2950048/repro.zip)
That is a repro of the issue.
username_0: This the plugin: https://pypi.org/project/mkdocs-markdownextradata-plugin/
username_1: Can you also explain more how your variables stopped working? I don't yet have a clear understanding of what you expect to happen vs what actually happens. Are you saying variable replacement doesn't work everywhere? Just in snippet content? Please provide specific details.
username_0: Sorry, to be clear the variables work normally everywhere except within the
snippet. If you give my repro a run I show a simple example of it rendering
the variable as it should and a snippet with the very same variable not
rendering.
It seems like a necessary pass isn't running within the snippet extension
processor from what I can tell.
Did the repro zip help?
username_1: From what you describe, it sounds like the variables are applied before the snippet is. I suspect they try to run their as early as possible, while we try to run early as well. Sounds like they are running first.
username_1: Ah, theirs is a mkdocs plugin. It is run before the markdown parser. I can never run before that. You'd need a mkdocs plugin that basically does the same thing as our snippet extension and have it run before the mkdocs plugin.
Status: Issue closed
username_0: FWIW for other readers passing by:
https://github.com/mkdocs/mkdocs/issues/304#issuecomment-460058741
There is a workaround.
username_1: Glad you got something working! |
suharev7/clickhouse-rs | 912793766 | Title: `GetHandle` is not a future
Question:
username_0: Is there anyway to solve this issue ?
35 | let mut client = pool.get_handle().await?;
| ^^^^^^^^^^^^^^^^^^^^^^^ `GetHandle` is not a future
|
= help: the trait `futures_util::Future` is not implemented for `GetHandle`
= note: required by `futures_util::Future::poll`
Answers:
username_1: Which versions of `tokio-rs` and `clickhouse-rs` do you use?
username_0: thank you
[dependencies]
clickhouse-rs = "^0.1.21"
tokio = {"version"= "^1.5", features=["full"]}
futures-util = "^0.3.12"
env_logger = "0.8.3"
username_0: I changed to below and worked. thank you
clickhouse-rs = "^1.0.0-alpha.1"
tokio = {"version"= "^1.5", features=["full"]}
futures-util = "^0.3.12"
env_logger = "0.8.3"
Status: Issue closed
username_0: issue resolved |
oscarngncc/COMP4111_project | 623908252 | Title: % character is not escaped in book lookup
Question:
username_0: **Describe the bug**
`%` character is not escaped in LIKE statements in book searching criteria.
**To Reproduce**
1. Login as usual
2. Add a book that title does not contains `%`
```json
{
"Title": "# Book!",
"Author": "Author",
"Publisher": "Publisher",
"Year": "2028"
}
```
3. Search book containing title `%`
`GET localhost:8080/BookManagementService/books?token={{token}}&title=%25`
**Expected behavior**
204 No Content
**What actually happens**
200 OK with the book returned
```json
{
"Title": "# Book!",
"Author": "Author",
"Publisher": "Publisher",
"Year": "2028"
}
``` |
krevis/MIDIApps | 433328 | Title: Need to be able to listen to a subset of available MIDI sources
Question:
username_0: Sometimes people have MIDI devices which send sysex at the wrong time, interfering with the sysex they really want to receive. Should add a way to choose which ports to listen to.
Answers:
username_0: This also came up with a specific MIDI interface (Retrokits RK-006) which has two midi input ports (IN_1 and IN_2) and an auto-merged input port (IN_ALL).
Workaround: use MIDI Monitor and
1. choose only the input you want
2. receive a sysex message
3. Double click it to see the details (opens a separate window)
4. Press the “Save As…” button and save it as “whatever.syx”
5. Then add the file to SysEx Librarian or do whatever you like with it
But if you want to record multiple sysex messages into one file, it's pretty awkward. |
railslink/railslink | 367506071 | Title: Create a users/channels admin interface
Question:
username_0: There isn't a users or channels admin interface at all. not that we should allow changing stuff, but it might be nice to flesh out an index and/or let admins search and stuff in preparation for future things
Answers:
username_1: A lot of the view code for approving membership submissions can probably be co-opted pretty easily for the user admin interface.
username_2: I'm not clear on what kinds of actions we're able to perform through the APIs yet, but just having a nice way to browse the data read-only might be handy. Eventually we'll probably start looking for ways to do management tasks beyond approving join requests, maybe look up inactives for removal, search profile details, etc.
username_1: For now I would focus on the index screen, searching, and a show/edit screen that's non-functional. Like the pending memberships. It's all there, but you can't change any of it.
I don't think it's worth trying to re-implement the things we could do in Slack like set a channel topic, but this lays the groundwork for future outside-of-slack things we may want to do. User profiles, some nifty graphs by channel, etc.
username_1: This is mostly all done via 1de1fb8a300219b06fdf890826f0c96b0e3a3eb1
Closing it and we can open new tickets for additional functionality.
Status: Issue closed
|
Azure/go-autorest | 868975947 | Title: [Event Grid Key Authorizer] Support aeg-sas-token for authentication
Question:
username_0: Azure Event Grid supports authentication via SAS token using aeg-sas-token header. See https://docs.microsoft.com/en-us/azure/event-grid/security-authentication for more details. But the EventGridKeyAuthorizer recognises only topicKey and sets only aeg-sas-key header here - https://github.com/Azure/go-autorest/blob/master/autorest/authorization.go#L266
Answers:
username_0: @username_1 Can I get an update on this Issue ?
username_1: I believe this is the correct link.
https://docs.microsoft.com/en-us/azure/event-grid/authenticate-with-access-keys-shared-access-signatures
We are no longer adding new features to our track 1 SDKs. However, you should be able to implement this yourself. All that's required is to create a type that satisfies the `autorest.Authorizer` interface. I took a stab at an initial implementation [here](https://gist.github.com/username_1/6647e7fb4c7fff8f228d076795d93aa5).
Status: Issue closed
|
jbake-org/jbake | 309948035 | Title: Option to generate relative image links in output
Question:
username_0: I want to generate an output site file tree that has _relative_ links to the images. I don't want to use server mode -- I want an _accurate_ preview loading from `file://` URLs. I also want the option to post to a staging and production server (using Gitlab CI) so the image links shouldn't be hardcoded to the production server.
But the images are pulled from the previously published version of the site running on `site.host`.
According to the documentation `site.host` should contain the URL of the server and is overridden for server mode. But there appears to be no (easy) way to request the generation of relative image links in the output.
Answers:
username_1: Thanks for raising this issue.
With 2.6.0, the image URL generation has changed (you may have already seen [this](https://jbake.org/docs/2.6.0/#images_path)). This change skips the images where URL explicitly starts with http:// or https://. I think the initial version of change also excluded modifying image URLs starting with `/`. I wonder if that would have helped to achieve this setup.
Were you able to achieve all this with 2.5.1 where image path's are not modified?
Running in server mode could be used while testing site in local.
username_0: Gradle plugin doesn't support server mode: https://github.com/jbake-org/jbake-gradle-plugin/issues/32
username_0: Thanks @ancho, I may try that.
username_0: How hard would this be to fix? Is it something that could be easily submitted as a PR?
username_0: @username_2 It seems to me a static site generator should be able to generate fully static sites that:
1. Are entirely self-contained
2. Can be opened and navigated by opening `index.html` in any root directory on the file system
3. Can be deployed to multiple servers on multiple URLs without any modification.
Any suggestions on what we can do got get his fixed? I'm willing to help if given some guidance and this is not too difficult.
Status: Issue closed
|
openjs-foundation/cross-project-council | 510168500 | Title: [ONBOARDING] Appium
Question:
username_0: ## Project Onboarding Checklist - Appium
<!--Copied from ../../PROJECT_PROGRESSION.md on 2019-08-23 -->
**Stage: Impact**
- [ ] Adopt the OpenJS Foundation Code of Conduct
- [ ] Update project CoC reporting methods to include OpenJS Foundation escalation path
- [ ] Transfer official domains to OpenJS Foundation
- [ ] Identify and document other core project infrastructure
- [ ] If choosing to use a Contributor License Agreement (CLA) or Developer Certificate of Origin (DCO), make selection and implement appropriate tool
- [ ] Add or Update Governance.md document (required for Impact stage)
- [ ] Confirm required files in place (CODE_OF_CONDUCT.md, LICENSE.md)
- [ ] Project Charter is published on website or github
- [ ] Update legal copyright notice on project website and github
- [ ] Add OpenJS Foundation logo to project website
- [ ] Add Project logo to OpenJS Foundation website; update PROJECTS.md file
- [ ] Transfer logomark to the OpenJS Foundation
- [ ] If project is using crowdfunding platforms, add disclaimer to platforms
- [ ] Identify individuals from the project to join the CPC
- [ ] Document project and foundation contacts for:
* marketing & social media
* infrastructure
* legal/governance help
Answers:
username_1: No crowdfunding used by Appium, so that's checked-off. |
riboseinc/rspec-pgp_matchers | 480744106 | Title: Check for a passphrase prompt
Question:
username_0: https://github.com/riboseinc/rspec-pgp_matchers/blob/89a02228550fef73ea378b39827dd8fd84624d05/lib/rspec/pgp_matchers/gpg_runner.rb#L16
Is this checking for a passphrase prompt? I'm trying to implement this gem, but it's failing to recognize a legitimate PGP message. I'm wondering if it's because my CLI is prompting me to enter a passphrase, and that doesn't match this checker. Thanks for clarification.
Answers:
username_1: @username_0 thanks for reporting this! @username_2 could you help check this?
username_2: @username_0 Yes, this is because of decryption key is password-protected. As a general advice, I recommend not to use such keys in tests. It saves a lot of hassle.
I am thinking about some support for that. I am against implementing it in gem's API (it would become very complicated when given message is encrypted for many recipients), however I have several other ideas already. Possible solutions include [presetting keys](https://wincent.com/wiki/Using_gpg-agent_on_OS_X), allow customization of GnuPG to be executed (which would effectively allow passing passwords via CLI arguments), or even implementing a custom pinentry.
username_2: @username_0 See https://github.com/riboseinc/rspec-pgp_matchers/blob/master/PROTECTED_KEYS.adoc.
Status: Issue closed
username_2: It is unlikely that this gem will be usable for anything but testing anytime soon. This is because it relies on parsing human-readable output from GnuPG, which is subject to changes and error-prone in general. There is the issue #16 all about migrating to machine-readable GnuPG output format, however this is going to be difficult and not going to be done soon. Main obstacles are that the machine-readable output is missing some information which has to be extracted in some other way, and that the format is poorly documented. |
cuicorey/FINA4350-GroupProject | 802724401 | Title: Type Error
Question:
username_0: Hey there,
Thank you for the codes. However, as I run the last part which is:
***
with torch.set_grad_enabled(False):
outputs = model(input_ids, token_type_ids, attention_mask)
outputs = F.softmax(outputs,dim=1)
print("s" + str(i), 'FinBert score: ', labels[torch.argmax(outputs).item()])
***
I repeatedly get the following error to which I couldn't find any solutions: "TypeError: 'NoneType' object is not callable".
Would you please provide me with some insight as to how i can tackle this?
Best, |
bioconda/bioconda-recipes | 364669479 | Title: Canu misssing canu.unitigs.bed
Question:
username_0: Does anyone get `canu.unitigs.bed` unsing bioconda's recipe?
Thank you in advance.
Michal
Answers:
username_1: I expect this is related to #11509, you don't really have 1.7.1. If you clean up any existing Canu install and start from scratch (or install to a different location) it would work. |
arximboldi/immer | 225296974 | Title: Restrictive licensing
Question:
username_0: Nobody can actually use this in non-GPL-compatible codebases, and it makes no sense for a library to be GPL'd.
Answers:
username_1: This, along with the chosen license, suggests that the author may want to be remunerated for commercial uses of the library. This seems reasonable; do YOU work for free?
Status: Issue closed
username_0: @username_1 Yes, I don't expect to be paid for every single thing I write, that would be unreasonable. Indeed the majority of libraries like this are actually free to use in most projects.
@username_2 Thank you, that answers it. Hopefully you'll get picked up by someone willing to sponsor it.
username_2: FYI, the license has been changed to LGPLv3+ |
android/media-samples | 497868630 | Title: Picture in Picture mode issue
Question:
username_0: I am trying to implement Picture in Picture mode independent from activities. For now i have pip functionality but its activity dependent.so i want to make it activity independent and it should continue for all over the app once this available for devices. |
andrewrk/libsoundio | 302119533 | Title: Should this be officially archived/abandoned?
Question:
username_0: Consider the number of unaddressed issues and the age of the `v2` branch. Has everyone involved moved on? The `v2` branch is the only one that supports my output hardware, but I wonder if JUCE's or QT's licenses would allow someone to look at how they handle more complex devices and continue making basic progress on something like this?
Answers:
username_1: no
Status: Issue closed
username_0: Yay!
What about the second question?
username_1: I don't know the answer to that question.
I understand that I've not been on top of maintenance work on this library, and I won't be for a while longer yet. But you can bet your bottom dollar this is not abandoned. Here's my plan:
* zig 0.2.0 release
* rewrite groove basin music player in zig
- this version will depend on libsoundio instead of SDL
- libsoundio will get rewritten in zig, with a C API exposed
- this actually makes it no longer depend on libc on windows, better for static executables.
* continue work on my DAW using libsoundio and zig
it's a bit of a roundabout journey, and the folks who want a single .c file to drop in their MSVC projects won't be happy.
but that's how it's gonna happen. those who want to be on board, great. those who don't, :sunglasses: bye! |
cltl/pepper | 654735359 | Title: Pose recognition
Question:
username_0: Use pose recognition from:
- https://github.com/tensorflow/tfjs-models/tree/master/posenet
- https://github.com/deephdc/posenet-tf
Available as pre-trained model: https://github.com/tensorflow/tfjs-models/tree/master/posenet#loading-a-pre-trained-posenet-model
Available as Docker: https://github.com/deephdc/posenet-tf#docker-installation
Answers:
username_1:  [Identify people by voice](https://trello.com/c/BD2G4ORM/1-identify-people-by-voice) |
kevinspider/kevinspider.github.io | 840622202 | Title: 逆向环境搭建 | 凡墙总是门
Question:
username_0: https://username_0.github.io/androidenv/
google 系统镜像 推荐8.1.0 OPM7.181205.001 下载地址: https://developers.google.com/android/images 国内镜像: http://aosp.opersys.com/ sdk 工具包 包含 adb, fastboot, systrace 等 下载地址: https://developer.android.com/studio/r |
juanifioren/django-oidc-provider | 337682822 | Title: Token Model crashes in Django Admin when a token has been given to a Client using Client Credentials
Question:
username_0: When a client request a token using client credentials, this is token is created with no user id associated, so in models.py, on line 152, when it tries to access self.user.email, it crashes because there is no associated user. It can be easily fixed with an if, to check first. For example:
`def __str__(self):
return u'{0} - {1}'.format(self.client, self.user.email if self.user else self.client.contact_email)`
Status: Issue closed
Answers:
username_1: Fix d825061508b485310c601844b6ecbe47f7c9721d
Thanks for reporting @username_0 |
brotherlogic/versiontracker | 811367917 | Title: Error for /versiontracker.VersionTrackerService/NewVersion
Question:
username_0: zero2 [fe9cfa01b5db5a39553ab2b007b2c1e3]: 51 calls 41 errors (rpc error: code = Unknown desc = Error waiting on copy: argon:/media/scratch/buildserver/builds/github.com/username_0/provisioner/provisioner-8e9e0427dd06f23739e88a300fc68415, /home/simon/gobuild/bin/provisioner.new -> exit status 1 (/home/simon/gobuild/bin/provisioner.new: set times: No such file or directory))<issue_closed>
Status: Issue closed |
dask/dask | 1071029276 | Title: Task stealing regression in 2021-11-0+ (preventing task load balancing)
Question:
username_0: <!-- Please include a self-contained copy-pastable example that generates the issue if possible.
Please be concise with code posted. See guidelines below on how to provide a good bug report:
- Craft Minimal Bug Reports http://matthewrocklin.com/blog/work/2018/02/28/minimal-bug-reports
- Minimal Complete Verifiable Examples https://stackoverflow.com/help/mcve
Bug reports that follow these guidelines are easier to diagnose, and so are often handled much more quickly.
-->
**What happened**: There seems to be an issue with the task scheduling. When the cluster starts, and the first worker grabs the first task, it appears to block the rest of the workers until it completes the first task. This is particularly an issue when large clusters are spun in the cloud, as the scheduler start to send tasks when not all the workers are ready. This minimum working example below also demonstrate how one of the workers appears to keep all the tasks assigned and only distribute them as needed throughout the length of the computation.
**What you expected to happen**: Workers should start processing tasks immediatly and be able to steal tasks from the 1st worker. This is appears to be a regresssion from pre dask-2021-11-0 (I confirmed for sure it works as expected in 2021-9-2)
**Minimal Complete Verifiable Example**:
```python
import numpy as np
import time
import dask.bag as db
def slow_function(input):
time.sleep(30)
return input
bag = db.from_sequence(np.random.rand(1000, 1), npartitions=1000)
bag.map(slow_function).compute()
```
**Anything else we need to know?**:
This is a minimized example of an issue our team ran into at scale on Coiled (cc @gjoseph92)
**Environment**:
- Dask version: 2021.11.2
- Python version: 3.9.7
- Operating System: Linux / Coiled
- Install method (conda, pip, source): conda |
renpy/renpy | 774910498 | Title: Bug: Container doesn't pass transform events to its children
Question:
username_0: Simple example where its most noticeable. ScreenDisplayable passes transform_event to children, bug Fixed does not.
```
image sample = Solid("#0ff", xysize=(100, 100))
transform alpha5():
on show:
alpha 0.0
linear 0.5 alpha 1.0
on hide:
alpha 1.0
linear 0.5 alpha 0.0
screen working():
add "sample" align (0.5, 0.5) at alpha5
screen broken():
fixed:
fit_first True
add "sample" align (0.5, 0.5) at alpha5
label start:
pause
show screen working
"working"
hide screen working
pause
show screen broken
"broken"
hide screen broken
pause
return
```
Answers:
username_1: Not a bug. Transform events don't propagate like this. At best this is a feature enhancement, but I need to work out if this is something that we'd want.
Status: Issue closed
username_1: Simple example where its most noticeable. ScreenDisplayable passes transform_event to children, bug Fixed does not.
```
image sample = Solid("#0ff", xysize=(100, 100))
transform alpha5():
on show:
alpha 0.0
linear 0.5 alpha 1.0
on hide:
alpha 1.0
linear 0.5 alpha 0.0
screen working():
add "sample" align (0.5, 0.5) at alpha5
screen broken():
fixed:
fit_first True
align (0.5, 0.5)
add "sample" at alpha5
label start:
pause
show screen working
"working"
hide screen working
pause
show screen broken
"broken"
hide screen broken
pause
return
```
username_0: Sure. Another place it's apply to - the layer at of layered image. |
dotnet/arcade | 997056103 | Title: Make sure XHarness telemetry is sent even when job times out
Question:
username_0: ### Context
When we call user commands inside of the Helix SDK, we have to make sure that when they time out, we still send the telemetry at the end of the job (we do it via `HelixPostCommands`).
### To be done
- Investigate behavior of `HelixPostCommands` and whether they get called even when command times out
- Make sure telemetry is sent, i.e. time-constrain user commands and leave buffer to send the telemetry at the end<issue_closed>
Status: Issue closed |
nining377/dolby_beta | 1027335215 | Title: 断网状态下软件闪退
Question:
username_0: 杜比大喇叭beta3.1.1
lsp1.6.1
miui12.5
安卓11
网易云8.0.3 32位

断网进入 到开屏界面卡几秒然后直接闪退
什么问题,,<issue_closed>
Status: Issue closed |
jingw/pyhdfs | 842553293 | Title: Help me,please . The second run of the function in the script results in an abnormal result
Question:
username_0: I am a rookie~~!!
The following code:
`list_info = [{"tenant": "coco", "hive_path": "/user/open_001_dev", "ftp_path": "/files/prov/001"},
{"tenant": "lili", "hive_path": "/user/open_002_dev", "ftp_path": "/files/prov/002"}]
result = 0
client=pyhdfs.HdfsClient(hosts="10.173.5.18:9000",user_name="hdfs",timeout=10,max_tries=3,randomize_hosts="false")
def hive_content_size():
global result
for item in range(2):
if "hive_path" in list_info[item]:
print(client.get_content_summary(list_info[item]["hive_path"]))
hive_content_size()`
The result of the first loop is output normally,but the output of the second loop is abnormal.
The bottom is the error report:
`ContentSummary(directoryCount=1258, fileCount=3773, length=141829751002, quota=4000000, spaceConsumed=425489253006, spaceQuota=659706976665600)
Failed to reach to 10.173.5.18:9000 (attempt 3/3)
Traceback (most recent call last):
File "/usr/local/python/lib/python3.9/site-packages/urllib3-1.26.4-py3.9.egg/urllib3/connectionpool.py", line 445, in _make_request
six.raise_from(e, None)
File "<string>", line 3, in raise_from
File "/usr/local/python/lib/python3.9/site-packages/urllib3-1.26.4-py3.9.egg/urllib3/connectionpool.py", line 440, in _make_request
httplib_response = conn.getresponse()
File "/usr/local/python/lib/python3.9/http/client.py", line 1347, in getresponse
response.begin()
File "/usr/local/python/lib/python3.9/http/client.py", line 307, in begin
version, status, reason = self._read_status()
File "/usr/local/python/lib/python3.9/http/client.py", line 268, in _read_status
line = str(self.fp.readline(_MAXLINE + 1), "iso-8859-1")
File "/usr/local/python/lib/python3.9/socket.py", line 704, in readinto
return self._sock.recv_into(b)
socket.timeout: timed out
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/python/lib/python3.9/site-packages/requests-2.25.1-py3.9.egg/requests/adapters.py", line 439, in send
resp = conn.urlopen(
File "/usr/local/python/lib/python3.9/site-packages/urllib3-1.26.4-py3.9.egg/urllib3/connectionpool.py", line 755, in urlopen
retries = retries.increment(
File "/usr/local/python/lib/python3.9/site-packages/urllib3-1.26.4-py3.9.egg/urllib3/util/retry.py", line 532, in increment
raise six.reraise(type(error), error, _stacktrace)
File "/usr/local/python/lib/python3.9/site-packages/urllib3-1.26.4-py3.9.egg/urllib3/packages/six.py", line 735, in reraise
raise value
File "/usr/local/python/lib/python3.9/site-packages/urllib3-1.26.4-py3.9.egg/urllib3/connectionpool.py", line 699, in urlopen
httplib_response = self._make_request(
File "/usr/local/python/lib/python3.9/site-packages/urllib3-1.26.4-py3.9.egg/urllib3/connectionpool.py", line 447, in _make_request
self._raise_timeout(err=e, url=url, timeout_value=read_timeout)
File "/usr/local/python/lib/python3.9/site-packages/urllib3-1.26.4-py3.9.egg/urllib3/connectionpool.py", line 336, in _raise_timeout
raise ReadTimeoutError(
urllib3.exceptions.ReadTimeoutError: HTTPConnectionPool(host='10.173.5.18', port=9000): Read timed out. (read timeout=10)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/python/lib/python3.9/site-packages/PyHDFS-0.3.1-py3.9.egg/pyhdfs/__init__.py", line 418, in _request
response = self._requests_session.request(
File "/usr/local/python/lib/python3.9/site-packages/requests-2.25.1-py3.9.egg/requests/sessions.py", line 542, in request
resp = self.send(prep, **send_kwargs)
File "/usr/local/python/lib/python3.9/site-packages/requests-2.25.1-py3.9.egg/requests/sessions.py", line 655, in send
r = adapter.send(request, **kwargs)
File "/usr/local/python/lib/python3.9/site-packages/requests-2.25.1-py3.9.egg/requests/adapters.py", line 529, in send
raise ReadTimeout(e, request=request)
requests.exceptions.ReadTimeout: HTTPConnectionPool(host='10.162.3.171', port=19888): Read timed out. (read timeout=10)
Traceback (most recent call last):
File "/home/hadoop/shay/monthly_report/test01.py", line 24, in <module>
print(hive_content_size())
File "/home/hadoop/shay/monthly_report/test01.py", line 22, in hive_content_size
print(client.get_content_summary(list_info[item]["hive_path"]))
File "/usr/local/python/lib/python3.9/site-packages/PyHDFS-0.3.1-py3.9.egg/pyhdfs/__init__.py", line 633, in get_content_summary
File "/usr/local/python/lib/python3.9/site-packages/PyHDFS-0.3.1-py3.9.egg/pyhdfs/__init__.py", line 450, in _get
File "/usr/local/python/lib/python3.9/site-packages/PyHDFS-0.3.1-py3.9.egg/pyhdfs/__init__.py", line 442, in _request
pyhdfs.HdfsNoServerException: Could not use any of the given hosts`
ask for help~~!!!
Status: Issue closed
Answers:
username_1: This part of the error means you set the timeout parameter too low. It's taking HDFS more than 10 seconds to reply to your request. Could you try increasing it?
username_0: Thank you for your reply.
I solved the problem in your way.
( ^_^ ) |
daumPostcode/QnA | 475012780 | Title: 영문 주소 선택 시 영문으로 표기 문의
Question:
username_0: Daum의 우편번호 API를 잘 사용하고 있습니다,
좋은 API 제공에 매우 감사합니다.
현재 서비스중인 페이지가 다국어를 지워해야할 상황에 놓여 문의드립니다.
한글주소 검색후 영문변환은 가능하나, 영문명 클릭 시 한글 값이 들어가는 것을
영문으로 값이 들어가게 할수 있나요?
Answers:
username_1: @username_0
안녕하세요~
공식가이드페이지 (http://postcode.map.daum.net/guide#attributes)
의 속성부분을 보시면, 각 주소정보를 담고 있는 변수
address
jibunAddress
roadAddress
autoJibunAddress
autoRoadAdress
는 모두 쌍으로 English라는 단어를 뒤에 붙히면 (ex. addressEnglish)
영문값을 얻으실 수 있습니다.
이부분 참고 부탁드립니다~
Status: Issue closed
|
pvorb/node-dive | 57372127 | Title: complete function never gets called in special case
Question:
username_0: If the directory parameter passed to dive is not a directory (it comes from user input) then the complete function of dive never gets called... :(
Note: I noticed that you call in this case the first callback with an error object, but still, I would expect the complete function to be called when dive is finished.
Answers:
username_1: That's correct. If you could create a pull request, I'd merge it. At the moment I'm to busy to fix it on my own.
username_0: PR is created. Let me know if you'd like to discuss it before merging.
Status: Issue closed
username_1: Closed. See #13. |
lydiaPenglish/CyChecks2 | 565104180 | Title: Why the cyf_ and cyd_ prefixes?
Question:
username_0: I'm curious why you decided to put cyf_ prefixes on functions and cyd_ prefixes on data. This just seems like extra type. If there was a concern over possible confusion with other functions or data, you can always use CyChecks2::<function> or CyChecks2::<data>.
Answers:
username_1: I'm happy to delete those prefixes, but I'll let @username_2 have the final say.
username_2: It is extra typing, but for me I don't always remember what we called the data (ex was it saldept or deptsals??), so when I type cyd_ it fills in possibilities for me to choose from. Same logic for the functions.
username_0: Ah. I didn't think about autocomplete.
Status: Issue closed
|
LSSTDESC/DC2-production | 480978218 | Title: Cannot load LensPop on cori
Question:
username_0: In today's episode of "never update anything":
After Cori's OS upgrade, loading LensPop from
```
/global/common/software/lsst/cori-haswell-gcc/LensPop/20180203
```
causes the following error
```
Traceback (most recent call last):
File "../../bin.src/generateInstCat.py", line 12, in <module>
import desc.sims.GCRCatSimInterface.validation as ic_valid
File "/global/homes/d/username_0/sims_GCRCatSimInterface_master/python/desc/sims/GCRCatSimInterface/__init__.py", line 9, in <module>
from .TwinklesClasses import *
File "/global/homes/d/username_0/sims_GCRCatSimInterface_master/python/desc/sims/GCRCatSimInterface/TwinklesClasses.py", line 3, in <module>
from desc.twinkles.twinklesVariabilityMixins import VariabilityTwinkles
File "/global/homes/d/username_0/Twinkles/python/desc/twinkles/__init__.py", line 11, in <module>
from .sprinkler import *
File "/global/homes/d/username_0/Twinkles/python/desc/twinkles/sprinkler.py", line 9, in <module>
import om10
File "/global/common/software/lsst/cori-haswell-gcc/OM10/20180203/om10/__init__.py", line 1, in <module>
from .db import *
File "/global/common/software/lsst/cori-haswell-gcc/OM10/20180203/om10/db.py", line 15, in <module>
from lenspop import population_functions, distances
File "/global/common/software/lsst/cori-haswell-gcc/LensPop/20180203/lenspop/__init__.py", line 2, in <module>
from .make_lens_pop import *
File "/global/common/software/lsst/cori-haswell-gcc/LensPop/20180203/lenspop/make_lens_pop.py", line 7, in <module>
matplotlib.use('TkAgg')
File "/opt/lsst/software/stack/python/miniconda3-4.5.12/envs/lsst-scipipe-1172c30/lib/python3.7/site-packages/matplotlib/__init__.py", line 1391, in use
switch_backend(name)
File "/opt/lsst/software/stack/python/miniconda3-4.5.12/envs/lsst-scipipe-1172c30/lib/python3.7/site-packages/matplotlib/pyplot.py", line 222, in switch_backend
newbackend, required_framework, current_framework))
ImportError: Cannot load backend 'TkAgg' which requires the 'tk' interactive framework, as 'headless' is currently running
```
LensPop is required for the sprinkler, which is required for sims_GCRCatSimInterface, which I use for InstanceCatalog, truth catalog, and reference catalog generation. It looks like the offending line has been fixed in @drphilmarshall 's fork (which is where this copy of LensPop was cloned from). Maybe we should just do a `git pull`? Someone with `desc` group user privileges will have to do that, though. @heather999<issue_closed>
Status: Issue closed |
ExtendRealityLtd/VRTK | 1089541130 | Title: [problem] How to solve this problem
Question:
username_0: 
Answers:
username_1: download the latest version of vrtk
(i fixed this in this PR https://github.com/ExtendRealityLtd/VRTK/pull/2137)
Status: Issue closed
|
olifolkerd/tabulator | 315740222 | Title: How to color column headers
Question:
username_0: Can you help me on coloring column headers in tabulator.
Answers:
username_1: Hey @username_0
Tabulator is primarily built to be styled with CSS, what is preventing you from adding some CSS to style the headers?
Cheers
Oli :)
username_0: I want to do background color and color to the column header and group header. so how can i do?

username_2: Adding today class for current month.
```javascript
renderComplete:function(){
var columns = $("#project-groups-plan").tabulator("getColumns");
if (currentYear == todayYear){
var todayColumns = columns.filter(function(obj){
var field = obj.column.field;
return field == "plan_" + todayMonth || field == "fact_" + todayMonth || field == "exp_" + todayMonth;
});
todayColumns.forEach(function(column){
column.getElement().addClass("today");
});
var todayColumn = todayColumns[0].getElement();
var normColumn = todayColumn.parent().siblings(".tabulator-col-content");
normColumn.addClass("today");
var monthColumn = normColumn.parent().parent().siblings(".tabulator-col-content");
monthColumn.addClass("today");
}
},
```

username_1: Hey @username_0
You can also use the ***cssClass*** column definition property to automatically assign a given class to each column:
```js
$("#example-table").tabulator({
columns:[
{title:"Name", field:"name", cssClass:"blue-background"},
]
});
```
Alternatively you can also use the ***tabulator-field*** which is present on all column header and cell elements, that can help you style a given column. in the case of the name field in the above example it would be:
```css
.tabulator .tabulator-header .tabulator-col[tabulator-filed="name"]{
//your styles go here
}
```
There is a full list of all tabulator classes in the [Theming Documentation](http://tabulator.info/docs/3.5#css)
I hope that helps
Cheers
Oli :)
Status: Issue closed
|
voxpupuli/plumbing | 396517839 | Title: Puppet etc. versions shipped with various OSes
Question:
username_0: It's useful to know what versions of various puppet related components ship with which OSes. In many cases AIO packages are available, but not always.
OS | Puppet | Ruby | Facter | Opensource/Free AIO? | Comment
| -| - | - | - | - | - |
EL6 (with EPEL) | ~ | 1.8.7 | 1.6.18 | :+1: | Use AIO |
EL7 (with EPEL) | 3.6.2 | 2.0.0 | 2.4.1 | :+1: | Use AIO |
Fedora 27 | 4.10.1 | 2.4.2 | 2.4.3 | :+1: | EOL since 2018-11-30 |
Fedora 28 | 4.10.10 | 2.5.1 | 3.9.3 | :+1: |
Fedora 29 | 5.5.1 | 2.5.1 | 3.9.3 | :+1: |
Solaris 11.4 | 5.5.0 | 2.3.1 | 2.5.1 | :-1: | 11.3 Only contained puppet 3. Old facter version could be an issue |
Answers:
username_1: I've updated the table and split it the AIO column into exact versions (PC1, P5, P6). Also added Debian.
username_1: Should this be a wiki page?
username_0: @username_1 Yeah maybe. Also the thumbs up/down aren't great, are they?! ;)
username_2: @username_0 Arch also packages Puppet 5, probably until it is end of life: https://www.archlinux.org/packages/community/any/puppet5/
username_1: This could use an update :) |
commons-app/apps-android-commons | 419230842 | Title: Effective UI in night mode
Question:
username_0: **Summary:**
UI of some activity is not effective in night mode
**Steps to reproduce:**
Go to setting. Switch on the night mode
1. Go to Home>Nearby then tab on any location. The bottom row buttons are hard to see in night mode.

2. Backgroundolour of navigation bar header doesn't changes in night mode.

3. Background colour of achievement_activity doesn't change in night mode. Texts are not clearly visible.

4. Background colour of tutorial activities does not change in night mode.

**Device and Android version:**
Realme 2 pro, Android Version 8.1.0, Stock Version from manufacturer
**Commons app version:**
2.10.1 prodDebug master
**Would you like to work on the issue?**
Yes!
Answers:
username_1: There are several issues about night mode already, would you mind removing from this issue the ones that are already covered at existing issues? (see #2180)
Thanks!
username_0: Hello @username_1,
I updated this issue, removing the one same as #2446.
Pull request #2574 fixes both this issue and #2446.
Thank you!
username_2: Fixed in #2574
Status: Issue closed
|
OrleansContrib/OrleansDashboard | 312679653 | Title: Orleans 2.0 cannot find grain implementation class when using Dashboard
Question:
username_0: I'm getting this exception when the Orleans Dashboard is configured with 2.0:

If I comment out `UseDashboard(...)`, it runs just fine. The dashboard seems to be somehow interfering with grain resolution.
Repro project attached.
[Orleans2GettingStarted.zip](https://github.com/OrleansContrib/OrleansDashboard/files/1891725/Orleans2GettingStarted.zip)
Answers:
username_1: Thanks for reporting @username_0, looking into it now...
username_1: Fixed in version 2.0.0 (PR #119)
Status: Issue closed
username_1: ...although I notice that reminders are returning a 500. I'll investigate that. |
pingcap/tidb | 277657436 | Title: Auth failed message show password instead of account.
Question:
username_0: 1. What did you do?
A Flink streaming job with Spring jdbc template to flush data to TiDB.
2. What did you expect to see?
```
Access denied for user 'tf_fuwu_rw'@'zzzzzzzzz' (using password: YES)
```
3. What did you see instead?
```
Access denied for user 'tf_fu201931wu_rw'@'zzzzzzzzz' (using password: YES)
```
4. What version of TiDB are you using (`tidb-server -V`)?
```
Release Version: 1.0.0
Git Commit Hash: 1a175b562bc679694f606c336131a9373064b324
Git Commit Branch: release-1.0
UTC Build Time: 2017-10-19 02:38:02
```
Actually, `tf_fu201931wu_rw` is the password, not the usename, and we did not do anything to this job ( it works for about five days and works well ), but today the worker throws this exception with the wrong message, then we restart this job and this message does not occur.
Answers:
username_1: @username_0 Thanks for your feed back! @username_2 PTAL.
username_2: Maybe something wrong with the String() method.
@username_3 PTAL
username_3: Seems that we cannot easily reproduce this issue:
```
# breezewish @ ~/Work/PingCAP/src/github.com/pingcap/tidb on git:release-1.0 x [19:04:43]
$ mysql -u root --column-type-info --host 127.0.0.1 --port 4000
Welcome to the MySQL monitor. Commands end with ; or \g.
Your MySQL connection id is 1
Server version: 5.7.1-TiDB-v1.0.3-1-ga80e796f MySQL Community Server (Apache License 2.0)
Copyright (c) 2000, 2017, Oracle and/or its affiliates. All rights reserved.
Oracle is a registered trademark of Oracle Corporation and/or its
affiliates. Other names may be trademarks of their respective
owners.
Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.
mysql> CREATE USER 'finley'@'localhost' IDENTIFIED BY 'password';
Query OK, 1 row affected (0.00 sec)
mysql> exit
Bye
# breezewish @ ~/Work/PingCAP/src/github.com/pingcap/tidb on git:release-1.0 x [19:06:27]
$ mysql -u finley --column-type-info --host 127.0.0.1 --port 4000
ERROR 1045 (28000): Access denied for user 'finley'@'127.0.0.1' (using password: YES)
# breezewish @ ~/Work/PingCAP/src/github.com/pingcap/tidb on git:release-1.0 x [19:06:34] C:1
$ mysql -u finley --column-type-info --host 127.0.0.1 --port 4000 -p
Enter password:
ERROR 1045 (28000): Access denied for user 'finley'@'127.0.0.1' (using password: YES)
# breezewish @ ~/Work/PingCAP/src/github.com/pingcap/tidb on git:release-1.0 x [19:06:43] C:1
$ mysql -u finley --column-type-info --host 127.0.0.1 --port 4000 -p
Enter password:
ERROR 1045 (28000): Access denied for user 'finley'@'127.0.0.1' (using password: YES)
# breezewish @ ~/Work/PingCAP/src/github.com/pingcap/tidb on git:release-1.0 x [19:06:46] C:1
$
```
@username_2 any ideas?
Status: Issue closed
username_2: I check the code, never find anywhere we use password as username in the error message. @username_3
Further more, in the handshake protocol, MySQL sends a MD5 hash encrypted password (not user's raw password), TiDB that password use a salt and encryption algorithm.
**TiDB never know or store user's raw password**, let alone return it as error message, so this issue must be a mistake, I'll close it.
If you can provide more detail to verify the issue, feel free to reopen it. @username_0
username_0: I never reproduce this issue these days. leave it as closed.
Thanks all of you. |
thomasloven/lovelace-slider-entity-row | 424085965 | Title: Custom element doesn't exist in Safari
Question:
username_0: And sure enough, it turns out your custom card works in Chrome, but not in Safari, which happens to be my main browser.
I'm not sure if it's feasible to add support for Safari, but perhaps you could consider adding a note to your readme. I can prepare a PR if you like.
Status: Issue closed
Answers:
username_1: I assure you custom elements (and slider-entity-row) works perfectly well in Safari

and have done so since version 10.1 (released 2016-10-24). Mobile Safari as well.
Please read this guide for installation and debugging instructions: https://github.com/username_1/hass-config/wiki/Lovelace-Plugins
username_0: Thank you, Thomas. And sorry, I'm not sure what happened, I swear it worked for me in Chrome but not Safari. Anyway, I tried again now and it works. I'm not aware of doing anything differently. Anyway, I must have done something wrong.
And BTW great job! This plugin is really awesome!
username_1: Eh. It happens. |
ganeti/ganeti | 238306274 | Title: harep manpage double listing of "repair disallowed"
Question:
username_0: Originally reported of Google Code with ID 922.
```
What software version are you running? Please provide the output of "gnt-
cluster --version", "gnt-cluster version", and "hspace --version".
gnt-cluster version
Software version: 2.11.3
Internode protocol: 2110000
Configuration format: 2110000
OS api version: 20
Export interface: 0
VCS version: (ganeti) version v2.11.3
<b>What distribution are you using?</b>
ubuntu 14.04
<b>Please provide any additional information below.</b>
manpage of harep states:
"Harep is able to recognize what state an instance is in (healthy, suspended, needs repair, repair disallowed, pending repair, repair disallowed, repair failed)"
it guess it should be:
"Harep is able to recognize what state an instance is in (healthy, suspended, needs repair, repair disallowed, pending repair, repair failed)"
```
Originally added on 2014-08-22 07:10:20 +0000 UTC.
Answers:
username_0: ```
theres another typo (is vs. it):
"and frequently using a cron job, so that is can actually follow the instance along all the process."
should be
"and frequently using a cron job, so that it can actually follow the instance along all the process."
```
Originally added on 2014-08-22 07:48:34 +0000 UTC.
username_0: ```
-- Empty comment --
```
Originally added on 2014-08-22 09:29:01 +0000 UTC.
Added Labels: Type-Documentation Priority-Medium SmallTask Component-htools
Changed State: Started
Added to Milestone: Release2.11
username_0: ```
-- Empty comment --
```
Originally added on 2014-08-22 09:34:21 +0000 UTC.
username_0: ```
Fixed in
commit 9ce2440d00846604804b958d14aed0cdff530ca6
Author: <NAME> <<EMAIL>>
Date: Fri Aug 22 11:32:46 2014 +0200
Fix typos in the harep man page
See issue #922 for the details.
```
Originally added on 2014-08-26 15:35:50 +0000 UTC.
Changed State: Fixed
username_0: ```
2.11.6 released
```
Originally added on 2014-09-22 15:56:02 +0000 UTC.
Changed State: Released
Status: Issue closed
|
pnp/pnpjs | 401533041 | Title: Bug within SharePoint REST API with managed metadata that affects pnpjs documentation
Question:
username_0: ### Category
- [ ] Enhancement
- [X] Bug
- [ ] Question
- [X] Documentation gap/issue
### Version
Please specify what version of the library you are using: [1.2.8]
Please specify what version(s) of SharePoint you are targeting: [SharePoint Online]
### Expected / Desired Behavior / Question
When updating a list item with multiple taxonomy fields and updating a single taxonomy field within that item with pnpjs documented taxonomy utilities, SharePoint Search should return all taxonomy field data in managed properties.
### Observed Behavior
When updating a list item with multiple taxonomy fields and updating a single taxonomy field with pnpjs documented taxonomy utilities, SharePoint Online search only returns the taxonomy fields that were updated in the update call. In this case the single taxonomy field.
### Steps to Reproduce
1. Create a team site
2. Create a new content type inheriting from item in the team site
3. Add two new Managed Metadata site columns to the created content type (e.g. named TEST_MM_1 and TEST_MM_2) with customized term set. You can use single value managed metadata field.
4. Add terms to the customized term sets via Term Store Management
5. Create a new custom list
6. Configure the list to allow management of content types (Advanced settings - Allow management of content types
7. Add the created content type to the list
8. Add a new item with added content type and fill in both the managed metadata fields
9. Wait for sharepoint search crawl to commence
10. Use e.g. SharePoint Search Query Tool to check that automatically created managed properties owstaxIdTESTMM1 and owstaxIdTESTMM2 have values. Note that with automatically created managed properties the managed properties are case sensitive.
11. Update the created items one managed metadata field as described in [here](https://github.com/pnp/pnpjs/blob/dev/packages/sp-taxonomy/docs/utilities.md).
12. Optional: You can also update the list item title field as documented [here]https://github.com/pnp/pnpjs/blob/dev/packages/sp/docs/items.md#Update. Makes it easier for you to check when sharepoint search has done it's crawling magic.
13. Wait for SharePoint search crawl the changes
14. Check via SharePoint Search Query Tool that you can only get the other owstaxIdTESTMM# managed property value for that item.
### Description of issue
Basically this seems to be an issue on SharePoint Online REST API update method what the taxonomy utility is using. So technically it's not a bug within pnpjs, but the documentation should note that you should not use the update api to update managed metadata fields if you need search functionality.
### Circumvention of issue
1. When doing updates on an item with taxonomy fields, you can use the update API, but you should always update all managed metadata fields in the same update call to keep search intact.
2. Use item.updateValidateUpdateListItem -method, which SharePoint Online modern list forms seem to use internally. This does not cause the aforementioned issue and will also fix a broken item in search to return all taxonomy managed properties.
### Further information to consider
I've opened a service request to notify Microsoft about this bug, but it might take a while as my development tenant is not a premier support tenant. I've managed to reproduce this in 3 different tenants already. I should be able to produce pnp schemas to simplify the creation of a test case if required. Would like to hear if you can reproduce this easily.
I would recommend updating the documentation about these "gotchas" that a SharePoint developer just "has to know" when doing development.
In general the "old-school" REST API's for getting an item and updating an item seem to be problematic when it comes to working with taxonomy fields. Item.get() will not return the managed metadata labels, but for some odd reason the WssId as the label. You can see the same odd behaviour when filtering taxonomy fields in a modern list view.
In my day to day development I try to avoid using item.add, item.get and item.update completely. I've noticed that using list.addValidateUpdateItemUsingPath, list.renderListDataAsStream and item.validateUpdateListItem cause no/less issues.
Answers:
username_0: This also seems to only affect the ows_taxId_TEST_MM_# crawled property, as ows_TEST_MM_# crawled property still seems to work after updating a taxonomy field via update API. If you are e.g. using the taxonomy label as a refiner (RefinableStringXX), you won't notice the issue and Search will get all (updated and not updated) labels correctly.
username_1: Hi @username_0 - thank you for the detailed description of the issue you have found. As you noted this doesn't appear to be an issue with this library, so we don't have any actions to take.
Documenting all the possible "gotchas" in SharePoint development falls well outside of the scope of this library.
You can report this as well to the [sp-dev-docs](https://github.com/sharepoint/sp-dev-docs/issues) issue list which is monitored by the SP engineering team.
Status: Issue closed
|
dotnet/csharplang | 343418825 | Title: [Proposal] if(;) and for(;)
Question:
username_0: `if(; )`:
many times we have to do:
```
var foo=bar();
if(foo!=null) {
//..........
}
```
where the variable foo is not useful outside of the `if` scope.
An intuitive syntax for this is:
```
if(var foo=bar(); foo!=null) { // just like the famous for loop
//..........
}
//same as, (not considering `continue`) :
for(var foo=bar(); foo!=null;) {
//..........
break;
}
```
`/**********************************************************************************/`
`for(; )`:
From time to time we may face the pattern:
```
for(string line = stream.ReadLine(); line != null; line = stream.ReadLine()) {
//..........
}
```
Which can be shorten if with have the syntax `for(_statement_; _condition_) ` :
```
for(string line = stream.ReadLine(); line != null) { // line is a new var in each loop
//..........
}
same as:
while(true) {
string line = stream.ReadLine();
if(!( line != null))break;
//..........
}
```
`/**********************************************************************************/`
Both are basically extended `for(;;)` loop, as a syntax sugar.
Answers:
username_1: Also, with C# 8.0 pattern matching, you will be able to write:
```c#
if (bar() is {} foo) {
//...
}
while (stream.ReadLine() is {} line) {
//...
}
```
The `is {}` syntax is a property pattern with no properties. This syntax is certainly confusing when you see it for the first time, but maybe it's good enough?
username_2: It _might_ work with C# 7 as well in your case if you're dealing with classes and non-`null` as condition:
```c#
if (bar() is var foo) { /* ... */ }
while (stream.ReadLine() is string line) { /* ... */ }
```
(except the fact that the declaration is _after_ the statement instead of _before_)
username_3: Note that `var` patterns will match `null`.
username_2: Oh, right, it should probably be `if (bar() is ReturnTypeOfBar foo)` then, right?
💭 _and I should go review some code I wrote earlier that made it through without blowing up_
username_4: @username_2 Using the mnemonic that `var` means `declare local variable`, I haven't really had any trouble spilling into thinking that it means `<infer type>`.
username_4: @username_2 Using the mnemonic that `var` means `<declare local variable>`, I haven't really had any trouble slipping into thinking that it means `<infer type>`.
username_2: I've commonly used it as "infer the type here and declare a variable", so I might have to change my way of thinking a little. Thanks.
username_5: @username_2 Most people think like that. Is was a completely wrong decision of the design team
username_4: @username_5 Whether or not you agree that the decisions in the design of C# 7.0 were good ones, they have been made. There's no point in stating your unhappiness because it isn't actionable for anyone.
Status: Issue closed
username_0: let me close it =D
username_5: @username_4 I just state that he is not the wrong one. It should not be like this. It counterintuitive from the start in every aspect. And if it really the point that majority of people using C# think it is a bug, not a design decision, then we should fix it now
username_5: @username_4 Actually because we try to say that it has been made. It has been done. It cannot fixed anymore. And so we left this bug as it is. Then there would be people create a subtle bug from non intuitiveness of this feature right now
username_3: @username_5 You not liking it doesn't make it a bug. The language team and many C# devs both understand this decision and agree with it. Sure, maybe it's a little unintuitive, the first time. You learn how it works once. This is true of most new language features.
username_5: @username_3 Right now I have more confident that it not just because I don't like it but it really not intuitive in general. Many C# dev agree with it *after* they have take part in the result. But if someone does not carefully read every aspect, like someone above, they would have misunderstand that aspect because it really a subtle bug, only when the value is `null`
It not intuitive because both `var` and `is` is not new syntax so we have some expectation over it, and the new feature just break this expectation
username_6: Evidence for this claim: "Most people think like that" would be appreciated.
I, personally, don't think that way. But i don't presume me personal beliefs necessarily apply to the majority of users wrt to this facet of the language :) |
pyenv/pyenv | 583725599 | Title: pyenv install 3.7.0 fails.
Question:
username_0: [python-build.20200318130512.18924.log](https://github.com/pyenv/pyenv/files/4349164/python-build.20200318130512.18924.log)
Windows 10
Answers:
username_1: Please check the README and common build problems wiki before creating a new issue, as stated in the issue template.
It is clearly stated in the README that pyenv does not work on windows outside the WSL: https://github.com/pyenv/pyenv#installation
Status: Issue closed
|
weaveworks/weave | 213111449 | Title: Zombie apocalypse with kubernetes and weave-kube
Question:
username_0: Kubernetes:
$ kubectl version --short
Client Version: v1.5.4
Server Version: v1.5.4
Weave:
$ kubectl get ds -n kube-system weave-net -o jsonpath='{.spec.template.spec.containers[*].image}'
weaveworks/weave-kube:1.9.3 weaveworks/weave-npc:1.9.3
Zombies:
$ knife ssh "role:kubernetes_node" "sudo ps aux|grep [l]aunch.sh"|sort
kube01 root 2230 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
kube02 root 7411 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
kube03 root 5007 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
kube04 root 6274 0.0 0.0 0 0 ? Z 19:19 0:00 [launch.sh] <defunct>
...
`uname -a`:
$ knife ssh "role:kubernetes_node" "uname -a"|sort
kube01 Linux kube01 4.4.0-51-generic #72-Ubuntu SMP Thu Nov 24 18:29:54 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
kube02 Linux kube02 4.4.0-51-generic #72-Ubuntu SMP Thu Nov 24 18:29:54 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
kube03 Linux kube03 4.4.0-51-generic #72-Ubuntu SMP Thu Nov 24 18:29:54 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
kube04 Linux kube04 4.4.0-64-generic #85-Ubuntu SMP Mon Feb 20 11:50:30 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux
Maybe it is connected with following issue in kubernetes: https://github.com/kubernetes/kubernetes/issues/39334
We experience this cluster-wide.
Answers:
username_0: I found it!
Attention to node `web05`
I downgrade weave version to one version at the time on this host.
weave-1.9.2:
$ knife ssh "role:kubernetes_node" "sudo ps aux|grep [l]aunch.sh"|sort
web01 root 639 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
web02 root 2863 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
web03 root 27055 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
web05 root 17381 0.0 0.0 0 0 ? Z 21:49 0:00 [launch.sh] <defunct>
weave-1.9.1:
$ knife ssh "role:kubernetes_node" "sudo ps aux|grep [l]aunch.sh"|sort
web01 root 639 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
web02 root 2863 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
web03 root 27055 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
web05 root 22508 0.0 0.0 0 0 ? Z 21:54 0:00 [launch.sh] <defunct>
weave-1.9.0:
$ knife ssh "role:kubernetes_node" "sudo ps aux|grep [l]aunch.sh"|sort
web01 root 639 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
web02 root 2863 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
web03 root 27055 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
web05 root 26266 0.0 0.0 0 0 ? Z 21:57 0:00 [launch.sh] <defunct>
weave-1.8.2:
$ knife ssh "role:kubernetes_node" "sudo ps aux|grep [l]aunch.sh"|sort
web01 root 639 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
web02 root 2863 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
web03 root 27055 0.0 0.0 0 0 ? Z Mar07 0:00 [launch.sh] <defunct>
web05 root 4712 0.0 0.0 1524 1016 ? Ss 22:01 0:00 /bin/sh /home/weave/launch.sh --host=10.83.8.203 --status-addr=10.83.8.203:6782
Tadaaa!
No zombies with 1.8.2.
username_1: Many thanks for reporting this @username_0!
I confirm:
1. I can systematically see `defunct` processes for `launch.sh`;
2. this behaviour seems to have been introduced by [this](https://github.com/weaveworks/weave/commit/bf59a003bbee14438e43163fafb3745b55f1e2cf) change;
3. that although not a proper fix, applying a patch removing `exec` to the `1.9.3` branch removes the symptoms;
4. [we're internally discussing](https://weaveworks.slack.com/archives/net/p1489401047279615) a proper fix/refactoring to resolve this issue.
Status: Issue closed
|
ds4dm/ecole | 1066484493 | Title: Load and solve VRP instances
Question:
username_0: ### Discussed in https://github.com/ds4dm/ecole/discussions/287
<div type='discussions-op-text'>
<sup>Originally posted by **ndrwnaguib** November 23, 2021</sup>
Hi,
I am trying to load [CVRPLIB](http://vrp.galgos.inf.puc-rio.br/index.php/en/) into Ecole. However, I do not think designing it as an instance generator is the way. I view it as a loader instead. I'm not sure what steps I should follow to achieve that, I'd appreciate any pointers.
Off topic: if you believe there would be considerable use for these datasets, I also would like to contribute it to Ecole problem instances.
Thank you. </div>
Need to investigate how to load and solve VRP files. A question I have is whether the problem in the example is fully represented in the SCIP model, or if it has implicit variables (the custom pricer). |
awslabs/aws-embedded-metrics-python | 532848650 | Title: Automatically flush for > 100 metrics per key
Question:
username_0: Currently, if putMetric is called > 100 times, it will fail silently on the backend. We should automatically flush client-side if this limit is hit.
Answers:
username_1: The previous PR resolve the case when the Unqiue Metrics are over 100, but it does not resolve the case, when the Metrics values are over 100. |
jqno/equalsverifier | 104561458 | Title: Suppress - Overloaded: More than one equals method found
Question:
username_0: Hello,
I would like to have possibility to avoid throw *java.lang.AssertionError: Overloaded: More than one equals method found.* when is it makes sense to have overloaded equals method.
Please see my example
```java
import nl.username_1.equalsverifier.EqualsVerifier;
import org.junit.Test;
public class VerifyOverloadFailed {
class Thing {
final int x;
Thing(int x) {
this.x = x;
}
@Override
public int hashCode() {
return x;
}
public boolean equals(Thing other) {
return this.x == other.x;
}
@Override
public boolean equals(Object o) {
return (o instanceof Thing) && (this.equals((Thing) o));
}
}
@Test
public void testEquals() {
EqualsVerifier.forClass(Thing.class).verify();
}
}
```
Thanks!
Answers:
username_1: Thanks for reporting this.
I've actually considered this before, but there's one question I can't really answer, and that is: what does the overload really bring? Java has no generic `Equatable` interface like C# has. Due to type erasure, collections will never call the overload. And even if they did, no matter what you do, you'll always have to write the `Object` overload with the type check and the cast. And if you do, why not have that method handle the case when it's a `Thing`? The only thing I can figure, is that it's a performance thing, but I've never seen any profiler data saying that it actually matters.
I'd be very interested to have an answer to this. If you can convince me, I'll add it to EqualsVerifier :).
username_0: Well, my argument is simple. We have a source code which we can't modify but need to cover it with tests. Without this thing I need to create so much tests manually. So it may be added as an *option* to EqualsVerifier if you in doubt regarding overall Language Design Philosophy.
Thanks!
username_2: I have a class that provides a `static` `equals` method that takes multiple parameters, and EqualsVerifier fails with the same error: `java.lang.AssertionError: Overloaded: More than one equals method found`.
Would it be possible to reopen this issue to allow suppressing _static_ equals methods? |
SpriteLink/NIPAP | 81764329 | Title: Separate LDAP search credentials
Question:
username_0: For some LDAP setups it's required to use separate credentials to perform the LDAP search for the user logging in. The search is performed after authentication to get the full name of the user and permissions (if used).<issue_closed>
Status: Issue closed |
frees-io/freestyle-microservices | 259000093 | Title: Use modules and algebras of frees-io
Question:
username_0: In the branch `ep-7-connect-backend-frontend`, we have a PoC of comunication between backend and clients through websocket.
We are using frees-io but we are not using algebras and modules.
In this issue we will have to move all the business logic to algebras and modules
Answers:
username_0: I have this issue solved in the branch `ep-14-use-frees-io` except a implicit that I am struggling against him. I hope end up this in the first day of the next sprint.
username_0: Move this issue to the frees-io/freestyle-opscenter repository
Status: Issue closed
|
trailofbits/ebpfpub | 766080500 | Title: 揭阳汽车站哪里有真实大保健(找特色服务d
Question:
username_0: 揭阳汽车站哪里有真实大保健(找特色服务(十微781372524) 由蔡洙应执导,董成明、瞿澳晖、陈意涵等主演,上海七贤影视文化传媒有限公司出品的科幻惊悚电影《超能事件》,今日宣布定档月日于爱奇艺独家上映,与此同时官方还发布了“失控版”先导预告以及“超能突发”定档海报。 该片讲述了母亲瘫痪在床家庭贫困的大学生成明,因一次意外获得了能操作物体的超能力。在与好友澳晖经历多次超能力实验后,成明对于如何使用超能力陷入了迷惘;并在好友的怂恿下利用超能力非法牟利、伤害他人。不加节制的两人越玩越过火,逐渐走向了一条不可挽回的毁灭之路,但随着成明母亲过世与过度使用超能力的副作用双重打击下成明幡然醒悟,终于在女友紫琳的鼓励下重新振作,勇敢面对一切并最终回归爱与希望。国内科幻片首次尝试“伪纪录片”拍摄手法千人特效团队打造极致真实感 相信在《超能事件》电影的开始,最先引观众的是它独特的拍摄手法,就像在预告中公开的一样,在镜头中时而是一个帅气的年轻人自拍,时而一变又会变成其他人或自拍或监控的不同角度。据悉,本次电影《超能事件》主创团队使用了全新的拍摄手法——伪纪录片的形式,这也是国内科幻片首次尝试“伪纪录片”的拍摄手法,可以说是又一次全新尝试与视觉升级。 作为第一个吃螃蟹的人,导演蔡洙应曾表示:“为了增加电影的真实性以及能够更好的调动观众们的好奇心,电影中的大多数镜头都应用了相对比较困难的拍摄手法,同时,也是希望观众们能通过这样一个真实还原的镜头,引起他们曾经对超能力幻想的共鸣,并能过足一把超能瘾。”导演蔡洙应在用丰富的镜头语言,在对影像世界进行着无尽探索,同时也会给观众带来更加真实的代入感,而伪纪录片的形式相信在未来的传播中能够为影片助力更多。 而蔡洙应导演作为曾经《狄仁杰之神都龙王》的监制,除去拍摄手法外,特效也是非常值得一提的。本次《超能事件》与其说是一场视觉盛宴,不如说是一场真实的特效秀,片中的各类特效非常完美的融入到了日常的生活之中,让观众近距离体验所有超能力发生的瞬间,打造出身临其境的真实感,这种真实感,正是其背后多达千人的特效团队,不顾日夜不断摸索、打磨,精心制作的结果,相信在电影真正上映之后,一定会让观众高呼精彩与刺激。从千人一面到千人千面“超能”事件背后表达现实主题 近些年来,超能力相关的电影其实非常多,特别是好莱坞各类“超人”、“英雄”电影层出不穷,然而无论是哪种英雄,似乎都在用西方主流的价值观对观众进行洗脑,用简单但粗暴的方式表达着西方世界心目中的真善美,但其实对于亚洲观众来说,在西方的故事体系中,我们虽然能够到感觉到相似性,但却很难找到共鸣感。《超能事件》在故事构建之初便不同于普通“超能”类电影,从预告片来看,他更加重了本土性与落地性。 故事发生在上海的一个普通贫困少年成明身上,同时作为一个不那么“阳光”的主角。成明有自己苦涩——母亲瘫痪在床,全靠自己拼命挣钱,也有自己的局限——性格有些懦弱,甚至有些自卑,他也有不靠谱的朋友,也有追不上的女神。在遇到问题的第一时间,他走了错误的路,甚至造成了难以想象的后果。他就是我们每个人的性格缩影,遇到霸凌会退缩,遇到挑拨会犯错,在电影中的成明更像是另外一个世界的我们,替我们演绎着一个普通的平常人,在遇到挫折时、困难时候挣扎与迷茫。 《超能事件》故事的真实性与落地性的背后正体现了主创团队其对于现实世界的敏锐观察,给予一些普通到不能再普通的人突如其来的变化,究竟会发生什么呢?就如同预告里所给出的疑问一样,当你拥有了超能力,你想要什么?是想要知道世界所有的真相,是拥有能够控制人心的能力,还是利用超能力获取巨大的金钱与权利?通过电影,主创团队想要表达是不仅是小人物性格命运的改变,更想要利用小人物的行动来表现社会背后所隐藏的方方面面。可以说,主创团队利用独特的镜头语言与故事视角为我们描绘了一个千人千面的新世界,展现了对于什么是真正的善与恶的无尽探索。当“超能”主演遇上“超强”内容团队打造顶尖院线影片水准 其实距离《超能事件》的杀青已经过去了将近三年的时间,在这段时间里主创团队还在不断的打磨电影,并希望最终打造出一部院线制作水准的电影,而出品这部电影的上海七贤影视文化也曾表达过对于自己制作电影的初衷:特立独行,不急功,不近利,做有价值、有品格、有气节、有理想的影像创作。从结果来看,《超能事件》可以说就是一部情感张力、影像表现力与现实控诉力三者兼具的优秀影片,然而这样优秀的影片自然离不开敬业的主演和水准一流的制作团队。提及主演,董成明、瞿澳晖、陈意涵三位“超能少年”也贡献了令人惊喜的演技。 在片中饰演穷困小子的董成明曾经凭借饰演《陪读妈妈》中罗盼而为观众所熟知,此次他扮演的同名男主,性格内敛不外放,是十分不好把握的角色,但成明的完成度极高,相信会给大家不少的惊喜;而试验成明“坏朋友”的瞿澳晖,则性格反差明显,需要有很强的表演张力,特别是最后的反转,瞿澳晖的表现令人震撼;而作为片中唯一女性“超能者”紫琳的扮演者陈意涵,则完美的表达出了女主身上兼具岁月静好与坚若磐石的美好品质,是片中美与希望的化身。纵观成片,三人都超能的完成了主创团队给予给角色的任务,并赋予了自己角色独有的气质。 真实与虚幻的交界,善良与邪恶的博弈,相信月日《超能事件》在爱奇艺上映时,一定交给观众一份满意的答卷。把瓮突樟倥https://github.com/trailofbits/ebpfpub/issues/317?55573 <br />https://github.com/trailofbits/ebpfpub/issues/4613?24689 <br />https://github.com/trailofbits/ebpfpub/issues/3233?61397 <br />https://github.com/trailofbits/ebpfpub/issues/1853?62405 <br />https://github.com/trailofbits/ebpfpub/issues/531?iomme <br />https://github.com/trailofbits/ebpfpub/issues/1840 <br />https://github.com/trailofbits/ebpfpub/issues/461 <br />gxqyrslmvqoxkdoertroofwmencksmhiwkt |
7026/Blog | 727960133 | Title: [Vssue]简单介绍 | 小人物
Question:
username_0: http://localhost:8081/blogs/briefNotes/2020/introduce.html
Answers:
username_0: hello world
username_0: # test
username_1: 写的很好 very good
username_0: 啊哈哈
username_1: 请把简历投到 https://www.alibaba.com 谢谢
username_0: <h1> 江城子 . 程序员之歌</h1>
十年生死两茫茫,写程序,到天亮。
千行代码,Bug何处藏。
纵使上线又怎样,朝令改,夕断肠。
领导每天新想法,天天改,日日忙。
相顾无言,惟有泪千行。
每晚灯火阑珊处,夜难寐,加班狂。
username_0: /*
* _oo0oo_
* o8888888o
* 88" . "88
* (| -_- |)
* 0\ = /0
* ___/`---'\___
* .' \\| |// '.
* / \\||| : |||// \
* / _||||| -:- |||||- \
* | | \\\ - /// | |
* | \_| ''\---/'' |_/ |
* \ .-\__ '-' ___/-. /
* ___'. .' /--.--\ `. .'___
* ."" '< `.___\_<|>_/___.' >' "".
* | | : `- \`.;`\ _ /`;.`/ - ` : | |
* \ \ `_. \_ __\ /__ _/ .-` / /
* =====`-.____`.___ \_____/___.-`___.-'=====
* `=---='
*
*
* ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
*
* 佛祖保佑 永不宕机 永无BUG
*
* 佛曰:
* 写字楼里写字间,写字间里程序员;
* 程序人员写程序,又拿程序换酒钱。
* 酒醒只在网上坐,酒醉还来网下眠;
* 酒醉酒醒日复日,网上网下年复年。
* 但愿老死电脑间,不愿鞠躬老板前;
* 奔驰宝马贵者趣,公交自行程序员。
* 别人笑我忒疯癫,我笑自己命太贱;
* 不见满街漂亮妹,哪个归得程序员?
*/ |
iosGroupProjectF18/howardTaskRabbit | 367855231 | Title: Project Feedback!
Question:
username_0: Looks like you are missing the following required user story:
- Create simple wireframes for your app (+6pts)
Your assignment is incomplete until all the [required user stories](https://courses.codepath.com/snippets/ios_university/readme_templates/assignment_1_readme.md) are complete. Once completed, please push your updates, update your README, update your gif and [submit your assignment](https://courses.codepath.com/courses/ios_university/pages/submitting_assignments) again using the submit button on the appropriate assignment page so we can regrade it.
Whenever you make updates to your project that require re-grading, you need to **re-submit** your project using the submit button on the associated assignment page in the course portal. This will flag your project as “updated” on our end and we know to re-grade.
You should re-submit your assignment anytime you:
- Update a previously incomplete assignment
- Add optional and additional features to an already completed assignment
Answers:
username_0: Looks like you are missing the following required user story:
- Using GitHub issues, break down the work that each team member is expected to complete this week (+2pts)
Your group milestone is incomplete until all the required user stories are complete. Once completed, please push your updates, update your README, update your gif and submit your group milestone again using the submit button on the appropriate assignment page so we can regrade it.
Whenever you make updates to your project that require re-grading, you need to re-submit your project using the submit button on the associated assignment page in the course portal. This will flag your project as “updated” on our end and we know to re-grade.
You should re-submit your assignment anytime you:
- Update a previously incomplete assignment
- Add optional and additional features to an already completed assignment
/cc @username_0
username_0: Looks like you are missing the following required user stories:
- Complete 1/2 of the optional stories (+2pts)
- Using GitHub issues, break down the work that each team member is expected to complete this week (+2pts)
Your group milestone is incomplete until all the required user stories are complete. Once completed, please push your updates, update your README, update your gif and submit your group milestone again using the submit button on the appropriate assignment page so we can regrade it.
Whenever you make updates to your project that require re-grading, you need to re-submit your project using the submit button on the associated assignment page in the course portal. This will flag your project as “updated” on our end and we know to re-grade.
You should re-submit your assignment anytime you:
- Update a previously incomplete assignment
- Add optional and additional features to an already completed assignment
/cc @username_0
username_0: Looks like you are missing the following required user stories:
- Complete ALL of the optional stories (+2pts)
- Using GitHub issues, break down the work that each team member is expected to complete this week (+2pts)
Your group milestone is incomplete until all the required user stories are complete. Once completed, please push your updates, update your README, update your gif and submit your group milestone again using the submit button on the appropriate assignment page so we can regrade it.
Whenever you make updates to your project that require re-grading, you need to re-submit your project using the submit button on the associated assignment page in the course portal. This will flag your project as “updated” on our end and we know to re-grade.
You should re-submit your assignment anytime you:
- Update a previously incomplete assignment
- Add optional and additional features to an already completed assignment
/cc @username_0 |
pamplemousser/1485-cupcake | 325434798 | Title: HTML Review
Question:
username_0: Cleanliness
- [ ] - Class need to be used more than once. Since you have IDs on each chunk of content for the navigation, utilize those IDs to apply the background color instead of creating an extra class.
- [ ] - Footer is a unique element, shouldn’t need a class
- [ ] - Reduce use of IDs and Classes
Semantics
- [ ] - Per the in class tutorial, you do not want to place the smooth scroll class around the whole website otherwise your non smooth scrolling links will not work
- [ ] - The whole website cannot be placed within one wrapper class as there are multiple background colors and images that need to span to the edge of the browser. Every time there is a change in background that goes to 100% width, there will need to be a different wrapper. Due to this, the overall structure is incorrect.
- [ ] - Alt attributes for images only need to communicate what the image say if it’s text. The logo doesn’t need to say ‘logo’
- [ ] - Code elements in the order in which they appear: top to bottom / left to right. The image of the logo needs to be within the navigation, within the ul/li if it will be sitting amongst those links
- [ ] - Line 26 is not an h2 or a headline
- [ ] - Line 31 is not the h1 for the website
- [ ] - Images that are not communicating information do not need to have an alt attribute value
- [ ] - Missing header tags around introductory copy throughout
- [ ] - Missing sections to group headlines with their corresponding paragraphs throughout
- [ ] - The cherries should not be placed in through HTML. These need to be placed in through CSS. See EX 8 - star gif for the nav example
- [ ] - Line 76 is not a headline
- [ ] - Images for our treats will need to be placed in through CSS to create the circle and white background
- [ ] - Daily flavors: there will need to be four columns per row. Your current groupings do not make sense for two rows and will be difficult to align per the different lengths of body copy
- [ ] - Missing caption on table
- [ ] - TH should be used within the THead
- [ ] - Remove empty th/td elements
- [ ] - Incorrect type attribute values on phone and email
- [ ] - Missing name and value attributes throughout select/options
- [ ] - Will want to use the same row/column set up on flavors for both type and quantity so that they align with minimal CSS |
chjj/bns | 350632560 | Title: Doc example doesn't seem to work
Question:
username_0: It looks like the readme examples are out of date.
In the **Base Server** example, `dns.Server` is undefined (it's not exported from `require("bns")`).
If I change it to `bns.DNSServer`, `server.on("query")` is never invoked and an assertion error is emitted from `server.on("error")`.
If I change it to `bns.AuthServer`, `server.on("query")` is invoked, but it looks like you shouldn't call `res.send()` ("res.send is not a function" is emitted). An assertion error happens in this case also.
If I change it to `bns.StubServer`, ditto, except `Error: No servers available` is emitted instead of the assertion failure.
Since you're apparently not supposed to call `res.send()`, does that mean it's not possible to have an async response?
([email protected])
Answers:
username_1: @chjj same issue here, `bns.Server` should be `bns.DNSServer` in docs and `server.on("query")` is never invoked...cannot get this library working. |
PointyCastle/pointycastle | 128876813 | Title: Dually license the project under Mozilla Public License as well as under LGPL
Question:
username_0: _From @izaera on May 11, 2014 18:26_
Initially the project was just released under LGPL 3.0 but due to the way dart2js works, it was necessary to license it under MPL 2.0. This is because LGPL 3.0 forces developers of closed source applications to provide some way to let the final user change the LGPL library by a different implementation which is impossible due to dart2js not supporting dynamic linking.
With MPL 2.0 the library can be compiled with dart2js and mixed with closed source without violating the license. The good thing is that, even with that possibility, modifications to any source file belonging to cipher library must be made available under MPL 2.0 and contributed to the original project so they will still remain free.
_Copied from original issue: izaera/cipher#84_<issue_closed>
Status: Issue closed |
elsa-workflows/elsa-core | 978083532 | Title: Failed to run workflow - Collection was modified; enumeration operation may not execute.
Question:
username_0: Rare one:
```
System.InvalidOperationException: Collection was modified; enumeration operation may not execute.
at System.Collections.Generic.Dictionary`2.ValueCollection.Enumerator.MoveNext()
at System.Linq.Enumerable.WhereEnumerableIterator`1.MoveNext()
at System.Collections.Generic.List`1..ctor(IEnumerable`1 collection)
at System.Linq.Enumerable.ToList[TSource](IEnumerable`1 source)
at Elsa.Persistence.InMemory.InMemoryStore`1.FindManyAsync(ISpecification`1 specification, IOrderBy`1 orderBy, IPaging paging, CancellationToken cancellationToken)
at Open.Linq.AsyncExtensions.Extensions.ToList[TSource](Task`1 source)
at Elsa.Persistence.InMemory.InMemoryStore`1.DeleteManyAsync(ISpecification`1 specification, CancellationToken cancellationToken)
at Elsa.Services.Bookmarks.BookmarkIndexer.IndexBookmarksAsync(IEnumerable`1 workflowInstances, CancellationToken cancellationToken)
at Elsa.Services.Bookmarks.BookmarkIndexer.IndexBookmarksAsync(WorkflowInstance workflowInstance, CancellationToken cancellationToken)
at Elsa.Handlers.UpdateBookmarks.Handle(WorkflowInstanceSaved notification, CancellationToken cancellationToken)
at MediatR.Mediator.PublishCore(IEnumerable`1 allHandlers, INotification notification, CancellationToken cancellationToken)
at Elsa.Persistence.Decorators.EventPublishingWorkflowInstanceStore.SaveAsync(WorkflowInstance entity, CancellationToken cancellationToken)
at Elsa.Handlers.PersistWorkflow.SaveWorkflowAsync(WorkflowInstance workflowInstance, CancellationToken cancellationToken)
at Elsa.Handlers.PersistWorkflow.Handle(WorkflowExecutionPassCompleted notification, CancellationToken cancellationToken)
at MediatR.Mediator.PublishCore(IEnumerable`1 allHandlers, INotification notification, CancellationToken cancellationToken)
at Elsa.Services.Workflows.WorkflowRunner.RunCoreAsync(WorkflowExecutionContext workflowExecutionContext, ActivityOperation activityOperation, CancellationToken cancellationToken)
at Elsa.Services.Workflows.WorkflowRunner.RunAsync(WorkflowExecutionContext workflowExecutionContext, ActivityOperation activityOperation, CancellationToken cancellationToken)
```
Answers:
username_1: Do you have a sample project I can use to reproduce? I just tried to run a workflow during server startup and using the in-memory provider, but I'm not getting this issue.
username_0: I imagine I could tempt one of the console samples to reproduce it, but I expect it will be far from consistent. Not sure when I'll get the chance to quite yet though.
username_1: Yeah I tried to do that as well. But I understand if you're busy so no rush. |
DevExpress/testcafe-hammerhead | 468322859 | Title: document.documentElement.querySelectorAll returns Shadow UI Elements
Question:
username_0: `document.documentElement.querySelectorAll` can return Shadow UI elements. It can cause subtle bugs when using `Selector.parent()` in TestCafe.
How to reproduce:
1. Open example.com in the playground.
2. Type `window['%hammerhead%'].shadowUI.getRoot()` in the DevTools console.
3. Type `document.documentElement.querySelectorAll('div')`.
4. The Shadow UI root element is the second item in the returned array. |
phalcon/cphalcon | 242602187 | Title: Uncaught Error: Access to undeclared static property: Phalcon\Di::$_default
Question:
username_0: build project by `phalcon project --type=modules`
in `IndexController ` `indexAction`
```php
for ($i=1;$i<2000;$i++){
file_get_contents(__DIR__.'/IndexController.php');
}
return 'test';
```
then I visit my website faster; Fatal error show;
but visit my website slowly ; Fatal error not show;
```php
Fatal error: Uncaught Error: Access to undeclared static property: Phalcon\Di::$_default in E:\mytest\PhalconAdmin\admin\public\default\app\bootstrap_web.php:17
Stack trace: #0 [internal function]: Phalcon\Di->__construct()
#1 E:\mytest\PhalconAdmin\admin\public\default\app\bootstrap_web.php(17): Phalcon\Di\FactoryDefault->__construct()
#2 E:\mytest\PhalconAdmin\admin\public\default\public\index.php(2): require('E:\\mytest\\Phalc...')
#3 {main} Next Error: Access to undeclared static property: Phalcon\Di::$_default in E:\mytest\PhalconAdmin\admin\public\default\app\bootstrap_web.php:17 Stack trace:
#0 [internal function]: Phalcon\Di->__construct()
#1 E:\mytest\PhalconAdmin\admin\public\default\app\bootstrap_web.php(17): Phalcon\Di\FactoryDefault->__construct()
#2 E:\mytest\PhalconAdmin\admin\public\default\public\index.php(2): require('E:\\mytest\\Phalc...')
#3 {main} thrown in E:\mytest\PhalconAdmin\admin\public\default\app\bootstrap_web.php on line 17
```
* Phalcon version: (`3.2.0`)
* PHP Version: (`PHP Version 7.1.7`)
* Operating System:window10
* Server: Apache
* Server API | Apache 2.0 Handler
*php opcache.enable=0
Answers:
username_1: Duplicate of https://github.com/phalcon/cphalcon/issues/12056
username_0: switch to php5.6 , I've had no problems.
username_1: There is solution provided there - switch to php-cgi + nginx/apache, and problem solved, even with php 7.1
Status: Issue closed
username_2: Close in favor of #12056 |
googleapis/google-cloud-ruby | 1133097119 | Title: Sample request: setting endpoint on storage client
Question:
username_0: We need samples added that describe how to set the endpoint for our storage clients.
Such as setting for localhost/emulator, or private/vpc, or for the soon to be added locational endpoints feature.
Sample tracker & Request:
https://docs.google.com/document/d/1aeL7jgs<KEY>/edit# |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.