repo_name
stringlengths 4
136
| issue_id
stringlengths 5
10
| text
stringlengths 37
4.84M
|
---|---|---|
telred/telred-overlay | 457596597 | Title: jpeg-62 has been removed
Question:
username_0: sky-ng dep is not available:
https://gitweb.gentoo.org/repo/gentoo.git/tree/virtual/jpeg?id=15f8c32a171cca665574bbfbd936a4dea34317ed
Answers:
username_0: I had to change also following:
`media-video/ffmpeg:0/55.58.58[X]`
But there is problem with openssl:
`sky: error while loading shared libraries: libcrypto.so.1.0.0: cannot open shared object file: No such file or directory`
I have following ssl:
Installed versions: 1.1.0k(0/1.1)^d(09:04:17 AM 06/06/2019)(asm zlib -bindist -rfc3779 -sctp -static-libs -test -tls-heartbeat -vanilla ABI_MIPS="-n32 -n64 -o32" ABI_RISCV="-lp64 -lp64d" ABI_S390="-32 -64" ABI_X86="32 64 -x32" CPU_FLAGS_X86="sse2" ELIBC="-musl")
username_0: Do you plan to upgrade to 1.1 ?
username_1: Same issues here with openssl and ffmpeg |
jaegertracing/jaeger | 598946897 | Title: storage/host requirements for Jaeger
Question:
username_0: Hi Team,
I have used Jaeger production templates for Openshift (not all-in-one) to deploy in our environments and I am not using Jaeger operator. Since our elasticsearch cluster is running externally (not on Openshift), I have tweaked the yamls to point to our elasticsearch cluster. I would like to know, is there scenario that might come up while running Jaeger in production that may need us to get some persistent storage in Openshift OR any extra host required for Jaeger. Where does Jaeger store the data that it shows on Jaeger-query URL? Is it the elasticsearch backend in my case?
Regards,
Deependra
Answers:
username_1: @username_0 All the data is stored in the external Elasticsearch in the scenario you described.
Is there a particular reason you didn't use the Jaeger operator?
username_0: @username_1 thanks for the response. There is no particular reason. I found it easy without operators. Is there something I am missing by not using Operators?
username_1: The operator aims to simplify configuration of different scenarios, and will help with upgrade situations.
Status: Issue closed
|
ModellingWebLab/project_issues | 252257708 | Title: Grant Task 3a: Quantifying and visualising uncertainty
Question:
username_0: We will implement sampling from the parameter distributions obtained in Tasks 2a and 2b when running virtual experiments, and display the resulting distributions of protocol outputs, rather than single traces as at present. Some work on computational optimisation (including using faster simplifications of models known as emulators, and running on cloud and GPU resources) will be included to speed up this process, since many more simulations will need to be run than currently.
Answers:
username_1: We hope to also experiment with visualisation in [Pints](https://github.com/martinjrobins/pints)!
username_0: Trac ticket on optimising the Python backend: https://chaste.cs.ox.ac.uk/trac/ticket/2459 |
juliomarcopineda/jdbc-stream | 731629690 | Title: Closing ResultSet
Question:
username_0: What about closing result set when stream is processed?
I don't see it is closed in JdbcStream, so shall it be handled by invoker?
Answers:
username_1: Hello, I think closing the result set has to be handled outside of `JdbcStream`.
My intent with this is class is to purely handle the conversion from JDBC `ResultSet` to a `Stream` object so anyone can easily use the Java Stream API (or even [StreamEx](https://github.com/amaembo/streamex)). Any connection management to the SQL database is left to the user.
So for example, you could use a try-with-resources to call the `close()` method of `ResultSet` automatically:
`java
try (ResultSet resultSet = preparedStatement.executeQuery()) {
long count = JdbcStream.stream(resultSet).count()
System.out.println(count)
}
`
I think this is for the best so that the user has the most flexibility handling the connections to the SQL database.
username_0: I think in your example resultSet won't be closed by default, you need add handler for Stream.onClose with logic that will close the ResultSet |
phetsims/scenery-phet | 172524916 | Title: LaserPointerNode creates unnecessary this.tandem
Question:
username_0: Noted while working on #256.
LaserPointerNode line 99:
```js
this.tandem = options.tandem; // @private
this.tandem && this.tandem.addInstance( this, TNode );
```
But `this.tandem` is not used outside of the constructor, so it's unnecessary.
Answers:
username_0: Looks like I added this code in 700d3f4a87ffd499fba6b12c28849bdb0f23d25f. Fixed up in the above commit. @username_1 would you mind reviewing, to make sure I didn't break any PhET-iO sims?
username_1: I fixed options.tandem.removeInstance in the above commit, tested BLL in instance-proxies wrapper. All seems well, closing.
Status: Issue closed
username_0: Doh. Maybe I should stay out of common code today. |
oliver006/rethinkdb_exporter | 828285140 | Title: do you have an example of connecting to aws rethinkdb?
Question:
username_0: does the command take a .json file for parameters db.user and db.pass, etc?
such like,
rethinkdb-exporter rethinkdb_config.json
thanks.
Answers:
username_0: Found the solution
rethinkdb_exporter -db.addr -db.user -db.pass
this will pull data to localhost:9123
then it should work.
Status: Issue closed
|
Masclins/signepedia | 326760537 | Title: Canviar taula views
Question:
username_0: Actualment la taula de views distingeix entre majúscules i minúscules. Guarda la paraula i el número.
Ho hauria de guardar sempre en minúscules per evitar repeticions i hauria de tenir l'opció ignorar per poder suprimir cerques sense sentit com per exemple: asdhasòh.
Answers:
username_1: Canviat a #167
Status: Issue closed
username_1: Resolved by #171 on develop |
barryvdh/laravel-dompdf | 241292847 | Title: Arabic RTL support is not working
Question:
username_0: While before using `loadHTML` echoing the output html its getting RTL alignment with it. After calling this function not getting RTL layout properly.
```
$pdf = App::make('dompdf.wrapper');
$pdf->loadHTML($html1);
return $pdf->stream('invoice.pdf',array('Attachment'=>0));
```
_Find screen shots for reference_


Answers:
username_1: i have the same problem ,please if you get any solution share with us
username_2: DOMPDF has an issue with both Persian and Arabic, it shows characters dispatched.
I found a solution that might be tough to be implemented in Laravel:
https://github.com/dompdf/dompdf/issues/712#issuecomment-171533682
You can use mPDF Library instead which its wrapper is provided for Laravel as well:
https://github.com/niklasravnsborg/laravel-pdf
username_3: Same issue here also
Are you planning to fix or move to another library ?!
username_1: i have the same issue
--
Cordialement
<NAME>
Ingénieur d Etat en informatique
+212627023256
username_4: I am able to fix the issue by using mpdf library.
1. Use the mpdf library. In laravel use package https://github.com/niklasravnsborg/laravel-pdf
2. Use the font-family: "XB Riyaz" with dir=”rtl” attributes
http://www.solutionsbased.in/laravel-arabic-content-display-issue-in-the-pdf/
username_5: Run
composer require kgcoder/laravel-ar-dompdf
it's a wrapper for laravel dompdf with arabic support addition.
See the Readme File here https://github.com/username_5/laravel-dompdf
username_6: same problem , what's different?
username_7: im working with laravel 5.7 & barryvdh/laravel-dompdf 0.8.4 fixed arabic letters as below:-
1- download the ar-php library for khaled.alshamaa from the following link
http://www.ar-php.org/ (its free arabic Library with LGPL license)
2 - extract and upload I18N folder inside dompdf folder
3- open Glyphs.php file in this path
vendor\dompdf\I18N\Arabic
search
public function utf8Glyphs($str, $max_chars = 50
replace to
public function utf8Glyphs($str, $max_chars = 150
4-in text_renderer.cls.php file
vendor/dompdf/dompdf/src/Render/Text.php
before class name class Text extends AbstractRenderer
add this
include DOMPDF_DIR . "/I18N/Arabic/Glyphs.php";
use I18N_Arabic_Glyphs;
before this line
$this->_canvas->text($x, $y, $text
add this code
if ( ! class_exists( 'I18N_Arabic' ) ){
$Arabic = new I18N_Arabic_Glyphs('Glyphs');
$text = $Arabic->utf8Glyphs($text);
}
in ur pdf file add this style
body{
font-family: DejaVu Sans, sans-serif;
}
username_8: @username_7
Many thanks IT work with me.
You are super star
username_9: you can use this package which supports rtl languages such as persian and arabic
https://github.com/barryvdh/laravel-snappy
username_10: use below package:
https://github.com/mccarlosen/laravel-mpdf
i tested and works with persian well.
username_11: How to DomPDF html to pdf direction LTR to RTL using codeigniter?
I've facing problems with DOMPDF. I'm running PHP 7.3.6 and have the newest DOMPDF version. my pdf view can't move Direction LTR to RTL.
my code :
```
$basurl=base_url();
$url="assets/images/logo.png";
$id = $this->uri->segment(3);
$dir=($ltype=="en"?"ltr":"rtl");
$html_content = '<!DOCTYPE html><html dir="'.$dir.'"><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8"></head><body>
<p align="center" style="margin:0"><img style="width:250px;border:0" src="https://mahlyat.com/images/logo.png"></p><br><br><h2 align="center" style="margin:0 0 10px 0">'.$lang1['orderview'].'</h2>
<h4 align="center" style="margin:0">'.$shopname.'</h4><p align="center" style="margin:0">+'.$q->spphone.'</p>';
$this->pdf->loadHtml($html_content,'UTF-8');
$this->pdf->set_option('isRemoteEnabled', true);
$this->pdf->set_paper(array(0, 0, 454, 841), 'portrait');
//$this->pdf->set_paper('a4', 'landscape');
$this->pdf->render();
$this->pdf->stream("".$id."_view.pdf", array("Attachment"=>0));
```
How i table data direction change LTR to RTL,

username_12: I followed username_7 solution and created a fork: https://github.com/username_12/dompdf
username_7: you welcome bro
username_7: thanks
username_13: Can you please share your code, thanks.
username_14: I have tried using this fork. It was working fine.
But from two days its not working. Let me know if its working at you side.
username_15: I followed with @username_7 and many thanks, it worked successfully 😄👏
username_16: @username_7
you made my day ..
thanks alot .
it work for me laravel --v 7.x
username_17: Hello.
I downloaded ar-php the file, but there is no the folder. could you explain about it in detail? |
lmammino/cube-daemons | 34003234 | Title: Daemon won't show default dashboard -fix
Question:
username_0: Default daemon scripts won't load default dashboard because Cube needs to be run from installed directory.
To fix this you need to edit both init scipts and add a node path value and the --chdir parameter:
```
...
NODE_DIR="$NODE_PREFIX/lib/node_modules/cube"
...
start() {
log_daemon_msg "Starting $NAME..."
mkdir -p /var/run/cube
if start-stop-daemon --background --chdir "$NODE_DIR" --make-pidfile --p
idfile $PIDFILE -o --chuid $USER:$GROUP --start --exec "$APP_DIR/$APP_BIN" -- $A
PP_ARGS ; then
log_end_msg 0
else
log_end_msg 1
fi
}
...
```<issue_closed>
Status: Issue closed |
symfony/symfony | 398521632 | Title: [WebProfilerBundle] toolbar: invisible route name in Firefox
Question:
username_0: **Symfony version(s) affected**: 4.2.2
**Description**
In Firefox, when route name is too long, it's invisible on the debug toolbar. Screenshots from Chrome and from Firefox below. I did also additional check without any extra stylesheets, with the same effect.
**How to reproduce**
Go to page with long route name. I used `user.my_account_settings` on my screenshot.
**Possible Solution**
Remove `text-overflow: ellipsis;` from `.sf-toolbar-block .sf-toolbar-icon` in `toolbar.css.twig`.
**Additional context**
https://i.imgur.com/a2g4fBK.png
Answers:
username_0: There is no problem with it. Without text-ellipsis route name is rendered without breaking UI. Look at this screenshot: https://i.imgur.com/Emjno6G.png
username_1: removine
```
.sf-toolbar-block .sf-toolbar-info-piece-additional, .sf-toolbar-block .sf-toolbar-info-piece-additional-detail {
display: inline-block;
}
```
seems to show the ellepsis both in chrome and FF, but it's colored blue.. so a bit hidden.
https://github.com/symfony/symfony/blob/master/src/Symfony/Bundle/WebProfilerBundle/Resources/views/Profiler/toolbar.css.twig#L517
should check for history, what this solved :)
username_2: putting `display: inline;` instead of `display: inline-block;` seems working to me
so the proposed modification is:
```
.sf-toolbar-block .sf-toolbar-info-piece-additional,
.sf-toolbar-block .sf-toolbar-info-piece-additional-detail {
display: inline;
}
```
at https://github.com/symfony/symfony/blob/master/src/Symfony/Bundle/WebProfilerBundle/Resources/views/Profiler/toolbar.css.twig#L517
I can produce a PR if needed
username_3: Please send a PR yes.
username_2: @username_3 here you are: https://github.com/symfony/symfony/pull/30383
Status: Issue closed
|
wanadev/holes-in | 218114662 | Title: Inside becomes outside when hole is aligned with outline
Question:
username_0: As title says, the hole becomes inverted when its border is aligned with parent mesh.

Answers:
username_1: Fixed in last version
Status: Issue closed
|
Ogidi/ogidi-java-learning | 105787913 | Title: Check which class names are valid
Question:
username_0: Check which of the following class names are valid in Java. Publish the results on the Class Name and File Name wiki page (create a section called Valid Class Names). Use a table to summarize your results. For each compilation that fails, try to explain why that class name is not valid.
1. T
2. cat
3. 4G
4. G4
5. _Basket
6. Play_Ground
7. *
8. $Account
9. Account$
10. Test.Case
11. &Car
12. Em@il
13. #Set
14. North-West
15. 文学
16. Object
17. if
18. _
19. W<T>
20 W[T]
Status: Issue closed
Answers:
username_1: Valid Class and File Names. Closes #10 |
collinss/Cinnamon-Command-Launcher | 425837685 | Title: Still need password to run as root (Mint Cinnamon 19.1)
Question:
username_0: I'm not sure if this is an error or if this is how it is meant to function, but even with 'run as root' option checked; I'm still asked to input my password each time before it runs the command. Is there anyway that I could include my password in the command or something so it will just run with root privileges when using hot keys or single button click? I'd greatly appreciate the help. I'm running Linux Mint Cinnamon 19.1, thanks!
Status: Issue closed
Answers:
username_1: Not that I know of, but there may be. The idea here was to allow the user to still be able to run commands that require root privileges and pop up a dialog that prompts for a password. Unfortunately, this is messy stuff, and I don't think I ever got this working as well as I wanted to.
If you do figure out a way to make it work better, you are always welcome to submit a pull request for it to https://github.com/linuxmint/cinnamon-spices-applets, but otherwise I don't plan to do much with this, as I spend most of my coding time on Mint itself now. |
KratosMultiphysics/GiDInterface | 490772681 | Title: [ContactStructuralMechanics] Missbehaviour of the normal definition
Question:
username_0: Hi!
Recently I've been generating cases with GiD for the ContactStructuralMechApp and, despite **I've checked that the normals are ok in the preprocess, in the postprocess the normals are backwards**....
I'm wondering if something has changed in the normal condition generation :)
PD: Currently the ptype is writing ""raileigh_alpha" instead of "rayleigh_alpha" which throws an error.
Thank you kindly,
Alejandro
Answers:
username_1: Hi @username_0
Could you attach an example of that backward normal issue?
I'm checking the `rayleigh_alpha `one
username_0: [contact_problem.gid.zip](https://github.com/KratosMultiphysics/GiDInterface/files/3591427/contact_problem.gid.zip)
here you have an example :)
username_1: Hi Alejandro, I've tested your case (changing a bit the end_time and the pressure and I got this contact

I don't know about that normal direction issue, but I think this is not the expected contact
username_0: Hi @username_1 ! This example has already 2 years so maybe I have to check everything (who knows). What do you prefer? I try to generate a new one or updating this one?
username_1: I was planning to create 2 examples (2D and 3D) for the Structural mechanics application where we can see the contact working.
Maybe it's a good moment to close this issue and create those examples from an existing and working test |
Azure/azure-sdk-for-java | 469447578 | Title: appconfiguration: ConfigurationService.java Does listKeyValues API need additional values for nextUrl calls
Question:
username_0: When we need to get listkeys for nextPage using nextUrl, do we need to pass additional query parameter to next page call ? Or nextUrl already have needed parameters.
These parameters are
QueryParam - String key,
QueryParam - String label,
QueryParam - String fields,
HeaderParam - String acceptDatetime,
HeaderParam - String range
Change following API if needed
ConfigurationService.java
Mono<PagedResponse<ConfigurationSetting>> listKeyValues(@HostParam("url") String url, @PathParam(value = "nextUrl", encoded = true) String nextUrl, Context context);<issue_closed>
Status: Issue closed |
ai-se/ML-assisted-SLR | 235874382 | Title: Miwa2014
Question:
username_0: [Miwa2014](https://scholar.google.com/scholar?q=Performance+of+support-vector-machine-based+classification+on+15+systematic+review+topics+evaluated+with+the+wss%4095+measure&btnG=&hl=en&as_sdt=0%2C34)
Evaluation: Utility, Coverage, and AUC
### Core:
- Sampling: certainty, uncertainty
- Data balancing: weighting, aggressive undersampling, no balancing
- Assumes: stop training at some point of time, has initial seed training set
Conclusion: certainty+weighting
### Enhancements:
- Ensemble classifiers: voting, PAL
- Covariate shift: two-stage approximation method
- Clustering before classification: LDA
Conclusion: LDA is promising |
nacho270/GTL | 229496876 | Title: Revisar ODTs con salida parcial
Question:
username_0: Revisar ODTs que están parcialmente en salida o bien que están completamente en salida y facturadas pero que no tienen ese estado.
Answers:
username_0: Uso este query para detectar esas ODTs:
select odt.a_codigo, count(*), sum(case when rel.F_PIEZA_ODT_P_ID is not null then 1 else 0 end)
from t_pieza_odt podt
inner join t_orden_de_trabajo odt on odt.p_id = podt.f_odt_p_id
inner join T_REMITO_SALIDA_ODT sal on sal.F_ODT_P_ID = odt.p_id
inner join t_remito rs on rs.p_id = sal.F_REMITO_SALIDA_P_ID
inner join t_factura f on f.p_id = rs.f_factura_p_id
left join T_PIEZA_REM_SAL_PIEZA_ODT rel on rel.F_PIEZA_ODT_P_ID = podt.p_id
where odt.a_id_estado !=7
group by odt.a_codigo
-- having count(*) > 0 && count(*) != sum(case when rel.F_PIEZA_ODT_P_ID is not null then 1 else 0 end)
having sum(case when rel.F_PIEZA_ODT_P_ID is not null then 1 else 0 end) > 0 && count(*) != sum(case when rel.F_PIEZA_ODT_P_ID is not null then 1 else 0 end)
order by odt.a_codigo desc;
Sumo 1h.
Total: 1h
Status: Issue closed
username_0: Sumo 2 hs de la implementación de la acción para autogestión, detección de casos en la B, armado de mail y despliegue de versión.
Total acumulado: 3h |
sebsylvester/parse-server-mailgun | 330143706 | Title: An appName, publicServerURL, and emailAdapter are required for password reset and email verification functionality.
Question:
username_0: I recently started getting this error when trying to reset the password in our app. I saw a previous issue about this however, I believe I have all the correct info and I am still getting this issue.
**The Error Message:**
```
error: Error handling request: ParseError {
code: 1,
'An appName, publicServerURL, and emailAdapter are required for password reset and email verification functionality.' } code=1, message=An appName, publicServerURL, and emailAdapter are required for password reset and email verification functionality.
```
**My Index.js**
```
emailAdapter: {
module: 'parse-server-mailgun',
options: {
appName: 'MyAppName',
fromAddress: process.env.EMAIL_FROM_ADDRESS,
domain: process.env.EMAIL_DOMAIN,
apiKey: process.env.EMAIL_KEY,
publicServerURL: process.env.SERVER_URL,
templates: {
passwordResetEmail: {
subject: 'Reset Your QuizBot Password',
pathPlainText: '/email/PasswordReset.txt',
pathHtml: '/email/PasswordReset.html'
}
}
}
},
```
Any ideas what is causing this? Thanks!
Status: Issue closed
Answers:
username_1: The error was not thrown by the `parse-server-mailgun` module, but the Parse Server. From what I see, you've misplaced the `appName` parameter. It does not belong in the emailAdapter config, but a level higher. I'm getting this from the [docs](https://github.com/parse-community/parse-server#email-verification-and-password-reset), which is supposed be the first thing you check before posting issues on GitHub. If this does not solve the problem, you might want to verify the values of the env vars.
username_0: Thank you. I do try and review the docs. It was an easy oversight to make. Thanks for your help! |
nextstrain/ncov | 593512868 | Title: Metadata: Host field has "Human" as well as "human"
Question:
username_0: **Current Behavior**
The metadata "host" field has "Human" and in a few cases "human". All other hosts start with an upper case character as well.
**Expected behavior**
Consistent labeling, to avoid that case-sensitive matching (often default) fails.
**Possible solution**
Switch all "human" to "Human" in host column.
Answers:
username_1: Hi @username_0 ! I just checked the `data/metadata.tsv` file which is online now, and I can't find lowercase 'human' instances in this copy. Can you check you're using a recently pulled (or downloaded) version of the file? If you're still finding lowercase 'human' instances, can you let us know the strain name they're associated with? Thanks!
username_0: Great, the newer versions don't have lowercase "human" anymore, it must have been fixed by someone already!
Status: Issue closed
|
Shopify/shipit-engine | 180541713 | Title: Reveal Shipit's authentication to outside of the engine
Question:
username_0: The use case:
I'm currently working on an application which includes both the Shipit engine and other endpoints/functionality not related to Shipit. Users of this "other functionality" should still be authenticated and it would be convenient to use the authentication system has already built into Shipit.
An example `routes.rb` with a proposed interface, inspired by devise:
```ruby
Rails.application.routes.draw do
mount Shipit::Engine, at: '/'
Shipit.authenticate do
resource :some_model
end
end
```
This API also allows for use with other engines as well as eliminating the need for all controllers to inherit from `Shipit::ShipitController` to gain access to the authentication.
I understand this use case is a bit niche, but I am curious about what the interest level is on a feature like this. This is definitely something I would be prepared to build out in a fork if it we like the idea.
Answers:
username_1: I wouldn't be opposed to such feature. But I have to admit I'm not sure how I'd go about implementing it.
But if you say Devise does it I can take a look.
username_1: Alternatively, would including a module in your controllers be ok?
I understand you don't want to have to inherit from `Shipit::ShipitController`, but if you could just include `Shipit::Authentication` would that work for you?
username_0: Hmm I'm not sure I follow, that sounds like a great solution, but I can't seem to find the module `Shipit::Authentication`?
username_1: It doesn't exist yet, but It could. That's why I'm asking if that would be a satisfactory solution for you?
username_0: Ah okay I see, then yes, this seems like a great solution.
Status: Issue closed
username_2: Done. Let me know if you have any issue with it. |
SignalR/SignalR | 533645497 | Title: server downgrades protocol version to 1.2 when going through nginx
Question:
username_0: I am using Microsoft.AspNet.SignalR.Core 2.4.1
Client is jquery.signalR-2.1.2
When I connect straight to my web app
```shell
$ curl 'http://192.168.1.235:83/signalr/negotiate?clientProtocol=1.4&connectionData=%5B%7B%22name%22%3A%22nodes%22%7D%5D&_=1575586094093' -H 'Accept: text/plain, */*; q=0.01' -H 'Referer: http://localhost:32770/' -H 'X-Requested-With: XMLHttpRequest' -H 'User-Agent: Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.132 Safari/537.36' -H 'Sec-Fetch-Mode: cors' -H 'Content-Type: application/json; charset=UTF-8' --compressed
{"Url":"/signalr","ConnectionToken":"<KEY>","ConnectionId":"4c11611e-25bb-4a10-8506-4a0bfffdfd1b","KeepAliveTimeout":20.0,"DisconnectTimeout":30.0,"ConnectionTimeout":110.0,"TryWebSockets":true,"ProtocolVersion":"1.4","TransportConnectTimeout":5.0,"LongPollDelay":0.0}
```
When I go through nginx acting as a reverse proxy
```shell
$ curl 'http://localhost:32770/hub//signalr/negotiate?clientProtocol=1.4&connectionData=%5B%7B%22name%22%3A%22nodes%22%7D%5D&_=1575586094093' -H 'Accept: text/plain, */*; q=0.01' -H 'Referer: http://localhost:32770/' -H 'X-Requested-With: XMLHttpRequest' -H 'User-Agent: Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.132 Safari/537.36' -H 'Sec-Fetch-Mode: cors' -H 'Content-Type: application/json; charset=UTF-8' --compressed
{"Url":"/signalr","ConnectionToken":"<KEY>","ConnectionId":"211aafbb-5b7c-44c3-9b92-3c4b0d8b80e9","KeepAliveTimeout":20.0,"DisconnectTimeout":30.0,"ConnectionTimeout":110.0,"TryWebSockets":true,"ProtocolVersion":"1.2","TransportConnectTimeout":5.0,"LongPollDelay":0.0}
```
nginx configuration is
```
location ~ ^/hub/(.*)$ {
proxy_pass http://192.168.1.235:83/$1;
proxy_http_version 1.1;
proxy_set_header Upgrade $http_upgrade;
proxy_set_header Connection "upgrade";
proxy_set_header Host $host;
proxy_cache_bypass $http_upgrade;
}
```
### Expected behavior
Server should use protocol 1.2 in both cases
### Actual behavior
server switches protocol to 1.2 when it goes through nginx - causing the client to kick it out
Answers:
username_0: ok well I see it's not p[assing querystring so I chnages nginx config to
```
location ~ ^/hub/(.*) {
proxy_pass http://192.168.1.235:83/$1$is_args$args;
proxy_http_version 1.1;
proxy_set_header Upgrade $http_upgrade;
proxy_set_header Connection "upgrade";
proxy_set_header Host $host;
proxy_cache_bypass $http_upgrade;
}
```
and I add another one for the web socket
```
location ~ ^/signalr-hub/(.*) {
proxy_pass http://192.168.1.235:83/signalr-hub$is_args$args;
proxy_http_version 1.1;
proxy_set_header Upgrade $http_upgrade;
proxy_set_header Connection "upgrade";
proxy_set_header Host $host;
proxy_cache_bypass $http_upgrade;
}
```
now it does get correct protocol and fires off websocket but I only get one message of {} back
a long shot but any guidance on how to set up nginx to reverse proxy signalR?
Status: Issue closed
|
adobe/helix-content-proxy | 859161337 | Title: Changes to reverse resolver break sidekick
Question:
username_0: See https://github.com/adobe/helix-content-proxy/pull/379/files#diff-41629dcc63276d799267ad24df29f434d9d5ebc007a00350f6cd839200b85e42L43-L54
This change in #379 (in anticipation of https://github.com/adobe/helix-home/issues/187) causes the sidekick in `spark-website` to break - see https://github.com/adobe/spark-website/pull/121#pullrequestreview-637017545 for a workaround.
See https://cq-dev.slack.com/archives/C01F5TQV0SF/p1618510521096600 for end-user bug report.
Answers:
username_1: sorry for that. but let's fix sidekick....
Status: Issue closed
username_2: I consider this a new feature 😉 See https://github.com/adobe/helix-pages/pull/867 |
msp-rs/multiwii_serial_protocol_v2.rs | 660041041 | Title: Offering a nicer API on std
Question:
username_0: Hi, what do you think of offering a nicer API for std builds? My ideas so far:
+ offering an API which turns anything which offers `Read` & `Write` into a MSP Connection:
```
pub trait MspConnection: Read + Write {}
impl<T: Read + Write> MspConnection for T {}
```
+ Having a fetch method on the MspMessage Type itself, hence one can use code likes this, getting rid of matching against a variant of an enum with all messages:
```
let my_message: MspAttitude = MspMessage::fetch(msp_conn)
```
+ Having a fetch method on the `MspConnection` as well, which returns an `enum MspMessages`
+ Offering an async interface. (This I don't know yet how to make it nice)
This are just some quick ideas of my head, what do you think about it?
Answers:
username_1: I like the idea.
I guess for sync-std we can offer another trait
```
pub trait MspConnectionAsync: async_std::io::Read + async_std::io::Write {}
impl<T: async_std::io::Read + async_std::io::Write> MspConnectionAsync for T {}
```
but then the library will dependent on std and async-std and won't compile at all for no_std, right?
do we want to support not_std at all?
maybe we should use the serde.rs trait also, https://docs.serde.rs/serde/ser/trait.Serialize.html after all its a serialization library.
because its already covers the `Write` & `Read` traits
username_0: The idea of using serde is interesting, indeed. Though replacing rudis struct de-/ser code will be a massive task. First I want to think about how async request shall work though, I'm still not at a point where I have a presentable solution.
username_1: Actually Rudis code used packet_struct library, which already uses serde, the Serialize and Deserialize are coming from there.
So ignore my comments :)
username_0: Yes! `std` shall be optional!
username_1: I am totally fine with that |
hasura/learn-graphql | 712610602 | Title: Multiple Typos in backend tutorials - access control section
Question:
username_0: permissions mispelled in 1-user-permissions.md
Could the team add hacktoberfest label here?
I already created PR fixing this.
Answers:
username_0: Checkout the PR - https://github.com/hasura/learn-graphql/pull/338
Status: Issue closed
username_1: Hi @username_0 - Thanks for your PR. Merged now. |
ItemConsulting/TeamUP | 172890145 | Title: Personas APP
Question:
username_0: Every team sites need a person directory.
This task is to get the basics done for personas APP.
- Create a personas app
- Create profile contenttype
- Create a simple profile view page
- Create a simple list of all persons in direcotry
- Create a simple search bar for search |
newinnovations/remote-edit-ni | 474735309 | Title: Uncaught Error: Unknown message digest
Question:
username_0: internal/crypto/sig.js:98
Hide Stack Trace
Error: Unknown message digest
at new Verify (internal/crypto/sig.js:98:16)
at Object.createVerify (crypto.js:133:10)
at onKEXDH_REPLY (/Users/pusingh/.atom/packages/remote-edit-ni/node_modules/ssh2-streams/lib/ssh.js:2580:25)
at SSH2Stream.<anonymous> (/Users/pusingh/.atom/packages/remote-edit-ni/node_modules/ssh2-streams/lib/ssh.js:188:45)
at SSH2Stream.emit (events.js:182:13)
at parse_KEXDH_REPLY (/Users/pusingh/.atom/packages/remote-edit-ni/node_modules/ssh2-streams/lib/ssh.js:3908:8)
at parse_KEX (/Users/pusingh/.atom/packages/remote-edit-ni/node_modules/ssh2-streams/lib/ssh.js:3856:16)
at parsePacket (/Users/pusingh/.atom/packages/remote-edit-ni/node_modules/ssh2-streams/lib/ssh.js:3729:12)
at SSH2Stream._transform (/Users/pusingh/.atom/packages/remote-edit-ni/node_modules/ssh2-streams/lib/ssh.js:551:13)
at SSH2Stream.Transform._read (_stream_transform.js:190:10)
at SSH2Stream._read (/Users/pusingh/.atom/packages/remote-edit-ni/node_modules/ssh2-streams/lib/ssh.js:212:15)
at SSH2Stream.Transform._write (_stream_transform.js:178:12)
at doWrite (_stream_writable.js:410:12)
at writeOrBuffer (_stream_writable.js:394:5)
at SSH2Stream.Writable.write (_stream_writable.js:294:11)
at Socket.ondata (_stream_readable.js:672:20)
at Socket.emit (events.js:182:13)
at addChunk (_stream_readable.js:279:12)
at readableAddChunk (_stream_readable.js:264:11)
at Socket.Readable.push (_stream_readable.js:219:10)
at TCP.onread (net.js:636:20)
My Key look like this :
-----BEGIN RSA PRIVATE KEY-----
Proc-Type: 4,ENCRYPTED
DEK-Info: AES-128-CBC,***....char_hidden***
KEY HERE
-----END RSA PRIVATE KEY----- |
react-component/tree-select | 214803410 | Title: Allow prop highlightTreeNode to be passed in
Question:
username_0: If I pass in `filterTreeNode`, I am able to customize which items are filtered. However, that is different from what is highlighted. I'd like for the highlighting to be consistent with the filtering. The easiest way would be to check if `filterTreeNode` is defined, and if so, apply highlighting based on that prop, rather than highlighting independently
https://github.com/react-component/tree-select/blob/master/src/SelectTrigger.jsx#L175
Answers:
username_1: + i alse need some hightlighter |
graphql-compose/graphql-compose-mongoose | 490799320 | Title: Cast to date failed error when using date value inside filter with _operators
Question:
username_0: I'm using a filter similar to the following and running into an error `"Cast to date failed for value \"{ '': 2019-01-01T00:00:00.000Z }\" at path \"beginsAt\" for model \"Event\""`.
```js
filter: {
_operators: { beginsAt: { gte: "2019-01-01T00:00:00.000Z" } }
}
```
It seems related to how the `_toMongoDottedObject` function builds up the query conditions.
I have a failing test that I'll submit as a PR. |
hellworld-io/hellworld-io.github.io | 123395800 | Title: Personal projects wish list
Question:
username_0: Number Baseball Game
Bingo Game
weather REST API
Lotto
Answers:
username_0: nBaseball (Number Baseball Game) : https://github.com/hellworld-io/nBaseball
I will start this project on 12/28, 2015
username_0: I will start nBaseball project on Wednesday.
username_0: bMark added |
bitmade/node-twig | 172883537 | Title: How to use @alias and ::namespace ?
Question:
username_0: Hi,
I don't find any things about how to define/use @alias and or namespace.
I found ```AliasLoader.php```, but how to use it ?
Thx.
Answers:
username_1: Hi,
you're right, we are missing some documentation here.
Since we are using this package in our dev-tools you can have a look there:
https://github.com/bitmade/dev-tools/blob/master/src/twigEngine.js
So aliases basically are just key value pairs passed into the options under the aliases key.
The values are paths to the folders you want to alias.
Let me here if that works for you.
username_0: Hi,
Yes, thanks !
Status: Issue closed
|
twitchdev/issues | 1183311309 | Title: Bits usage not working on Android anymore
Question:
username_0: **Brief description**
Since the Twitch Android v12.6 update, when calling the Twitch extension function ``Twitch.ext.bits.useBits(sku)`` inside Twitch extensions, the Bits confirmation dialog is not working anymore. It also invokes the ``window.Twitch.ext.onError`` callback.
This completely breaks Bits usage inside Twitch extensions on Android devices.
**How to reproduce**
Call Twitch extension function ``Twitch.ext.bits.useBits(sku)`` inside Twitch extensions on mobile Twitch app v12.6 on Android.
**Expected behavior**
A Bits confirmation dialog should appear where the user can confirm their Bits usage within the Twitch extension
Answers:
username_1: I'm able to reproduce this issue using the beta Android app with the following build:
```
v12.7.0_BETA (c)
March 25, 2022 at 15:35
```
username_2: Thanks for reporting and sharing details. This is being tracked internally on INC-2101.
username_2: Please confirm that the issue as been resolved. It may take some time for changes to propagate, but folks believe this to be corrected.
username_1: Can confirm this is resolved on my build (Android app didn't update, still on the same version as mentioned above), thank you for passing this on and to the devs for the prompt resolution!
username_0: Can confirm too it's working again on 12.6. Bits confirmation window appears again. Thank you very much for resolving this so quickly!
Status: Issue closed
|
gophercloud/gophercloud | 249895672 | Title: Add support for OS_PROJECT_* variables
Question:
username_0: `openstack.AuthOptionsFromEnv` doesn't support the `OS_PROJECT_ID` and `OS_PROJECT_NAME` environment variables. A standard `openrc` file that a user downloads from Horizon (when using Keystone v3) will have `PROJECT` set instead of `TENANT`.
Adding support for `PROJECT` will make it possible for users to not have to modify their `openrc` file when using `openstack.AuthOptionsFromEnv`.
Answers:
username_1: I am applying for outreachy round 15 and starting to work on this issue as advised by <NAME> the mentor. I hope no one is working on it for now.
username_0: @username_1 Thank you for your interest in this! However, there is a PR already in place: #462 (referenced above).
username_1: @username_0 Thanks for the feedback I will let him know its been worked on so he can advise on another. If you have ideas for any first time bugs I can work on. You can also advise. Thanks
username_0: @username_1 Sure, I can try to find something. What's your current skill level with each:
* OpenStack APIs
* Go
brand new, some experience, familiar, very familiar?
username_1: @username_0 I am familiar with Go but brand new to OpenStack APIs but am flexible and willing to learn.
username_2: closed by #462
Status: Issue closed
|
dundalek/GrammKit | 212898280 | Title: Generate SVGs as separate files
Question:
username_0: Just a suggestion, but GrammKit might be more useful if SVGs were generated as separate files from `<output>.html`. They could still be included in the output file, but then people who wanted to use them differently (say by linking to them, possibly out of order, in a README) could easily do so.
Answers:
username_1: Thanks, that's a good idea. Do you think it would make sense to add option to export to markdown which creates those separate files and then a markdown file that includes them in order? That way people could use them in READMEs or pick out individually.
username_0: @username_1: that sounds fantastic. It's a little more than I need for my situation (since I'll be writing my own markdown for the README), but I could definitely see people finding it useful.
username_1: I published a new version `v0.5.0`. You can run `grammkit -t md yourgrammar.peg` which will output a markdown file and separate SVG files. If you want to extract the list of files in the original order you can use unix friends like `sed -n 's/## \(.*\)/\1.svg/p' yourgrammar.md`. Let me know if you find any issues.
username_0: 🎉 🎉 🎉
Works great! Also, I added you to my [BNF-tools summary](https://github.com/username_0/state-of-haskell-bnf) page. It's only supposed to be Haskell programs, but since there isn't a Haskell lib to do this I put GrammKit on anyway.
Status: Issue closed
username_1: Cool, thanks. |
tleunen/babel-plugin-module-resolver | 191686594 | Title: Is there support for using with `System.import()` (or alternative?)
Question:
username_0: I am using `System.import()` in my React project for lazy loading components and data to reduce initial bundle size.
```javascript
["module-resolver", {"alias": { "data": "./src/data/"}}]
```
Sample code
```javascript
System.import('data/faqs')
.then(response => response.default)
.then(faqs => this.setState({faqs}))
```
Error
```shell
ERROR in ./src/scripts/app/pages/FAQ.js
Module not found: Error: Can't resolve 'data/faqs' in '<truncated-path>/src/scripts/app/pages'
@ ./src/scripts/app/pages/FAQ.js 42:6-39
@ ./src/scripts/bundles/app.js
@ multi app
```
Regular imports (`import faq from 'data'`) are working as expected, but not when using `System.import()`
If `System.import()` is not supported is there an alternative I can use?
For now I am just creating aliases within webpack config for the affected imports like below
```javascript
resolve: {
alias: {
data: path.join(__dirname, 'src/data/')
}
}
```
Would be nice if babel could handle all the aliases if possible :)
Answers:
username_1: Yes, as this time, the System.import syntax is not supported. But it would be great to add support for it. I'm adding this to the list of improvement.
It's a functionality similar to #97
Status: Issue closed
|
PirvuCatalin/JFXOpenCV | 317218994 | Title: [Database] Only works on local machine
Question:
username_0: As for now, the database querry (Oracle one) only works on a local machine. For the moment, we don't have any reference to a cloud database such as Azure.
Will look into it to create a database on the web and access it via Java code. |
MrTJP/ProjectRed | 354570771 | Title: Block Breakers don't drop an item when mined
Question:
username_0: When you mine them, they just disappear. They also don't have block breaking particles and don't have a prefered tool. I assume these are the result of the same bug. If not, I can resubmit them as separate issues.
Answers:
username_1: Block breakers drop correctly when mined as well as any block they have broken and is still stuck inside. No issues that I could find.
Status: Issue closed
|
ISPP-LinkedPet/backend_isppet | 596737940 | Title: Vistas Panel - Mis Peticiones
Question:
username_0: ## Mis Peticiones
En esta vista aparecerán todas las peticiones que ha realizado el usuario logeado a anuncios de otros usuarios. Para cada petición debe indicarse:
- Si la petición es a un anuncio de Crianza o Adopción
- El estado en el que se encuentra la petición
- El usuario al que pertenece el anuncio al que se le hace la petición
Aparecen nuevos estados:
- Las peticiones a anuncios que no han sido moderadas.
- Las peticiones a anuncios que aún no han sido aceptadas o rechazadas por el propietario.
- Las peticiones a anuncios que me han rechazado.
<issue_closed>
Status: Issue closed |
PaddlePaddle/Paddle | 291480589 | Title: fluid.nets.simple_img_conv_pool error TypeError: set_attr(): incompatible function arguments. The following argument types are supported
Question:
username_0: Run below code error on develop branch:
```python
conv_pool_1 = fluid.nets.simple_img_conv_pool(
input=images,
filter_size=5,
num_filters=20,
pool_size=2,
pool_stride=2,
act="relu")
```
```
Traceback (most recent call last):
File "dist_test.py", line 29, in <module>
act="relu")
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/nets.py", line 39, in simple_img_conv_pool
use_cudnn=use_cudnn)
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/layers/nn.py", line 1042, in conv2d
default_initializer=_get_default_param_initializer())
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/layer_helper.py", line 129, in create_parameter
dtype=dtype, shape=shape, **attr.to_kwargs(with_initializer=True))
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/framework.py", line 686, in create_parameter
kwargs['initializer'](param, self)
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/initializer.py", line 214, in __call__
"seed": self._seed
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/framework.py", line 707, in prepend_op
op = Operator(self, op_desc, *args, **kwargs)
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/framework.py", line 475, in __init__
self.desc.set_attr(attr_name, attrs[attr_name])
TypeError: set_attr(): incompatible function arguments. The following argument types are supported:
1. (self: paddle.v2.fluid.core.OpDesc, arg0: unicode, arg1: boost::variant<boost::blank, int, float, std::string, std::vector<int, std::allocator<int> >, std::vector<float, std::allocator<float> >, std::vector<std::string, std::allocator<std::string> >, bool, std::vector<bool, std::allocator<bool> >, paddle::framework::BlockDesc*, long, std::vector<long, std::allocator<long> > >) -> None
Invoked with: <paddle.v2.fluid.core.OpDesc object at 0x7f3da9ff5990>, u'shape', (20L, 1L, 5L, 5L)
```
Status: Issue closed
Answers:
username_0: Need to find why. I suppose this issue may due to incompatible gcc or pybind versions.
username_0: Run below code error on develop branch, build using docker image: `paddlepaddle/paddle_manylinux_devel:cuda7.5_cudnn5`, with options: `WITH_GPU=OFF WITH_DISTRIBUTED=ON`
```python
conv_pool_1 = fluid.nets.simple_img_conv_pool(
input=images,
filter_size=5,
num_filters=20,
pool_size=2,
pool_stride=2,
act="relu")
```
```
Traceback (most recent call last):
File "dist_test.py", line 29, in <module>
act="relu")
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/nets.py", line 39, in simple_img_conv_pool
use_cudnn=use_cudnn)
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/layers/nn.py", line 1042, in conv2d
default_initializer=_get_default_param_initializer())
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/layer_helper.py", line 129, in create_parameter
dtype=dtype, shape=shape, **attr.to_kwargs(with_initializer=True))
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/framework.py", line 686, in create_parameter
kwargs['initializer'](param, self)
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/initializer.py", line 214, in __call__
"seed": self._seed
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/framework.py", line 707, in prepend_op
op = Operator(self, op_desc, *args, **kwargs)
File "/paddle/build/python/build/lib-python/paddle/v2/fluid/framework.py", line 475, in __init__
self.desc.set_attr(attr_name, attrs[attr_name])
TypeError: set_attr(): incompatible function arguments. The following argument types are supported:
1. (self: paddle.v2.fluid.core.OpDesc, arg0: unicode, arg1: boost::variant<boost::blank, int, float, std::string, std::vector<int, std::allocator<int> >, std::vector<float, std::allocator<float> >, std::vector<std::string, std::allocator<std::string> >, bool, std::vector<bool, std::allocator<bool> >, paddle::framework::BlockDesc*, long, std::vector<long, std::allocator<long> > >) -> None
Invoked with: <paddle.v2.fluid.core.OpDesc object at 0x7f3da9ff5990>, u'shape', (20L, 1L, 5L, 5L)
```
username_1: pybind itself does not handle boost variant. It might be because the variant_caster at paddle/pybind/protobuf.cc has some bug. A related issue #7369
username_2: 您好,此issue在近一个月内暂无更新,我们将于今天内关闭。若在关闭后您仍需跟进提问,可重新开启此问题,我们将在24小时内回复您。因关闭带来的不便我们深表歉意,请您谅解~感谢您对PaddlePaddle的支持!
Hello, this issue has not been updated in the past month. We will close it today for the sake of other user‘s experience. If you still need to follow up on this question after closing, please feel free to reopen it. In that case, we will get back to you within 24 hours. We apologize for the inconvenience caused by the closure and thank you so much for your support of PaddlePaddle Group!
Status: Issue closed
|
scikit-rf/scikit-rf | 656880542 | Title: Network converted to mixed mode and back to single ended has 180 degree phase shift
Question:
username_0: A network passed through se2gmm followed by gmm2se will display the correct magnitude for the s-parameters, but the phase will be shifted by plus/minus 180 degrees.
```
import skrf as rf
import matplotlib.pyplot as plt
ntwk1 = rf.Network(sedatafile)
ntwk2 = rf.Network(sedatafile)
ntwk2.se2gmm(p=2)
ntwk2.gmm2se(p=2)
```
## Magnitude Difference:
```
for m in range(4):
for n in range(4):
plt.plot(sedata.f, abs(ntwk1.s)[:,m,n]-abs(ntwk2.s)[:,m,n], label=f'S{m+1}{n+1}')
plt.title('Magnitude Error between Measurements')
plt.legend(bbox_to_anchor=(1.1, 1.05))
```

## Phase by Parameter:
```
fig, axes = plt.subplots(4,4, sharex=True, figsize=(14,8))
for m in range(4):
for n in range(4):
ntwk1.plot_s_deg_unwrap(m=m, n=n, ax=axes[m][n])
ntwk2.plot_s_deg_unwrap(m=m, n=n, ax=axes[m][n], ls='--')
axes[m][n].get_legend().remove()
fig.tight_layout()
```

## Phase Difference by Parameter:
```
fig, axes = plt.subplots(4,4, sharex=True, figsize=(14,8))
for m in range(4):
for n in range(4):
axes[m][n].plot(ntwk1.f, ntwk2.s_deg_unwrap[:,m,n]-ntwk1.s_deg_unwrap[:,m,n])
axes[m][n].set_ylim([-200,200])
fig.tight_layout()
```

Answers:
username_0: This is likely at least partly a mathematical limitation rather than a bug, since going from mixed mode to single ended you would only have the differential and common phase information between ports, so you would expect to be able to obtain some arbitrary phase shift of all ports such that the relative phases were the same. I tried the following:
```
import numpy as npy
def se2gmm(ntwk, p, z0_mm=None):
# XXX: assumes 'proper' order (first differential ports, then single ended ports)
if z0_mm is None:
z0_mm = ntwk.z0.copy()
z0_mm[:, 0:p] = 100 # differential mode impedance
z0_mm[:, p:2 * p] = 25 # common mode impedance
Xi_tilde_11, Xi_tilde_12, Xi_tilde_21, Xi_tilde_22 = ntwk._Xi_tilde(p, ntwk.z0, z0_mm)
A = Xi_tilde_21 + npy.einsum('...ij,...jk->...ik', Xi_tilde_22, ntwk.s)
B = Xi_tilde_11 + npy.einsum('...ij,...jk->...ik', Xi_tilde_12, ntwk.s)
ntwk.s = npy.transpose(npy.linalg.solve(npy.transpose(B, (0, 2, 1)).conj(), npy.transpose(A, (0, 2, 1)).conj()),
(0, 2, 1)).conj() # (34)
ntwk.z0 = z0_mm
return ntwk
def gmm2se(ntwk, p, z0_se=None):
# TODO: testing of reverse transformation
# XXX: assumes 'proper' order (differential ports, single ended ports)
if z0_se is None:
z0_se = ntwk.z0.copy()
z0_se[:] = 50
Xi_tilde_11, Xi_tilde_12, Xi_tilde_21, Xi_tilde_22 = ntwk._Xi_tilde(p, z0_se, ntwk.z0)
#A = Xi_tilde_22 - npy.einsum('...ij,...jk->...ik', ntwk.s, Xi_tilde_12)
#B = Xi_tilde_21 - npy.einsum('...ij,...jk->...ik', ntwk.s, Xi_tilde_11)
A = Xi_tilde_22 - npy.matmul(ntwk.s, Xi_tilde_12)
B = Xi_tilde_21 - npy.matmul(ntwk.s, Xi_tilde_11)
ntwk.s = npy.linalg.solve(A, B) # (35)
ntwk.z0 = z0_se
return ntwk
ntwk1 = rf.Network(sedatafile)
ntwk2 = rf.Network(sedatafile)
print(ntwk1 == ntwk2)
print(ntwk1 == gmm2se(se2gmm(ntwk2, p=2), p=2))
print(se2gmm(ntwk1, p=2) == se2gmm(gmm2se(se2gmm(ntwk2, p=2), p=2), p=2))
print(ntwk1 == se2gmm(gmm2se(gmm2se(se2gmm(ntwk2, p=2), p=2), p=2), p=2))
print(gmm2se(se2gmm(ntwk2, p=2), p=2) == se2gmm(gmm2se(gmm2se(se2gmm(ntwk2, p=2), p=2), p=2), p=2))
```
```
True
False
True
False
True
```
As you can see, while a network transformed to mixed-mode and back to single ended is not equal to itself, those two unequal single-ended measurements transformed back to mixed-mode are in fact considered equal as far as skrf is concerned. As for why two of the parameters have a negative phase shift compared to the rest of the 4x4 matrix, I still have to investigate.
Status: Issue closed
username_0: Closing this for now as I believe it is not a bug in the code, but rather a mathematical limitation of how phase is relative. I am still slightly skeptical of why some ports are +180 and some are -180 degrees, but If this causes an issue for someone in the future we can revisit it.
username_0: A network passed through se2gmm followed by gmm2se will display the correct magnitude for the s-parameters, but the phase will be shifted by plus/minus 180 degrees.
```
import skrf as rf
import matplotlib.pyplot as plt
ntwk1 = rf.Network(sedatafile)
ntwk2 = rf.Network(sedatafile)
ntwk2.se2gmm(p=2)
ntwk2.gmm2se(p=2)
```
## Magnitude Difference:
```
for m in range(4):
for n in range(4):
plt.plot(sedata.f, abs(ntwk1.s)[:,m,n]-abs(ntwk2.s)[:,m,n], label=f'S{m+1}{n+1}')
plt.title('Magnitude Error between Measurements')
plt.legend(bbox_to_anchor=(1.1, 1.05))
```

## Phase by Parameter:
```
fig, axes = plt.subplots(4,4, sharex=True, figsize=(14,8))
for m in range(4):
for n in range(4):
ntwk1.plot_s_deg_unwrap(m=m, n=n, ax=axes[m][n])
ntwk2.plot_s_deg_unwrap(m=m, n=n, ax=axes[m][n], ls='--')
axes[m][n].get_legend().remove()
fig.tight_layout()
```

## Phase Difference by Parameter:
```
fig, axes = plt.subplots(4,4, sharex=True, figsize=(14,8))
for m in range(4):
for n in range(4):
axes[m][n].plot(ntwk1.f, ntwk2.s_deg_unwrap[:,m,n]-ntwk1.s_deg_unwrap[:,m,n])
axes[m][n].set_ylim([-200,200])
fig.tight_layout()
```

username_0: As pointed out in the mailing list by <NAME>, the generated phase error on return to single ended can be corrected by transforming the parameters to differential and back to single ended an additional time. For example, in the below approach the red line shows 180 degree phase error after transforming se -> mm -> se. **The green line has gone se -> mm -> se -> mm -> se and has the phase of the original sample**. The yellow line has gone se -> mm -> se -> mm -> se -> mm -> se and has the same phase error as if it had gone through one cycle.
```
ntwk10 = rf.Network(sedatafile)
ntwk11 = rf.deepcopy(ntwk10)
ntwk11.se2gmm(p=2)
ntwk11.gmm2se(p=2)
ntwk12 = rf.deepcopy(ntwk11)
ntwk12.se2gmm(p=2)
ntwk12.gmm2se(p=2)
ntwk13 = rf.deepcopy(ntwk12)
ntwk13.se2gmm(p=2)
ntwk13.gmm2se(p=2)
fig, axes = plt.subplots(4,4, sharex=True, figsize=(14,8))
for m in range(4):
for n in range(4):
axes[m][n].plot(sedata.f, ntwk11.s_deg_unwrap[:, m, n] - ntwk10.s_deg_unwrap[:, m, n], ls='-.', color='red')
axes[m][n].plot(sedata.f, ntwk12.s_deg_unwrap[:, m, n] - ntwk10.s_deg_unwrap[:, m, n], ls='--', color='green')
axes[m][n].plot(sedata.f, ntwk13.s_deg_unwrap[:, m, n] - ntwk10.s_deg_unwrap[:, m, n], ls=':', color='yellow')
fig.tight_layout()
```

Magnitude error for all three is very small:
```
fig2, axes2 = plt.subplots(4,4, sharex=True, figsize=(14,8))
for m in range(4):
for n in range(4):
axes2[m][n].plot(sedata.f, ntwk11.s_mag[:, m, n] - ntwk10.s_mag[:, m, n], ls='-.', color='red')
axes2[m][n].plot(sedata.f, ntwk12.s_mag[:, m, n] - ntwk10.s_mag[:, m, n], ls='--', color='green')
axes2[m][n].plot(sedata.f, ntwk13.s_mag[:, m, n] - ntwk10.s_mag[:, m, n], ls=':', color='yellow')
```

username_1: I think this issue should be solved. I encountered the same problem, when calculating RLGC parameters for differential transmission lines. After back-conversion, the results were wrong, since 180° shifted S-Parameters represent something completely different.
I have not checked other source files, but in my case, all parameters were shifted and I could simply multiply the whole S matrix with -1. Is this factor maybe missing at some point in the calculation?
@username_0 The +-180° are likely a result from wrapping, since +-180° shift are the same from a mathematical point of view. Some parameters are shifted slightly more than 180.0, which is wrapped to -179.9...9, while others are not.
For now I am applying the transformation twice for future compatibility, because it should also work, once the issue is fixed. |
awslabs/aws-mobile-appsync-sdk-android | 850884246 | Title: No callback triggered after executing query
Question:
username_0: **Describe the bug**
A clear and concise description of what the bug is.
We run a version check by querying appsync when the onUserStateChanged userstate == SIGNED_IN.
Sometimes this version check fails to complete, no failure, no timeout, no response. This usually only happens once when the app is first started, when closing and reopening the app it works.
This problem typically happens when a user opens the app for the first time
**To Reproduce**
Steps to reproduce the behavior:
```
override fun onUserStateChanged(details: UserStateDetails) {
when (details.userState) {
UserState.SIGNED_IN -> {
versionCheck() // queries appsyync for version info
}
// more cases
}
}
fun versionCheck() {
let versionCode = 1
let versionName = 1.0.0
val query = GetVersionInfoQuery.builder().versionCode(versionCode).versionName(versionName).build()
client.query(query)
.responseFetcher(NETWORK_ONLY)
.enqueue(object : GraphQLCall.Callback<GetVersionInfoQuery.Data?>() {
override fun onResponse(@Nonnull response: Response<GetVersionInfoQuery.Data?>) {
// Handle Success (never called)
}
override fun onFailure(@Nonnull e: ApolloException) {
// Handle Fail (never called)
}
})
}
```
**Expected behavior**
A clear and concise description of what you expected to happen.
Execute the query, and get a response either success or failure.
**Environment(please complete the following information):**
- 3.1.3
**Device Information (please complete the following information):**
- Pixel 2 Android 21-30 (simulator)
- Samsung S10+ Real device |
electron/electron | 198790802 | Title: Failed at the [email protected] postinstall script 'node install.js'
Question:
username_0: 14410 error npm v3.10.10
14411 error code ELIFECYCLE
14412 error [email protected] postinstall: `node install.js`
14412 error Exit status 1
14413 error Failed at the [email protected] postinstall script 'node install.js'.
14413 error Make sure you have the latest version of node.js and npm installed.
14413 error If you do, this is most likely a problem with the electron package,
14413 error not with npm itself.
14413 error Tell the author that this fails on your system:
14413 error node install.js
14413 error You can get information on how to open an issue for this project with:
14413 error npm bugs electron
14413 error Or if that isn't available, you can get their info via:
14413 error npm owner ls electron
14413 error There is likely additional logging output above.
14414 verbose exit [ 1, true ]
Answers:
username_1: Can you include the console output when running `npm install electron`?
username_2: Hi
i got a similar problem on ubuntu 16.04 . while trying to npm install i got this
`26351 silly rollbackFailedOptional Starting
26352 silly rollbackFailedOptional Finishing
26353 silly runTopLevelLifecycles Starting
26354 silly runTopLevelLifecycles Finishing
26355 silly install printInstalled
26356 warn [email protected] No description
26357 warn [email protected] No repository field.
26358 warn [email protected] No license field.
26359 verbose stack Error: [email protected] postinstall: `node install.js`
26359 verbose stack spawn ENOENT
26359 verbose stack at ChildProcess.<anonymous> (/usr/share/npm/lib/utils/spawn.js:17:16)
26359 verbose stack at emitTwo (events.js:87:13)
26359 verbose stack at ChildProcess.emit (events.js:172:7)
26359 verbose stack at maybeClose (internal/child_process.js:821:16)
26359 verbose stack at Process.ChildProcess._handle.onexit (internal/child_process.js:211:5)
26360 verbose pkgid [email protected]
26361 verbose cwd /home/chen/dev/nodejs/electron_twitter
26362 error Linux 4.8.0-30-generic
26363 error argv "/usr/bin/nodejs" "/usr/bin/npm" "install"
26364 error node v4.2.6
26365 error npm v3.5.2
26366 error file sh
26367 error code ELIFECYCLE
26368 error errno ENOENT
26369 error syscall spawn
26370 error [email protected] postinstall: `node install.js`
26370 error spawn ENOENT
26371 error Failed at the [email protected] postinstall script 'node install.js'.
26371 error Make sure you have the latest version of node.js and npm installed.
26371 error If you do, this is most likely a problem with the electron-prebuilt package,
26371 error not with npm itself.
26371 error Tell the author that this fails on your system:
26371 error node install.js
26371 error You can get information on how to open an issue for this project with:
26371 error npm bugs electron-prebuilt
26371 error Or if that isn't available, you can get their info via:
26371 error npm owner ls electron-prebuilt
26371 error There is likely additional logging output above.
26372 verbose exit [ 1, true ]
`
username_3: 47652 verbose stack Error: [email protected] postinstall: `node install.js`
47652 verbose stack Exit status 1
47652 verbose stack at EventEmitter.<anonymous> (/usr/share/npm/lib/utils/lifecycle.js:232:16)
47652 verbose stack at emitTwo (events.js:87:13)
47652 verbose stack at EventEmitter.emit (events.js:172:7)
47652 verbose stack at ChildProcess.<anonymous> (/usr/share/npm/lib/utils/spawn.js:24:14)
47652 verbose stack at emitTwo (events.js:87:13)
47652 verbose stack at ChildProcess.emit (events.js:172:7)
47652 verbose stack at maybeClose (internal/child_process.js:821:16)
47652 verbose stack at Process.ChildProcess._handle.onexit (internal/child_process.js:211:5)
47653 verbose pkgid [email protected]
47654 verbose cwd /home/username_3/Descargas/anatine-1.1.0
47655 error Linux 4.4.0-57-generic
47656 error argv "/usr/bin/nodejs" "/usr/bin/npm" "install"
47657 error node v4.2.6
47658 error npm v3.5.2
47659 error code ELIFECYCLE
47660 error [email protected] postinstall: `node install.js`
47660 error Exit status 1
47661 error Failed at the [email protected] postinstall script 'node install.js'.
47661 error Make sure you have the latest version of node.js and npm installed.
47661 error If you do, this is most likely a problem with the electron package,
47661 error not with npm itself.
47661 error Tell the author that this fails on your system:
47661 error node install.js
47661 error You can get information on how to open an issue for this project with:
47661 error npm bugs electron
47661 error Or if that isn't available, you can get their info via:
47661 error npm owner ls electron
47661 error There is likely additional logging output above.
47662 verbose exit [ 1, true ]
username_1: @username_2 This is because Node is installed on your machine as `nodejs` instead of node. Upgrading to npm `3.7.0` or above should address this.
https://github.com/npm/npm/releases/v3.7.0
Status: Issue closed
username_4: Same problem here. Failed at `node install.js`.
- macOS 10.12.2 Sierra
- node 6.9.2
- npm 4.1.1 (using registry: https://registry.npm.taobao.org)
The console shows some error about permission denied to link the release package. But I did used `sudo` command. I tried to manually perform the link and then run `sudo node install.js`. It put me on hold for over 10 mins showing me nothing, until I hit `Ctrl + C` to cancel.
The release package `electron-v1.4.13-darwin-x64.zip` is found under `~/.electron` with size 42.1MB so the network should be fine. I unzipped the file and copied Electron.app to `/Applications` folder. It works just fine. Launch the app and it tells me "To run your app with Electron, execute the following command in your Console (or Terminal):" `/Applications/Electron.app/Contents/MacOS/Electron path-to-your-app`. Run it with no path and it opened up another electron window.
Guess a manually link to global CLI command is all I need to do next. But this doesn't sound like a long-term solution for future upgrade with npm.
I also tried with the latest version 1.4.1. Nothing changed. Hope the team can fixed this ASAP. thx ahead.

username_5: Same here 🌵
username_6: same here
username_7: same here
win 10
node v6.9.4
username_8: same here
node v7.7.2
npm v4.1.2
macOs 10.11.6
username_9: Same here:
macOs 10.12.3
node v7.7.4
npm 4.4.1
username_10: `7495 error Windows_NT 10.0.14393
7496 error argv "C:\\Program Files\\nodejs\\node.exe" "C:\\Program Files\\nodejs\\node_modules\\npm\\bin\\npm-cli.js" "install" "-g" "electron"
7497 error node v7.7.4
7498 error npm v4.1.2
7499 error code ELIFECYCLE
7500 error [email protected] postinstall: `node install.js`
7500 error Exit status 1
7501 error Failed at the [email protected] postinstall script 'node install.js'.
7501 error Make sure you have the latest version of node.js and npm installed.
7501 error If you do, this is most likely a problem with the electron package,
7501 error not with npm itself.
7501 error Tell the author that this fails on your system:
7501 error node install.js
7501 error You can get information on how to open an issue for this project with:
7501 error npm bugs electron
7501 error Or if that isn't available, you can get their info via:
7501 error npm owner ls electron
7501 error There is likely additional logging output above.
7502 verbose exit [ 1, true ]
`
Same problem. Can't get past node install.js
Win 10 x64
node v7.7.4
npm v4.1.2
username_11: anybody solved this?
username_12: same here
```
npm verb node v7.9.0
npm verb npm v4.5.0
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! [email protected] postinstall: `node install.js`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the [email protected] postinstall script 'node install.js'.
npm ERR! Make sure you have the latest version of node.js and npm installed.
npm ERR! If you do, this is most likely a problem with the electron package,
npm ERR! not with npm itself.
npm ERR! Tell the author that this fails on your system:
npm ERR! node install.js
npm ERR! You can get information on how to open an issue for this project with:
npm ERR! npm bugs electron
npm ERR! Or if that isn't available, you can get their info via:
npm ERR! npm owner ls electron
npm ERR! There is likely additional logging output above.
npm verb exit [ 1, true ]
npm ERR! A complete log of this run can be found in:
npm ERR! /Users/saekitakahiro/.npm/_logs/2017-04-29T14_30_54_433Z-debug.log
```
I just did it.
`sudo npm i electron -g --verbose`
username_4: I've got a solution. Not the best way but should be enough if you need to install electron right now and don't have time to wait for the official fix.
Everyone here got into trouble when it comes to `postinstall script 'node install.js'`. So let's start with that.
Run `npm install --save electron` for your project and hit `Ctrl + C` when it comes to this sentence. The download has not complete yet but we've got what we need.
Under `YOUR_PROJECT_PATH/node_modules/electron` you can find this `install.js` file. If you open it and take a look. It's simply some javascript code to download the release package (which you can download directly from Github) and extract it. That's all. So **either something went wrong with the network so the download took too long, or the script do not have permission to extract zip file and rename it**.
Since it crashes when running automatically, let's do it manually.
Just follow what the file says:
1. Download the zip file (from Github using any way you like)
2. extract it into `YOUR_PROJECT_PATH/node_modules/electron`.
3. Rename the extracted folder to 'dist'.
4. Create a `path.txt` file here. Write `dist/Electron.app/Contents/MacOS/Electron` and save it. (This is for mac, For Windows write `dist/electron.exe`, and for linux write `dist/electron`. You can find these in the end of 'install.js')
Now you've installed it locally.
By running `./node_modules/electron/dist/Electron.app/Contents/MacOS/Electron .`, your project should start up. (For non-MacOS environment. the path might be a little different. I trust you can find the right one.)
Next, you can either use npm script to make a alias for that loooooooong command. Or install electron globally so you can run `electron` from the terminal. (Same process. Only this time run `npm install -g electron` and do it under `/usr/local/lib/node_modules/electron`)
username_13: facing the same issue, tried the last answer did not work, please help.
username_14: This worked for me! Do this at your own risk.
First delete any folders called electron in (node_modules) in your project.
Then delete electron like this
`npm uninstall electron`
and lastly run this command to make sure permissions 'refresh' and you can install it correctly.
`npm install electron --save-dev`
username_13: Thanks for the reply. Sorry, I am just a beginner, what are the potential risks of installing like this?
Thanks in advance.
username_14: Yes. The only risks involved is causing your project to not be compatible with the latest version of electron that installs. You don't need to be too worried. This will just ensure that you erase electron from your project then install all the files fresh to ensure a clean install. Give it a try! Maybe backup your project before continuing. Maybe a thanks if it works? <3
username_15: username_4, I don't find where to download the package if you stop it like this. What release panel? What release package. The javascript has no direct download link in the script so am unsure what to download. Mine tries to download an armv7 package and my electron fails here like everyone else's did in this thread, this is the package "Downloading tmp-20048-0-electron-v3.1.9-linux-arm64.zip" so where can I go find this package to try this process.
username_15: I can't find where to download the release packages? Where are they?
username_15: Why is this issue closed
username_16: **Machine Specs:**
<ins>Operating System:</ins>
macOS Big Sur v11.6
<ins>NPM version:</ins>
v6.14.13
<ins>Node version :</ins>
v14.17.0
After some digging, I found [this post](https://stackoverflow.com/a/48515453/4433536) that allowed me to get it installed globally.
**TL;TR**
Open your terminal and execute the follow command
`sudo npm install -g electron --unsafe-perm=true --allow-root`
Hope it helps!
Happy coding =) |
roopakv/google-photos | 1004657059 | Title: code is not defined
Question:
username_0: Sorry in advance, im new to this...
i have a nextja app and am trying to use google-photos package together with the googleapis package (as recomeded for authorization) ... I am recieveing the following error: "ReferenceError: code is not defined" on:
const {tokens} = await oauth2Client.getToken(code);
Any help would be much appreicated.
Jon.
Answers:
username_1: It's because it has been written in javascript and there's no support for typescript. You should use `require` insted of `import`.
```
✅ const Photos = require("googlephotos");
❌ import Photos from "googlephotos";
```
username_2: yeah this is something i've been meaning to fix. I need to add TS support 🤔 |
meddelare/meddelare-node-server | 144837567 | Title: Google + counter
Question:
username_0: Google + always return "-1" value.
Answers:
username_1: @username_0: you're right. Thought it was a temporary issue, but perhaps not. Let's hope it's not as "bad" as #2.
Thanks for reporting this! No time estimate for completion yet. Patches are always welcome =)
username_1: @username_0: actually, this was easy to fix -- it was just a regexp that was outdated. Fixed in meddelare/meddelare-node-counters#6.
Released as meddelare-counters v1.0.2 on NPM. The new version should be used automatically next time you upgrade/install/reinstall medddelare-node-server.
https://www.npmjs.com/package/meddelare-counters
Thanks for reporting the issue!
Status: Issue closed
|
rancher/rancher | 342679899 | Title: AKS Advanced Network Config Misses essential network options
Question:
username_0: Rancher v2 / current

1) Failure to allow control over the K8 service range on an existing network will cause all deployments to fail as the default range is huge and will overlap other ranges.
2) Http Application routing isn't always desirable, the ability to kill it is important.
Answers:
username_1: Fixed in [rancher/kontainer-engine#107](https://github.com/rancher/kontainer-engine/pull/107)
Close in favor of #15697
Status: Issue closed
|
ridedott/merge-me-action | 854646040 | Title: Doesn't merge pull requests with default settings
Question:
username_0: **How would you describe the issue?**
I'm using github's built-in dependabot and added merge-me according to documentation. The action doesn't seem to properly recognise dependabot as commits author out of the box.
**How can we reproduce the issue?**
1. Create new job using default settings
```
jobs:
merge-me:
name: Merge me!
runs-on: ubuntu-latest
steps:
- name: Merge me!
uses: ridedott/[email protected]
with:
GITHUB_TOKEN: ${{ secrets.AUTO_MERGE_TOKEN }}
```
**What are the expected results?**
dependabot PRs get merged
**What are the actual results?**
Job fails with the following output:
```
Automatic merges enabled for GitHub login: dependabot.
Found pull request information: {"authorLogin":"dependabot","commitAuthorName":"dependabot[bot]","commitMessage":"Bump @sentry/node from 6.2.4 to 6.2.5\n\nBumps [@sentry/node](https://github.com/getsentry/sentry-javascript) from 6.2.4 to 6.2.5.\n- [Release notes](https://github.com/getsentry/sentry-javascript/releases)\n- [Changelog](https://github.com/getsentry/sentry-javascript/blob/master/CHANGELOG.md)\n- [Commits](https://github.com/getsentry/sentry-javascript/compare/6.2.4...6.2.5)\n\n","commitMessageHeadline":"Bump @sentry/node from 6.2.4 to 6.2.5","mergeableState":"MERGEABLE","merged":false,"pullRequestId":"MDExOlB1bGxSZXF1ZXN0NjExMjQ4NzIz","pullRequestState":"OPEN","pullRequestTitle":"Bump @sentry/node from 6.2.4 to 6.2.5","reviewEdges":[]}.
Pull request changes were not made by dependabot.
```
**How much does it hurt?**
Just setting GITHUB_LOGIN to `dependabot[bot]` (which is `commitAuthorName`) prevents the job from running since PR author doesn't match anymore:
```
Automatic merges enabled for GitHub login: dependabot[bot].
Pull request created by dependabot, not dependabot[bot], skipping.
```
However, it's possible to workaround this with micromatch:
```
GITHUB_LOGIN: '(dependabot|dependabot\[bot\])'
```
---
Here's an example workflow that failed https://github.com/mavend/octoboard/runs/2307117681?check_suite_focus=true and it's job definition https://github.com/mavend/octoboard/actions/runs/733497847/workflow
And a workflow with workaround that merged the PR https://github.com/mavend/octoboard/runs/2307312762?check_suite_focus=true and it's job definition https://github.com/mavend/octoboard/actions/runs/733570773/workflow
Answers:
username_1: Thanks for the report, this should be fixed with https://github.com/ridedott/merge-me-action/pull/861. Could you confirm?
username_0: Yup, works great with default settings now 👍
Status: Issue closed
|
kyma-project/kyma | 786103078 | Title: Upgrade and self-manage grafana image
Question:
username_0: <!-- Thank you for your contribution. Before you submit the issue:
1. Search open and closed issues for duplicates.
2. Read the contributing guidelines.
-->
**Description**
The recent grafana version 7.3.7 fixes several security issues compared to the bundled version 7.3.3
Furthermore, the default docker image of grafana introduced a big bunch of alpine packages for supporting a scenario not relevant for kyma, see also https://github.com/grafana/grafana/pull/20214
We should introduce a custom image having the related block in the Dockerfile removed giving us the chance to also update the base image whenever we need without upgrading the actual grafana version.
<!-- Provide a clear and concise description of the feature. -->
**Reasons**
Staying up-to-date with dependencies to be fine from known security vulnerability perspective
<!-- Explain why we should add this feature. Provide use cases to illustrate its benefits. -->
**Attachments**
<!-- Attach any files, links, code samples, or screenshots that will convince us to your idea. -->
Status: Issue closed
Answers:
username_1: I created a custom image excluding the non-relevant parts of the original image https://github.com/kyma-incubator/third-party-images/tree/master/grafana. It is also up-to-date with the latest version of Grafana (v7.4.0) and uses the latest alpine version (v3.13) as the base image. The change to use this custom image on the Grafana chart has also been merged: https://github.com/kyma-project/kyma/commit/0f0d5a232f7d46ef38485a365f5e645101aea133 |
fossasia/pslab-android | 236710066 | Title: Handling faulty hardware attached
Question:
username_0: **Actual Behaviour**
If for some reason the processor in the PSLab hardware does not respond, the app freezes for over a minute or two instead of detecting a fault in the hardware
A few minutes later it says 'initialization complete' without actually realising the error
**Expected Behaviour**
report a hardware error. Do not try to load calibration.
**Steps to reproduce it**
Reprogram the PSLab with some firmware that does not implement the PSLab state machine
**LogCat for the issue**
```
D/dalvikvm(22783): VFY: replacing opcode 0x6f at 0x0008
D/dalvikvm(22783): GC_FOR_ALLOC freed 119K, 3% free 7698K/7879K, paused 41ms
I/dalvikvm-heap(22783): Grow heap (frag case) to 8.547MB for 1048592-byte allocation
D/dalvikvm(22783): GC_CONCURRENT freed 10K, 3% free 8711K/8967K, paused 2ms+2ms
D/dalvikvm(22783): GC_FOR_ALLOC freed 2K, 3% free 8718K/8967K, paused 17ms
I/dalvikvm-heap(22783): Grow heap (frag case) to 9.543MB for 1048592-byte allocation
D/dalvikvm(22783): GC_FOR_ALLOC freed <1K, 4% free 9741K/10055K, paused 18ms
I/ActivityManager( 158): Displayed org.fossasia.pslab/.activity.MainActivity: +1s534ms
E/CommunicationHandler(22783): Read Error: 0
D/dalvikvm(22783): GC_FOR_ALLOC freed 1071K, 12% free 9610K/10887K, paused 19ms
D/dalvikvm(22783): GC_CONCURRENT freed 2K, 3% free 10624K/10887K, paused 1ms+2ms
E/CommunicationHandler(22783): Read Error: 0
E/CommunicationHandler(22783): Read Error: 0
E/CommunicationHandler(22783): Read Error: 0
D/audio_hw_primary( 90): start_output_stream: card:0, port:0, rate:44100
D/OpenGLRenderer( 223): Flushing caches (mode 0)
E/CommunicationHandler(22783): Read Error: 0
I/Process ( 158): Sending signal. PID: 22783 SIG: 3
I/dalvikvm(22783): threadid=3: reacting to signal 3
I/dalvikvm(22783): Wrote stack traces to '/data/anr/traces.txt'
E/CommunicationHandler(22783): Read Error: 0
I/Process ( 158): Sending signal. PID: 22783 SIG: 3
I/dalvikvm(22783): threadid=3: reacting to signal 3
I/dalvikvm(22783): Wrote stack traces to '/data/anr/traces.txt'
E/CommunicationHandler(22783): Read Error: 0
I/Process ( 158): Sending signal. PID: 22783 SIG: 3
I/dalvikvm(22783): threadid=3: reacting to signal 3
I/dalvikvm(22783): Wrote stack traces to '/data/anr/traces.txt'
E/CommunicationHandler(22783): Read Error: 0
E/PacketHandler(22783): Error in reading byte
I/Process ( 158): Sending signal. PID: 22783 SIG: 3
I/dalvikvm(22783): threadid=3: reacting to signal 3
I/dalvikvm(22783): Wrote stack traces to '/data/anr/traces.txt'
E/CommunicationHandler(22783): Read Error: 0
E/NRF24L01(22783): Radio transceiver not installed/not found
I/Process ( 158): Sending signal. PID: 22783 SIG: 3
I/dalvikvm(22783): threadid=3: reacting to signal 3
I/dalvikvm(22783): Wrote stack traces to '/data/anr/traces.txt'
E/CommunicationHandler(22783): Read Error: 0
E/PacketHandler(22783): Error in packetHandler Reading
I/Process ( 158): Sending signal. PID: 22783 SIG: 3
I/dalvikvm(22783): threadid=3: reacting to signal 3
I/dalvikvm(22783): Wrote stack traces to '/data/anr/traces.txt'
E/CommunicationHandler(22783): Read Error: 0
E/PacketHandler(22783): Error in packetHandler Reading
I/Process ( 158): Sending signal. PID: 22783 SIG: 3
[Truncated]
I/PlayCommon(21207): [546] com.google.android.play.a.g.e(224): No file ready to send
E/CommunicationHandler(22783): Read Error: 0
E/PacketHandler(22783): Error in packetHandler Reading
E/CommunicationHandler(22783): Read Error: 0
E/PacketHandler(22783): Error in packetHandler Reading
E/CommunicationHandler(22783): Read Error: 0
E/PacketHandler(22783): Error in packetHandler Reading
E/CommunicationHandler(22783): Read Error: 0
E/PacketHandler(22783): Error in packetHandler Reading
and so on for a long while...
^C
```
**Screenshots of the issue**
start screen stuck at 'initializing ..'
**Would you like to work on the issue?**
nope. I can help with testing newer versions.
Answers:
username_1: @username_2 can you elaborate on this a little more ?
username_2: Sure,
Most of it is covered in issue description, if some faulty device is connected, then app should be capable of telling that the device connected is not functioning as intended and provide user with appropriate message/instructions.
username_3: Is this issue still open , I want to fix this issue.
username_2: @username_3 yeah its still open. Go ahead 👍
username_4: @username_2 since there is no work on the issue, may I fix it?
username_2: @username_4 sure go ahead 👍
username_5: Current firmware does not go into state machine unless it is fully initiated.
Status: Issue closed
|
drdhaval2785/SanskritVerb | 112733751 | Title: pages 379 to 381 panding to code
Question:
username_0: p. 379 to p. 381 pending. It is difficult to incorporate these exceptions in liT. Also akaH savarNe dIghaH and Gittva have to be separately explained to the computer. Pending.
Answers:
username_0: kaumudIkAra doesn't approve of optional forms of SranTa! and granTa!.
So only damBa! has to be corrected.
and 'zvaYja!' has to be corrected.
Status: Issue closed
|
coatk1/playground | 628868769 | Title: [FEATURE] CK Django
Question:
username_0: - [ ] I have searched the [issues](https://github.com/username_0/playground/issues) of this repository and believe that this is not a duplicate.
**Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
Answers:
username_0: closing this
Status: Issue closed
|
mzlee/makefiles | 59755914 | Title: Add more to step 0 - how do we even compile?
Question:
username_0: It's good, but I don't think it's sufficient. Long ago, I was under the delusion that Makefiles were a magical box that g++ lived inside of. I did not realize g++ could be called outside of Makefiles... or that I was calling it. There are some missing steps/assumptions taken for granted. |
microsoft/verona | 553366099 | Title: Can't build it on centos7
Question:
username_0: Thank you very much for your reply, I can't build it in centos7.4,But I can create it in centos7.7,
```
➜ ~ cmake --version
cmake version 3.15.5
CMake suite maintained and supported by Kitware (kitware.com/cmake).
➜ ~ clang --version
clang version 10.0.0 (trunk 375507)
Target: x86_64-unknown-linux-gnu
Thread model: posix
Scanning dependencies of target veronac-sys
Scanning dependencies of target veronac
[ 93%] Building CXX object src/compiler/CMakeFiles/veronac-sys.dir/main.cc.o
[ 95%] Building CXX object src/compiler/CMakeFiles/veronac.dir/main.cc.o
[ 96%] Linking CXX executable interpreter-sys
[ 96%] Built target interpreter-sys-bin
[ 98%] Linking CXX executable veronac
[100%] Linking CXX executable veronac-sys
[100%] Built target veronac
[100%] Built target veronac-sys
```
Status: Issue closed
Answers:
username_1: Please provide more information. Centos is not a platform we have tested on.
Have you installed the listed dependencies? Which version of Clang/GCC are you using? Which version of CMake are you using?
Can you provide the output of the failure?
username_0: Thank you very much for your reply, I can't build it in centos7.4,But I can create it in centos7.7,
```
➜ ~ cmake --version
cmake version 3.15.5
CMake suite maintained and supported by Kitware (kitware.com/cmake).
➜ ~ clang --version
clang version 10.0.0 (trunk 375507)
Target: x86_64-unknown-linux-gnu
Thread model: posix
Scanning dependencies of target veronac-sys
Scanning dependencies of target veronac
[ 93%] Building CXX object src/compiler/CMakeFiles/veronac-sys.dir/main.cc.o
[ 95%] Building CXX object src/compiler/CMakeFiles/veronac.dir/main.cc.o
[ 96%] Linking CXX executable interpreter-sys
[ 96%] Built target interpreter-sys-bin
[ 98%] Linking CXX executable veronac
[100%] Linking CXX executable veronac-sys
[100%] Built target veronac
[100%] Built target veronac-sys
```
Status: Issue closed
|
roswell/roswell | 695477334 | Title: ros install sbcl fails when shell command exits with error
Question:
username_0: `ros install sbcl` failed because a shell command exited with error 2.
I issued the offending command in a shell, and got an error from `cc`, `unrecognized command line option '-Wimplicit-fallthrough'.
Details follow:
## Environment
```
$ uname -a
Linux hal9000.us.tel.com 3.10.0-514.26.2.el7.x86_64 #1 SMP Tue Jul 4 15:04:05 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux
$ros --version
roswell 20.06.14.107(72b2d66)
$ros config
ccl-bin.version=1.11.5
sbcl.version=2.0.2
setup.time=3793532798
sbcl-bin.version=2.0.2
default.lisp=sbcl-bin
```
## How to reproduce the issue
```sh
ros install sbcl
```
## Expected behavior
Successful SBCL update from 2.02 to 2.08
## Actual result
```sh
$ ros install sbcl
Checking version to install....
Checking for installed implementation in /export/home/977315/.roswell/impls/x86-64/linux/sbcl/2.0.8/
Skip downloading https://github.com/sbcl/sbcl/archive/sbcl-2.0.8.tar.gz
specify download.force=t to download it again.
Extracting archive:/export/home/977315/.roswell/archives/sbcl-2.0.8.tar.gz
Applying patch:sbcl-1.3.11.patch
chdir /export/home/977315/.roswell/src/sbcl-2.0.8/
patching file make-genesis-2.sh
Reversed (or previously applied) patch detected! Skipping patch.
1 out of 1 hunk ignored
patching file make-host-1.sh
Reversed (or previously applied) patch detected! Skipping patch.
1 out of 1 hunk ignored
patching file make-host-2.sh
Reversed (or previously applied) patch detected! Skipping patch.
1 out of 1 hunk ignored
patching file make-target-1.sh
Reversed (or previously applied) patch detected! Skipping patch.
1 out of 1 hunk ignored
patching file make-target-2.sh
Reversed (or previously applied) patch detected! Skipping patch.
1 out of 1 hunk ignored
0 ....Unhandled UIOP/RUN-PROGRAM:SUBPROCESS-ERROR in thread #<SB-THREAD:THREAD "main thread" RUNNING
{1000518083}>:
[Truncated]
rm -f variables.texinfo
rm -f sbcl.ps asdf.ps sbcl.pdf asdf.pdf html-stamp tempfiles-stamp
rm -f asdf.aux asdf.cp asdf.cps asdf.fn asdf.fns asdf.ky asdf.log asdf.pg asdf.toc asdf.tp asdf.tps asdf.vr asdf.vrs sbcl.aux sbcl.cp sbcl.cps sbcl.fn sbcl.fns sbcl.ky sbcl.log sbcl.pg sbcl.toc sbcl.tp sbcl.tps sbcl.vr sbcl.vrs
rm -f sbcl.info sbcl.info-* asdf.info
rm -rf *.include *.info *.pdf *~ *.cp *.fn *.ky *.log *.pg *.toc \
*.tp *.vr *.aux *.eps *.png *.dvi *.ps *.txt *.fns \
html-stamp sbcl-internals/
//entering make-config.sh
//ensuring the existence of output/ directory
//guessing default target CPU architecture from host architecture
//setting up CPU-architecture-dependent information
sbcl_arch="x86-64"
//initializing /export/home/977315/.roswell/src/sbcl-2.0.8/local-target-features.lisp-expr
//setting up OS-dependent information
make: Entering directory `/export/home/977315/.roswell/src/sbcl-2.0.8/tools-for-build'
cc -Wunused-parameter -Wimplicit-fallthrough -fno-omit-frame-pointer -momit-leaf-frame-pointer -I../src/runtime determine-endianness.c -ldl -o determine-endianness
cc: error: unrecognized command line option '-Wimplicit-fallthrough'
make: *** [determine-endianness] Error 1
make: Leaving directory `/export/home/977315/.roswell/src/sbcl-2.0.8/tools-for-build'
```
Answers:
username_0: For what it's worth, `ros install sbcl-bin/2.0.7` succesfully installs sbcl 2.0.7. I will close this issue because I can update sbcl.
Status: Issue closed
|
babel/babel | 115601276 | Title: babylon should handle duplicate declaration errors (?)
Question:
username_0: Asked a bit in the IRC channel and it was suggested that this should be the case. Basically, if you imagine a real time linter (kind of like Babel's Try It) that just wants to show errors, or a syntax highlighter that also shows errors, it would be nice to show Duplicate Declaration Errors (let x = 10; let x = 10;).
As of right now it appears that the only way to get this is through going through a transform (as opposed to babylon.parse just throwing it). This is unfortunate since its a lot of extra work since I won't actually be using the transform, just wanting to get this one error. I might be wrong here and there is some in-between step you can do to get this but I don't know.
Answers:
username_1: This requires scope tracking and isn't something that belongs in a parser. You need to do a full tree traversal anyway in order to do scope tracking due to function hoisting, TDZ and more. You can get this behaviour with a noop traverse:
```javascript
require("babel-traverse").default(yourAstHere);
```
Status: Issue closed
username_2: I get the reasoning behind this, but it really is too bad. The spec defines these things as early errors, I don't think I'd expect to have to traverse in order to get early errors.
username_1: Early errors don't have to happen in the parser, it's an implementation detail. |
hankcs/HanLP | 951304952 | Title: cannot import name 'albert_models_tfhub' from 'bert'
Question:
username_0: <!--
提问请上论坛,不要发这里!
提问请上论坛,不要发这里!
提问请上论坛,不要发这里!
以下必填,否则恕不受理。
-->
**Describe the bug**
bert包里没有albert_models_tfhub
File "C:\ProgramData\Anaconda3\envs\nlp\lib\site-packages\hanlp\layers\transformers\loader_tf.py", line 9, in <module>
from bert import albert_models_tfhub, fetch_tfhub_albert_model, load_stock_weights
ImportError: cannot import name 'albert_models_tfhub' from 'bert' (C:\ProgramData\Anaconda3\envs\nlp\lib\site-packages\bert\__init__.py)
**Code to reproduce the issue**
import hanlp
import bert
from hanlp_restful import HanLPClient
test1 = '葛鹤军先生:博士学位。2008年至2011年就职于中诚信证券评估有限公司、中诚信国际信用评级有限公司。2011年11月加盟银华基金管理有限公司,历任研究员、基金经理助理。现任投资管理三部基金经理。自2014年10月8日起担任“银华中证中票50指数债券型证券投资基金(LOF)”“银华信用债券型证券投资基金(LOF)”的基金经理;自2014年10月8日起至2016年4月25日担任“银华中证成长股债恒定组合30/70指数证券投资基金”基金经理;自2015年4月24日起担任“银华泰利灵活配置混合型证券投资基金”基金经理;自2015年5月6日起担任“银华恒利灵活配置混合型证券投资基金”基金经理;自2015年6月17日起担任“银华信用双利债券型证券投资基金”基金经理;自2016年8月5日起担任“银华通利灵活配置混合型证券投资基金”基金经理;自2016年12月5日开始担任“银华上证10年期国债指数证券投资基金”基金经理;自2016年12月5日开始担任“银华上证5年期国债指数证券投资基金”基金经理;自2017年4月17日开始担任“银华中债-10年期国债期货期限匹配金融债指数证券投资基金”基金经理;自2017年4月17日开始担任“银华中债-5年期国债期货期限匹配金融债指数证券投资基金”基金经理。2018年7月加入长盛基金管理有限公司,现任固定收益部执行总监,自2018年12月6日起任长盛盛琪一年期定期开放债券型证券投资基金基金经理。拟任长盛安鑫中短债债券型证券投资基金基金经理。2020年9月29日起担任长盛盛裕纯债债券型证券投资基金基金经理。'
HanLP = hanlp.load(hanlp.pretrained.ner.MSRA_NER_ALBERT_BASE_ZH)
HanLP(test1)
```python
```
**Describe the current behavior**
启动不了
**Expected behavior**
正常启动
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04):
- Python version:
- HanLP version:
**Other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
* [x] I've completed this form and searched the web for solutions.
<!-- 发表前先搜索,此处一定要勾选! -->
<!-- 发表前先搜索,此处一定要勾选! -->
<!-- 发表前先搜索,此处一定要勾选! -->
Status: Issue closed
Answers:
username_1: 请不要自己安装第三方包,版本不对。请参考:https://hanlp.username_1.com/docs/install.html#import-error |
verdel/ipamcli | 236205137 | Title: Network prefix in add subcommand output
Question:
username_0: **Current output:**
The entry for ip 10.33.68.160/255.255.252.0 has been successfully created. The entry ID: 13019.
**Need output:**
The entry for ip 10.33.68.160/22 (255.255.252.0) has been successfully created. The entry ID: 13019.<issue_closed>
Status: Issue closed |
hpi-swt2/connections-portal | 752214448 | Title: Setup meeting with contact
Question:
username_0: Requires #36
- [ ] create invitation page
- [ ] invitations from other users are displayed with name and zoom-link
- [ ] add 'start meeting'-Button to send user a invitation
- [ ] on contact page #26
- [ ] on profile page #30<issue_closed>
Status: Issue closed |
dotnet/roslyn | 194895959 | Title: Performing a Find All References before another has finished, doesn't cancel existing search
Question:
username_0: 1. Clone this repo to this commit: https://github.com/dotnet/roslyn-project-system/commit/e8f444a1ebe3d78b6b494d5dc355e95b7da19086
2. From a VS command prompt, run build.cmd from the root
3. Open src\ProjectSystem.sln (build 26005.0 D15PreRel)
4. Open src\Microsoft.VisualStudio.ProjectSystem.Managed\Threading\Tasks\TaskResult.cs
5. Right-click on TaskResult and choose Find All References
6. Right-click on TaskResult and choose Find All References
7. Right-click on TaskResult and choose Find All References
Notice that the progress bar completes to 100% and takes 3 times as long as a single search.
Answers:
username_1: Hrmm. According to the docs the editor should be unsubscribing from the table source. However, looking through their code, i can't see where they would do that. So it might be right for us to preemptively remove those sources.
I'm guessing this regressed when the editor added multi-instancing to find-refs. They probably didn't want to cancel the existing work if the user locked the current window and a new window was being created. However, when that's not hte user flow we should def be cancelling. Will talk to Daofa about this.
username_1: I debugged through this and talked to Daofo. The actual issue is that the progress bar is not implemented well on their side.
Each time we call ReportProgress, they enqueue a UI request to update the progress bar. If we search N files, we'll report progress N times. Just processing all those messages takes forever, and also interferes when a new search starts.
FindRefs window needs to use a better approach for aggregating and reporting progress results that is not a 1:1 processing of each REportProgress call on the UI thread.
username_1: https://devdiv.visualstudio.com/DevDiv/_workitems?id=359162&fullScreen=false&_a=edit
Opened to track this.
Status: Issue closed
username_3: Mitigated with https://github.com/dotnet/roslyn/pull/15873 |
Hodge0699/McompProject | 391812332 | Title: Creating boss rooms
Question:
username_0: The procedural room generation script can now instantiate any boss room prefabs after a certain amount of rooms have been beaten but they need certain requirements to work.
1. A Room script attached to it with the dimensions set. (The more accurate the dimensions the more seamless the connection between rooms).
2. A 6 unit wide (z axis) hole in the left wall for the doorway.
3. A child object to the object Room is attached to named "Doors".
Other recommended things to make the room more uniform is for the walls to be 5 units high (y axis) and have y position of 2.5 units.
Once this is set up just drag the prefab into the procedural script and set the "Rooms" value to however many rooms you want the player to beat before it is spawned.
Answers:
username_0: All boss rooms made
Status: Issue closed
|
Cycling74/min-api | 1073756039 | Title: Expose more dictionary methods
Question:
username_0: It doesn't seem possible right now to get a count of keys in a dictionary (e.g. with `max::dictionary_getentrycount()`). It's possible from a _subdict_ because you can get an atom_reference and then get the size of that, but it's not straightforward for a plain dict.
I'm sure there are more methods that would be useful to wrap: https://cycling74.com/sdk/max-sdk-8.2.0/group__dictionary.html |
City-of-Helsinki/kerrokantasi-ui | 244408567 | Title: Display labels in sensible order
Question:
username_0: Ah, that's a good point. Will do it in the backend instead, so this isn't a frontend issue at all.
Status: Issue closed
Answers:
username_1: Does sensible mean in this case e.g. alphabetical? Is this sorting possible to be done in the backend?
username_0: Ah, that's a good point. Will do it in the backend instead, so this isn't a frontend issue at all.
Status: Issue closed
|
baaahs/sparklemotion | 480532464 | Title: Eye Controls pubSub crashes after UI changes tabs
Question:
username_0: **Steps to reproduce:**
1. Run `./gradlew runPinkyJvm` and `./gradlew runBrainJvm`
1. Go to localhost:8004
1. Click on the tab: `Eye Controls`
1. Click on the `Shows` tab
1. Click back to the `Eyes Controls` tab, then try to control the eye
**Expected Result:**
Should move the eyes
**Actual Result:**
UI freezes, some infinite loop sometimes occurs, then the UI crashes |
ie3-institute/PowerSystemDataModel | 852469981 | Title: Implement support for scenario files
Question:
username_0: At the moment we do not support different grid scenarios e.g. with different combinations of enabled or disabled assets within the same grid model. However, this functionality might become a requirement soon.
One way to implement a scenario support would be to post-process a `GridContainer` with additional information about the asset operation times for different scenarios. Such information could be structured as
`uuid, operatesFrom, operatesUntil`.
This could then be read in after the `GridContainer` has been build and manipulate all assets accordingly. To avoid the overhead of rebuilding the whole grid, one could also consider include this scenario modification information in the general reading process. However, this would require all connectors to be altered.
Feel free to discuss different approaches below here as well. |
simapetr/uart_terminal | 569547151 | Title: Application memory overload
Question:
username_0: # Fix Memory expansion
After receive UART Application alocate memory but do not release it.
## 1. Find error
Find module with bug unconditionally require check
- JS receive function call.
- Application to UART interface
- Uart DRIVER
## 2. Fix error and update driver interrface
- Update method name in UART driver.
- Implement new functionality<issue_closed>
Status: Issue closed |
philipperemy/keras-tcn | 363993430 | Title: compatibility with plaidml backend
Question:
username_0: Hello i use plaidml as keras backend because it supports AMD graphics cards
there are two issues with your implementation
1) spatialdropout is not supported by plaidml so
x = SpatialDropout1D(dropout_rate, name='spatial_dropout1d_%d_s%d_%f' % (i, s, dropout_rate))(x)
will not work
https://github.com/plaidml/plaidml/issues/165
2) it seems that keras.backend.max isn't implemented too
If you don't have time to fix it can you suggest a workaround?
Answers:
username_1: @username_0 yes sure.
1. Instead of SpatialDropout1D you can simply use Dropout. It will be slightly different but results should be the same.
2. We apply max() in the channel normalization. In that case you may use BatchNormalization!
Status: Issue closed
username_1: Inactivity. Please re-open when you have something new! |
hyperledger/besu | 1122624281 | Title: SPIKE: add metrics for RocksDB
Question:
username_0: Currently we have no metrics around RocksDB. Investigate what would be involved. What metrics make sense?
### Description
As an [Actor], I want [feature] so that [why].
### Acceptance Criteria
* [Criteria 1]
### Steps to Reproduce (Bug)
1. [Step 1]
2. [Step 2]
3. [Step ...]
**Expected behavior:** [What you expect to happen]
**Actual behavior:** [What actually happens]
**Frequency:** [What percentage of the time does it occur?]
### Versions (Add all that apply)
* Software version: [`besu --version`]
* Java version: [`java -version`]
* OS Name & Version: [`cat /etc/*release`]
* Kernel Version: [`uname -a`]
* Virtual Machine software & version: [`vmware -v`]
* Docker Version: [`docker version`]
* Cloud VM, type, size: [Amazon Web Services I3-large]
### Additional Information
Answers:
username_1: @username_0 it seems that RocksDB metrics are already collected, see `RocksDBMetricsFactory` for example, but I am not able to see them exposed by Prometheus, so there could be some issues on exposing these metrics.
username_2: That category may just be disabled by default. From memory there were some categories disabled because they included timers I think which had a significant performance impact. Quite possibly we should just remove the timers and enable the rest.
username_0: ok thanks @username_1 and @username_2 - that gives us somewhere to start! |
2sic/2sxc | 643183992 | Title: Razor Snippet for List Presentation Fields Uses ListContent.Presentation instead of Header.Presentation
Question:
username_0: <!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION WE MIGHT CLOSE YOUR ISSUE WITHOUT INVESTIGATING
-->
**I'm submitting a ...**
<!-- remove lines which don't apply -->
[x] bug report => search github for a similar issue before submitting
**...about** <!-- remove lines which don't apply -->
[x] Razor templating
**Current behavior**
<!-- Describe how the bug manifests. -->
When you're working with a list and you use the snippets to insert the Presentation fields for the list, the snippets are using the depricated ListContent instead of Header object, which causes errors in the template display. See video below
**Expected behavior**
<!-- Describe what the behavior would be without the bug. -->
The snippet should use Header.Presentation.FIELD_NAME
**Minimal reproduction of the problem with instructions**
<!--
If the current behavior is a bug or you can illustrate your feature request better with an example,
please provide the *STEPS TO REPRODUCE* and/or a *MINIMAL DEMO* like a screenshot or screencast
-->
https://www.screencast.com/t/lyn1pEnu
**What is the motivation / use case for changing the behavior?**
<!-- Describe the motivation or the concrete use case -->
Causes confusion to the developer/programmer. The predefined snippets should insert the correct object
**Please tell us about your environment:**
<!-- Operating system, IDE, package manager, HTTP server, ... -->
Windows Server 2012, SQL Server 2016
* **2sxc version(s):** 11.2.0 <!-- Check whether this is still an issue in the most recent version -->
* **Browser:** [all ]
* **DNN:** [9.6.1 ]
* **Language:** [any]
**Anything you would like to add**
Answers:
username_1: Thanks for the video, that helped a lot. Love these kinds of issues 👍
Status: Issue closed
|
diffblue/cbmc | 192854575 | Title: CBMC crashing with byte_extract flatting with non-constant size: byte_extract_little_endian
Question:
username_0: When running with a sufficently large `--depth` argument, CBMC can produce the following error:
`byte_extract flatting with non-constant size: byte_extract_little_endian`
Full log: [json_decode_error.txt](https://github.com/diffblue/cbmc/files/624752/json_decode_error.txt)
Source file: https://github.com/jpmens/jo/blob/2bedfd486f8f4a79b1865e370e6c858eb04257f5/json.c
Command line: `cbmc json.c --function json_decode --depth 150 --cover location`
Answers:
username_1: This has been fixed in #795 (237d8be12d).
Status: Issue closed
|
adobe/aem-core-wcm-components | 378639537 | Title: [Tabs] Tab focus behaviour too invasive
Question:
username_0: The focus handling in the tab component is too invasive. Tabs are focused each time the component is initialised and navigated. This leads to a situation where the tabs grab the focus for a page on first load.
Focus should only be managed for a tab selection change.
CQ-4257090
Status: Issue closed
Answers:
username_0: This was added with the merge of https://github.com/adobe/aem-core-wcm-components/pull/370 and is on master. It will be available in the upcoming [2.2.2](https://github.com/adobe/aem-core-wcm-components/milestone/9) release. |
naser44/1 | 88291627 | Title: هل_تعرف أحد الأنشطة الاجتماعية في اليابان هو البكاء الجماعي ! حيث يجتمع الأصد...
Question:
username_0: إقرأ المزيد<br>
.<br>
.<br>
.<br>
http://ift.tt/1MCB2gp<br>
المصدر<br><a href="http://ift.tt/1MpfvtA">عمو</a> |
aws/aws-sdk-js-v3 | 794641892 | Title: Kinesis video signaling
Question:
username_0: **Describe the bug**
Kinesis video signaling client `GetIceServerConfigCommand` http request blocked by CORS policy
**SDK version number**
3.3.0
**Is the issue in the browser/Node.js/ReactNative?**
Browser
**Details of the browser/Node.js/ReactNative version**
"Request header field amz-sdk-invocation-id is not allowed by Access-Control-Allow-Headers in preflight response."
**To Reproduce (observed behavior)**
Send `GetIceServerConfigCommand` from `KinesisVideoSignalingClient` in Chrome browser.
**Expected behavior**
Request should succeed.
Answers:
username_1: Getting this issue as well, any fixes in the works? It looks like this was mentioned in #1376 but was never resolved.
username_1: #1789 solved this for other services but KVS was not included. @AllanZhengYP I see KVS was mentioned in #1376 but was perhaps forgotten when implementing the fix
username_0: @username_1 I opened a PR to address this: #1963
Status: Issue closed
username_2: **Describe the bug**
Kinesis video signaling client `GetIceServerConfigCommand` http request blocked by CORS policy
**SDK version number**
3.3.0
**Is the issue in the browser/Node.js/ReactNative?**
Browser
**Details of the browser/Node.js/ReactNative version**
"Request header field amz-sdk-invocation-id is not allowed by Access-Control-Allow-Headers in preflight response."
**To Reproduce (observed behavior)**
Send `GetIceServerConfigCommand` from `KinesisVideoSignalingClient` in Chrome browser.
**Expected behavior**
Request should succeed.
username_2: Reopening as call to `get-ice-server-config` supports retry headers as per CORS URL in [cors-test](https://github.com/AllanZhengYP/cors-test/blob/master/output.md)
```console
$ curl --insecure -i -X OPTIONS https://kinesisvideo.us-east-1.amazonaws.com/v1/get-ice-server-config -H "User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/84.0.4147.135 Safari/537.36" -H "Origin: https://foo.example" -H "Access-Control-Request-Method: POST" -H "Access-Control-Request-Headers: content-type,content-length,host,x-amz-user-agent,user-agent,Authorization,amz-sdk-invocation-id,amz-sdk-request,x-amz-security-token" -H "host: kinesisvideo.us-east-1.amazonaws.com"
HTTP/1.1 200 OK
x-amzn-RequestId: 1ccc9c64-9c4d-4077-823d-234a33c155b8
Access-Control-Allow-Origin: *
Access-Control-Allow-Headers: content-type,content-length,host,x-amz-user-agent,user-agent,Authorization,amz-sdk-invocation-id,amz-sdk-request,x-amz-security-token
Access-Control-Allow-Methods: POST
Access-Control-Expose-Headers: x-amzn-RequestId,x-amzn-ErrorType,x-amzn-ErrorMessage,Date
Access-Control-Max-Age: 172800
Content-Length: 0
Date: Tue, 16 Feb 2021 21:50:05 GMT
```
@username_0 Do you have a minimal repro code to reproduce this issue with `@aws-sdk/[email protected]`?
username_0: @username_2 Here's minimal code that should re-pro:
```javascript
import {
GetIceServerConfigCommand,
KinesisVideoSignalingClient
} from "@aws-sdk/client-kinesis-video-signaling";
const region = "us-west-2";
const channelARN = "<REPLACE>";
const endpoint = "<REPLACE>";
const credentials = {
accessKeyId: "<REPLACE>",
secretAccessKey: "<REPLACE>"
};
const kinesisVideoSignalingChannelsClient = new KinesisVideoSignalingClient({
region,
credentials,
endpoint
});
const getIceServerConfigCommand = new GetIceServerConfigCommand({
ChannelARN: channelARN
});
// This operation is blocked by CORS policy
const sigChannelResponse = kinesisVideoSignalingChannelsClient
.send(getIceServerConfigCommand);
``` |
theCrag/website | 628796506 | Title: Topo editor: right click menu not open reliably / closes instantly
Question:
username_0: **What happened?**
While editing a topo, some actions like moving the route number is only possible in the right click menu, other actions like 'delete point' easier to control. While the right click menu does show reliably, it does not stay open/visible reliably, but in some combinations close instantly (e.g. Firefox with easyGestures N).
**What you expected:**
* [ ] Avoid closing the context menu
and/or
* [ ] Open the context menu not only on right klick but also double clik or long click (relates to https://github.com/theCrag/website/issues/2345) |
johanneswigh/-miniex | 313224190 | Title: Feedback on mini_ex8 by <NAME>
Question:
username_0: Hey Johannes and Mikael.
I decided to give both of you feedback, but I didn’t realize that you two worked together. So, I’m going to give you feedback together but then comment separately on what you would like to work with further on.
It is unfortunately that your link doesn’t work. I tried downloading your code, but I still couldn’t get it to work, but I still really enjoyed reading your readme and looking at your screenshots. You have created a program, where the user can produce a new picture for every different search word. I thought that was a funny idea, and it fix your guys personality perfectly. I like that you had a deeper meaning to the program, by adding the biggest company names to each search word. You are trying to show the lack for fluency in for example google searches. That is a great element to the program.
You mention that you had problems with your API, which I found is a common thing for most of the students in the class. I like how you briefly addressed the problem but tried to find a solution by suggesting an alternative route to run your program.
Things you could improve:
- Looking at your screenshots; your program looks pretty plain. Maybe you could make it more interesting or “fun” looking, so it would match the function of your program.
Johannes, I like your idea of researching more into APIs and their possibilities for creating new content, but I really think you should focus on what impact the sharing of data, especially personal data, has on the users. I believe that would be a great topic.
Great job guys! |
IFRCGo/go-frontend | 876032948 | Title: Refactor Field Report Form
Question:
username_0: **Is your feature request related to a problem? Please describe.**
Most of the Field Report Form code is obsolete. It uses method of coding that was used few years ago. Since then, React itself has updated a lot and has introduced better and easier way, which allows better, cleaner and less error-prone coding. If the Field Report Form is left as-is, It would be a bit difficult and time consuming to add features or modify.
**Describe the solution you'd like**
Refactor the Field Report Form to use better code, form validation
Answers:
username_0: Available for testing on: https://ifrc-go-feature-new-field-report-form.surge.sh/reports/new
username_1: @username_0 this is some really amazing work here.
@username_2 - would you be able to spend some time testing https://ifrc-go-feature-new-field-report-form.surge.sh/reports/new to create field reports, and verify that all the information on the created field report is correct? Will be good to try a few combinations of things, there likely will be minor issues.
username_1: @username_2 to note: @username_0 is working on fixing the visual / CSS issues to make it look like the previous form, but it will be great if you can test the functionality, the validation, saving, etc. Thanks!
username_2: @username_0 @username_1, here my findings testing the creation of different Field Reports
- The required title field validation is giving a error even when the title field is filled.

- In a event related to Covid, in the Situation tab, when the "Number of new cases since last Field Report" is filled, the "Number of new deaths since last Field Report" field is also auto filled with the same above input number.

- In a event related to Covid, in situation tab, should the Source of figures and Date of data be required fields when the numeric fields is filled? If so, the validation of required fields is not working here.
- When a FR is created, these fields are not displayed on the FR page. It seems that the imputs have not been saved, because when the FR is edited, those fields appear clean too.

In a FR no related to covid and a FR of Early Warning also there are numbers that are not been displayed.

- In an Early Warning FR, the sources for the numerical detail appear as not required unlike the sources in a event no related to covid where they are required.

username_0: Could you please take a look at it again? Thanks again!
username_2: @username_0, all the above things is working good now :+1:
username_0: # Summary of the refactor
### Identified Report types (EW, EVT, EPI, COV)
- Early Warning / Early Action
- Event
- Epidemic
- Covid
- Other epidemic
- Other event
### Form Steps
#### Step1: Context
- The type of report will be determined from this step with combination of various fields
- Conditions for different report types
- EW: `Status` is `Early Warning / Early Action`
- COV: `COVID-19 Related Event` is `Yes`
- EPI: `Disaster type` is `Epidemic`
- EVT: `Status` is `Event`
- `COV` is subtype of `EPI` and `EPI` is subtype of `EVT`
#### Step2: Risk Analysis / Situation
- `Risk Analysis` will appear for `EW` reports
- Mostly contains numeric details / sources, source details and risk analysis
- `Situation` will appear for other report types (`EVT, EPI, COV`)
- Mostly contains numeric details / source
- Fields will depend on type of report
- For epidemic: Cumulative cases, suspected cases, confirmed cases, etc
- For covid: Fields will be slightly different than epidemic
- For event: injured, dead, missing, etc
#### Step3: Early Action / Action
- `Early Action` will appear for report type `EW`
- `Action` will appear for other types
- Will have similar type of fields for all report type with some added / removed fields
#### Step4: Response
- Will have similar type of fields for all report type with some added / removed fields
### Known issues in old system (Currently at staging server)
- `EPI` type report doesn't show intended fields for the Step2: Situation
- Any other `Disaster type` than the `Epidemic` can be selected for `COV` type of report along with the `Early Warning / Early Action` as `Status` which breaks the form when we proceed to Step2
- A report that is selected to be a `COV` type initially breaks the Step3: Action when we switch the report to `EPI` (i.e. set `COVID-19 Related Event` to `No`)
- When edition a field report from the field report detail page, sometimes previous data doesn't appear (It appears as new field report form)
### Notes for testing
- Check if the new system (Test server) matches closely with old system (Staging server). The known issues, mentioned above, should also be fixed
- Check if the above mentioned logic, steps for the form is functional
- Some of the labels, description of the fields change according to value of some other field (For example: In Step1: Context, date field has different label and description when disaster type is epidemic)
- Disaster type is not `Epidemic` 
- Disaster type is `Epidemic`

- Some fields become conditionally required depending on the value of some other field (For example: As shown in the figure below, estimation becomes required when source is filled or vice-versa) 
- Some fields become conditionally optional when other field is filled. For example: In Step1: Context, `Title` becomes optional when we link the report to an existing emergency.
- Title is required here 
- Title is optional here 
username_3: When changing from Epidemic type of Event to EW/EA and back to Event, then a error message indicates the missing Disaster type. Disaster type should be cleared, but without error message. Error message should only appear when user tries to step to the next page.


username_1: @username_2 could you give this another test, based on @username_0's comment above?
username_2: @username_1 , all is working as @username_0's comments and the field required message when a epidemic report is changed to Early warning reported by @username_3 is still there.
username_3: @username_0, here is a text duplication on Action tab, only 'Information Bulletin' should appear in bold text in the first line:

username_3: @username_0, noticed that some of the field values are not saved:
Situation tab:
- Source details
- Situation overview
Actions tab:
- Actions Taken by Others (Governments, UN)
username_0: @username_3 This is expected behaviour (i.e. we set the validation to immediately trigger). This behaviour is consistent with other places as well, where the validation is immediately triggered when related fields are changed. Let me know if we should change this behaviour.
username_1: @username_2 could you do one final test of the Field Report form functionality that everything looks good after all the latest changes?
username_4: Some feedback from the UAT regarding Field Reports:
- [ ] In the Epidemic FR the actions taken by the IFRC is showing only 3 actions, this list should match the list used for Actions taken by National Society and for any other RCRC actor
- [ ] When No on international assistance requested is checked, it doesn't shows up on the review page, the Yes is working correctly
- [ ]
@username_1
username_2: @username_1 this is what I found
- In EW field report, the title Affected Pop Centres that have as source "Government" appears only as "Affected".

- In the edition mode of a FR, the data of "Region / Province" and "Date of Data" does not load.
username_0: @username_4 Thanks for the catch. I've fixed the "No" not being showed for the "International assistance requested". For the Epidemic FR actions taken by the IFRC, it appears different in the staging because it's loaded from the API. It'll be same as other 2 when it goes to production. I verified this in my local setup using the prod API.
Status: Issue closed
|
lbryio/lbry-desktop | 1157805230 | Title: Option to display blacklisted videos
Question:
username_0: **Is your feature request related to a problem? Please describe.**
As far as I checked the master branch, the LBRY desktop get the data for blacklistes videos, but then with a conditionnal component in the UI, it doesn't displays it. Blacklisted videos being available, it would be great to have an option to have it disabled by default for standard users, and enabled for users who would like to see the content being shared.
**Describe the solution you'd like**
I see this feature as the option for mature content. Off, we do not see mature content, on we see it. Off we don't see blacklisted content, on we see it. A simple checkbox would be great.
**Describe alternatives you've considered**
An alternative would be for example share their metadata with additional servers added by a user. This option of running our own server to enable people to get blacklisted content would be more difficult than using a checkbox though, and would divide the network like on Peertube (where people have a different view of the network according to "friend" servers)
**Additional context**
No additional context. |
CyclopsMC/FluidConverters | 256455851 | Title: [1.12] Default recipe missing
Question:
username_0: FluidConverters-1.12-2.2.6
Using default config for blood.json with EvilCraft and Tinker's Construct loaded.
```
],
"hasRecipe": true,
"hasDefaultRecipe": true
}
```
No recipe in JEI. Assembling the default recipe on the wiki using a blood bucket from either mod does not produce a recipe output.
-----
`FluidConvertersRecipeHandler` calls a function `FluidGroup.isHasDefaultRecipe()`, but I can't find that function in `FluidGroup.java`. The `FluidGroup` class does have a private boolean `hasDefaultRecipe`, but I don't see where it is set or used.
-----
Aside: The example configs `blood.json` and `example.json.template` contain an element `hasRecipe`, which looks like it was removed [here](https://github.com/CyclopsMC/FluidConverters/commit/d22deadc65ad5eca7fc7eaf88346140db622697a). Can you update the examples to the newest config format? Thanks.
Answers:
username_0: I was also unable to get a custom recipe working. The log contained this error.
```
[main/ERROR] [FluidConverters/]: Could not find the predefined item for key 'fluidconverters:fluidconverter_blood'.
```
Status: Issue closed
username_1: New release with a fix for this should appear on CurseForge soon. |
phenopackets/phenopacket-schema | 419531573 | Title: Add surgical procedure type for biosample
Question:
username_0: For instance, prostatocystectomy is used to remove bladder cancer, and a biopsy is taken from the resulting specimen. On the other hand, a smaller biopsy could be taken by cystsocopy.
There are NCI thesaurus terms to represent these procedures.
@cmungall @mellybelly @username_2 @username_1 @drseb
Answers:
username_1: @username_0 Very good point for improvement. We're playing around with designing a first level `provenance` attribute, where this could be part of, as an optional named object. Example here, using EFO:
```
"provenance" : {
"material" : {
"description" : "normal kidney",
"type" : {
"id" : "EFO:0009654",
"label" : "reference sample"
}
},
"collection_process" : {
"description" : "percutaneous needle biopsy from left kidney",
"type" : {
"id" : "EFO:0009293",
"label" : "percutaneous kidney biopsy"
}
},
"geo" : {
"label" : "New York City, United States",
"precision" : "city",
"city" : "New York City",
"country" : "United States",
"latitude" : 40.71,
"longitude" : -74.01
}
}
```
The example doesn't include a time component etc. Also, looking at EFO, there are __gaps__ to fill...
Only first ideas here; so please comment!
username_1: (The `collection_process` was written _ad hoc_, based on Peter's creation of the issue. Other instantiations & examples encouraged!)
username_2: @username_1 where are you discussing the provenance attribute? Are you referring to [this issue](https://github.com/ga4gh-metadata/SchemaBlocks/issues/7) if so, could you or @mcourtot add the information in the Google doc to the issue, or open up the Google doc.
username_2: Commenting on the ```collection_process```
```json
"collection_process" : {
"description" : "percutaneous needle biopsy from left kidney",
"type" : {
"id" : "EFO:0009293",
"label" : "percutaneous kidney biopsy"
}
}
```
The 'left' is getting lost in a comment. This is likely important, so ought there not be a ```laterality``` field in there? Or perhaps relative location, to take into account front/back, medial/lateral.
username_1: @username_2 I used `provenance` here in a physical sense (a.k.a. "sample origin"); this is just meant as an example - @mcoutot 's issue is about evidence modelling.
Re. kidney example: Yes, that is intentional - I actually did put in the "left" for making a case about incomplete ontologies ... See note about "gaps".
(IMO the "left" could be expressed by having a subclass.)
username_2: @username_0 surgical procedure sounds like it ought to be a message in its own right.
Status: Issue closed
username_2: https://github.com/phenopackets/phenopacket-schema/blob/696d4faeac5d99ad337a8801319f7b4ea125c7cf/src/main/proto/org/phenopackets/schema/v1/core/base.proto#L261-L275
username_2: For instance, prostatocystectomy is used to remove bladder cancer, and a biopsy is taken from the resulting specimen. On the other hand, a smaller biopsy could be taken by cystsocopy.
There are NCI thesaurus terms to represent these procedures.
@cmungall @mellybelly @username_2 @username_1 @drseb
username_2: Might want to add some docs about the preferred ontologies for each field?
username_1: @username_2 @username_0 So my example would translate to:
```
"collection_process" : {
"description" : "percutaneous needle biopsy from left kidney",
"procedure" : {
"code" : {
"id" : "EFO:0009293",
"label" : "percutaneous kidney biopsy"
},
"body_site" : {
"id" : "UBERON:0004538",
"label" : "left kidney"
},
}
}
```
Looks good to me...
username_2: Cool
Status: Issue closed
|
wujichao/hipda_ios_client_v3 | 233348111 | Title: FrontBoardServices
Question:
username_0: #### in __destroy_helper_block_.387
* Number of crashes: 1
* Impacted devices: 1
There's a lot more information about this crash on crashlytics.com:
[https://fabric.io/solo2/ios/apps/username_0.hipda/issues/59324b53be077a4dcc9b8878](https://fabric.io/solo2/ios/apps/username_0.hipda/issues/59324b53be077a4dcc9b8878) |
facebook/react | 1158260319 | Title: Bug: react stay with Ukraine, who stay with the open source?react支持乌克兰,谁支持开源技术?
Question:
username_0: <!--
Please provide a clear and concise description of what the bug is. Include
screenshots if needed. Please test using the latest version of the relevant
React packages to make sure your issue has not already been fixed.
-->
React version:
## Steps To Reproduce
1.
2.
<!--
Your bug will get fixed much faster if we can run your code and it doesn't
have dependencies other than React. Issues without reproduction steps or
code examples may be immediately closed as not actionable.
-->
Link to code example:
<!--
Please provide a CodeSandbox (https://codesandbox.io/s/new), a link to a
repository on GitHub, or provide a minimal code example that reproduces the
problem. You may provide a screenshot of the application if you think it is
relevant to your bug report. Here are some tips for providing a minimal
example: https://stackoverflow.com/help/mcve.
-->
## The current behavior
## The expected behavior
Answers:
username_1: 可笑的是react还在使用nginx
username_0: react作为优秀的技术被大家所认可,现在却被当做一个政治宣传的工具,何其悲哀? |
dart-lang/dart_style | 533399281 | Title: Keep collapsed blocks
Question:
username_0: When I format the code, everything that is collapsed just expands again. Which isn't convenient when trying to focus on one block of code, and you just wanna format that block without having to search that block again, since expanding everything will scroll the editor.
Answers:
username_1: Please give an example of code that you try to format, and your expected result, and the actual result.
Status: Issue closed
username_2: Duplicate of https://github.com/dart-lang/dart_style/issues/876
Code folding in your IDE isn't something that can be controlled by `dartfmt`. Please follow up with the issue tracker for the Dart plugin on whichever IDE you are using. See the linked issue for issue tracker links for intellij and VS code. |
Azure/azure-cli | 658822601 | Title: Error in the description
Question:
username_0: [Enter feedback here]
Error is the description:
1. **--endpoint-dest-address**
Error in description, it should be "Address of the destination of connection monitor endpoint"
2. **--endpoint-source-resource-id**
If endpoint is intended to used as source, this option is required. The resource ID should be the source of connection monitor endpoint instead of destination as documented.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 0800fdc3-05ed-9fa8-888a-a8997c6a5a8e
* Version Independent ID: 5d6188b7-5eb4-835d-d63a-32aa1a1f49ca
* Content: [az network watcher connection-monitor](https://docs.microsoft.com/en-us/cli/azure/network/watcher/connection-monitor?view=azure-cli-latest)
* Content Source: [src/azure-cli/azure/cli/command_modules/network/_help.py](https://github.com/Azure/azure-cli/blob/dev/src/azure-cli/azure/cli/command_modules/network/_help.py)
* Service: **virtual-network**
* GitHub Login: @rloutlaw
* Microsoft Alias: **routlaw**
Answers:
username_1: add to S174
Status: Issue closed
|
trdischat/grouproll | 911277393 | Title: PF2 1.13+ onward changed strings, need to be updated
Question:
username_0: Strings under PF2 system are broken as of 1.13. This is because the system changed how they name their strings and they now appear with their technical names instead. Otherwise functionality is intact.
Status: Issue closed
Answers:
username_1: I think I fixed this. Please test new version 0.7.3 and let me know if it works for PF2e now. Thanks. |
episphere/quest | 643060327 | Title: If-then-else syntax
Question:
username_0: 3. Can you please add syntax of if-then-else statements in WiKi?
[IF CIG5 = 01 or NO RESPONSE, fill “smoke” and “do”
IF CIG5 = 02 or 03, fill “smoked” and “did”]
[CIG9A?] On days that you smoke|displayif=Or(equals(CIG5,2), equals(CIG5,3))|d|, - this does not work
Answers:
username_1: Thank you @username_0 checking it out
username_1: Hi @username_0 ,
The documentation is correct, however, I think there is a bug in the code.
I was able to reproduce your error, but first, the function is "or" not "Or",
In your example there is whitespace in between the comma and the second equals in the or:
`|displayif=or(equals(CIG5,2), equals(CIG5,3))|d|`
When I remove the space, it works. However, this is a bug as the space should not matter.
Here is a minimal working example for your bug...
`
[CIG5] enter a number |__|__|
[CIG9A?] On days that you smoke|displayif=or(equals(CIG5,2),equals(CIG5,3))|d| this does not work
[END,end]
`
Nick should be able to get this up and running quickly...
username_0: [IF CIG5 = 01 or NO RESPONSE, fill “smoke” and “do”
IF CIG5 = 02 or 03, fill “smoked” and “did”]
I still do not know how to script the code for do/did.
username_0: Another complex logic I am struggling with is
IF (CIG4 = 01) OR ((CIG4 = 02, 03, OR 04) AND CIG5 = 02 AND CIG7B = 01) OR ((CIG4 = 02, 03, OR 04) AND CIG5 = 03 AND CIG8C = 01), GO TO CIG10
ELSE, GO TO CIG9B
username_1: |displayif=equals(isDefined(CIG5,1),1)|smoke||displayif=or(equals(isDefined(CIG5,1),2),equals(isDefined(CIG5,1),3))|smoked|
username_1: Can you open that next question as a new issue?
username_2: Conditional IF added to the wiki
Status: Issue closed
|
inspector-apm/inspector-laravel | 1162363667 | Title: v4.7.20 deprecated in PHP 8.1.3
Question:
username_0: **Describe the bug**
On PHP 8.1.3, when using inspector-laravel 4.7.20 we're getting PHP Deprecated in log:
[08-Mar-2022 09:20:26 Europe/Warsaw] PHP Deprecated: Return type of Inspector\Models\Arrayable::offsetExists($key) should either be compatible with ArrayAccess::offsetExists(mixed $offset): bool, or the #[\ReturnTypeWillChange] attribute should be used to temporarily suppress the notice in /[...]/vendor/inspector-apm/inspector-php/src/Models/Arrayable.php on line 112
**To Reproduce**
Steps to reproduce the behavior:
Install inspector-laravel on Laravel 9 with PHP 8.1.3, look for error log
Answers:
username_1: Hi @username_0 thank you for reporting the issue. I'll update you when the fix is released.
username_2: @username_1 +1
We cannot use inspector service because of the issue. Please fix it ASAP
username_1: Hi, the issue was caused by the @inspector-apm/inspector-php basic library. It was fixed by the latest reòease 3.7.9
Run composer update to get the latest version.
Status: Issue closed
|
asascience/ott-coops | 1083447104 | Title: Add site metadata
Question:
username_0: As we're approaching a public release, we should start setting up metadata for OM3.
This is a big nice to have, especially for sharing and seo.
<img width="835" alt="Screen Shot 2021-12-17 at 11 49 27 AM" src="https://user-images.githubusercontent.com/43226410/146579416-54b3ebae-3519-420f-ae12-499e9b046ff2.png">
Since we're not doing SSR atm, this should be relatively quick and simple - and generic'ish
Answers:
username_1: Yeah this also leads into the fact that we need some better infrastructure for swapping metadata, and themes (colors) on the fly
username_0: Putting this on hold til we have SSR in place.
username_1: Been fighting nextjs all day and I think long term it's a good solution to our injecting configs problem, but it's very hard...
In the meantime, I did find this, and with us being able to control colors dynamically now we can specify our metadata in env vars that gets us to where we need to be for the maracoos release https://create-react-app.dev/docs/adding-custom-environment-variables/
username_1: Nextjs is super SSR dependent and we dont benefit enough for the effort needed. Instead we can use .env to mess wiht most of the metadata and it works pretty well.
Maybe long term we get SSR going but its not worth the time sink right now
username_2: Is this issue all set @username_1 ? |
kubernetes/cloud-provider-aws | 992608384 | Title: EKS 1.21 Service type=LoadBalancer does not support ServiceAppProtocol
Question:
username_0: `ServiceAppProtocol` is supposed to be `stable` in [1.20+](https://kubernetes.io/docs/concepts/services-networking/service/#application-protocol).
**Scenario**:
Applying
```yaml
apiVersion: v1
kind: Service
metadata:
name: test-lb
annotations:
#service.beta.kubernetes.io/aws-load-balancer-backend-protocol: http
service.beta.kubernetes.io/aws-load-balancer-ssl-cert: arn:aws:acm:us-east-1:<redacted>:certificate/<redacted>
service.beta.kubernetes.io/aws-load-balancer-ssl-ports: https
spec:
type: LoadBalancer
ports:
- name: http
port: 80
targetPort: http
appProtocol: http
- name: https
port: 443
targetPort: 80
appProtocol: https
- name: 5671-tcp
port: 5671
targetPort: 5671-tcp
appProtocol: tcp
```
creates a Classic LB with the following listeners:

If `service.beta.kubernetes.io/aws-load-balancer-backend-protocol: http` is uncommented:

**Expected**:
`appProtocol` takes care of LB Listener protocol, something like:

- Kubernetes version:
```
Client Version: version.Info{Major:"1", Minor:"22", GitVersion:"v1.22.1", GitCommit:"<PASSWORD>", GitTreeState:"clean", BuildDate:"2021-08-19T15:38:26Z", GoVersion:"go1.16.6", Compiler:"gc", Platform:"darwin/amd64"}
Server Version: version.Info{Major:"1", Minor:"21+", GitVersion:"v1.21.2-eks-0389ca3", GitCommit:"<PASSWORD>", GitTreeState:"clean", BuildDate:"2021-07-31T01:34:46Z", GoVersion:"go1.16.5", Compiler:"gc", Platform:"linux/amd64"}
```
<!-- DO NOT EDIT BELOW THIS LINE -->
/kind bug |
h4cc/awesome-elixir | 130824884 | Title: Add Package "xe"
Question:
username_0: Add the Package "xe" from hex.pm
Link: https://hex.pm/packages/xe
Description:
Real time conversion for currencies
Answers:
username_1: What is the proper category for this package `Third Party APIs` or `Text and Numbers` :thought_balloon:
username_0: Since its calling a remote service, i would think so too.
username_1: :ok_hand:
Status: Issue closed
|
flutter/flutter | 485623540 | Title: cupertino_icons is not all
Question:
username_0: 
just a few icons in the icons.dart, not all.
This is just a case that not completed for the flutter package. and so many package is not completed.
still a
Answers:
username_1: CC @willlarche
username_0: maybe we can just write a script loop the icons to the icons.dart, not naming every icon
username_2: Merging into #16102
Status: Issue closed
|
costoffreedom/costoffreedom.cc | 119105932 | Title: Handle book links to main site after book moves to subdomain
Question:
username_0: Two obvious approaches:
* [404](https://help.github.com/articles/custom-404-pages/) page, might be able to have javascript redirect to pages on book subdomain
* Brute force, stub for each book page on main site with js redirect to corresponding book subdomain page.
As mentioned in #34
Answers:
username_1: Don't use Javascript: let the server do it and keep track of URLs.
https://help.github.com/articles/redirects-on-github-pages/
It's not a nice nginx rewrite one-liner, but we can degrade gracefully
without having to resort to javascript in this case.
==
hk
--
_ _ We are free to share code and we code to share freedom
(_X_)yne Foundation, Free Culture Foundry * https://www.dyne.org/donate/
username_0: That uses javascript and only works within one jekyll site.
username_2: Structure is similar on both meaning that for instance http://costoffreedom.cc/book/opening:freedom/ is now http://book.costoffreedom.cc/book/opening:freedom/
username_0: This should be done in 990484524cd5071f34d49f5fad67f9ed16361898 and fa9dd7456d2c294baf8be33861be7892509f56ee with a note added in 07b6ab73dda8a5716fce562c0d844849ada29722
I didn't notice til now 35d5e0b and #50 ... I like the 404 handler as a backup. I'd rather not completely remove the initial book URLs, especially the PDF.
username_0: That said I don't care that much. @username_2 feel free to just switch gh-pages to the 'clean' branch if that's what you think is best.
username_2: I added the 404 as well, so we have both :) I think we can close this now.
Status: Issue closed
|
tomasbjerre/simple-bitbucket-commit-checker | 572724369 | Title: Error: The settings must only have values of type String, Boolean or Number.
Question:
username_0: SBCC: 3.12
BitBucket Server: 5.7.0
After the "Enable" button is clicked, we get the error: "The settings must only have values of type String, Boolean or Number."
 |
schnack/pdu_sms | 304299208 | Title: ArgumentError: The message is too long
Question:
username_0: Hello again,
This time I have the following PDU:
07912470338016004404B96430001181301190404340A0050003CE0201886FB13C0F2297DD2C903E0C9A81C8EE3CC81A6E83DC61D0383D6797413618ED668BD96035D07CFD768FD3A037B95D768741FA3088FC16A7E969D0B40E2FCB41F3B05E1C76819A757D995E06CDD320751A14669741737798EC7E83F4E17919A40FC3DDEF3ADD055286D73FD0941E1EA741F33488FC16A7E9A0B5BC4C4ED341F0375D5E16BBDF
Decoding attempt produces the following error in the latest version of your gem (1.1.2):
ArgumentError: The message is too long
/home/pawelpacholek/.rbenv/versions/2.4.0/lib/ruby/gems/2.4.0/gems/pdu_sms-1.1.2/lib/pdu_sms/user_data.rb:196:in `_check_message'
/home/pawelpacholek/.rbenv/versions/2.4.0/lib/ruby/gems/2.4.0/gems/pdu_sms-1.1.2/lib/pdu_sms/user_data.rb:12:in `initialize'
/home/pawelpacholek/.rbenv/versions/2.4.0/lib/ruby/gems/2.4.0/gems/pdu_sms-1.1.2/lib/pdu_sms/user_data.rb:104:in `new'
/home/pawelpacholek/.rbenv/versions/2.4.0/lib/ruby/gems/2.4.0/gems/pdu_sms-1.1.2/lib/pdu_sms/user_data.rb:104:in `decode_sc'
/home/pawelpacholek/.rbenv/versions/2.4.0/lib/ruby/gems/2.4.0/gems/pdu_sms-1.1.2/lib/pdu_sms/packet_data_unit.rb:85:in `decode_sc'
/home/pawelpacholek/.rbenv/versions/2.4.0/lib/ruby/gems/2.4.0/gems/pdu_sms-1.1.2/lib/pdu_sms/packet_data_unit.rb:92:in `decode'
The given PDU can be decoded here:
https://www.diafaan.com/sms-tutorials/gsm-modem-tutorial/online-sms-pdu-decoder/
As usually, I will be very grateful for fixing this bug in the next version of this gem.
Answers:
username_1: Добрый день!
Исправлено в версии 1.1.4.
Ошибка в проверке длины сообщения в 7 битной кодировке.
Спасибо за обращение.
username_0: Hello,
It works now in version 1.1.5. Thank you very much once again.
Status: Issue closed
|
algorithm014-algorithm014/algorithm014-algorithm014 | 679698137 | Title: week1
Question:
username_0: #学号: G20200343110051
#姓名: 周宗波
#班级: 14期1班1组
#语言: Java
#作业&总结链接: https://github.com/username_0/algorithm014-algorithm014/tree/master/Week_01
Answers:
username_1: #学号: G20200343110032
#姓名: 程浩
#班级: 14期2班1组
#语言: Java
#作业&总结链接: https://github.com/username_1/algorithm014-algorithm014/tree/master/Week_01
username_2: 亲 你的作业交错地方啦 要放到班班建立的 Week 01 作业链接合集👈作业交这里那里
Status: Issue closed
|
jlippold/tweakCompatible | 306197248 | Title: `CircleIcons` working on iOS 11.0
Question:
username_0: ```
{
"packageId": "com.vedboon.circleicons",
"action": "working",
"userInfo": {
"arch32": false,
"packageId": "com.vedboon.circleicons",
"deviceId": "iPhone9,3",
"url": "http://cydia.saurik.com/package/com.vedboon.circleicons/",
"iOSVersion": "11.0",
"packageVersionIndexed": false,
"packageName": "CircleIcons",
"category": "Tweaks",
"repository": "BigBoss",
"name": "CircleIcons",
"packageIndexed": true,
"packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.",
"id": "com.vedboon.circleicons",
"commercial": false,
"packageInstalled": true,
"tweakCompatVersion": "0.0.6",
"shortDescription": "circle shaped icons in Settings app",
"latest": "1.1.3-15",
"author": "Javier (VedBoon) Brito",
"packageStatus": "Unknown"
},
"base64": "<KEY>
"chosenStatus": "working",
"notes": ""
}
```<issue_closed>
Status: Issue closed |
njleonzhang/vue-data-tables | 308258162 | Title: Use "Button Click" instead of "Keypress" to search in Data table server?
Question:
username_0: I would like to use `Button click` instead of `keypress` to search in `data table server`.
Could example?
Answers:
username_1: I suggest you define your own tool bar to implement this feature.
https://username_1.github.io/vue-data-tables/#/defineActionBar
Status: Issue closed
username_0: Thanks for your reply.
But I am not sure, how to trigger `queryChange()` method when `Button` clicked
```
<data-tables-server :data="data" :total="total" :loading="loading" @query-change="queryChange">
............
=============
queryChange(info) {
this.loadData(info)
}
btnSearchClick(){
this.queryChange()
}
```
Could example
username_1: Maybe you need to store the last `queryInfo`.
username_0: I tried to use `Custom action bar`, but work on any key press of `Search Textbox`(`queryChange called`).
username_1: [Define you own action bar](https://username_1.github.io/vue-data-tables/#/defineActionBar), You can control everything yourself.
username_0: </el-table-column>
</data-tables-server>
</div>
---------------------
data(){
tableCustomFilters: [
{ vals: [] },
],
},
methods:{
tableQueryChange(query) {
this.getData(query)
},
handleSearch(){
......
}
}
``
Please help me:
- Stop auto call `tableQueryChange` method when press any key on `Search Textbox`
- Call `tableQueryChange` method with `current query` when click `Search Btn`
username_2: 我明白他的意思, 就是目前的触发搜索是只要输入有变化,就自动触发loadData,他的意思是按一个按钮才触发loadData
```js
queryChange(type) {
let info = {
type,
...this.queryInfo
}
this.$emit('query-change', info)
this.loadData && this.innerLoadData(info) // 在组件内部就当搜索有变化就自动请求数据了
}
```
username_1: ```
<data-tables-server
:data="tableData"
:total="tableTotal"
:loading="loading"
:table-props="tableProps"
:show-action-bar="false"
:custom-filters="tableCustomFilters"
```
`tableCustomFilters` change cause the data-table change, if you want some button to trigger the change, then you can change `tableCustomFilters` when button click. It's all about your implementation, not controlled by `vue-data-tables`.
We need to understand how things work, not just that things can work in some way. |
pjbriggs/nebulizer | 115487158 | Title: manage_users: problems report admin and percent_quota in long listing format (-l)
Question:
username_0: When more extensive user information, it appears that the `is_admin` and `percent_quota` values that are displayed are all the same.
Status: Issue closed
Answers:
username_0: `manage_users` was deprecated and has now been removed in PR #61, so closing this issue. |
ContinuumIO/anaconda-issues | 243805236 | Title: Navigator Error
Question:
username_0: ## Main error
Application <b>notebook</b> launch may have produced errors.
## Traceback
```
[I 21:14:04.659 NotebookApp] Serving notebooks from local directory: C:\Users\orly
[I 21:14:04.660 NotebookApp] 0 active kernels
[I 21:14:04.660 NotebookApp] The Jupyter Notebook is running at: http://localhost:8888/?token=<PASSWORD>
[I 21:14:04.660 NotebookApp] Use Control-C to stop this server and shut down all kernels (twice to skip confirmation).
[C 21:14:04.662 NotebookApp]
Copy/paste this URL into your browser when you connect for the first time,
to login with a token:
http://localhost:8888/?token=<PASSWORD>
[I 21:14:05.171 NotebookApp] Accepting one-time-token-authenticated connection from ::1
[E 21:14:05.932 NotebookApp] Unhandled error in API request
Traceback (most recent call last):
File "C:\Users\orly\Anaconda3\lib\site-packages\notebook\base\handlers.py", line 516, in wrapper
result = yield gen.maybe_future(method(self, *args, **kwargs))
File "C:\Users\orly\Anaconda3\lib\site-packages\tornado\gen.py", line 1055, in run
value = future.result()
File "C:\Users\orly\Anaconda3\lib\site-packages\tornado\concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "C:\Users\orly\Anaconda3\lib\site-packages\tornado\gen.py", line 307, in wrapper
yielded = next(result)
File "C:\Users\orly\Anaconda3\lib\site-packages\notebook\services\contents\handlers.py", line 124, in<issue_closed>
Status: Issue closed |
jacksonc1123/ThrowsFoodException | 284821402 | Title: Register/Update
Question:
username_0: I should be able to register a new account, and also update my profile.
Status: Issue closed
Answers:
username_0: updated. login and registration components can be plugged into other components, but confirm profile and update profile components should be a separate page or view |
frontendbr/forum | 308702495 | Title: Sobre trabalhar sozinho: organização e inovação
Question:
username_0: Creio que muitos de vocês, assim como eu, em freelas ou empresas já trabalharam sozinhos num projeto ou eram o(a) único(a) dev da empresa, certo?
Nessa situação, como vocês se organizam desde o Git até a escolha e tecnologias/frameworks e/ou inovam na escolha de uma tecnologia nova sendo que possui um prazo que talvez não permita essa ousadia?
Answers:
username_0: pergunto isso justamente por trabalhar como dev único nas últimas empresas que passei e muitas vezes me sentir desatualizado com o resto do mercado, por não ousar na escolha de alguma tecnologia por conta de prazo, etc.
username_1: Comigo acontece algo simples: tem que desenvolver o site/tema em WordPress e pronto. É praticamente sempre o mesmo, apesar de aqui onde trabalho atualmente o cliente nunca tem contato com o WordPress. Mas a ideia é deixar mais genérico para que outros Devs futuros possam dar manutenção.
A forma como eu me organizo não tem nada demais pois tudo gira em torno de mim. Então o uso do Git é algo que nem o meu chefe sabe se eu faço ou não pois ele não sabe o que é, mas eu uso para a minha própria segurança.
Quando eu prestava serviços para uma agência que já trabalhei e lá o trabalho era em equipe, aí a situação muda da água pro vinho, não que seja melhor ou pior, depende do ponto de vista. MAs o fluxo do uso do Git era bem maior com certeza.
Com respeito a tecnologias não tenho muito o que dizer. Só usamos o WordPress.
username_2: Mas nem empresa grande ousa usar novas tecnologias para clientes, isso tem que ser feito de maneira controlada. Se você trabalha sozinho, fica até mais fácil pois você pode treinar/estudar fazendo uns projetos paralelos ou reproduzindo algum que você já fez usando uma nova tecnologia que você se interessou. Assim você ganha confiança para usa-la em produção.
Eu trabalho solo como freelancer desenvolvendo plugins para WordPress, tenho várias VM e eu para ser sincero quase não uso git. Trabalho mais é criando backups regurlamente. Geralmente só crio um repo privado quando o job já foi finalizado. Uso o git constante apenas em projetos meus.
username_3: Olá @username_0 , eu nunca cheguei a trabalhar sozinha em uma empresa mas eu acredito que a mesma organização deve ser mantida. Eu me organizaria definindo e separando as atividades pra implementar tal funcionalidade, talvez utilizando algo como o trello, e criaria uma branch no git pra cada funcionalidade implementada. Talvez eu gaste um tempo com isso no início mas eu acho que seria mais fácil depois eu focar no desenvolvimento depois de ter tudo o que precisa ser feito definido.
Uma coisa legal que em um time de uma pessoa só se consegue fazer é usar a técnica do pomodoro, aquela onde você foca 25 minutos produzindo algo sem distrações e para 5 para descansar a fim de aumentar a produtividade. Hoje eu não consigo fazer isso no trabalho pois todos do time estão sempre interagindo um com o outro para tirar dúvidas etc e acaba um interrompendo o outro.
username_4: Fala @username_0. Sempre trabalhei com outros devs, mas por vezes fui o único Frontender da empresa, e de fato isso pode ser um problema, dada a natureza do mundo de desenvolvimento, frontenders podem ter problemas muito especificos para resolver (performance web, acessibilidade, ux...).
Portanto as minhas soluções foram:
- Participar de fóruns / grupos de discussões e compartilhar minhas dores
- Fazer muitos projetos experimentais (que no passado resultaram num [belo Codepen](https://codepen.io/lfeh) alias)
- Iniciativas open-source me ajudavam a ter contato com outras metodologias e trabalhar em equipe
- Ir em eventos e debater sobre esses temas com outros developers
A própria criação do MeetupCSS foi uma tentativa de ter um espaço para debater CSS e melhorar como profissional... 😄
username_5: Vou deixar minha experiência rápida. Foi quando eu trabalhei por conta e sozinho que mais aprendi. Eu precisava ser ágil, precisava de ferramentas que fossem boas e fáceis. E o melhor, eu não tinha que pedir permissão para ninguém, eu apenas usava e aprendia. Evoluí muito nessa época. Claro, isso tudo já com toda a base teórica, então eu tentava manter o olhar crítico. Foi nessa época que, por exemplo, encontrei o Laravel (ainda na v3 na época) e eu sabia que aquilo iria ser grande, estava na cara que havia um trabalho excepcional ali. Hoje em dia eu ouço, num corredor ou outro, desenvolvedor falando: "Nossa, descobri um framework, o Laravel, vai mudar nossa vida aqui na empresa".
Agora em empresas grandes, a história é outra. Muito software legado (forçando invlusive você a não atualizar o seu SO em casos extremos) - O cliente não quer mudar porque "tá funcionando". O cliente não tem budget para montar um time de pesquisa para migração para outro stack ou framework. É muito mais burocrático. E ao meu ver, esse cenário é o mais "tóxico" - Pode criar os grandes "seniors" que só sabem Drupal 7 (Que nem é orientado a objeto, por exemplo) - só usando um exemplo. Você sempre se vê acuado porque tudo o que você dá de ideia é sempre "vamos ver, vamos ver" - Eu me policio bastante para não me "contaminar" continuo fazendo meus projetos paralelos, continuo lendo e seguindo muita gente na comunidade e estou sempre de olho no que está vindo pro mercado, ainda mais no frontend, o qual sentimos na pelo sua velocidade.
username_6: No início dos meus freelas eu fazia muitos sites institucionais simples, coisas de 10 páginas no máximo. Então eu acabei desenvolvendo um boilerplate que eu pudesse agilizar o processo, usando preprocessador (comecei com Less, depois Sass e por fim Stylus <3) e linguagem de template (jade/pug e o nunjucks).
Cada vez que eu ia fazendo um projeto, eu ia aprimorando a estrutura e separando também os plugins que eu ia criando. Com isso eu comecei a "ganhar tempo" nos projetos e isso me fez abrir espaço para experimentar.
Ao invés de falar para o cliente que eu faria o site dele em X dias e não mais X+10 dias, eu continuei dizendo que ia durar o mesmo de antes e aí nesse tempo extra, eu ia estudando coisas novas para os próximos clientes.
Hoje, dentro da Toptal (pelo menos no time que eu estou), apesar de ser um projeto relativamente novo, ele segue alguns padrões mais antigos e não tem tanta novidade... Em empresa realmente é mais complicado de inovar, mas é por isso que eu faço projetos pessoais por fora, assim aprendo e compartilho =D
username_7: Trabalhei sozinho por 2 anos, era o único dev da empresa. Usava git para tudo, versionar projetos é muito importante e evita muitas cagadas (principalmente porque não tem ninguém para revisar o que você faz).
Tentava automatizar tarefas frequentes, principalmente deploy (isso economizava muitas horas do dia e evitava dores de cabeça).
Sempre organizava o que precisava fazer com quadros no trello, isso evitava de esquecer coisas.
Quanto a escolha de tecnologias, dependia muito do prazo que eu tinha pra fazer a tarefa e o quão seguro eu estava para usar a tecnologia naquele momento.
O que sentia ser mais difícil era se atualizar, eu acompanhava alguns devs no twitter e muitos repositórios no github, principalmente esse aqui :D. Mesmo assim sentia que não era o suficiente, a troca de experiência faz muita diferença na carreira, tanto que acabei optando por sair e trabalhar em empresas com equipes maiores.
username_8: Tinha o exato mesmo problema do @username_0, estudando em paralelo e tentando evoluir por fora do trabalho. Funcionou ano passado, mas esse ano tava batendo uma sensação de atraso, que não estava aprendendo o suficiente para evoluir como profissional.
Acabei tendo a mesma atitude do @username_7, me demiti sexta pra conseguir algo numa empresa com equipe e focada em Desenvolvimento. Enquanto não acho, vou estudando e ampliando meu portfolio.
Status: Issue closed
|
BlackGekko/blackgekkodevelopment.github.io | 110036266 | Title: Logo/favicon
Question:
username_0: 
Answers:
username_1: I don't know why but I hate it. I hate it and I created it. its like hating your child
username_0: Then remake it today, we have a few days to get these little things done, take a step back and think how to make it better, ill add this into a task for you |
roshub/vapor_master | 401267201 | Title: [Bug] Loading params from yaml via roslaunch and <rosparam> not consistent with roscore
Question:
username_0: Hi folks,
looks like I have found another bug. I have a YAML file that looks like this:
joints:
- (JointA, 0)
- (JointB, 1)
- (JointC, 16)
...basically.
When I load this YAML file in a launch file by means of `<rosparam command="load" file="...` and fetch the parameters afterwards with `rosparam get joints` I get the following returned from `roscore`:
['(JointA, 0)', '(JointB, 1)', '(JointC, 16)']
... which obviously is a list.
But when I try to do that with `vapor_master` I get the following:
{'0': '(JointA, 0)', '1': '(JointB, 1)', '2': '(JointC, 16)'}
... which undoubtedly is a dictionary.
In order to use `vapor_master` as a true drop-in replacement, these things must be consistent.
Status: Issue closed
Answers:
username_1: Fix has been merged and is now available in the [snap store as version `0.3`](https://snapcraft.io/vapor-master) |
TDAmeritrade/stumpy | 618144098 | Title: Add AB joins for SCRIMP
Question:
username_0: A reference/discussion for adding AB joins for SCRIMP (`stumpy.scrump`).
This can be done without changing much, we only have to keep in mind to calculate the whole distance matrix instead of the upper triangle.
Answers:
username_1: I'm really excited about this!
Status: Issue closed
username_1: [This PR](742babaefca827e898992663ac3d12bc74c7db8) fixes some of this to be consistent with having `T_B` containing the query subsequences |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.