repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
realfagstermer/realfagstermer
436543282
Title: Miljøforvaltning Question: username_0: Jeg synes at _[Økosystembasert forvaltning](https://www.miljostatus.no/tema/hav-og-kyst/okosystembasert-forvaltning/) EN Ecosystem based management_ kan være greit å ha (jobber med [denne](https://bora.uib.no/handle/1956/13156)) - den kunne legges som BF under _miljøforvaltning_? Answers: username_1: Gjort: Vi hadde allerede Økosystemforvaltning som BF, men jeg la til Økosystembasert forvaltning også. Status: Issue closed
oppia/oppia-android
756932681
Title: Profile Name is going out of bounds Question: username_0: **Describe the bug** ![Screenshot_1607072477](https://user-images.githubusercontent.com/9396084/101143945-a25f6d80-363d-11eb-9097-4a77c6cee195.png) In the above image we can see that the Profile name is going out of screen. **Device** Use `Nexus 9` in portrait mode **To Reproduce** Steps to reproduce the behavior: 1. Add a profile with max characters. 2. Click on AdminControls 3. Click on Edit Profiles 4. Select the added profile. 5. Check that name is going out of bounds **Expected behavior** The text should shift to next line. The text should be fully visible. Also, check this on tablet landscape and mobile device too. Answers: username_1: Hey can I work on this issue? username_2: @username_1 Please check the Onboarding Instructions before starting with the contribution. https://github.com/oppia/oppia-android/wiki#onboarding-instructions username_2: @username_1 Any update on this? username_1: I am still working on this. Will try to make a PR in a day or two. username_2: Thanks for the update. username_3: Can I work upon this? username_2: @username_3 As you are already on other good-first-issues, I suggest looking for harder issues. username_3: @username_2 Sure, I will! I was trying to get comfortable with project's codebase and grasp Kotlin, coming from Java background. username_4: Hey, can I be assigned this issue username_0: Duplicate #2529 Status: Issue closed
AutomatedLab/AutomatedLab
377957544
Title: Question: Any documentation on how to use AL in a DevOps release pipeline ? Question: username_0: I am looking if there are any docs or ideas on how to use AL in a DevOps release pipe line. For example how can i pass the credentials to the Login-AzureRmAccount , does the module install ect.... ? Will referencing to the file from Save-AzureRmContext do the job ? Or can i better use DevLabs for this with a predefined image ? I like the AL script way of doing because i can use this onprem and in azure using the same script. Any way, just asking if there are already ideas on this scenario before i spend time on it . Thanks, michel Answers: username_1: Hi @username_0 , We removed using the AzureRM context file, since it was only trouble with multiple issues during a deployment with expired credentials. The persistent credentials were used instead, see here for details: https://docs.microsoft.com/en-us/powershell/azure/context-persistence In a pipeline (I assume you want to create a lab during e.g. integration tests), you can use the persisted context for example. This assumes that your build workers are not created and destroyed permanently so that the context can potentially be persisted. Status: Issue closed
google-research/bleurt
866848394
Title: Is text truncation to 512 tokens handled automatically for both candidate and reference texts? Question: username_0: Let's say I am using the following model - **BLEURT-Base, 512 (max #tokens)**. In my case, both generated text and reference text are longer than 512 tokens. While computing the BLEURT, will it automatically truncate both generated text and reference text to fit the requirement and then calculate the score between truncated versions of generated text and reference text? Or do I need to cut the length of generated text and reference text manually before calling the function to calculate BLEURT? Many thanks in advance, Ruslan
snowplow/iglu-central
58960345
Title: Explore application context ideas Question: username_0: As per discussion in https://github.com/snowplow/iglu-central/pull/135 Answers: username_1: Based on the things we collect, a first draft would be something like: ``` { "application": "Wunderlist", "version": "2.3.4", "build": "s9f2k2d", "androidFlavor" : "china", "buildType": "release" } ``` Notes: I'm suggesting `build` instead of our own `githash` as git is too specific. We use git hashes, but others would use a linear build id like `432` or even like Mac OS X’s `14C109`. For a 1.0.0 version, the above would capture what we need and I'd love to get something like it into our next release which will be the first with our Snowplow collectors. username_0: It's taking shape! What about: ```json { "name": "Wunderlist", "version": "2.3.4", "buildId": "s9f2k2d", "buildFlavor" : "china", "buildType": "release" } ``` `buildFlavor` would make the context a little less mobile-specific... username_0: Or event `buildStage: release` as type is a bit vague? username_1: I'm good with `buildFlavor`. Especially since on iOS, we could use this for "App Store", "Ad Hoc" etc based on provisioning profile. As to the `buildType/Stage`, I wonder if a boolean is more appropriate. Something like a `"developmentBuild": "true"` or a `"releaseBuild": "false"`... Hrmmmm... username_0: Hmm, I quite like "buildIsRelease": bool. Release is perhaps less ambiguous than buildIsDebug, buildIsDevelopment... username_0: Random thought - there could be something interesting (more for server-side apps) in capturing the log level (debug / warn / info kind of thing) - that's a pretty common unit of currency... username_1: `buildIsRelease` sounds good. Or, `buildIsProduction`? They're pretty equivalent—just thinking through how Rails has the concept of environment and how that'd map to this. I can see that a `logLevel` property could be useful. It could probably be picked up pretty easily from Log4J or whatever configuration. But, I passed the idea by a few of our server peeps and they were lukewarm on the idea. So, for now, I think I'm pretty neutral on that one. username_0: Yes - the thing is, once we get into `buildIsProduction`, we are starting to get into the language of different environments. I think we should distinguish between the use case the app is built for (debug versus release), versus the environment an app (perhaps server-side) is running in (dev / staging / ua / prod)... username_1: So, then that would argue for `buildIsProduction` as a boolean and `buildEnv` as an arbitrary string? username_0: Hmm - I reckon more `buildIsRelease` (boolean) as that's closer to the terminology for Mac apps. I reckon lets leave off \*environment\* as that may not be either hardwired into the app or obtainable from the app. username_1: Cool. Should I draft up a schema and PR it? username_0: Sounds good! username_2: Pushing back username_2: Hey @username_0! What should we do with this one? It has broader scope than https://github.com/snowplow/iglu-central/pull/843 (mobile only). Do we still need generic application schema? /cc @username_3 username_3: @username_2 coming at this only from the perspective of what I've seen our users ask for, I'd say that a generic application schema would be unnecessary. username_0: Let's close Status: Issue closed
chriskacerguis/codeigniter-restserver
127474809
Title: Send an md5 variable as header in each api request Question: username_0: Hi I'd like to know if there's a way to access these protected arrays _args _get_args so I can get a custom header value from each api request I wanna identify userid in each request without sending it in the url for all the methods that require it. I'm not sure if storing as session variable because it is being accessed from other domain, android app, etc. and don't know how I should work with that http://i.imgur.com/WZgkKEk.jpg Thanks a lot in advance Answers: username_0: If you see the image url referenced, I've tried all the following sentences to get the userid value from headers $x1 = $this->get(NULL); $x2 = $this->options(NULL); $x3 = $this->head(NULL); $x4 = $this->query(NULL); print_r($x1); print_r($x2); print_r($x3); print_r($x4); But output is always empty array. The apikey is correct, auth is ok, etc. username_1: It is not clear where you want to read the header value - on the client or on the server. Anyway, let me guess: ```php $userid = $this->input->get_request_header('userid'); ``` Is this what you want? It is a CodeIgniter's feature. username_0: Wow, that worked Ivan. However, I thought I could read those header values with the same Rest Server library I thought one of these methods was for that... am I wrong? $x1 = $this->get(NULL); $x2 = $this->options(NULL); $x3 = $this->head(NULL); $x7 = $this->query(NULL); Thanks a lot in advance! Saludos, <NAME> Freelancer Web Application developer http://www.freelancer.com/u/username_0.html San Juan // Argentina username_1: Yes, you are wrong, none of these methods serve this purpose. На 19.01.2016 в 22:20, username_0 написа: > Wow, that worked Ivan. > > However, I thought I could read those header values with the same Rest > Server library > > I thought one of these methods was for that... am I wrong? > > $x1 = $this->get(NULL); > $x2 = $this->options(NULL); > $x3 = $this->head(NULL); > $x7 = $this->query(NULL); > > Thanks a lot in advance! > > > > Saludos, > > > <NAME> > Freelancer Web Application developer > http://www.freelancer.com/u/username_0.html > San Juan // Argentina > > username_0: I thought the head method was for that, as stated in the docs /* * Retrieve a value from a HEAD request * * @access public * @param NULL $key Key to retrieve from the HEAD request * If NULL an array of arguments is returned * @param NULL $xss_clean Whether to apply XSS filtering * @return array|string|NULL Value from the HEAD request; otherwise, NULL */ public function head($key = NULL, $xss_clean = NULL) { if ($key === NULL) { return $this->head_args; } return isset($this->head_args[$key]) ? $this->_xss_clean($this->head_args[$key], $xss_clean) : NULL; } username_2: https://github.com/username_3/codeigniter-restserver/issues/605#issuecomment-172973284 solved my curiosity completely. Thanks. And one more thing, I am accessing the post values like `$post = json_decode($this->input->raw_input_stream, true);` as `$post = $this->input->post('post_key');` returns `null`. Is there any way to access posted value easier? username_1: @username_2 https://github.com/username_3/codeigniter-restserver/blob/2.7.2/application/controllers/api/Example.php#L109 http://iridadesign.com/codeigniter-restserver-test/post-test Status: Issue closed
almond-sh/almond
436763463
Title: Can't define type of var Question: username_0: I'm making some class notes using this kernel and I found the following issue: If I try the following code (which was working at this kernel when it was named jupyter-scala), I get an error: ``` var x: Any = 1 x = 1.5 x = true x = 'c' x = "string" --- cmd2.sc:4: type mismatch; found : Double(1.5) required: Int val res2_1 = x = 1.5 ^cmd2.sc:5: type mismatch; found : Boolean(true) required: Int val res2_2 = x = true ^cmd2.sc:7: type mismatch; found : String("string") required: Int val res2_4 = x = "string" ^Compilation Failed ``` I thought that the problem may be related to the Any type, so I made another test: ``` var i: Int i = 10 --- cmd2.sc:117: class Helper needs to be abstract, since variable inteiro is not defined (Note that variables need to be initialized to be defined) final class Helper extends java.io.Serializable{ ^Compilation Failed ``` Also, if I try something like initializing a var with null, the type is ignored: ``` var i: Int = null i = 10 --- cmd2.sc:4: type mismatch; found : Int(10) required: Null val res2_1 = i = 10 ^Compilation Failed ``` I do not know what is causing this, because, inside the project sbt execution I can correctly perform those actions. Answers: username_1: Thanks for the report! That seems to be a regression in the current master. The latest release (`0.4.0`) seems fine. If you are running the kernel from sources, just check out `0.4.0` with `git checkout v0.4.0` before running or installing the kernel. username_1: As a workaround, putting the type ascription with the value works fine, like ```scala var x = 1: Any x = 1.5 x = true x = 'c' x = "string" ``` username_0: Thank you @username_1, it worked perfectly for me! Status: Issue closed username_1: Closing this, this should be fine with `0.6.0`.
department-of-veterans-affairs/va.gov-team
621902557
Title: Coronavirus content review: Review and edit Annie protocol messaging Question: username_0: # User story As a Veteran enrolled in the Annie text messaging protocol, I need to receive messages that are helpful, accurate, and empathetic. # Details Request from stakeholders to review protocol content. Provided review and substantive editing. Answers: username_0: @stephorkandatx Complete and closing Status: Issue closed
aryeelab/hichipper
270672110
Title: ERROR: Missing hic_results or bowtie_results files for sample Question: username_0: hello, I meet a error when I used hichipper. This is my hichipper Command Line and error: ``` hichipper --out hichipper_res hichipper_config.yaml Thu Nov 02 22:05:39 CST 2017: Starting hichipper pipeline v0.6.1 Thu Nov 02 22:05:39 CST 2017: Parsing user parameters ERROR: Missing hic_results or bowtie_results files for sample; either exclude the sample or manage the file architecture/input ``` and this is my hichipper_config.yaml: ``` peaks: - ChIP-Seq_human_SRR502135_36_SMC3_bowtie2_merged_hichipper_peaks.narrowPeak resfrags: - MboI_resfrag_hg19.bed hicpro_output: - hicpro_simplify ``` and this is hicpro_simplify file: ``` hicpro_simplify/bowtie_results/bwt2/sample: SRR3467175_hg19.bwt2pairs.pairstat hicpro_simplify/hic_results/data/sample: SRR3467175_allValidPairs SRR3467175_hg19.bwt2pairs.REPairs SRR3467175_hg19.bwt2pairs.SinglePairs SRR3467175_hg19.bwt2pairs.DEPairs SRR3467175_hg19.bwt2pairs.RSstat SRR3467175_hg19.bwt2pairs.validPairs SRR3467175_hg19.bwt2pairs.DumpPairs SRR3467175_hg19.bwt2pairs.SCPairs ``` I checked my directory stucture with your manual, but I don't know where is wrong. could you help me to solve it? Thank you! Answers: username_1: Could you try updating your install to the most recent version? (V0.7.0) I think that we resolved this error since 0.6.1 username_0: It works. I updated to v0.7.0. Thanks for your prompt answer. Status: Issue closed
pykeen/pykeen
785736044
Title: ValueError: Cannot continue_training without being trained once. Question: username_0: Grouping triples: 100%|██████████████████████████████████████████████████████████████████████████████████████████| 150k/150k [00:00<00:00, 778ktriple/s] INFO:pykeen.training.training_loop:Starting batch_size search for training now... INFO:pykeen.training.training_loop:Concluded batch_size search with batch_size=32. Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/data/hadmap-big/nicoleqye/kge/lib/python3.7/site-packages/pykeen/training/training_loop.py", line 196, in train num_workers=num_workers, File "/data/hadmap-big/nicoleqye/kge/lib/python3.7/site-packages/pykeen/training/training_loop.py", line 300, in _train raise ValueError('Cannot continue_training without being trained once.') ValueError: Cannot continue_training without being trained once. ``` **Expected behavior** Can I fix this error? **Environment (please complete the following information):** - OS: [Linux] - Python version: [e.g. 3.7] - Version of this software: [pykeen 1.0.5] Answers: username_1: Hi @username_0 , Thanks for reporting this bug. I have looked into it and the reason for the error is that you have to set the `batch_size` parameter of the training loop when you want to continue training. We will soon release a fix for this bug. For now you can already get going by changing your code above to: ``` from pykeen.triples import TriplesFactory from pykeen.models import ComplEx from torch.optim import Adam from pykeen.training import LCWATrainingLoop,SLCWATrainingLoop from pykeen.stoppers import EarlyStopper tf = TriplesFactory(path='data.tsv') training, testing = tf.split([.9,.1]) model = ComplEx(triples_factory=training) optimizer = Adam(params=model.get_grad_params()) training_loop = LCWATrainingLoop(model=model, optimizer=optimizer) training_loop.train(batch_size=32, num_epochs=20,use_tqdm_batch=False) training_loop.train(batch_size=32, num_epochs=20,use_tqdm_batch=False, continue_training=True) ``` username_0: Grouping triples: 100%|██████████████████████████████████████████████████████████████████████████████████████████| 150k/150k [00:00<00:00, 752ktriple/s] Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/data/hadmap-big/nicoleqye/kge/lib/python3.7/site-packages/pykeen/training/training_loop.py", line 196, in train num_workers=num_workers, File "/data/hadmap-big/nicoleqye/kge/lib/python3.7/site-packages/pykeen/training/training_loop.py", line 300, in _train raise ValueError('Cannot continue_training without being trained once.') ValueError: Cannot continue_training without being trained once. ``` username_1: @username_0 Sorry for the inconvenience, in the current master it's fixed already. Given that it's PyKEEN 1.0.5, you also have to deactivate automatic memory optimization explicitly. This code should work for you 😊 ``` from pykeen.triples import TriplesFactory from pykeen.models import ComplEx from torch.optim import Adam from pykeen.training import LCWATrainingLoop, SLCWATrainingLoop from pykeen.stoppers import EarlyStopper tf = TriplesFactory(path='data.tsv') training, testing = tf.split([.9, .1]) model = ComplEx(triples_factory=training, automatic_memory_optimization=False) optimizer = Adam(params=model.get_grad_params()) training_loop = LCWATrainingLoop(model=model, optimizer=optimizer) training_loop.train(batch_size=32, num_epochs=20, use_tqdm_batch=False) training_loop.train(batch_size=32, num_epochs=20, use_tqdm_batch=False, continue_training=True) ``` username_0: Cool! Thanks! It worked well! 👍 username_2: @username_0 glad to hear things have worked out! I'm going to close this issue, but feel free to open another one if you've got more questions :) Status: Issue closed
trikset/trik-studio
451855005
Title: Different type multiplication in EV3 Question: username_0: Умножение целого числа на вещественное ошибочно производится целочисленным умножением в байткоде ![image](https://user-images.githubusercontent.com/8670331/58862378-3e12c500-86b9-11e9-915e-ff42ff404b20.png) Answers: username_0: Не воспроизводится, нужен пример.
rivafarabi/deckboard
480915406
Title: Sugguestion: Allow the use of clickable hotkeys from the same device. Question: username_0: Have an edit mode, and then a use mode. When in the edit mode, the buttons can be edited. When in the use mode and a button is clicked, the action is performed. Answers: username_1: Actually not a bad idea! I sometimes click on a button but then forget it's not like that lol
soflyy/oxygen-bugs-and-features
432619403
Title: Table of Contents (TOC) Question: username_0: **Describe the feature you'd like to see included in Oxygen.** An easy possiblity to add a table of contents to a post. It seems that typical TOC-plugins don't work together with oxygen (tested with Lucky WP TOC, EasyTOC) - at least as long you have headings not directly in the content, but in the oxygen-frontend. _Why not just write content with Gutenberg and import to Oxygen via Dynamic Data?_ Well, I have currently written most of my content inside an Oxygen-RTF-widget. I even developed my own pseudo-shortcodes to render images and other specific data inside the RTF-block (wrapped my own render-shortcode around the RTF for that). I just feel more at home within the Oxygen-environment. **What are the use cases for this feature?** Exhaustive articles **Examples of this feature or functionality.** See [EasyTOC](https://de.wordpress.org/plugins/easy-table-of-contents/) for example. how TOC could look like. Functionality: Either parse all `h`-Tags and generate TOC from that, or provide content-wrapper that defines what the relevant content is for the TOC. Answers: username_1: If you're manually creating long form articles not through the CMS but the Design Editor then why not manually create the TOC too? I'd call this a rare edge case. username_2: +1
scopt/scopt
48420471
Title: checkConfig blocks in one command are run in all other commands as well Question: username_0: Check out this example code: ``` object ScoptBug extends App { case class Config( command: Config => Unit = null, requiredParameter: String = null) lazy val parser = new scopt.OptionParser[Config]("ScoptBug") { cmd("commandWithoutParameter") action { (_, c) => c.copy(command = (c) => println("without")) } cmd("commandWithParameter") action { (_, c) => c.copy(command = (c) => println("with " + c.requiredParameter)) } children { opt[String]("requiredParameter") action { (id, c) => c.copy(requiredParameter = id) } checkConfig { c => if (c.requiredParameter == null) failure("You need to specify requiredParameter") else success } } } parser.parse(args, Config()) map { config => config.command(config) } } ``` This is how I run it: `ScoptBug commandWithoutParameter` This is the result: ``` Error: You need to specify requiredParameter Usage: ScoptBug [commandWithoutParameter|commandWithParameter] [options] Command: commandWithoutParameter Command: commandWithParameter --requiredParameter <value> ``` I expect it to simply run the command and print "without". I know I can use the `required()` option, but this is a toy example. Answers: username_1: Also ran into this. Have a checkConfig in the children section of command a. Run command b which doesn't even have those options and it still checks them. username_1: Saw the comment on the other post so I'll post my structure ```scala //other commands cmd("info") action { (_,params) => params.copy(mode = MyCommand.Info) } text("some action text") children( opt[String]("option-name") optional() action { (input, params) => params.copy(name = optional(input)) } text("some option text"), checkConfig {params => someCustomValidate(params) } ) // other comands ... ``` username_2: 👍
testcontainers/testcontainers-java
751470783
Title: Cannot use withFileSystemBind on Windows Question: username_0: When I try to use `withFileSystemBind` on Windows I get an parse error. After some debugging I found out that the problem is the `new File(..).toURI().toString()` in the class `MountableFileSystem` and also there was a bug in the class `Bind` from docker-java. The second one is fixed in release 3.2.6. But even with the new version it does not work because the `toURI()` changes all `\` from the windows path to `/` and after that `Bind.parse(String)` throws an exception because the regex pattern looks for `\` The problem is that `Bind.parse(String)` uses `:` as separator and this is also used in normal windows paths, e.g. `c:\temp\mypath`. Summary: * I call `withFileSystemBind("c:\temp\mypath", "/mypath", BindMode.READ_WRITE)` * In `MountableFile.forHostPath()` this is converted to `file:/c:/temp/mypath` * `Bind.parse(String)` throws an exception Possible solution: Use `new File(..).toString()` in `MountableFile.forHostPath()` or maybe `new File(path).getAbsolutePath().toString()`. The second solution is simular to the code path with the `Path` parameter. As far as I understand it is not possible to change the regex pattern in `Bind.parse()` because `..:/...` is the normal docker syntax for a volume mapping and you cannot use a negative lookahead for `/`. Testcontainers: 1.15.0 Docker Desktop with Docker Engine 19.03.13 Answers: username_1: Duplicate of #3493 Status: Issue closed
ant-media/Ant-Media-Server
976121874
Title: Broadcasting showing as play if I stop OBS or other stream application without stopping stream Question: username_0: **Is your feature request related to a problem? Please describe.** Need manual option to stop broadcasting channel if stream application getting closed without stopping stream. This close may be human mistake, Laptop Issue or some other reason. In this kind of situation I have only one option restart my server to connect broadcasting again. The main irritation points here is that if i didnot restart server Ant Media software is not allowing other broadcasting channel. So it will be good if there is a manual option I got stopping broadcasting. **Describe the solution you'd like** A clear and concise description of what you want to happen. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here. Answers: username_1: Hi @username_0 , There is an option to stop broadcasting on the REST Method. Check this out https://antmedia.io/rest/#/BroadcastRestService/stopStreamingV2 If you would like to have this option on the web panel, yes it's not available. For a quick workaround solution, there is a `Delete Broadcast` menu, it both stops and deletes the stream. You can then create the same stream with the same `stream id`. Ant Media Server web panel starts to support creating streams by specifying their stream id in the 2.4 ![delete_broadcast](https://user-images.githubusercontent.com/3456251/130325809-7dd72f3c-8f2b-4be6-8a1f-092f4f48fd00.png) Delete Broadcast Button ![stream_id](https://user-images.githubusercontent.com/3456251/130325812-7836b0ed-bd3b-441c-9071-5c104e7f16b6.png) Create Stream by specifying Stream Id(Optional)
pwa-builder/PWABuilder-CLI
884995925
Title: How can I build my PWA for ios Question: username_0: Hi, I would prepare my app for ios, I remember PWABuilder had that option before, but I see theoption is removed, Please provide me with that matter and let me know how can I prepare my app for ios. Thank you, Answers: username_1: @username_0 See https://github.com/pwa-builder/PWABuilder-CLI/issues/354 Status: Issue closed
wincent/loupe
760544691
Title: Feature Request: auto nohl on cursor move. Question: username_0: Was curious if implementing something similar to vim-slash and is.vim in regards to automatically calling :nohl after you move your cursor after n or N. I appreciate the mapping to be able to call :nohl, but I get tired of having to do that all the time. Answers: username_1: I just took a look at [the code of vim-slash](https://github.com/junegunn/vim-slash/blob/31aee09b7ea8893a18fa34f65e63e364fc998444/plugin/slash.vim#L41) and I came up with ```vim augroup Loupe autocmd! " autocmd CursorMoved,CursorMovedI * set nohlsearch | autocmd! Loupe autocmd CursorMoved,CursorMovedI * execute "normal \<Plug>(LoupeClearHighlight)" augroup END ``` But this doesn't quite work. The commented line works a bit too good: since `n` moves the cursor, it removes the highlighting everywhere. The second line removes the "current" highlight just fine, but the rest stays, although manually calling `:execute "normal \<Plug>(LoupeClearHighlight)"` works as expected :thinking: Maybe someone else has an idea on how to improve this username_2: I'm trying this out and have a similar feeling. I coded something like this ```lua function _G.coh() if vim.o.hlsearch then vim.fn['loupe#hlmatch']() else vim.fn['loupe#private#clear_highlight']() end end vim.cmd('autocmd OptionSet hlsearch call v:lua.coh()') ``` @username_1 this is neovim's lua, it's mostly the same in viml. Feels hacky though, calling `loupe#private...`, Is there any chance to get this into the plugin instead? username_1: @username_2 Can't you "type" `<Plug>(LoupeClearHighlight)`? This will call that function for you without that hack. At some point I wanted to take a look at lua, but so far my neovim-config is all vimscript...
VinylRecords/Vinyl
62580395
Title: Subset Instance? Question: username_0: I'd like to have something like the following: instance (is' ~ ???, RSubset rs ss is) => RSubset rs (s ': ss) is' Is this possible? Answers: username_1: There are two instances of RSubset in vinyl: instance RSubset '[] ss '[] instance (RElem r ss i , RSubset rs ss is) => RSubset (r ': rs) ss (i ': is) I'm not sure what you want your instance to do, and you would probably need OverlappingInstances for it to work. You will need to clarify what you are trying to do. username_0: I think it's pretty obvious what I'm trying to express: that if rs is a subset of ss, and we add another element s to ss, then rs will be a subset of s ': ss. For implementation, casting should certainly be easy: just take the rtail first. I'm not sure about the other class methods though, or what to do with the index type variable. username_2: Hi @username_0—frustratingly, the weakening judgement `RSubset rs ss _ => RSubset rs (s ': ss)` is admissible, but not derivable by the constraint solver given the current definitions. Also, you cannot add another instance, since the ones given are meant to constitute a syntax-directed algorithm for deciding subsethood. In the past, when I needed something like this, I was able to get away with using proof terms to demonstrate the validity of the weakening judgement. For instance, you can write a program which derives a subset lens for your use-case, and then use it in place of the automatic resolution provided by the type classes. Specifically, you may write a function of the following type: ```haskell weaken :: Lens (Rec f ss) (Rec f rs) -> Lens (Rec f (s ': ss)) (Rec f rs) ``` I do not have time today to demonstrate the technique; if my explanation has been sufficient, I am very pleased, but otherwise, feel free to leave this issue open and keep bugging me until I have been able to demonstrate it properly. Status: Issue closed username_0: My type-fu has increased to the point where all this makes sense. Sometimes I feel like scrapping all my typeclasses and using singletons instead.
Nosskirneh/Apace-public
433068952
Title: lost songs Question: username_0: first off, i wish there was a clear explanation for how to use the tweak in the first place... took me much longer than i wanted to figure out that i had to press and hold .. .. *on the song name* to bring up apace controls. second, i know it's only one step to get there, by pressing and holding on the song name but i really really wish i just could add the repeat, repeat 1 and shuffle buttons straight to the music player in Control centre.... it's something i would use so often that it's an issue for me. third, while trying to figure out how to use the tweak, i managed to watch 3 of my songs disappear from my phone... not even realising whath the 'library' button did, no explanation of it anywhere... now itunes says they are on my phone but my music app and siri say otherwise. i cannot get them back , even after i uninstalled apace2 and rebooted. can you help me get them back? Answers: username_1: This explanation is what it says at Packix. It doesn't have to be on the song name, it could be anywhere on the widget. I tried hard for **several weeks** to get the buttons to show up like in the initial Apace (iOS 10). However, I couldn't figure it out at that time, and that's why I went for this solution instead. Sure, I probably could explain a lot better how to use Apace 2 and what each action does, but people tend to ignore reading long explanations and 'just get on with it'. As for your songs, I'm terribly sorry to hear that happened. I don't use the Music app myself. Are these local tracks synced from iTunes or are they Apple Music synced tracks? You should be able to download/sync them again. username_0: thank you for the reply. the tracks are synced from itunes on my computer, just raw mp3 files, i have a rule setup to sync any song with higher than 3 play count and these songs are either in that category or manually added to a synced playlist. i don't use apple music. username_0: i'm sorry i missed the explanation in the packix page, that one is on me. i was only looking within the tweak for the instructions. also, i'm sorry and appreciate that you spent all that time trying to get the controls to appear at the initial widget. however i am finding the system in apace as it stands to be quite a drag to use. if i want to toggle repeat, repeat one and repeat off, i have to double tap every single time, and that excludes my dislike of having to bring up the apace option toggles by tap and hold. i have to request a refund and it brings me no pleasure in doing it because either way i appreciate the time and effort you put into it. username_0: on second thoughts. just keep the money. i don't need it and you deserve it for your work regardless. cancel the refund request i put through. username_0: i found a fix for the missing songs. i removed the *playlist* which the songs were in from my iphone via itunes by desyncing that playlist, then re-added the playlist to synced playlists again. :) Status: Issue closed username_1: I agree wholeheartedly – I am not entirely happy with the system either. Maybe I'll do a partial rewrite and change it. I'm grateful for your gratitude. However, I wouldn't want anyone to pay for something they won't be using. As such, I've of course refunded you. I'm glad you were able to solve the lost tracks!
F5Networks/f5-azure-arm-templates
231245014
Title: Deployment into existing Azure network stack Question: username_0: Hi team, I need to do a multi-nic(3) standalone into an existing Azure network stack, which I see is currently under construction in the experimental folder. I am doing a test deployment of the experimental 3nic standalone in a test vnet at the momet, however I realise it's not going to be supported. Is there a documented manual (or otherwise) process to achieve this deployment which is supported? I've read through https://support.f5.com/kb/en-us/products/big-ip_ltm/manuals/product/bigip-ve-setup-msft-azure-12-1-0.html but it doesn't provide clear instructions for a multi nic deployment. Answers: username_0: Hi, the 3 nic experimental template looks to have deployed fine, will license it up and check the F5 out next but the network configuration appears ok. So far great work guys! I also managed to manually deploy a 3nic VM from image which is up but am not happy with the result. It was fiddly and i need a (supported) repeatable process for scaling, best practice config options and repeatable redeployment such as the template provides. I am doing this in advance of a production deployment. Appreciate and advice username_1: Hi Blake, I'll contact you via email shortly. Mike
angular-klingon/klingon
405985734
Title: Revamp angular klingon design for better user experience Question: username_0: **Is your feature request related to a problem? Please describe.** Currently it feels very inconvenient as a user that i have to import or create new project every time i open the project. Although the empty around different tabs can be utilized in better ways. Hence this issue will cover and track those improvements which will improve/overhaul overall user experience of angular developer. Answers: username_1: I would love to get involved with this project, can try and help with the ui/ux?? username_0: That'd be great @username_1 . Thanks for showing your interest. Let me explain what problems we are facing now and how we are planning to take it forward. Current state of the UI is, we can easily create any angular project under **New** tab.. But all other options such as Build, serve, test etc. are also visible or disabled even if no project is created or imported. It confuses us a bit as a developer. Logically, options such as serve, build, test etc. should be visible only if an angular project is open, right? But currently it is not happening. We want to change that experience. In order to improve that experience, we thought of creating a **workspace** kind of UI which would allow user to :- 1. Create or import a workspace(angular project) first. 2. Allow to open recently accessed workspace. 3. All other **angular required** options (serve, build, test etc.) would be visible only after user creates or opens angular workspace by following step 1 or 2. This workspace can also be called a **startup page** which will be displayed every time user opens klingon (But we can also provide an option to not display startup screen next time, in future..) I've personally thought of a workspace design in which, apart from create and import option, project & contributors details would also be displayed. We can divide it in 2 column layout:- 1. First column would contain option to create or import angular project and also display a list of recently accessed projects. 2. Second column would contain summarized klingon details along with log and would also display list of [great people](https://github.com/angular-klingon/klingon/graphs/contributors) who have so far contributed to it. ### Workspace screen should not have scroll bar because, psychologically, scroll bar in a startup screen, irritates a bit. It should look as straightforward as it can be. Let's discuss what you think about above points and what are your ideas to improve its user experience. username_0: Taking this forward.
codestudiohq/laravel-totem
319850102
Title: Import Form has different action than used route Question: username_0: According to ```artisan route:list``` the allowed methods for ```task/import``` are GET and HEAD but the form uses POST. Is there reasoning for this? If so please let me know otherwise i will bring up a PR for this later. Answers: username_1: I'm interested in this too. Throws an error `MethodNotAllowedHttpException` on upload. username_2: This is now fixed in the latest release. Status: Issue closed
spring-cloud/spring-cloud-consul
339672721
Title: When I add spring-cloud-starter-consul-discovery,the app report bugs Question: username_0: When I add spring-cloud-starter-consul-discovery: <dependency> <groupId>org.springframework.cloud</groupId> <artifactId>spring-cloud-starter-consul-discovery</artifactId> </dependency> run the springboot application,it reports: org.springframework.beans.factory.UnsatisfiedDependencyException: Error creating bean with name 'documentationPluginsBootstrapper' defined in URL [jar:file:/D:/Program%20Files/.m2/io/springfox/springfox-spring-web/2.9.2/springfox-spring-web-2.9.2.jar!/springfox/documentation/spring/web/plugins/DocumentationPluginsBootstrapper.class]: Unsatisfied dependency expressed through constructor parameter 1; nested exception is org.springframework.beans.factory.UnsatisfiedDependencyException: Error creating bean with name 'webMvcRequestHandlerProvider' defined in URL [jar:file:/D:/Program%20Files/.m2/io/springfox/springfox-spring-web/2.9.2/springfox-spring-web-2.9.2.jar!/springfox/documentation/spring/web/plugins/WebMvcRequestHandlerProvider.class]: Unsatisfied dependency expressed through constructor parameter 1; nested exception is org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'requestMappingHandlerMapping' defined in class path resource [org/springframework/web/servlet/config/annotation/DelegatingWebMvcConfiguration.class]: Invocation of init method failed; nested exception is java.lang.ArrayStoreException: sun.reflect.annotation.TypeNotPresentExceptionProxy the follow is my spring cloud config: <properties> <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding> <project.reporting.outputEncoding>UTF-8</project.reporting.outputEncoding> <java.version>1.8</java.version> <spring-cloud.version>Finchley.M9</spring-cloud.version> <!-- <spring-cloud.version>Edgware.SR2</spring-cloud.version> --> <spring-cloud.component.version>1.4.4.RELEASE</spring-cloud.component.version> <pagehelper.version>5.1.4</pagehelper.version> </properties> <dependencyManagement> <dependencies> <dependency> <groupId>org.springframework.cloud</groupId> <artifactId>spring-cloud-consul-dependencies</artifactId> <version>2.0.0.RELEASE</version> <type>pom</type> <scope>import</scope> </dependency> </dependencies> </dependencyManagement> Answers: username_1: Finchley.RELEASE is the latest spring cloud release. I have no idea if spring fox is compatible. You'll have to take it up with the maintainers. Status: Issue closed
crystal-lang/crystal
429132672
Title: Incorrect comments indentation formatting in `select` block Question: username_0: ```crystal case when 1 # This is OK foo when 2 # This is OK bar end select when x # Not OK foo when y # Still not OK bar end ``` Answers: username_0: This could be put in one box I guess. Note `then` indentation: ```crystal case when x then foo when yz then bar end select when x then foo when yz then bar end ```
jhipster/generator-jhipster
191602514
Title: javadoc generation for test classes failes Question: username_0: tions' * Try: Run with --info or --debug option to get more log output. * Exception is: org.gradle.api.tasks.TaskExecutionException: Execution failed for task ':javadoc'. at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeActions(ExecuteActionsTaskExecuter.java:84) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.execute(ExecuteActionsTaskExecuter.java:55) at org.gradle.api.internal.tasks.execution.SkipUpToDateTaskExecuter.execute(SkipUpToDateTaskExecuter.java:61) at org.gradle.api.internal.tasks.execution.ValidatingTaskExecuter.execute(ValidatingTaskExecuter.java:58) at org.gradle.api.internal.tasks.execution.SkipEmptySourceFilesTaskExecuter.execute(SkipEmptySourceFilesTaskExecuter.java:88) at org.gradle.api.internal.tasks.execution.ResolveTaskArtifactStateTaskExecuter.execute(ResolveTaskArtifactStateTaskExecuter.java:45) at org.gradle.api.internal.tasks.execution.SkipTaskWithNoActionsExecuter.execute(SkipTaskWithNoActionsExecuter.java:51) at org.gradle.api.internal.tasks.execution.SkipOnlyIfTaskExecuter.execute(SkipOnlyIfTaskExecuter.java:54) at org.gradle.api.internal.tasks.execution.ExecuteAtMostOnceTaskExecuter.execute(ExecuteAtMostOnceTaskExecuter.java:43) at org.gradle.api.internal.tasks.execution.CatchExceptionTaskExecuter.execute(CatchExceptionTaskExecuter.java:34) at org.gradle.execution.taskgraph.DefaultTaskGraphExecuter$EventFiringTaskWorker.execute(DefaultTaskGraphExecuter.java:233) at org.gradle.execution.taskgraph.DefaultTaskGraphExecuter$EventFiringTaskWorker.execute(DefaultTaskGraphExecuter.java:215) at org.gradle.execution.taskgraph.AbstractTaskPlanExecutor$TaskExecutorWorker.processTask(AbstractTaskPlanExecutor.java:74) at org.gradle.execution.taskgraph.AbstractTaskPlanExecutor$TaskExecutorWorker.run(AbstractTaskPlanExecutor.java:55) at org.gradle.execution.taskgraph.DefaultTaskPlanExecutor.process(DefaultTaskPlanExecutor.java:32) at org.gradle.execution.taskgraph.DefaultTaskGraphExecuter.execute(DefaultTaskGraphExecuter.java:113) at org.gradle.execution.SelectedTaskExecutionAction.execute(SelectedTaskExecutionAction.java:37) at org.gradle.execution.DefaultBuildExecuter.execute(DefaultBuildExecuter.java:37) at org.gradle.execution.DefaultBuildExecuter.access$000(DefaultBuildExecuter.java:23) at org.gradle.execution.DefaultBuildExecuter$1.proceed(DefaultBuildExecuter.java:43) at org.gradle.execution.DryRunBuildExecutionAction.execute(DryRunBuildExecutionAction.java:32) at org.gradle.execution.DefaultBuildExecuter.execute(DefaultBuildExecuter.java:37) at org.gradle.execution.DefaultBuildExecuter.execute(DefaultBuildExecuter.java:30) at org.gradle.initialization.DefaultGradleLauncher$4.run(DefaultGradleLauncher.java:197) at org.gradle.internal.Factories$1.create(Factories.java:25) at org.gradle.internal.progress.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:91) at org.gradle.internal.progress.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:53) at org.gradle.initialization.DefaultGradleLauncher.doBuildStages(DefaultGradleLauncher.java:194) at org.gradle.initialization.DefaultGradleLauncher.access$200(DefaultGradleLauncher.java:36) at org.gradle.initialization.DefaultGradleLauncher$1.create(DefaultGradleLauncher.java:118) at org.gradle.initialization.DefaultGradleLauncher$1.create(DefaultGradleLauncher.java:112) at org.gradle.internal.progress.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:91) at org.gradle.internal.progress.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:63) at org.gradle.initialization.DefaultGradleLauncher.doBuild(DefaultGradleLauncher.java:112) at org.gradle.initialization.DefaultGradleLauncher.run(DefaultGradleLauncher.java:98) at org.gradle.launcher.exec.GradleBuildController.run(GradleBuildController.java:66) at org.gradle.tooling.internal.provider.ExecuteBuildActionRunner.run(ExecuteBuildActionRunner.java:28) at org.gradle.launcher.exec.ChainingBuildActionRunner.run(ChainingBuildActionRunner.java:35) at org.gradle.launcher.exec.InProcessBuildActionExecuter.execute(InProcessBuildActionExecuter.java:41) at org.gradle.launcher.exec.InProcessBuildActionExecuter.execute(InProcessBuildActionExecuter.java:26) at org.gradle.tooling.internal.provider.ContinuousBuildActionExecuter.execute(ContinuousBuildActionExecuter.java:75) at org.gradle.tooling.internal.provider.ContinuousBuildActionExecuter.execute(ContinuousBuildActionExecuter.java:49) at org.gradle.tooling.internal.provider.ServicesSetupBuildActionExecuter.execute(ServicesSetupBuildActionExecuter.java:44) at org.gradle.tooling.internal.provider.ServicesSetupBuildActionExecuter.execute(ServicesSetupBuildActionExecuter.java:29) at org.gradle.launcher.daemon.server.exec.ExecuteBuild.doBuild(ExecuteBuild.java:67) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:36) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:120) at org.gradle.launcher.daemon.server.exec.WatchForDisconnection.execute(WatchForDisconnection.java:47) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:120) at org.gradle.launcher.daemon.server.exec.ResetDeprecationLogger.execute(ResetDeprecationLogger.java:26) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:120) at org.gradle.launcher.daemon.server.exec.RequestStopIfSingleUsedDaemon.execute(RequestStopIfSingleUsedDaemon.java:34) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:120) [Truncated] ls: no such file or directory: .jhipster/*.json ##### **Browsers and Operating System** java version "1.8.0_111" Java(TM) SE Runtime Environment (build 1.8.0_111-b14) Java HotSpot(TM) 64-Bit Server VM (build 25.111-b14, mixed mode) git version 2.7.0.windows.1 node: v4.4.7 npm: 3.10.9 bower: 1.8.0 gulp: [01:10:21] CLI version 1.2.2 [01:10:21] Local version 3.9.1 yeoman: 1.8.5<issue_closed> Status: Issue closed
AlertingAvian/MinecraftServer
863402469
Title: me no have all shops available to me[BUG] Question: username_0: **Describe the bug** A clear and concise description of what the bug is. **To Reproduce** Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** If applicable, add screenshots to help explain your problem. **Desktop (please complete the following information):** - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] **Smartphone (please complete the following information):** - Device: [e.g. iPhone6] - OS: [e.g. iOS8.1] - Browser [e.g. stock browser, safari] - Version [e.g. 22] **Additional context** Add any other context about the problem here. Answers: username_1: Missing information for the bug report. username_1: More information required. @username_0 Status: Issue closed username_1: Closed until more info is provided. username_1: @username_0 Issue closed until more information is provided. username_1: Issue requires more information. Will reopen provided report is correct. @username_0 username_1: Issue requires more information. Will reopen provided report is correct. @username_0
JohannGuadalupe/AS91893
826818434
Title: Research at least 2 or 3 similar website in PMI Question: username_0: In a PMI chart research at least 2 or 3 websites that are similar in your chosen genre P.OSITIVE : Stuff I like M.INUS : Features I dislike about the site I.NTERSTING : Stuff I will want to add into my website that is cool
goohooh/noreeter
145900933
Title: users앱 초기화 Question: username_0: - AbstractUser를 상속받은 User 모델 생성 - LoginView, LogoutView, SignupView, ProfileView 구현 Answers: username_0: FBV로 이미지를 받아올때 request.POST.get 이 아니라 request.FILES.get 으로 받아옴 username_0: FBV로 이미지를 받아올때 request.POST.get 이 아니라 request.FILES.get 으로 받아옴 Status: Issue closed username_0: - AbstractUser를 상속받은 User 모델 생성 - LoginView, LogoutView, SignupView, ProfileView 구현 Status: Issue closed
diatomic/diy
286380286
Title: support windows Question: username_0: Windows builds fail as follows: ``` c:\...diy\storage.hpp(8): fatal error C1083: Cannot open include file: 'unistd.h': No such file or directory ``` Answers: username_1: The comment in `storage.hpp` says that it's using that file to get `mkstemp()` function on a Mac. So the easy part of the problem: we can just remove that include on Windows. The more complicated part of the problem is how to get some equivalent of `mkstemp()`. I don't use Windows (and I don't even have any Windows machines), so it's difficult for me to debug this. username_0: working on it. I'll have an pr when ready. username_1: Did #59 close this? @username_0 username_0: yes..I am sure there'll be more iteration to fix warnings etc, but this can be closed. Status: Issue closed
florisboard/florisboard
759875148
Title: Add Flow layout when implementing glide/swipe typing Question: username_0: I just learned about the Flow keyboard after seeing [this issue](https://github.com/AnySoftKeyboard/AnySoftKeyboard/issues/2545) on ASK, and while I haven't spent a ton of time on it, my initial thought is it's super promising and if the reviews on google play are anything to go by, it's really good. That said, the emoji keyboard is seriously lacking, and there's no dark theme option. Perhaps that layout could be added to florisboard, and the glide typing as done in Flow could be used as an inspiration for implementing it in florisboard? I'd love to help, however I'm not an android dev so I may not be able to help as much as I'd like :( https://sourceforge.net/projects/flowkeyboard/ https://play.google.com/store/apps/details?id=com.enchantedcode.flow <!-- - Describe your idea in a short but concise way. - If you have multiple ideas which are not directly connected to each other, file an issue per idea. This makes it easy to implement one feature proposal at a time. - If you have any examples, e.g. screenshots or other keyboards which have the proposed feature implemented, link them here. - Please search existing proposals to avoid creating duplicates. - Thank you for your help in making FlorisBoard better! --> Answers: username_1: Even if you are not an Android dev, you can still help by testing out new features, suggest ideas, report bugs and crashes, etc. This help is incredibly useful in improving FlorisBoard and for me it would not be possible to do this alone :) username_0: Yeah, I tried looking through the source code of flow and florisboard to see if I could do it myself but I got lost real quick. I was hoping that since glide typing hasn't been implemented yet that it would be easy to use a similar algorithm for all the swipe typing in florisboard regardless of layout, but it sounds like that won't be possible which is unfortunate. I'll keep an eye out for anything, it's already a pretty solid keyboard from what I've seen so keep up the good work :+1: username_1: Thanks for the positive feedback! :) username_0: That would be pretty cool, and would make it easier to add all sorts of new layouts in the future. But yeah that is one big issue with it, unfortunately I think for other languages (even ignoring the lack of special characters) you'd need to regenerate the optimal layout based on the common patterns of that language. I'm not sure how <NAME>man generated flow for english, but I'm sure it could be done with an ai like how this guy made [halmak](https://github.com/MadRabbit/halmak) (also English, but the process could be followed for other languages) username_2: It would be awesome for the ability to choose between which layout you wanted to use was implemented in settings, so you could switch back and forth between the layouts if you needed to. Considering that flow is open source wouldn't it be possible to input/implement some of it's code into Florisboard? username_2: It looks like Flow keyboard supports French, German, Spanish, Portuguese, and American and British English. https://sourceforge.net/p/flowkeyboard/code/HEAD/tree/trunk/res/raw/ username_0: @username_1 I finally got around to trying out the latest beta and I just wanted to say great work on the gesture/swipe typing! A couple other potential layouts that could be good to add for swipe typing are the ones found here http://sangaline.com/post/finding-an-optimal-keyboard-layout-for-swype/ and here https://jasmcole.com/2017/06/04/swype-right/ If you'd be ok with adding those layouts I could probably go ahead and make a pull request with them username_1: @username_0 Definitely interesting but I'd wait for the new layout logic to be merged into master (#734). This PR abstracts the layout definition and should make implementing such a layout a bit easier. username_0: I think with #1015 merged, there's no longer any reason to bring in the flow keyboard layout. The amount of effort to get it to work seems excessive to me, and after getting used to the NALMY layout for several weeks after using flow regularly, it's about the same if not better in terms of efficiency anyways. Of course if you want to add it eventually that's up to you, but I think it makes the most sense to close this issue username_1: @username_0 If you think so then we can close this issue. As there's already a lot of development stuff to do in so many different areas I think this is the best option. Status: Issue closed
openshift/origin
266922481
Title: openshift_sdn_pod_teardown_latency reports NaN for all metrics on ca-central-1 Question: username_0: Not sure if this had the fixes we put into it, or what, but it's coming out NaN for everything. @openshift/sig-networking Answers: username_0: It doesn't look like the second round of metrics fixes made it to this release, but recording it here anyway to double check next time
OpenSourceEconomics/respy
427293540
Title: Use conda everywhere Question: username_0: - make package conda installable - remove requirements.txt - use conda environments in appveyor, ... This is easy after Fortran is removed Answers: username_1: This issue is a little bit dated, but the main point of making respy installable via conda still holds. username_2: @username_1 , we can close this, right? Status: Issue closed
regro/regolith
297997077
Title: problem merging dbs Question: username_0: not sure if this is a bug or some local problem, but I think it is bugg-ish so reporting it. Here is more or less what happened I think. 1) I pushed my local dbs to remote 2) I ran a file that I had made that called `chained_db` 3) everything worked fine 4) I made some edits to my local dbs 5) I went to push my local dbs upstream 6) I got a message that my local repo was out of date 7) I tried to pull (should have used `--ff-only` but didn't....still not used to that) 8) got a bunch of merge conflicts. 9) looked in the files and there were a bunch of changes....empty lines deleted, some keys quoted etc.. It looked as if regolith had done some internal validation thing and updated the dbs automatically on remote. 10) all of this ok in general, except that the merge was a shitshow. Basically <<<<HEAD contained the entire file to ====== and then the entire file from remote to >>>>>. This may be a windows thing? UNIX windows line-ends or sthg? 11) anyway, bottom line, the files are essentially unmergable. update: actually, it looks as if regolith did some horrible things to the projects db file.....actually broke it.....so maybe it is a bug Answers: username_0: this seems to be the culprit: commit ac7c20f394c51464bbf52584a2e1ff6a14a076bd Author: <NAME> <<EMAIL>> Date: Sat Feb 17 05:41:28 2018 -0500 added Julien to people, changed gitlaburl to url commit 69861bf5cfc418d721af91015732f327e22a4df0 Author: <NAME> <<EMAIL>> Date: Sat Feb 17 05:22:43 2018 -0500 regolith auto-commit username_0: forgot to tag @username_1 on this.....this is a show-stopper till we fix it. I had a look to see if I could do a quick PR myself, but this is buried a bit deep in the machinery. I am happy for regolith to do auto-commits....I want the regolith machinery to be maintaining the integrity of the dbs as much as possible....but if it makes any tiny changes to my local db unmergable it will lead to endless problems for me I am afraid. username_1: Thank you for reporting, I'll see what I can do. username_0: btw, even using yaml it is very easy to break the dbs. I am thinking we will need helper functions for entering stuff in the dbs before rolling it out to the broader group. username_1: Ok. The app does this to some degree (since it checks if it is valid json before putting records in). username_0: I know (or would if it were working) and I would like more validating, but it is free edits and I am finding the yaml to be picky. OK, I can get used to it, but I want people flowing through the group, and don't want to go through the pain of teaching each and every one of them correct yaml etiquette. I spend too much of my time already teaching them about hard-spaces in latex..... I will experiment with some helper functions to see if they ease or make harder the data-entry activities. S
axios/axios
541984345
Title: axios load error. 429 Question: username_0: I am loading axios like below `<script src="https://unpkg.com/axios/dist/axios.min.js"></script> ` and below error is caused. <img width="737" alt="Screen Shot 2019-12-24 at 1 21 57 PM" src="https://user-images.githubusercontent.com/20388249/71393570-80e5bb80-2650-11ea-8320-209e432e7e5e.png"> I didn't have this issue until just few hours ago. Is there a bug int he update? Answers: username_1: Try downloading the actual .js file. username_2: me too, hope this can be fix username_0: confirmation that it is the statics server problem. yarn is still able to download the package so I ended up serving my own. Close the issue when it's fixed. Thanks. username_3: me too....we service not work;;;; username_4: me too. Though, I download the js file( npm install axios ) now. It's working now. username_5: You are able to resolve this issue by installing axios module with npm (like **npm install axios**). however, for me(or if you can't use npm or something), I just have inserted a raw js file into <head></head> to fix this issue immediately. https://github.com/axios/axios/blob/master/dist/axios.min.js It is not a best choice but hope this can be helpful! username_2: i fix it using another cdn https://cdnjs.com/libraries/axios Status: Issue closed username_6: looks like it's fixed and it's an issue with unpkg, thank you everyone!
ballerina-platform/ballerina-distribution
534855819
Title: Inconsistent version info in 1.1.0-alpha and 1.0.x distributions Question: username_0: When you run `ballerina version` on 1.1.0-alpha, ``` $ ballerina version jBallerina 1.1.0-alpha Language specification 2019R3 Command 1.1.0-alpha ``` And when you run it on 1.0.4, ``` $ ballerina version Ballerina 1.0.4 Language specification 2019R3 Command 1.1.0-alpha ``` In the above it says `Ballerina 1.0.4` while in the former it says `jBallerina 1.1.0-alpha`. Wouldn't this be inconsistent, changing it to jBallerina all of a sudden? Answers: username_0: Just realized, to fetch the 1.0.4 version, I used the command `ballerina dist fetch jballerina-1.0.4`. It's inconsistent there as well isn't it? In the CLI tool itself it refers to the 1.0.4 version as version `jballerina-1.0.4` while the output of the `ballerina version` command says `Ballerina 1.0.4`. username_1: From the 1.0.0 onwards it has to be jBallerina as we have introduced that. Unfortunately it has not updated in the those releasees (1.0.0 to 1.0.4). From 1.1.0 onwards we specifically mention it as jBallerina to be consistent with our references. For an instance when we run `ballerina dist pull jballerina-1.0.4`, we specify `jballerina` along with the version. Status: Issue closed username_1: When you run `ballerina version` on 1.1.0-alpha, ``` $ ballerina version jBallerina 1.1.0-alpha Language specification 2019R3 Command 1.1.0-alpha ``` And when you run it on 1.0.4, ``` $ ballerina version Ballerina 1.0.4 Language specification 2019R3 Command 1.1.0-alpha ``` In the above it says `Ballerina 1.0.4` while in the former it says `jBallerina 1.1.0-alpha`. Wouldn't this be inconsistent, changing it to jBallerina all of a sudden? Status: Issue closed
Cazadorsniper/PO3
416521697
Title: Crash With extra utilities gen Question: username_0: **Pack Version:** 3.0.38 **Server Version:** 3.0.38 I used the official files **Description:** Ender io pipes fed into all 14 item based extra utilities gens Server crashed when i put a bone into the system. note this did not happen when i put food in the system it worked perfectly then. My server now crashes when booting, **Steps to reproduce:** link 14 gens together with ender io conduit. insert bone **Crash log:** Initial crash: https://pastebin.com/tDEPqbny Server boot crash: https://pastebin.com/Rz35EYLM Another server boot crash: https://pastebin.com/RkgDDEqc Answers: username_0: After looking at the logs myself I saw the potion gen being referenced so I used McEdit to remove it and now the server will boot. I am going to apply filters now so that the conduits only try to place items where they are suppose to go. Status: Issue closed
opnsense/core
1117390310
Title: MVC page using legacy config generation does not properly update service configuration Question: username_0: **Important notices** Before you add a new report, we ask you kindly to acknowledge the following: - [x] I have read the contributing guide lines at https://github.com/opnsense/core/blob/master/CONTRIBUTING.md - [x] I am convinced that my issue is new after having checked both open and closed issues at https://github.com/opnsense/core/issues?q=is%3Aissue **Describe the bug** Right now the legacy restart/reconfigure(apply) buttons in the webgui do not function the same as the MVC restart/configure(apply) buttons. The MVC status widget actions and the reconfigure button assume that all configuration is generated via the template system, while the legacy restart/reconfigure actions configure *both* legacy and templated configuration data. Since it is sometimes necessary to prevent large pieces of logic cluttering up a template or minimizing impact, legacy config generation might be utilized instead, while the MVC model is used to do model validation and page rendering. Currently, this is the case for Unbound -> Overrides -> Host entries, which I migrated last week. The overall effect is that both applying and restarting the service from this MVC page (in this example Host Overrides) does not actually update the configuration in /var/unbound/* unless (from a users' perspective): 1. You restart the service from any legacy page. 2. You reboot the system. Which in both case activate the relevant pluginctl hooks. The question now becomes how we might do this correctly. It seems https://github.com/opnsense/core/blob/master/src/opnsense/scripts/unbound/start.sh must always be called to properly start Unbound, which is called in both the MVC and legacy cases (https://github.com/opnsense/core/blob/master/src/etc/inc/plugins.inc.d/unbound.inc#L419 and https://github.com/opnsense/core/blob/master/src/opnsense/mvc/app/controllers/OPNsense/Base/ApiMutableServiceControllerBase.php#L187), so I suspect interference here is not necessary. I would say that overriding the reconfigureAction() method and calling a pluginctl hook in Unbounds' ServiceController might be a solution. This provides the added benefit of having a single legacy hook to signal that a transition in the future might be necessary. The downside is that it's not a generic solution - meaning one could add a host entry, switch to another MVC page for the same service and restart it from there, which would also not properly update the configuration. Please correct me if I'm wrong in any of these statements @AdSchellevis @username_1. **To Reproduce** Steps to reproduce the behavior: 1. Go to Unbound->Overrides->Host entries (default page) 2. Add a host entry 3. Click Apply and/or restart the service 4. Observe /var/unbound/host_entries.conf has not changed 5. Go to Unbound->General 6. Click apply and/or restart 7. Observe /var/unbound/host_entries.conf has updated correctly. **Expected behavior** I would expect the configuration file to properly update. **Describe alternatives you considered** None **Screenshots** None **Relevant log files** None **Additional context** None **Environment** OPNsense 21.7.a_1178-amd64 (amd64, OpenSSL). Answers: username_1: Yes, these differences exist due to the hybrid nature of the Unbound integration. Moving all to MVC would be the most consistent fix. traditionally legacy services have a single configure function that does „templating“ and service restart/stop depending on the configuration. the mvc code decoupled this behaviour. username_0: Since legacy code is not to be used in the mvc code, I'm unsure of how to move forward for now unless we all agree to move the whole thing to a template. username_1: We should fix existing problems within this, but I think it works ok at the moment the way it is unless I’m missing something. As I said best way to deal with this is get rid of legacy code. That doesn’t mean we have to at this point in time. 😊 username_0: Well, the main issue still remains, 'apply' and 'restart' basically does nothing in the hosts section. Whether or not this is a major issue is up to you of course :) worst case scenario it might lead to some annoyance. username_1: Will take a closer look on Monday. Status: Issue closed
bwssytems/ha-bridge
290660041
Title: Can't control WEMO Insight Switch Question: username_0: Hi, with the following call from postman I can control my Belkin WEMO Insight Switch: ![image](https://user-images.githubusercontent.com/35510003/35250437-e1d31a2a-ffd6-11e7-83b7-2dc77634e245.png) ![image](https://user-images.githubusercontent.com/35510003/35250451-f89682ec-ffd6-11e7-9419-86d171d42982.png) I tried to implement this in HA-Bridge like this: ![image](https://user-images.githubusercontent.com/35510003/35250523-4b41cefc-ffd7-11e7-9ba3-7384984824e3.png) But it's not working. Here's the log with http plugins set to DEBUG: [habridge.log](https://github.com/username_1/ha-bridge/files/1653992/habridge.log) I don't see the header name and value pairs in the log. Am I missing something? Cheers username_0 Answers: username_0: BTW: Since using RC7, my Aurora works fine but I don't get any error messages at all ;-) username_1: @username_0 The headers don't get dumped on debug. I should add that. But I do see you now have duplicate content type statements as you selected that from the drop down and then put it in the header as well. Take out the content type on your header and try it. username_0: I did but it doesn't work username_0: Tried to do it again with RC13 but still no luck. ![image](https://user-images.githubusercontent.com/35510003/35767313-6afc3856-08ea-11e8-9f44-26ba4b948a2b.png) This is how I configured the device and in the screenshots above you can see how it works with postman. [habridge.log](https://github.com/username_1/ha-bridge/files/1691777/habridge.log) username_1: @username_0 The only thing I can see from Wemo examples on the web, is that you have slash after BasicEvent1. Remove that slash, sometimes it makes a difference, sometimes not. username_0: It does! Removing the slash in the URL did the trick. Thank you very much! Status: Issue closed
eslint/eslint
1065363241
Title: [flat config] `ecmaVersion: "latest"` Question: username_0: ### ESLint version `flat-config-linter` branch ### What problem do you want to solve? ref: https://github.com/eslint/eslint/pull/15185#discussion_r757536540 With this configuration, ESLint will use `espree.latestEcmaVersion` as ecmaVersion: ```js { languageOptions: { ecmaVersion: "latest" } } ``` ### What do you think is the correct solution? ESLint should maintain its own `latestEcmaVersion` instead of using `espree.latestEcmaVersion`. `languageOptions.ecmaVersion` is an ESLint option, and it is used not only for parsing. An in-range version of `espree` can support `ecmaVersion` that is not supported by currently used version of ESLint. If we use `espree.latestEcmaVersion`, parsing will work well, but other features that are based on `ecmaVersion` may not. For example, in the mentioned scenario, ESLint will not add any ES globals ### Participation - [X] I am willing to submit a pull request for this change. ### Additional comments _No response_ Answers: username_1: Good call. Is there any discussion we need to have here before marking as accepted and moving to Ready to Implement? username_2: Good idea. 👍
jembi/openhim-mediator-mapping
690956964
Title: Multiple Instances API Endpoint update not reflecting in other instances Question: username_0: If the mapping mediator has been configured to scale horizontally, there is more than one instance connect to the same database. If an API endpoint update is sent to one of these services it creates the DB entry correctly but doesnt update the cache on the other services with the new endpoint changes Answers: username_1: We'd need DB listeners implemented to repopulate the local cache
eslint/eslint
1123950949
Title: New Rule: Transactions in a function should be passed through all queries. Question: username_0: ### Rule details It should make sure that once the transaction is declared on top, it should pass all the queries in that function else it can choke the node server and has done multiple times. ### Related ECMAScript feature Not sure about this! ### What type of rule is this? Warns about a potential problem ### Example code ```js create: async (req, res) => { const transaction = await sequelize.transaction(); try { let user = await User.findOne(req.body.id); // no transaction passed should throw the error. if (!user) await User.create(user, { transaction }); } catch { await transaction.rollback(); res.status(400) } ``` ### Why should this rule be in the core instead of a plugin? I am not sure about this. ### Participation - [ ] I am willing to submit a pull request to implement this rule. ### Additional comments _No response_ Answers: username_1: This a good rule but it should be implemented in a plugin like `eslint-plugin-sequalize`. username_2: Yes, ESLint core rules can’t use any libraries or frameworks. Any rules specific to libraries or frameworks should be in a plugin. Status: Issue closed username_0: @username_2 , Can you kindly guide me on how i can initiate a request for that?
ballerina-platform/ballerina-lang
531741337
Title: Issue in building ballerina standard library module with external dependencies Question: username_0: **Description:** Build fails when building the ballerina-xslt module after migrating the native functions to java interop, although the dependencies are added to the toml file. ``` [[platform.libraries]] artifactId = "axiom-api" version = "1.2.22" path = "./lib/axiom-api-1.2.22.jar" groupId = "org.apache.ws.commons.axiom" modules = ["xslt"] [[platform.libraries]] artifactId = "axiom-impl" version = "1.2.22" path = "./lib/axiom-impl-1.2.22.jar" groupId = "org.apache.ws.commons.axiom" modules = ["xslt"] [[platform.libraries]] artifactId = "axiom-dom" version = "1.2.22" path = "./lib/axiom-dom-1.2.22.jar" groupId = "org.apache.ws.commons.axiom" modules = ["xslt"] ``` ``` [2019-12-03 10:30:26,486] SEVERE {b7a.log.crash} - org/apache/axiom/om/OMDocument java.lang.NoClassDefFoundError: org/apache/axiom/om/OMDocument at java.lang.Class.getDeclaredMethods0(Native Method) at java.lang.Class.privateGetDeclaredMethods(Class.java:2701) at java.lang.Class.privateGetPublicMethods(Class.java:2902) at java.lang.Class.getMethods(Class.java:1615) at org.ballerinalang.nativeimpl.jvm.interop.JMethodResolver.getExecutables(JMethodResolver.java:527) at org.ballerinalang.nativeimpl.jvm.interop.JMethodResolver.resolveByMethodName(JMethodResolver.java:114) at org.ballerinalang.nativeimpl.jvm.interop.JMethodResolver.resolve(JMethodResolver.java:84) at org.ballerinalang.nativeimpl.jvm.interop.JInteropMethodValidator.resolveJMethod(JInteropMethodValidator.java:74) at org.ballerinalang.nativeimpl.jvm.interop.JInteropMethodValidator.validateAndGetJMethod(JInteropMethodValidator.java:61) at ballerina.jvm.$value$InteropValidator.validateAndGetJMethod(interop.bal:90) at ballerina.jvm.$value$InteropValidator.call(interop.bal) at ballerina.compiler_backend_jvm.interop.interop_method_gen.createJMethodWrapper(interop/interop_method_gen.bal:135) at ballerina.compiler_backend_jvm.interop.interop_method_gen.createJInteropFunctionWrapper(interop/interop_method_gen.bal:127) at ballerina.compiler_backend_jvm.interop.external_method_gen.createExternalFunctionWrapper(interop/external_method_gen.bal:180) at ballerina.compiler_backend_jvm.jvm_package_gen.generateClassNameMappings(jvm_package_gen.bal:477) at ballerina.compiler_backend_jvm.jvm_package_gen.generatePackage(jvm_package_gen.bal:149) at ballerina.compiler_backend_jvm.main.generateJarBinary(main.bal:86) at ballerina.compiler_backend_jvm.main.main(main.bal:60) at ballerina.compiler_backend_jvm.___init.$lambda$main$(compiler_backend_jvm) at org.ballerinalang.jvm.scheduling.SchedulerItem.execute(Scheduler.java:426) at org.ballerinalang.jvm.scheduling.Scheduler.run(Scheduler.java:218) at org.ballerinalang.jvm.scheduling.Scheduler.runSafely(Scheduler.java:191) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.ClassNotFoundException: org.apache.axiom.om.OMDocument at java.net.URLClassLoader.findClass(URLClassLoader.java:381) at java.lang.ClassLoader.loadClass(ClassLoader.java:424) at java.lang.ClassLoader.loadClass(ClassLoader.java:357) ... 23 more ``` Answers: username_1: This was due to missing some dependencies in libs. Once you add those axiom libs issue got resolved, Status: Issue closed
chrisdev/wagtail-cookiecutter-foundation
160251768
Title: Nginx configuration should follow industry best practices Question: username_0: We can find a good example/tool here: https://wiki.mozilla.org/Security/Server_Side_TLS Answers: username_0: Modern or Intermediate? username_0: See #275 Status: Issue closed username_0: Ok we've used https://mozilla.github.io/server-side-tls/ssl-config-generator/?server=nginx-1.10.1&openssl=1.0.1e&hsts=yes&profile=intermediate for the SSL stuff closing
BLAKE2/libb2
133559220
Title: API docs? Question: username_0: Thanks. Answers: username_1: The complete public API lives in [blake2.h](https://github.com/BLAKE2/libb2/blob/master/src/blake2.h) and is self-explanatory (scroll to the bottom). You may also want to take a look at [RFC7693](https://tools.ietf.org/html/rfc7693). username_0: No problem, I figured it out from the python binding. Here's a LuaJIT binding if you want to add it to the website: https://luapower.com/blake2 username_2: There's nothing leading to `blake2.h` and the The API isn't self explanatory, for example: what's `outlen` in the `*_init` functions? Simple usage example with comments would help a lot instead of relying on guesswork. username_3: That's a good point. Most of what these values mean lives on the BLAKE2 paper instead of API code. For the record, `outlen` in the `_init` funcions refers to the desired hash output length. This is to differentiate, say, BLAKE2s with a 128-bit output from BLAKE2s with a 256-bit output; they'll effectively be different hash functions.
google/s2geometry
835194819
Title: Please make python link conditional on an cmake variable Question: username_0: It should be possible to turn off python bindings. Further, python bindings should be buildable separately from the C++ library, because system can have several python versions installed at the same time, and each one can have s2geometry bindings built for it. Answers: username_1: I wanted to package s2geometry as a library for Debian, and requiring python include files even when not needing the python interface was a killer. Some inspiration to make python optional can be found here: https://github.com/rinigus/s2geometry/blob/main/CMakeLists.txt Sorry, I don't know enough CMake magic to really provide a nice patch. username_1: Sorry, I see there is even a MR to provide this at https://github.com/google/s2geometry/pull/170 username_2: Fixed by #189 Status: Issue closed
platform9/express
527288826
Title: lvm fails to create on RHEL Question: username_0: there is a line in /etc/lvm/lvm.conf that prevents express from creating a lvm on RHEL filter = [ "a/dev/sd.*/", "r/.*/" ] this above line needs to be commented out because the below happens: TASK [map-role : debug] *********************************************************************************************************************************************** Friday 22 November 2019 16:14:07 +0000 (0:00:00.693) 0:25:40.441 ******* ok: [h1-rhel-4] => { "config_lvm": { "changed": true, "failed": false, "rc": 0, "stderr": "Shared connection to 10.128.233.73 closed.\r\n", "stderr_lines": [ "Shared connection to 10.128.233.73 closed." ], "stdout": "n=2, args=cinder-volumes ,[/dev/vdb,/dev/vdc,/dev/vdd]\r\n--> vgcreate cinder-volumes /dev/vdb\r\n Device /dev/vdb excluded by a filter.\r\n--> vgcreate cinder-volumes /dev/vdc\r\n Device /dev/vdc excluded by a filter.\r\n--> vgcreate cinder-volumes /dev/vdd\r\n Device /dev/vdd excluded by a filter.\r\n--- VG INFO ------------------------------------------\r\n No volume groups found.\r\n------------------------------------------------------\r\n", "stdout_lines": [ "n=2, args=cinder-volumes ,[/dev/vdb,/dev/vdc,/dev/vdd]", "--> vgcreate cinder-volumes /dev/vdb", " Device /dev/vdb excluded by a filter.", "--> vgcreate cinder-volumes /dev/vdc", " Device /dev/vdc excluded by a filter.", "--> vgcreate cinder-volumes /dev/vdd", " Device /dev/vdd excluded by a filter.", "--- VG INFO ------------------------------------------", " No volume groups found.", "------------------------------------------------------" ] } } Answers: username_1: Feedback from @tomchris `Can we mod this filter instead to accept /dev/sd. or /dev/vd.`
serge1/ELFIO
1013155866
Title: Read only required section information in the elfio class 'load' function Question: username_0: Can only the required sections of the executable be read into the memory instead of reading the complete binary/all the sections and segments of the binary in the 'class elfio' load() function? With the complete executable being read into the memory, there is a peak in memory usage when binary sizes are in GB. Can you please consider this enhancement? Answers: username_1: Thank you for your suggestion. The current design of the library based on the fact that lifetime of the file stream cannot be garanteed after `elfio::load()` function finishes. I'll see whether a lazy section/segment data loading is possible as far as user garantees proper lifetime of the stream. username_2: beside the peak in the memory, there is another reason to take this approach. some of linux distribution have live memory dump file in `/proc/kcore`, in this elf file you can't always read from all of his segments(some of them are not readable although the file is with read permissions). this scenario lead to error in the stream and later to mistakes in the data elfio parse from this file. username_1: Hi @username_2, it is interesting aspect. How do you urge the library to treat /proc/kcore being an ELF file? username_2: This file is in elf formate. It's looks like a core dump but his size is enormous and as I said not all the data is readable. You can read more about kcore file in [here](https://schlafwandler.github.io/posts/dumping-/proc/kcore/).
sfepy/sfepy
290817611
Title: update TetgenMeshIO to read vertex and cell groups Question: username_0: Currently, the additional data in the .node, .ele are just thrown away. Answers: username_1: Using these numbers to define vertex groups doesn't seem right to me. [1] http://wias-berlin.de/software/tetgen/fformats.examples.html [2] http://wias-berlin.de/software/tetgen/fformats.node.html username_0: @username_2 was this solved by using meshio? username_2: Not yet. I will prepare commit to meshio to fix this. username_0: OK, thanks! username_2: Resolved by nschloe/meshio#900. username_0: OK, let us close this. Thanks! Status: Issue closed
10gen/mongo-orchestration
52520317
Title: Cannot create initial user when clusterauthmode set Question: username_0: clusterAuthMode needs to be stripped out when creating the initial user :] ``` { "name": "mongod", "id" : "STANDALONE-X509", "auth_key": "secret", "login": "root", "password": "<PASSWORD>", "procParams": { "dbpath": "/tmp/standalone-x509/", "ipv6": true, "logappend": true, "logpath": "/tmp/standalone-x509/m.log", "journal": true, "port": 2300, "setParameter": {"enableTestCommands": 1} }, "sslParams": { "clusterAuthMode": "x509", "sslMode": "requireSSL", "sslCAFile": "/phongo/scripts/ssl/ca.pem", "sslPEMKeyFile": "/phongo/scripts/ssl/server.pem", "sslWeakCertificateValidation": true } } ``` results in: 2014-12-19T10:27:52.220-0800 [conn1] end connection 192.168.112.10:54698 (1 connection now open) 2014-12-19T10:27:52.228-0800 [conn2] warning: no SSL certificate provided by peer 2014-12-19T10:27:52.231-0800 [initandlisten] connection accepted from 192.168.112.10:54700 #3 (2 connections now open) 2014-12-19T10:27:52.232-0800 [conn2] end connection 192.168.112.10:54699 (1 connection now open) 2014-12-19T10:27:52.238-0800 [conn3] warning: no SSL certificate provided by peer 2014-12-19T10:27:52.239-0800 [conn3] Unauthorized not authorized on admin to execute command { usersInfo: "root" } 2014-12-19T10:27:52.240-0800 [conn3] Unauthorized not authorized on admin to execute command { createUser: "root", pwd: "xxx", digestPassword: false, roles: [ "__system", "clusterAdmin", "dbAdminAnyDatabase", "readWriteAnyDatabase", "userAdminAnyDatabase" ], writeConcern: { fsync: true } } 2014-12-19T10:27:52.241-0800 [signalProcessingThread] got signal 15 (Terminated), will terminate after current cmd ends 2014-12-19T10:27:52.241-0800 [conn3] end connection 192.168.112.10:54700 (0 connections now open) Answers: username_1: resolved in 9aa35e28487ce87da61996a25c9dccbf1367ee98 Status: Issue closed
junstyle/vscode-php-cs-fixer
951423394
Title: Latest version (0.2.5) - Undefined method PhpCsFixer\Config::create() Question: username_0: In the latest version (0.2.5) the extension stopped working for me. The following console error is logged to the vscode developer tools console. ```sh PHP Fatal error: Uncaught Error: Call to undefined method PhpCsFixer\Config::create() in /home/jannick/Workspace/verloning/.php-cs-fixer.php:139 Stack trace: #0 phar:///home/jannick/.vscode/extensions/username_1.php-cs-fixer-0.2.5/php-cs-fixer.phar/src/Console/ConfigurationResolver.php(836): include() #1 phar:///home/jannick/.vscode/extensions/username_1.php-cs-fixer-0.2.5/php-cs-fixer.phar/src/Console/ConfigurationResolver.php(229): PhpCsFixer\Console\ConfigurationResolver::separatedContextLessInclude() #2 phar:///home/jannick/.vscode/extensions/username_1.php-cs-fixer-0.2.5/php-cs-fixer.phar/src/Console/ConfigurationResolver.php(561): PhpCsFixer\Console\ConfigurationResolver->getConfig() #3 phar:///home/jannick/.vscode/extensions/username_1.php-cs-fixer-0.2.5/php-cs-fixer.phar/src/Console/ConfigurationResolver.php(406): PhpCsFixer\Console\ConfigurationResolver->getFormat() #4 phar:///home/jannick/.vscode/extensions/username_1.php-cs-fixer-0.2.5/php-cs-fixer.phar/src/Console/Command/FixCommand.php(258): PhpCsFixer\Console\ConfigurationResolver->getReporter() #5 phar:///home/jannick/.vscode/extensions/username_1.php-cs-fixer-0.2.5/php-cs-fixer.phar/vendor/symfony/console/Command/Command.php(255): PhpCsFixer\Console\Command\FixCommand->execute() #6 phar:///home/jannick/.vscode/extensions/username_1.php-cs-fixer-0.2.5/php-cs-fixer.phar/vendor/symfony/console/Application.php(1009): Symfony\Component\Console\Command\Command->run() #7 phar:///home/jannick/.vscode/extensions/username_1.php-cs-fixer-0.2.5/php-cs-fixer.phar/vendor/symfony/console/Application.php(273): Symfony\Component\Console\Application->doRunCommand() #8 phar:///home/jannick/.vscode/extensions/username_1.php-cs-fixer-0.2.5/php-cs-fixer.phar/src/Console/Application.php(97): Symfony\Component\Console\Application->doRun() #9 phar:///home/jannick/.vscode/extensions/username_1.php-cs-fixer-0.2.5/php-cs-fixer.phar/vendor/symfony/console/Application.php(149): PhpCsFixer\Console\Application->doRun() #10 /home/jannick/.vscode/extensions/username_1.php-cs-fixer-0.2.5/php-cs-fixer.phar(113): Symfony\Component\Console\Application->run() #11 {main} thrown in /home/jannick/Workspace/verloning/.php-cs-fixer.php on line 139 ``` When I downgrade to 0.2.2 the extension works as normal again. Answers: username_1: .php-cs-fixer.php this file is config file, please go to [php-cs-fixer](https://github.com/FriendsOfPHP/PHP-CS-Fixer) and see how to config it correctly .php-cs-fixer.php is config file of v3 PhpCsFixer\Config::create() not exists in v3, new PhpCsFixer\Config() instead https://github.com/FriendsOfPHP/PHP-CS-Fixer see detail Status: Issue closed
asmith4299/asmith4299
218068710
Title: Sửa Cửa Cuốn Không Xuống Được| CUACUON.ORG.VN | Chỉnh Hành Trình Cửa Cuốn | LH 094 331 6065 Question: username_0: S&#7917;a C&#7917;a Cu&#7889;n Kh&ocirc;ng Xu&#7889;ng &#272;&#432;&#7907;c| CUACUON.ORG.VN | Ch&#7881;nh H&agrave;nh Tr&igrave;nh C&#7917;a Cu&#7889;n | LH 094 331 6065<br> http://www.youtube.com/watch?v=OvZ60-QrWWk<br><br><br> via S&#7917;a C&#7917;a Cu&#7889;n Chuy&ecirc;n Nghi&#7879;p http://www.youtube.com/channel/UCbgEJCUDHEWAijVqo9U1xzg<br> March 30, 2017 at 09:57AM
mtp401/protoc-grpcio
328348267
Title: Update to protobuf 2.0. Question: username_0: https://github.com/pingcap/grpc-rs/pull/187 was recently merged into `grpc-rs` to move to `protobuf` ~2.0. This issue is a request for `protoc-grpcio` to update to `protobuf` 2.0 as well. This would be dependent upon a new release of the `grpcio-compiler` crate, which hasn't yet been pushed. I've pushed up commit https://github.com/username_2/protoc-grpcio/commit/a99c4d24f35a9a6a5793159686713f91865011d4 to my fork that, right now, is rev-pinned because the updated `grpcio-compiler` crate hasn't yet been released. I've tested it against my `build.rs` script to generate from my protos and it seems to work just fine thus far (my `grpc-rs` dependency was also rev-pinned to the merge commit of https://github.com/pingcap/grpc-rs/pull/187). Thank you! Answers: username_1: `grpcio-compile` version 0.3.0 was released and contains `protobuf` 2.0, so what about release new version of `protoc-grpcio`? username_2: Hello, I just pushed an updated version of protoc-grpcio which supports grpcio 0.3 and protobuf 2.0. Cheers, Matt username_0: Thanks @username_2!
pelias/api
99710104
Title: HTTP GET vs GET+POST for /search, /place, /reverse place endpoints Question: username_0: # The Question at Hand For API 1.0.0, should our core endpoints, `/search`, `/reverse`, and `/place` offer HTTP POST requests, or only accept GET requests at this time? # Summary of points ## Exclusively GET When we launch 1.0.0 of the API, we won't be supporting batch geocoding, and so for `/search`/ and `/reverse`, API users will be requesting information for a single place per call. As this is just requesting information about a record (or records) and not affecting the records themselves, they fall along the lines of what should be considered a classic HTTP GET request. While URLs may get somewhat long, the syntax will be clear and straightforward for structuring one of these calls. We are not likely to exceed the [~2000c suggestion for a URI](http://stackoverflow.com/questions/417142/what-is-the-maximum-length-of-a-url-in-different-browsers). Also, keeping in line with comparable geocoders and search services, most only offer GET requests for these functions. Meanwhile for `/place`, a user may request the records more than one single location, but even then, we can compact the results into a single JSON FeatureCollection with each place its own feature. ## GET + POST The arguments for why to support GET are sound (:+1: me), but we also are creating a pretty complex query structure that may get overwhelming for many types of queries. There's a lot you can do in this system, for instance, the passing of geometries, which are done through verbose statements of geometries and parameters. It would be a lot nicer to just pass these in as JSON objects. We'll have to write a schema for accepting these queries as JSON no matter what for batch geocoding, why not get a head start now? Answers: username_1: I'm a bit late to the party on this one but I agree we should support both `GET` and `POST` in some capacity there are some technical issues which might effect which gets used in which circumstance: - for performance critical endpoints such as `/autocomplete` it would be best to *only* allow `GET`, this will make caching easier for this endpoint; which doesn't allow complex nested input parameters. - for batch geocoding it is redundant to support `GET` at all, as `POST` will clearly be used in the vast majority of cases, the user agent for batch requests will usually be another server. it's a little trickier for `/search`, it's handy to have support for both `GET` and `POST` and for them to be used interoperably. we will need to be clear in the documentation on which variables take precedence over each other, eg in the request below *both* the `GET` param and the `POST` param are set with the same name: ```bash POST /v1/search?size=1 HTTP/1.1 Content-Type: application/x-www-form-urlencoded Content-Length: 32 size=2 ``` For content types, there are a bunch of options we can support: - application/x-www-form-urlencoded - multipart/form-data - application/json - application/xml - etc. I would suggest we **only** support `application/json` and we **do not** support `application/x-www-form-urlencoded`
alibaba/nacos
706073456
Title: nacos 启动失败 Question: username_0: org.springframework.context.ApplicationContextException: Unable to start web server; nested exception is org.springframework.boot.web.server.WebServerException: Unable to start embedded Tomcat at org.springframework.boot.web.servlet.context.ServletWebServerApplicationContext.onRefresh(ServletWebServerApplicationContext.java:157) at org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:540) at org.springframework.boot.web.servlet.context.ServletWebServerApplicationContext.refresh(ServletWebServerApplicationContext.java:142) at org.springframework.boot.SpringApplication.refresh(SpringApplication.java:775) at org.springframework.boot.SpringApplication.refreshContext(SpringApplication.java:397) at org.springframework.boot.SpringApplication.run(SpringApplication.java:316) at org.springframework.boot.SpringApplication.run(SpringApplication.java:1260) at org.springframework.boot.SpringApplication.run(SpringApplication.java:1248) at com.alibaba.nacos.Nacos.main(Nacos.java:35) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49) at org.springframework.boot.loader.Launcher.launch(Launcher.java:109) at org.springframework.boot.loader.Launcher.launch(Launcher.java:58) at org.springframework.boot.loader.PropertiesLauncher.main(PropertiesLauncher.java:604) Caused by: org.springframework.boot.web.server.WebServerException: Unable to start embedded Tomcat at org.springframework.boot.web.embedded.tomcat.TomcatWebServer.initialize(TomcatWebServer.java:125) at org.springframework.boot.web.embedded.tomcat.TomcatWebServer.<init>(TomcatWebServer.java:86) at org.springframework.boot.web.embedded.tomcat.TomcatServletWebServerFactory.getTomcatWebServer(TomcatServletWebServerFactory.java:414) at org.springframework.boot.web.embedded.tomcat.TomcatServletWebServerFactory.getWebServer(TomcatServletWebServerFactory.java:174) at org.springframework.boot.web.servlet.context.ServletWebServerApplicationContext.createWebServer(ServletWebServerApplicationContext.java:181) at org.springframework.boot.web.servlet.context.ServletWebServerApplicationContext.onRefresh(ServletWebServerApplicationContext.java:154) ... 16 common frames omitted Caused by: org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'distroFilterRegistration' defined in class path resource [com/alibaba/nacos/naming/web/NamingConfig.class]: Bean instantiation via factory method failed; nested exception is org.springframework.beans.BeanInstantiationException: Failed to instantiate [org.springframework.boot.web.servlet.FilterRegistrationBean]: Factory method 'distroFilterRegistration' threw exception; nested exception is org.springframework.beans.factory.UnsatisfiedDependencyException: Error creating bean with name 'distroFilter': Unsatisfied dependency expressed through field 'distroMapper'; nested exception is org.springframework.beans.factory.UnsatisfiedDependencyException: Error creating bean with name 'distroMapper' defined in URL [jar:file:/home/nacos/target/nacos-server.jar!/BOOT-INF/lib/nacos-naming-1.3.0.jar!/com/alibaba/nacos/naming/core/DistroMapper.class]: Unsatisfied dependency expressed through constructor parameter 0; nested exception is org.springframework.beans.factory.BeanCreationException: Error creating bean with name 'serverMemberManager' defined in URL [jar:file:/home/nacos/target/nacos-server.jar!/BOOT-INF/lib/nacos-core-1.3.0.jar!/com/alibaba/nacos/core/cluster/ServerMemberManager.class]: Bean instantiation via constructor failed; nested exception is org.springframework.beans.BeanInstantiationException: Failed to instantiate [com.alibaba.nacos.core.cluster.ServerMemberManager]: Constructor threw exception; nested exception is ErrCode:500, ErrMsg:jmenv.tbsite.net at org.springframework.beans.factory.support.ConstructorResolver.instantiate(ConstructorResolver.java:627) at org.springframework.beans.factory.support.ConstructorResolver.instantiateUsingFactoryMethod(ConstructorResolver.java:456) at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.instantiateUsingFactoryMethod(AbstractAutowireCapableBeanFactory.java:1288) at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1127) at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:538) at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:498) at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:320) at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:222) at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:318) at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:204) at org.springframework.boot.web.servlet.ServletContextInitializerBeans.getOrderedBeansOfType(ServletContextInitializerBeans.java:235) at org.springframework.boot.web.servlet.ServletContextInitializerBeans.getOrderedBeansOfType(ServletContextInitializerBeans.java:226) at org.springframework.boot.web.servlet.ServletContextInitializerBeans.addServletContextInitializerBeans(ServletContextInitializerBeans.java:101) at org.springframework.boot.web.servlet.ServletContextInitializerBeans.<init>(ServletContextInitializerBeans.java:88) at org.springframework.boot.web.servlet.context.ServletWebServerApplicationContext.getServletContextInitializerBeans(ServletWebServerApplicationContext.java:261) at org.springframework.boot.web.servlet.context.ServletWebServerApplicationContext.selfInitialize(ServletWebServerApplicationContext.java:234) at org.springframework.boot.web.embedded.tomcat.TomcatStarter.onStartup(TomcatStarter.java:54) at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5098) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183) at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1432) at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1422) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at org.apache.tomcat.util.threads.InlineExecutorService.execute(InlineExecutorService.java:75) at java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:134) at org.apache.catalina.core.ContainerBase.startInternal(ContainerBase.java:944) at org.apache.catalina.core.StandardHost.startInternal(StandardHost.java:831) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183) at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1432) at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1422) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at org.apache.tomcat.util.threads.InlineExecutorService.execute(InlineExecutorService.java:75) at java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:134) at org.apache.catalina.core.ContainerBase.startInternal(ContainerBase.java:944) [Truncated] at java.net.InetAddress.getAllByName(InetAddress.java:1193) at java.net.InetAddress.getAllByName(InetAddress.java:1127) at org.apache.http.impl.conn.SystemDefaultDnsResolver.resolve(SystemDefaultDnsResolver.java:45) at org.apache.http.impl.conn.DefaultHttpClientConnectionOperator.connect(DefaultHttpClientConnectionOperator.java:111) at org.apache.http.impl.conn.PoolingHttpClientConnectionManager.connect(PoolingHttpClientConnectionManager.java:353) at org.apache.http.impl.execchain.MainClientExec.establishRoute(MainClientExec.java:380) at org.apache.http.impl.execchain.MainClientExec.execute(MainClientExec.java:236) at org.apache.http.impl.execchain.ProtocolExec.execute(ProtocolExec.java:184) at org.apache.http.impl.execchain.RetryExec.execute(RetryExec.java:88) at org.apache.http.impl.execchain.RedirectExec.execute(RedirectExec.java:110) at org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:184) at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82) at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:107) at com.alibaba.nacos.common.http.BaseHttpClient.execute(BaseHttpClient.java:45) at com.alibaba.nacos.common.http.NacosSyncHttpClient.get(NacosSyncHttpClient.java:46) at com.alibaba.nacos.core.cluster.lookup.AddressServerMemberLookup.syncFromAddressUrl(AddressServerMemberLookup.java:155) at com.alibaba.nacos.core.cluster.lookup.AddressServerMemberLookup.run(AddressServerMemberLookup.java:101) ... 123 common frames omitted 2020-09-22 12:39:42,043 WARN [WatchFileCenter] start close [nacos-启动报错.txt](https://github.com/alibaba/nacos/files/5259208/nacos-.txt) Answers: username_0: 集群模式,其中一个pod副本启动总是报错这个。conf目录下面没有cluster.conf文件。 username_1: conf下加上cluster.conf就好了,通过operator去修改其中的内容就可以了 username_0: 我使用的镜像docker.io/nacos/nacos-server:1.3.0 进入这个镜像启动的容器内部查看如下,是有cluster.conf的 [root@4acf46f0b322 conf]# pwd /home/nacos/conf [root@4acf46f0b322 conf]# ll total 44 -rw-r--r-- 1 root root 2180 Jun 5 18:47 application.properties -rw-r--r-- 1 root root 1 Sep 22 17:43 cluster.conf -rw-r--r-- 1 501 games 25080 Jun 5 16:33 nacos-logback.xml -rw-r--r-- 1 501 games 7456 May 15 18:35 schema.sql 我的nacos是K8S部署的,通过helm的方式。我没有明白conf下加上cluster.conf如何加? username_1: 这就是k8s和operator的使用问题了,对于nacos而言,只需要在cluster.conf中添加正确的ip即可 username_0: 集群模式3个副本。可是有时候1个pod或者2个pod可以启动成功,启动不成功的。pod的容器里面就没有cluster.conf文件。 username_1: 这个和nacos就没有关系了,应该就是k8s或者operator有问题。 Status: Issue closed
ELVIS-Project/elvis-database
121966177
Title: Collections Users Question: username_0: We need to have the capability of assigning users to a specific collection (it does not matter whether the collection is *private* or *public*). The group of users should be able to add or delete files to a collection. The deletion process should move deleted files to a trash folder that can only be emptied by a super user of the editing group. This is an *essential* functionality when creating a data set collaboratively. Answers: username_1: We talked about this in Monday's meeting. This is not actually an issue. Status: Issue closed username_0: :-) username_0: We need to have the capability of assigning users to a specific collection (it does not matter whether the collection is *private* or *public*). The group of users should be able to add or delete files to a collection. The deletion process should move deleted files to a trash folder that can only be emptied by a super user of the editing group. This is an *essential* functionality when creating a data set collaboratively. username_0: This is not marked as a bug, just as an enhancement, and no we cannot assign a group of users to work on a specific data set. By it being public anybody with an account can work on the data set, and I'm not sure if that is desirable. Public or non public is not granular enough. username_2: Duplicate of #167 Status: Issue closed
sebastianbergmann/phpunit
117131996
Title: PHP7 syntax error causes "must be an instance of Exception, instance of Error given" error Question: username_0: Due to PHP7's new parent of Exception (called "Throwable") and because syntax errors are now catchable and thrown as type "Error", running phpunit (as of 5.0.9) on a script with a syntax error causes a rather unhelpful error message of ``` PHP Fatal error: Uncaught TypeError: Argument 2 passed to PHPUnit_Framework_TestResult::addError() must be an instance of Exception, instance of Error given, called in phar:///usr/local/bin/phpunit/phpunit/Framework/TestSuite.php on line 724 and defined in phar:///usr/local/bin/phpunit/phpunit/Framework/TestResult.php:202 Stack trace: #0 phar:///usr/local/bin/phpunit/phpunit/Framework/TestSuite.php(724): PHPUnit_Framework_TestResult->addError(Object(PHPUnit_Framework_TestSuite), Object(Error), 0) #1 phar:///usr/local/bin/phpunit/phpunit/Framework/TestSuite.php(747): PHPUnit_Framework_TestSuite->run(Object(PHPUnit_Framework_TestResult)) #2 phar:///usr/local/bin/phpunit/phpunit/TextUI/TestRunner.php(425): PHPUnit_Framework_TestSuite->run(Object(PHPUnit_Framework_TestResult)) #3 phar:///usr/local/bin/phpunit/phpunit/TextUI/Command.php(154): PHPUnit_TextUI_TestRunner->doRun(Object(PHPUnit_Framework_TestSuite), Array) #4 phar:///usr/local/bin/phpunit/phpunit/TextUI/Command.php(105): PHPUnit_TextUI_Command->run(Array, true) #5 /usr in phar:///usr/local/bin/phpunit/phpunit/Framework/TestResult.php on line 202 ``` Currently in order to see what the error actually is, we need to then run the code outside of phpunit to see the actual syntax error messaging. The request for this ticket is to handle the error gracefully and output PHP's error stack so that we may fix the syntax error. Thanks! Answers: username_1: Which version of PHPUnit are you using? username_1: I cannot reproduce this. Please provide a minimal, self-containted, recproducing test case. username_2: Sounds like an old version of phpunit is being used here. username_0: Thanks for looking into this. I caused the error with this code in a PHP file: ``` class Abc { public static function FunA() { var_dump(static::$xxxx); } } class SomeTest extends PHPUnit_Framework_TestCase { public static function setUpBeforeClass() { Abc::FunA(); } } ``` The syntax error in the above snippet is that $xxxx is not declared as a static property, therefore the PHP parse/compilation step fails. I'm not sure if this error is particular to statics, I only got my own case down to the above small example. Here is an output log: ``` $ php --version PHP 7.0.0RC5 (cli) ( NTS ) Copyright (c) 1997-2015 The PHP Group Zend Engine v3.0.0-dev, Copyright (c) 1998-2015 Zend Technologies $ phpunit --version PHPUnit 5.0.9 by <NAME> and contributors. $ phpunit abcTest.php PHPUnit 5.0.9 by <NAME> and contributors. PHP Fatal error: Uncaught TypeError: Argument 2 passed to PHPUnit_Framework_TestResult::addError() must be an instance of Exception, instance of Error given, called in phar:///usr/local/bin/phpunit/phpunit/Framework/TestSuite.php on line 724 and defined in phar:///usr/local/bin/phpunit/phpunit/Framework/TestResult.php:202 Stack trace: #0 phar:///usr/local/bin/phpunit/phpunit/Framework/TestSuite.php(724): PHPUnit_Framework_TestResult->addError(Object(PHPUnit_Framework_TestSuite), Object(Error), 0) #1 phar:///usr/local/bin/phpunit/phpunit/TextUI/TestRunner.php(425): PHPUnit_Framework_TestSuite->run(Object(PHPUnit_Framework_TestResult)) #2 phar:///usr/local/bin/phpunit/phpunit/TextUI/Command.php(154): PHPUnit_TextUI_TestRunner->doRun(Object(PHPUnit_Framework_TestSuite), Array) #3 phar:///usr/local/bin/phpunit/phpunit/TextUI/Command.php(105): PHPUnit_TextUI_Command->run(Array, true) #4 /usr/local/bin/phpunit(513): PHPUnit_TextUI_Command::main() #5 {main} thrown in phar:///usr/local/bin/phpunit/phpunit/Framework/TestResult.php on line 202 ``` username_1: `Error: Access to undeclared static property: Abc::$xxxx` is not a syntax but a runtime error. Will investigate the PHPUnit problem, though. username_1: Seems to only happen for errors triggered `setUpBeforeClass()`, not in `setUp()` or `test*()`. username_0: Ah thanks, sorry for my describing the problem wrong; you are correct. username_1: c3db067ba2f64843d6038c6dc9c51eca794f5e06 should fix this but it breaks backwards compatibility, meaning it cannot got into 5.x :-( Maybe I (or somebody else) can come up with a different solution that does not break BC. username_0: I was able to get a band aid working that works on both PHP 7 and PHP 5.6. IN my example above, adding this try/catch gives me the right error message on both versions: ``` class Abc { public static function FunA() { var_dump(static::$xxxx); } } class SomeTest extends PHPUnit_Framework_TestCase { public static function setUpBeforeClass() { try { Abc::FunA(); } catch (Throwable $e) { throw new Exception($e); } } } ``` I was half expecting PHP 5.6 to complain that Throwable is not defined, but it ran fine... Anyway, this is my temporary fix. Hopefully it helps people running into this. Status: Issue closed username_0: Thanks so much for fixing this! username_3: But like @username_0 said, this is a bandaid: you're just catching and ignoring the underlying error. I'm curious what the root cause is? Unless I'm misunderstanding the issue... Thanks username_0: @username_3 I think this has been fixed since that https://github.com/username_1/phpunit/commit/ac2e32c1d87cd8d2689a25e344a87de93858608c commit username_4: Looking for help on how to fix these 3 errors: Can someone provide instructions Notice: Constant JROUTER_MODE_RAW already defined in libraries/cms/router/router.php on line 17 Notice: Constant JROUTER_MODE_SEF already defined in libraries/cms/router/router.php on line 25 Fatal error: Uncaught TypeError: Argument 1 passed to JErrorPage::render() must be an instance of Exception, instance of ParseError given in libraries/cms/error/page.php:28 Stack trace: #0 [internal function]: JErrorPage::render(Object(ParseError)) #1 {main} thrown in libraries/cms/error/page.php on line 28 username_4: where is the solution should be for the libraries/cms/error/page.php on line 28?
alibaba/sentinel-golang
784260672
Title: [BUG] Dead Lock between Entry and LoadRules in v1.0.1 Question: username_0: ## Issue Description Type: *bug report* ### Describe what happened version: 1.0.1 ### Describe what you expected to happen ### How to reproduce it (as minimally and precisely as possible) 1. 2. 3. ### Tell us your environment ### Anything else we need to know?<issue_closed> Status: Issue closed
vmck/acs-interface
508473010
Title: Students List Question: username_0: Currently, you can only see all the submissions, for every homework, in one place. What I'm suggesting is: add a page with each student and his latest submission for each homework or class Something like the first vmchecker had. It would make grading easier and it would make it easier for students to access their review, too. Answers: username_1: Fixed by https://github.com/vmck/acs-interface/pull/81 with improvements tracked in https://github.com/vmck/acs-interface/issues/100 Status: Issue closed
33cn/plugin
585518008
Title: 平行链空块间隔参数配置优化 Question: username_0: 之前空块间隔的配置在批量读写toml配置文件时候支持比较复杂,改为简单的单kv 类型 Status: Issue closed Answers: username_1: :tada: This issue has been resolved in version 6.1.1 :tada: The release is available on [GitHub release](https://github.com/username_1/plugin/releases/tag/v6.1.1) Your **[semantic-release](https://github.com/semantic-release/semantic-release)** bot :package::rocket:
gsbelarus/check-and-cash
172126415
Title: Неправильно печатается ФИО кассира Question: username_0: POSitive Cash - при печати X, Z-отчетов на чеке вместо Кассир - <NAME>, печатается Кассир - кассир. При этом на фискальных чеках оплаты данные кассира печатаются правильно. Answers: username_1: Скачал версию с сайта. Проверил. Всё нормально печатается. Попробуйте просто обновить базу. Status: Issue closed
project-koku/koku-ui
501931743
Title: Cost models fit & finish: create cost model - select source step Question: username_0: **Describe the bug** Fit and finish **Expected behavior** - [ ] Description field should be a text area - [ ] Padding around fields - [ ] Length of fields shouldn't fill the space. Set them to the length of the longest value in the source type option. **Screenshots** ![image](https://user-images.githubusercontent.com/2453279/66112053-cacf8000-e5d2-11e9-8428-9eaec91d93c6.png)<issue_closed> Status: Issue closed
marta-file-manager/marta-issues
267236662
Title: Blurred text in panel after PageDown or Down keys Question: username_0: In 4.6 still persist this issue. This bug appear in directories where a lot of files, after passing "first portion" With mouse/touchpad scrolling font is normal. Answers: username_1: Unfortunately, I can't reproduce this behavior. Does this issue still persist in 0.4.6? username_0: In 4.6 still persist this issue. This bug appear in directories where a lot of files, after passing "first portion" With mouse/touchpad scrolling font is normal. username_1: @username_0 What kind of display do you use (Retina/non-retina)? username_1: I was able to reproduce the issue (without Retina display). ![screen shot 2018-01-03 at 13 56 06](https://user-images.githubusercontent.com/95996/34509511-24794d2a-f08e-11e7-912a-4c0ccc1130b3.png) ![screen shot 2018-01-03 at 13 56 16](https://user-images.githubusercontent.com/95996/34509512-289101aa-f08e-11e7-8d59-ab9de3b0657b.png) Status: Issue closed
gurman1995/capstone-project-embt-2020
729695497
Title: Schematic design Question: username_0: Abdul - What issues did you face when working with EasyEDA? Answers: username_1: Well, there were not many issues as the software was pretty easy and simple to use. The following are a couple of minor issues. 1. I personally feel that there are no standard components in the "systems" section of the search page of software for some components rather we had to search from the User-contributed section. 2. The 3D view isn't as expected. Status: Issue closed
jhermsmeier/node-vcf
460477255
Title: Expected VERSION Question: username_0: I am testing the parsing of vcard files with a bunch of sample vcard examples I found online. The birthday on them seems to come up before the version number, which I would assumed wouldn't break the parse. I'm just concerned if one of my users uploads a similar vcard with this similar format of NOT having the version number as the first data field, this parser cannot parse it and will fail. Is this a standardized failure by the example site I got these sample vcards from that is never to be expected with a vcard? I was thinking that this package would still parse through and search for a version number, but it appears not. Let me know what y'all think Error: ``` SyntaxError: Invalid vCard: Expected "VERSION:\d.\d" but found "BDAY;VALUE=DATE:1963-09-21" ``` This is the one I'm trying to parse: ``` BEGIN:VCARD BDAY;VALUE=DATE:1963-09-21 VERSION:3.0 N:Stenerson;Derik FN:<NAME> ORG:Microsoft Corporation ADR;TYPE=WORK,POSTAL,PARCEL:;;One Microsoft Way;Redmond;WA;98052-6399;USA TEL;TYPE=WORK,MSG:+1-425-936-5522 TEL;TYPE=WORK,FAX:+1-425-936-7329 EMAIL;TYPE=INTERNET:<EMAIL> END:VCARD BEGIN:VCARD VERSION:3.0 N:Ganguly;Anik FN:<NAME> ORG: Open Text Inc. ADR;TYPE=WORK,POSTAL,PARCEL:;Suite 101;38777 West Six Mile Road;Livonia;MI;48152;USA TEL;TYPE=WORK,MSG:+1-734-542-5955 EMAIL;TYPE=INTERNET:<EMAIL> END:VCARD BEGIN:VCARD VERSION:3.0 N:Moskowitz;Robert FN:<NAME> EMAIL;TYPE=INTERNET:<EMAIL> END:VCARD ``` Answers: username_1: Hi @username_0, the reason for that error is that the [specification mandates](https://tools.ietf.org/html/rfc6350#section-6.7.9) that the version must be the second value in a vCard – except in version 2.1, where it may appear anywhere in the vCard. This implementation currently also rejects this for version 2.1 though, which is something that'll need fixing: https://github.com/username_1/node-vcf/blob/e0a984fb294fb2b76f0e6bba009357f15195b92d/lib/vcard.js#L271-L273 Status: Issue closed
Alamofire/Alamofire
258788879
Title: Getting Data after Suspend Download (or quit app) in Mac OS Question: username_0: In the currently, I implement download for multi-file (pdf, image, txt...). When I quit app in Mac OS, I can not save current data (for example: I download 25% data) for download file. And when I launch app again, I want to the file continue downloading (with this data, and it will download with 25% data). Beside, I also have other issue when I pause (suspend download), I was been time out (default: 60s) and I want to pause download until user resume it. ## Alamofire Environment **Alamofire version: 4.5 **Xcode version: 8.3.3 **Swift version: 3.1 **Platform(s) running Alamofire: MacOS **macOS version running Xcode: 10.12.6 Answers: username_1: Pausing is usually better handled as a cancellation of the current request, in which you write whatever data already downloaded to disk (or if it's an Alamofire request started using `download`, this happens automatically). Then, when the user resumes, you can make the request again with the resume data, which should allow the download to pick up where it left off, if the server support resumption. Alamofire automatically supports this with the `download( resumingWith:to)` method, which is just a wrapper for the existing `URLSession` APIs which do the same thing. Status: Issue closed
notnoop/java-apns
769707
Title: Couldn't connec to APNS server java.net.UnknownHostException: gateway.sandbox.push.apple.com? Question: username_0: Any one has any idea about this exception. It was working unto yerterday. Couldn't connec to APNS server java.net.UnknownHostException: gateway.sandbox.push.apple.com Answers: username_1: I got same error today, I tried the ping command, and I get: "Ping request could not find host gateway.sandbox.push.apple.com. Please check the name and try again." Do you know why I cannot connect? username_2: Have you checked the port? The ping command normally uses port 80, but gateway.sandbox.push.apple.com listens on another port (sorry, I currently does not remember which port). Von meinem Samsung Gerät gesendet. -------- Ursprüngliche Nachricht -------- Firma: Capgemini Deutschland GmbH Aufsichtsratsvorsitzender: <NAME> • Geschäftsführer: Dr. <NAME> (Sprecher) • <NAME> • Dr. <NAME> • Dr. <NAME> Amtsgericht Berlin-Charlottenburg, HRB 98814 This message contains information that may be privileged or confidential and is the property of the Capgemini Group. It is intended only for the person to whom it is addressed. If you are not the intended recipient, you are not authorized to read, print, retain, copy, disseminate, distribute, or use this message or any part thereof. If you receive this message in error, please notify the sender immediately and delete all copies of this message. username_1: For testing I use port 2195, I tried also with ping gateway.sandbox.push.apple.com:2195, but got same message. Should I try with a different ping command ? Maybe it's from apple? username_3: @username_1 Got the same error today. Did you figured out what went wrong here? username_1: No, I keep getting Request timed out when trying to ping the address. I have not managed to solve it, will try the following days to check if I need to open a report at apple or something. username_4: We are also currently seeing this error intermittently which makes it hard to debug and difficult to diagnose what the problem actually is. username_5: I am facing the same issue. Anyone able to get this resolved?
mwaskom/seaborn
818296657
Title: Feature request: scatterplot, allow facecolor="none" and map the hue variable to edgecolor Question: username_0: It would be nice to have a simple way to specify hollow markers in `seaborn.scatterplot()`. One could map the `hue` variable to `edgecolor` by default and add a `facecolor` parameter. Currently, adding a `facecolor` parameter has no effect (probably because hue overrides?). Adding `edgecolor` works only when "a valid single color nor a color sequence consisting of single character color" is specified. Answers: username_1: I had been thinking about supporting this by adding a `hue_target` or similar parameter that could take `face` or `edge`, but this is a feasible alternate suggestion. I'd say it could go a bit further and map hue to the edges whenever `facecolor` is specified at all. (I guess I can't think of an obvious case where that's necessary, but it seems like supporting the general case has no cost, sooo). Another option would be to add `fill={bool}`, which is how the distribution plots work. I think the implementation would be a bit simpler. It wouldn't let you map edge with a different face color, but maybe that isn't useful enough. I'd want this to work consistently across all plots with artists that have facecolor/edgecolor, so `barplot`, the categorical scatter plots, and others should be included in any work. A relevant upstream issue is https://github.com/matplotlib/matplotlib/pull/17850. There are some significant complexities in matplotlib here to be aware of.
tealinuxos/lapor
145351308
Title: blackscreen Question: username_0: saat saya booting tealinux setelah loading daun tea, di layar hanya muncul cursor. tapi saya tekan win+r muncul pencarial, saya coba membuka aplikasi firefox bisa. mohon solusinya. trims Answers: username_1: terima kasih telah melapor. kalau boleh tahu dapat media instalasi dari mana? username_0: saya mendapatkannya dari doscom, lebih tepatnya http://pinguin.dinus.ac.id/iso/tealinuxos/TeaLinuxOS-7.1-i386.iso blackscreen itu langsung muncul, pdahal saya belum memilih live cd.. username_1: Coba booting ke bootable media instalasi TeaLinuxOS 7 anda, saat muncul icon keyboard di bawah tekan sembarang tombol sehingga muncul menu. Pilih "Try TeaLinuxOS without installing". Seharusnya sistem akan berjalan dengan baik. username_0: itu gak ada icon keyboard.. ak kmaren coba download yg amd64, sya coba bsa normal..tpi processor sya intel, bgus egk y d.install itu?? username_1: amd64 bisa dijalankan di prosesor intel yang mendukung arsitektur 64bit. saya pribadi biasa install linux 64bit di laptop intel saya. username_0: oke, trimakasih.. ditunggu tealinux 8 nya, :+1: :) username_1: sama sama. kalau bisa datang ya ke release party-nya. info http://tealinuxos.org/release username_0: pengen kk.. tpi sya tinggal di Imogiri, Bantul, Yogyakarta -_- klo tdak kberatan sya krimi link download iso TeaLinuxos 8, lgsung sya instal :) username_0: Kk tnya lg, tealinuxos sya eror stelah instal apliksi.. Jdi gni stelh booting tealnux kn hrus memsukkn pasword, bsa memilh adminstator,guest, tau other . Nah ketika sya sudh login administator hnya muncl walpaper, tdk ad icon, menu yg ad di atas, sya mengetk sesuatu tdk b.pengarh,.tpi klo sya login sbgai GUEST, tampilnnya normal bsa untk bkak apliksi,dll. username_1: aplikasi apa yang diinstall dan via apa installnya? username_0: sya install aplikasi Zekr dan elforkane melalui terminal semua . username_2: ka tealinuxOS 8 install printer canon gimana ? mohon bantuannya | TY username_1: cannon seri apa? username_1: @username_0 bisa dicoba dengan membuat user baru. apakah masih blank? username_1: @username_2 cannon seri apa? username_1: Harap diperhatikan, jika topik yang akan anda bahas berbeda dengan topik yang sudah ada, **mohon membuat topik/issue yang baru** untuk menjaga agar mudah terbaca. terima kasih. Status: Issue closed
xmake-io/xmake
854996124
Title: Provide option to remove CMAKE_INSTALL_LIBDIR setting Question: username_0: <!-- 所有 xmake 相关项目都属于个人开源项目,新特性的支持,取决于作者(ruki)自身的时间安排,即使被采纳,也不能确保一定会立即支持。 提交请求后,您可以耐心等待一段时间,当然,我们也欢迎您直接提交 pull request 来参与代码贡献。 另一种选择是通过公司赞助为项目提供财务支持,也能促使我们有足够的资金和时间去快速完成新特性的开发。不过前提是此特性已经被采纳,因此在赞助前请先联系作者进一步讨论相关特性需求。 --> ### 你在什么场景下需要该功能? ``` import("package.tools.cmake").install(package) ``` 上面的命令会自动设置CMAKE_INSTALL_LIBDIR为绝对路径。然而CMAKE_INSTALL_LIBDIR并不是一个规范的路径,一些库使用CMAKE_INSTALL_LIBDIR="lib",另一些使用CMAKE_INSTALL_LIBDIR="${CMAKE_INSTALL_PREFIX}/lib",官方文档未进行指定。设置此选项将导致使用前者定义的库无法正常编译 ### 描述可能的解决方案 提供选项, ``` import("package.tools.cmake").install(package, {}, {libdir = false}) ``` ### 描述你认为的候选方案 直接去掉这个设置,如果需要再手动加 Answers: username_1: 能给个具体的包case 和错误信息么 username_0: vtk包 https://github.com/Kitware/VTK/ 错误信息 ``` CMake Error at CMake/vtkModule.cmake:1829 (message): The `ARCHIVE_DESTINATION` must not be an absolute path. Use `CMAKE_INSTALL_PREFIX` to keep everything in a single installation prefix. Call Stack (most recent call first): CMake/vtkModule.cmake:2319 (_vtk_module_check_destinations) CMakeLists.txt:332 (vtk_module_build) ``` 相关文件 https://github.com/Kitware/VTK/blob/4f21cad8e3726f88c2cef519425c2c0284a212c2/CMake/vtkModule.cmake#L2320-L2322 顺带一提,VTK和cmake来自同一个组织Kitware username_1: 我改了下,你可以在外面重写 CMAKE_INSTALL_LIBDIR 试试,比如 `table.insert(configs, "-DCMAKE_INSTALL_LIBDIR=lib")` 会优先用外面设置的 Status: Issue closed
w3ctag/design-reviews
716206092
Title: APA Pronunciation Explainer Question: username_0: HIQaH! QaH! TAG! I'm requesting a TAG review of APA Pronunciation Explainer on behalf of the Pronunciation Task Force. This issue is just to ensure we schedule time to discuss this, and to track our thoughts on it. ... a standard mechanism to allow content authors to include spoken presentation guidance in HTML content. Also, [...] two identified approaches and enumerates their advantages and disadvantages. - Explainer¹ (minimally containing user needs and example code): https://w3c.github.io/pronunciation/explainer/ - Security and Privacy self-review²: n/a - GitHub repo (if you prefer feedback filed there): https://github.com/w3c/pronunciation - Organization/project driving the design: W3C Pronunciation TF We'd prefer the TAG provide feedback as (please delete all but the desired option): 💬 leave review feedback as a **comment in this issue** Answers: username_1: We had a joint meeting with APA about this during TPAC. I'm not sure how to find the minutes from that meeting—I'll keep looking—but it was a very productive hour IIRC, and I don't think there's much left to be done on our part. We should consider closing this issue once I've linked the minutes from here. username_0: As @username_1 mentioned above, we gave feedback on this directly at TPAC last year. We are expecting the proposal will likely change shape as a result of that proposal, so we're going to close this review for now and look forward to either re-opening that review or receiving a new review request once that process has completed. Thank you for working with us! Status: Issue closed
amhk/ash
441891196
Title: Add support for 'ash cmd' and [[command]] in the config file Question: username_0: Add support for specifying arbitrary shell commands in the config file, to be executed by 'ash cmd <name>'. This is a convenience for any loose shell scripts that I tend to migrate from Android repo to Android repo, such as launching the emulator.
bcurran3/ChocolateyPackages
316572468
Title: [Question] choco-cleaner Question: username_0: #username_1 You choco-cleaner is really nice and does a great job. Although, have you considered creating a CCleaner entry alongside choco-cleaner for users who want to clean Chocolatey through CCleaner instead. Thanks https://github.com/MoscaDotTo/Winapp2 Answers: username_0: Yes, its possible but, I don’t know how widely known it is, but its all explained in the [reed me](https://github.com/MoscaDotTo/Winapp2#winapp2). I created a [PR](https://github.com/MoscaDotTo/Winapp2/pull/256/files) but I don’t know if I covered everything. I am not a licenced user but thanks any way, yes, I noticed that it sadly never got published. username_1: @username_0 FYI: There might still be some changes enforced by ferventcoder, but I'm close to releasing a new "sanctified" version of choco-cleaner. https://chocolatey.org/packages/choco-cleaner/0.0.3-pre
simonw/simonwillisonblog
655131915
Title: Fix all the old simon.incutio.com internal links Question: username_0: 177 results for https://username_0illison.net/admin/blog/entry/?q=simon.incutio - most of these are internal links between my blog entries which are now broken. I could write a database migration script to fix these. Answers: username_0: These are links like `http://simon.incutio.com/archive/2003/07/30/python23` - on https://username_0illison.net/2006/Sep/22/excited/ - which need to be redirected to convert the 07 into Jul so that link becomes https://username_0illison.net/2003/Jul/30/python23/ Status: Issue closed
vitessio/vitess
1052215408
Title: panic: runtime error: slice bounds out of range [-20:] / Execute: unsupported query / types does not support hashcode yet: VARCHAR Question: username_0: When I start the app and it does those queries, the vtgate crashes and it logs the following: ```bash I1112 17:20:42.335863 1 vtgate.go:559] Execute: unsupported query: "select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted2 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted3 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted4 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted5 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted6 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted7 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted8 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted9 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted10 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted11" panic: runtime error: slice bounds out of range [-20:] ``` I also noticed an error on the nodejs side: ``` code: 'ER_NOT_SUPPORTED_YET', errno: 1235, sqlMessage: 'types does not support hashcode yet: VARCHAR', sqlState: '42000', index: 0, ``` Do you have any idea why this happens? Here are the logs: - [start_logs.log](https://github.com/vitessio/vitess/files/7529414/start_logs.log) - Queries sent to vtgate - [vtgate.log](https://github.com/vitessio/vitess/files/7529415/vtgate.log) - [vttablet.log](https://github.com/vitessio/vitess/files/7529416/vttablet.log) - [mysqld.log](https://github.com/vitessio/vitess/files/7529413/mysqld.log) I'm using the `vitess/lite:latest` docker image. Answers: username_1: I could not immediately reproduce this. Obviously the panic isn't correct; but I suspect there's something lower-level going on. Do we know what is causing all the vttablet errors like: ``` W1112 17:20:42.418812 1 tabletserver.go:1392] Code: ABORTED ``` ? Getting a general log from MySQL might be useful. username_0: I'm trying to activate the general logs, but I can't find a way to do it. Like I said, I'm running this in a cluster using the operator from planetscale. I've passed the following as an extra arg: ```bash --mycnf_general_log_path=/vt/vtdataroot/general_log.log ``` This was the only one found in [this documentation](https://vitess.io/docs/reference/programs/vttablet/). But it doesn't show up in `my.cnf`. I've also tried activating the general logs at run time, using the following queries: ```sql SET global general_log = ON; SET global log_output = 'file'; SET global general_log_file='/vt/vtdataroot/vt_1009888160/general_log.log'; ``` But the variables do not change, the `general_log` remains set to `OFF`. Do you know a way to activate those logs? username_0: @username_1 These are the params set to the vttablet: ``` vttablet: Container ID: containerd://81c42f30d6ed2850ddef3afeb1b17394dcbe9d10ff94272f24d8892168f4e403 Image: vitess/lite:latest Image ID: docker.io/vitess/lite@sha256:e2de042ceca8e0d2095f514ab0ffd98558d7030a80ea191e531f07b8b2013f35 Ports: 15000/TCP, 15999/TCP Host Ports: 0/TCP, 0/TCP Command: /vt/bin/vttablet Args: --backup_engine_implementation=xtrabackup --backup_storage_compress=true --backup_storage_implementation=s3 --binlog_use_v3_resharding_mode=true --client-found-rows-pool-size=300 --db-config-app-uname=vt_app --db-config-dba-uname=vt_dba --db-config-filtered-uname=vt_filtered --db-config-repl-uname=vt_repl --db_charset=utf8mb4 --enable_replication_reporter=true --enable_semi_sync=false --enforce_strict_trans_tables=true --grpc_max_message_size=67108864 --grpc_port=15999 --health_check_interval=5s --init_db_name_override=vt_sefsantier --init_keyspace=sefsantier --init_shard=- --init_tablet_type=replica --logtostderr=true --mycnf_error_log_path=/vt/vtdataroot/error_log.log --mycnf_general_log_path=/vt/vtdataroot/general_log.log --mycnf_socket_file=/vt/socket/mysql.sock --mysqlctl_socket=/vt/socket/mysqlctl.sock --port=15000 --queryserver-config-max-result-size=100000 --queryserver-config-pool-size=96 --queryserver-config-query-timeout=900 --queryserver-config-stream-pool-size=96 --queryserver-config-transaction-cap=300 --restore_concurrency=10 --restore_from_backup=true --s3_backup_aws_endpoint=s3.aws.com --s3_backup_aws_region=eu-central-003 --s3_backup_storage_bucket=bucket --s3_backup_storage_root=vt --service_map=grpc-queryservice,grpc-tabletmanager,grpc-updatestream --tablet-path=decontabodusseldorf-1009888160 --tablet_hostname=$(POD_IP) --topo_global_root=/vitess/vt/global --topo_global_server_address=vt-etcd-ec571fc7-client.default.svc:2379 --topo_implementation=etcd2 --wait_for_backup_interval=10s --xbstream_restore_flags=--parallel=2 --xtrabackup_backup_flags=--parallel=1 --xtrabackup_stream_mode=xbstream --xtrabackup_stripes=8 [Truncated] # If the -enable_semi_sync flag is used, VTTablet will enable semi-sync # at the proper time when replication is set up, or when a primary is # promoted or demoted. plugin-load = rpl_semi_sync_master=semisync_master.so;rpl_semi_sync_slave=semisync_slave.so # When semi-sync is enabled, don't allow fallback to async # if you get no ack, or have no replicas. This is necessary to # prevent alternate futures when doing a failover in response to # a primary that becomes unresponsive. rpl_semi_sync_master_timeout = 1000000000000000000 rpl_semi_sync_master_wait_no_slave = 1 ## /vt/config/mycnf/rbr.cnf binlog_format=row ## /vt/config/mycnf/log-error.cnf log-error = /vt/config/stderr.symlink ## /vt/config/mycnf/socket.cnf socket = /vt/socket/mysql.sock ``` username_0: I think those are generated by the logged query. The query is right below that error. It can also be found in the [start_logs.log](https://github.com/vitessio/vitess/files/7529414/start_logs.log) file. Which contains all queries sent to the vtgate. username_2: With Gen4 or Gen4Fallback you should not see issue with inforamtion_schema query. The fix was merged 5 days back with https://github.com/vitessio/vitess/pull/9183 username_0: @username_2 I was using the latest version(the main branch), 3 days ago. Which contained that, if it was integrated 5 days ago. Should I pass something to vitess to activate that? Is that an option? username_2: To be sure, share the commit SHA, one of the place to find is on VTGate UI, or run `show variables like '%version%'` username_0: @username_2 ```sql MySQL [(none)]> show variables like '%version%' -> ; +-------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------+ | Variable_name | Value | +-------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------+ | innodb_version | 5.7.31 | | protocol_version | 10 | | slave_type_conversions | | | tls_version | TLSv1,TLSv1.1,TLSv1.2 | | version | 5.7.9-vitess-13.0.0-SNAPSHOT | | version_comment | Version: 13.0.0-SNAPSHOT (Git revision 77b980630c branch 'main') built on Fri Nov 12 15:50:52 UTC 2021 by vitess@buildkitsandbox using go1.17 linux/amd64 | | version_compile_machine | x86_64 | | version_compile_os | Linux | +-------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------+ ``` But, shouldn't it be activated with a param first? username_2: also can you post `explain format=vitess <select_statement` username_0: ```sql MySQL [(none)]> explain format=vitess select * from company; +----------+-----------------+------------+-------------+------------+-----------------------+ | operator | variant | keyspace | destination | tabletType | query | +----------+-----------------+------------+-------------+------------+-----------------------+ | Route | SelectUnsharded | ionescu | | UNKNOWN | select * from company | +----------+-----------------+------------+-------------+------------+-----------------------+ 1 row in set (0.050 sec) ``` username_2: For the failing query :) username_2: have you set `-planner_version` on VTGate flag? username_0: ```sql explain format=vitess SELECT `TABLE_SCHEMA`, `TABLE_NAME` FROM `INFORMATION_SCHEMA`.`TABLES` WHERE `TABLE_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'company_invite_code' UNION SELECT `TABLE_SCHEMA`, `TABLE_NAME` FROM `INFORMATION_SCHEMA`.`TABLES` WHERE `TABLE_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'site_role' UNION SELECT `TABLE_SCHEMA`, `TABLE_NAME` FROM `INFORMATION_SCHEMA`.`TABLES` WHERE `TABLE_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'item' UNION SELECT `TABLE_SCHEMA`, `TABLE_NAME` FROM `INFORMATION_SCHEMA`.`TABLES` WHERE `TABLE_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'site_item_urgent' UNION SELECT `TABLE_SCHEMA`, `TABLE_NAME` FROM `INFORMATION_SCHEMA`.`TABLES` WHERE `TABLE_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'site_item_event' UNION SELECT `TABLE_SCHEMA`, `TABLE_NAME` FROM `INFORMATION_SCHEMA`.`TABLES` WHERE `TABLE_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'site_item' [Truncated] | │ │ │ │ │ └─ Concatenate | | | | UNKNOWN | | | │ │ │ │ │ ├─ Distinct | | | | UNKNOWN | | | │ │ │ │ │ │ └─ Concatenate | | | | UNKNOWN | | | │ │ │ │ │ │ ├─ Distinct | | | | UNKNOWN | | | │ │ │ │ │ │ │ └─ Concatenate | | | | UNKNOWN | | | │ │ │ │ │ │ │ ├─ Distinct | | | | UNKNOWN | | | │ │ │ │ │ │ │ │ └─ Concatenate | | | | UNKNOWN | | | │ │ │ │ │ │ │ │ ├─ Route | SelectDBA | ionescu | | UNKNOWN | select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :__vtschemaname and TABLE_NAME = :TABLE_NAME | | │ │ │ │ │ │ │ │ └─ Route | SelectDBA | ionescu | | UNKNOWN | select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :__vtschemaname and TABLE_NAME = :TABLE_NAME1 | | │ │ │ │ │ │ │ └─ Route | SelectDBA | ionescu | | UNKNOWN | select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :__vtschemaname and TABLE_NAME = :TABLE_NAME2 | | │ │ │ │ │ │ └─ Route | SelectDBA | ionescu | | UNKNOWN | select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :__vtschemaname and TABLE_NAME = :TABLE_NAME3 | | │ │ │ │ │ └─ Route | SelectDBA | ionescu | | UNKNOWN | select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :__vtschemaname and TABLE_NAME = :TABLE_NAME4 | | │ │ │ │ └─ Route | SelectDBA | ionescu | | UNKNOWN | select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :__vtschemaname and TABLE_NAME = :TABLE_NAME5 | | │ │ │ └─ Route | SelectDBA | ionescu | | UNKNOWN | select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :__vtschemaname and TABLE_NAME = :TABLE_NAME6 | | │ │ └─ Route | SelectDBA | ionescu | | UNKNOWN | select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :__vtschemaname and TABLE_NAME = :TABLE_NAME7 | | │ └─ Route | SelectDBA | ionescu | | UNKNOWN | select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :__vtschemaname and TABLE_NAME = :TABLE_NAME8 | | └─ Route | SelectDBA | ionescu | | UNKNOWN | select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :__vtschemaname and TABLE_NAME = :TABLE_NAME9 | +------------------------------------------------------------------------------------+-----------+------------+-------------+------------+-------------------------------------------------------------------------------------------------------------------------------------+ 28 rows in set (0.052 sec) ``` username_0: @username_2 no, I haven't. That's why I was asking above. I searched in the vtgate flags, and I couldn't find that flag: https://vitess.io/docs/reference/programs/vtgate/ username_2: we will update the docs, kindly set it to `Gen4Fallback` username_2: does that solve for you? username_0: @username_2 Last night it was late and I postponed it for today. I've added the flag to the vtgate only, and retried. But now I receive another error: ```sql E1116 08:01:43.506683 1 vtgate.go:542] Execute: symbol CONSTRAINT_NAME not found in table or subquery, request: map[BindVariables:map[vtg1:type:VARBINARY value:"ionescu" vtg10:type:VARBINARY value:"user_company" vtg11:type:VARBINARY value:"user" vtg12:type:VARBINARY value:"PRIMARY" vtg2:type:VARBINARY value:"company_invite_code" vtg3:type:VARBINARY value:"site_role" vtg4:type:VARBINARY value:"item" vtg5:type:VARBINARY value:"site_item_urgent" vtg6:type:VARBINARY value:"site_item_event" vtg7:type:VARBINARY value:"site_item" vtg8:type:VARBINARY value:"site" vtg9:type:VARBINARY value:"company"] Session:in_transaction:true shard_sessions:{target:{keyspace:"ionescu" shard:"-" tablet_type:PRIMARY} transaction_id:1636990091136287261 tablet_alias:{cell:"decontabodusseldorf" uid:1009888160}} autocommit:true target_string:"ionescu" options:{included_fields:ALL client_found_rows:true workload:OLTP} found_rows:59 row_count:-1 DDLStrategy:"direct" SessionUUID:"6fb7225b-46b3-11ec-87f8-2afd29d0a3f4" enable_system_settings:true Sql:SELECT * FROM ( SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE `kcu`.`TABLE_SCHEMA` = 'ionescu' AND `kcu`.`TABLE_NAME` = 'company_invite_code' UNION SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE `kcu`.`TABLE_SCHEMA` = 'ionescu' AND `kcu`.`TABLE_NAME` = 'site_role' UNION SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE `kcu`.`TABLE_SCHEMA` = 'ionescu' AND `kcu`.`TABLE_NAME` = 'item' UNION SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE `kcu`.`TABLE_SCHEMA` = 'ionescu' AND `kcu`.`TABLE_NAME` = 'site_item_urgent' UNION SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE `kcu`.`TABLE_SCHEMA` = 'ionescu' AND `kcu`.`TABLE_NAME` = 'site_item_event' UNION SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE `kcu`.`TABLE_SCHEMA` = 'ionescu' AND `kcu`.`TABLE_NAME` = 'site_item' UNION SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE [Truncated] --cell=decontabodusseldorf --cells_to_watch=decontabodusseldorf --enable_buffer=true --grpc_max_message_size=67108864 --grpc_port=15999 --logtostderr=true --mysql_auth_server_impl=static --mysql_auth_server_static_file=/vt/secrets/vtgate-static-auth/users.json --mysql_auth_static_reload_interval=30s --mysql_server_port=3306 --planner_version=Gen4Fallback --port=15000 --service_map=grpc-vtgateservice --tablet_types_to_wait=MASTER,REPLICA --topo_global_root=/vitess/vt/global --topo_global_server_address=vt-etcd-ec571fc7-client.default.svc:2379 --topo_implementation=etcd2 ``` If I run it with `explain`, I receive the same error. username_0: THese are the logs when the vtgate starts: ``` ERROR: logging before flag.Parse: E1116 07:57:12.232758 1 syslogger.go:149] can't connect to syslog I1116 07:57:12.400618 1 servenv.go:97] Version: 13.0.0-SNAPSHOT (Git revision 77b980630c branch 'main') built on Fri Nov 12 15:50:52 UTC 2021 by <EMAIL>andbox using go1.17 linux/amd64 I1116 07:57:12.441947 1 healthcheck.go:262] loading tablets for cells: decontabodusseldorf I1116 07:57:12.441979 1 healthcheck.go:282] Setting up healthcheck for cell: decontabodusseldorf I1116 07:57:12.442342 1 flags.go:160] vtgate buffer enabled. PRIMARY requests will be buffered during detected failovers. I1116 07:57:12.443038 1 gateway.go:103] Gateway waiting for serving tablets of types [PRIMARY REPLICA] ... I1116 07:57:12.469031 1 healthcheck.go:324] Adding tablet to healthcheck: alias:{cell:"decontabodusseldorf" uid:1009888160} hostname:"10.42.1.183" port_map:{key:"grpc" value:15999} port_map:{key:"vt" value:15000} keyspace:"ionescu" shard:"-" key_range:{} type:PRIMARY db_name_override:"vt_ionescu" mysql_hostname:"10.42.1.183" mysql_port:3306 primary_term_start_time:{seconds:1636748889 nanoseconds:620052460} I1116 07:57:12.474901 1 tablet_health_check.go:111] HealthCheckUpdate(Serving State): tablet: decontabodusseldorf-1009888160 (10.42.1.183) serving false => true for ionescu/- (PRIMARY) reason: healthCheck update I1116 07:57:42.444673 1 healthcheck.go:695] couldn't find tablets for target: keyspace:"ionescu" shard:"-" tablet_type:REPLICA cell:"decontabodusseldorf" W1116 07:57:42.447100 1 gateway.go:116] Timeout waiting for all keyspaces / shards to have healthy tablets of types [PRIMARY REPLICA], may be in degraded mode I1116 07:57:42.447687 1 vtgate.go:106] Transaction mode: 'MULTI' I1116 07:57:42.463573 1 vschema_manager.go:102] Received vschema update I1116 07:57:42.465909 1 vschema_manager.go:102] Received vschema update I1116 07:57:42.466866 1 streamlog.go:158] Streaming logs from VTGate at /debug/querylog. I1116 07:57:42.467501 1 grpc_server.go:159] Setting grpc max message size to 67108864 I1116 07:57:42.467660 1 grpc_server.go:330] Building interceptors with 0 unary interceptors and 0 stream interceptors I1116 07:57:42.469327 1 auth_server_clientcert.go:39] Not configuring AuthServerClientCert because mysql_server_ssl_ca is empty I1116 07:57:42.470047 1 statsd.go:71] statsdAddress is empty I1116 07:57:42.470451 1 service_map.go:64] Registering vtgateservice for grpc, disable it with -grpc-vtgateservice service_map parameter I1116 07:57:42.471212 1 auth_server_ldap.go:59] Not configuring AuthServerLdap because mysql_ldap_auth_config_file and mysql_ldap_auth_config_string are empty I1116 07:57:42.473545 1 auth_server_vault.go:70] Not configuring AuthServerVault, as -mysql_auth_vault_addr is empty. I1116 07:57:42.481345 1 grpc_server.go:232] Listening for gRPC calls on port 15999 I1116 07:57:42.482208 1 unix_socket.go:36] Not listening on socket file E1116 08:01:43.506683 1 vtgate.go:542] ``` username_2: thanks for sharing the exact query, will fix this and update on the issue username_0: @username_2 please notify me via tag when you managed to add a fix for it and it was integrated in main. I will use the latest docker image for it and I will be able to test it. Thanks again for everything. :bowing_man: Status: Issue closed username_2: @username_0 the fix is merged, you can try now and let us know. username_0: @username_2 I will test it, do you have a link where I can follow when the new `latest` image was built? I'm not sure if this build is generated once per day or for every commit on `main`. username_2: I also do not have clarity on this topic cc: @deepthi username_0: @username_2 I've noticed that a new build was made for `latest`. I've used that one: ```bash Your MySQL connection id is 4 Server version: 5.7.9-vitess-13.0.0-SNAPSHOT Version: 13.0.0-SNAPSHOT (Git revision d6b22587e0 branch 'main') built on Fri Nov 19 08:22:31 UTC 2021 by vitess@<EMAIL> using go1.17 linux/amd64 ``` The `d6b22587e0` matches the latest commit from main. So it contains your fix. I've tested, but now I've encountered another error: **symbol s.INDEX_NAME not found in table or subquery** ```sql E1119 08:55:09.344136 1 vtgate.go:553] Execute: symbol s.INDEX_NAME not found in table or subquery, request: map[BindVariables:map[vtg1:type:VARBINARY value:"ionescu" vtg10:type:VARBINARY value:"user_company" vtg11:type:VARBINARY value:"user" vtg12:type:VARBINARY value:"PRIMARY" vtg2:type:VARBINARY value:"company_invite_code" vtg3:type:VARBINARY value:"site_role" vtg4:type:VARBINARY value:"item" vtg5:type:VARBINARY value:"site_item_urgent" vtg6:type:VARBINARY value:"site_item_event" vtg7:type:VARBINARY value:"site_item" vtg8:type:VARBINARY value:"site" vtg9:type:VARBINARY value:"company"] Session:in_transaction:true shard_sessions:{target:{keyspace:"ionescu" shard:"-" tablet_type:PRIMARY} transaction_id:1637311473806207639 tablet_alias:{cell:"decontabodusseldorf" uid:1009888160}} autocommit:true target_string:"ionescu" options:{included_fields:ALL client_found_rows:true workload:OLTP} found_rows:6 row_count:-1 DDLStrategy:"direct" SessionUUID:"65ca253f-4916-11ec-be7e-5ec2bb1c4e0c" enable_system_settings:true Sql: SELECT `s`.* FROM ( SELECT * FROM `INFORMATION_SCHEMA`.`STATISTICS` WHERE `TABLE_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'company_invite_code' UNION SELECT * FROM `INFORMATION_SCHEMA`.`STATISTICS` WHERE `TABLE_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'site_role' UNION SELECT * FROM `INFORMATION_SCHEMA`.`STATISTICS` WHERE `TABLE_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'item' UNION SELECT * FROM `INFORMATION_SCHEMA`.`STATISTICS` WHERE `TABLE_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'site_item_urgent' UNION SELECT * FROM `INFORMATION_SCHEMA`.`STATISTICS` WHERE `TABLE_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'site_item_event' UNION SELECT * FROM `INFORMATION_SCHEMA`.`STATISTICS` WHERE [Truncated] `TABLE_NAME` = 'user_company' UNION SELECT * FROM `INFORMATION_SCHEMA`.`REFERENTIAL_CONSTRAINTS` WHERE `CONSTRAINT_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'user' ) `rc` ON `s`.`INDEX_NAME` = `rc`.`CONSTRAINT_NAME` AND `s`.`TABLE_SCHEMA` = `rc`.`CONSTRAINT_SCHEMA` WHERE `s`.`INDEX_NAME` != 'PRIMARY' AND `rc`.`CONSTRAINT_NAME` IS NULL] E1119 08:55:14.345330 1 throttled.go:77] Execute: skipped 1 log messages ``` username_2: When I start the app and it does those queries, the vtgate crashes and it logs the following: ```bash I1112 17:20:42.335863 1 vtgate.go:559] Execute: unsupported query: "select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted2 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted3 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted4 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted5 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted6 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted7 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted8 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted9 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted10 union select TABLE_SCHEMA, TABLE_NAME from INFORMATION_SCHEMA.`TABLES` where TABLE_SCHEMA = :redacted1 and TABLE_NAME = :redacted11" panic: runtime error: slice bounds out of range [-20:] ``` I also noticed an error on the nodejs side: ``` code: 'ER_NOT_SUPPORTED_YET', errno: 1235, sqlMessage: 'types does not support hashcode yet: VARCHAR', sqlState: '42000', index: 0, ``` Do you have any idea why this happens? Here are the logs: - [start_logs.log](https://github.com/vitessio/vitess/files/7529414/start_logs.log) - Queries sent to vtgate - [vtgate.log](https://github.com/vitessio/vitess/files/7529415/vtgate.log) - [vttablet.log](https://github.com/vitessio/vitess/files/7529416/vttablet.log) - [mysqld.log](https://github.com/vitessio/vitess/files/7529413/mysqld.log) I'm using the `vitess/lite:latest` docker image. username_0: @username_2 I also received a similar error for the following query: `Error: ER_UNKNOWN_ERROR: symbol rc.CONSTRAINT_SCHEMA not found in table or subquery` Maybe it will be easier to investigate if you have multiple queries. ```sql SELECT `kcu`.`TABLE_SCHEMA`, `kcu`.`TABLE_NAME`, `kcu`.`CONSTRAINT_NAME`, `kcu`.`COLUMN_NAME`, `kcu`.`REFERENCED_TABLE_SCHEMA`, `kcu`.`REFERENCED_TABLE_NAME`, `kcu`.`REFERENCED_COLUMN_NAME`, `rc`.`DELETE_RULE` `ON_DELETE`, `rc`.`UPDATE_RULE` `ON_UPDATE` FROM ( SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE `kcu`.`TABLE_SCHEMA` = 'ionescu' AND `kcu`.`TABLE_NAME` = 'company_invite_code' UNION SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE `kcu`.`TABLE_SCHEMA` = 'ionescu' AND `kcu`.`TABLE_NAME` = 'site_role' UNION SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE `kcu`.`TABLE_SCHEMA` = 'ionescu' AND `kcu`.`TABLE_NAME` = 'item' UNION SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE `kcu`.`TABLE_SCHEMA` = 'ionescu' AND `kcu`.`TABLE_NAME` = 'site_item_urgent' UNION SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE `kcu`.`TABLE_SCHEMA` = 'ionescu' AND `kcu`.`TABLE_NAME` = 'site_item_event' UNION SELECT * FROM `INFORMATION_SCHEMA`.`KEY_COLUMN_USAGE` `kcu` WHERE [Truncated] `CONSTRAINT_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'user_company' UNION SELECT * FROM `INFORMATION_SCHEMA`.`REFERENTIAL_CONSTRAINTS` WHERE `CONSTRAINT_SCHEMA` = 'ionescu' AND `TABLE_NAME` = 'user' ) `rc` ON `rc`.`CONSTRAINT_SCHEMA` = `kcu`.`CONSTRAINT_SCHEMA` AND `rc`.`TABLE_NAME` = `kcu`.`TABLE_NAME` AND `rc`.`CONSTRAINT_NAME` = `kcu`.`CONSTRAINT_NAME` ``` username_0: Hi @username_2, any news on this? username_2: I could not take a look at it, I have been busy with other commitments. I will update here when I will pick this up.
asyncapi/parser-js
672676681
Title: Make sure new parser releases are not introducing any regression Question: username_0: #### Reason/Context Parser has many tests but not all cases are covered. https://github.com/asyncapi/tck/ was created to hold all the possible valid and invalid cases. We should make sure that with new releases the percentage of the parser compatibility doesn't decrease. #### Description Maybe we can add an action that runs tck on a PR and validates the numbers and block merge? tck outputs JSON with results so this should be relatively easy Answers: username_1: @username_0 I want to work on this. I think _tck_ is not available as an npm repository. If you make it available then we can install it, generate the results and write a test to validate the result? or something? If you could show me the way(or some similar thing that has been done) I can start working on it. username_0: The trick here is to get `tck` and run against parser from PR and not the one specified in dependencies. We didn't do it anywhere yet. This is what needs to be figured out and I don't have a clear recommendation there as I haven't look into it that deeply, I'm not an author of tck. The options we have here: - tck is not only a set of test documents but also a project where a number of parsers are integrated and there are runners for those parsers that generate overall reports. Easiest would be to reuse the runner and use parser, but not sure if it is possible without some improvement in the runner - most probably we need a custom runner (which I don't think is super bad) specially designed for testing on PRs. Then on NPM we would only publish this directory https://github.com/asyncapi/tck/tree/master/tests/asyncapi-2.0. The advantage of publishing only asyncapi files is that we could also reuse the package for running snapshot testing for templates. Where to put this custom runner? 🤷🏼 this is open topic, to decide what is the best. Disadvantage of custom runner is that we would reinvent again the percentage calculation that we already have in the current runner So as you can see there are 2 different solutions and you would have to first dig into the `tck` (not much code there really) and see what is better, improve existing code or actually write something new. Would be awesome if you decide working on that, that would help us a lot to improve testing. Once you decide working on it please first start with drafting a proposal on what would be the solution in your opinion
desktop/desktop
435467986
Title: I can't upload my node_modules into my github repository Question: username_0: hello , i tried so hard to upload my node_modules folder into my github repository but it keep showing those errors .. please help me ![image](https://user-images.githubusercontent.com/40015058/56464616-71d5ac00-63e5-11e9-99b2-dde61eca9bfe.png) ![image](https://user-images.githubusercontent.com/40015058/56464621-7e5a0480-63e5-11e9-8867-82ca80b1b773.png) Status: Issue closed Answers: username_1: @username_0 you need to make a commit to the current branch: ``` $ git add [paths] $ git commit -m "[some message]" ``` Then you should be able to push to the remote branch.
chaintip/bounties
318918757
Title: [omersiar/esp-rfid]: Led and Buzzer Support Question: username_0: **Repository**: **[omersiar/esp-rfid](http://github.username_0.org/omersiar/esp-rfid)** **Issue [#114](http://github.username_0.org/omersiar/esp-rfid/issues/114)**: **Led and Buzzer Support** Hi. Can you add a buzzer controller pin in hardware menu? (Hocam kartın sistemi çalıştırdığını anlamak için buzzer eklenmesi mümkün mü? Bu güzel bir ek özellik olacaktır. Röleyi seçtiğimiz gibi buzzer için de bir pin seçilmesi mümkün olur mu?) *** **Total Bounty**: `0.00358035 BCH`| `~ 4.97 USD` To claim this bounty, get a pull request merged with ```@username_0 fixes #114``` in the creation comment. **On-Chain Tips**: - `0.00358035 BCH`| `~ 4.97 USD` [from](https://explorer.bitcoin.com/bch/address/bitcoincash:qqh0ld4v2fsutwt6c4n932w4j6wy20emkucv2nyqw3) **omersiar** to return to sender in 4 weeks from now. Answers: username_1: Thank
facebook/react-native
511371000
Title: Fresh Project fail to start Question: username_0: The current character read is 'E' with an int value of 69 Unable to determine the current character, it is not a string, number, array, or object line number 1 index number 0` Answers: username_1: Could you share your project via a git repo so we can investigate it further? username_0: i am going to closed that issue it's due to node version i upgraded from 8.0.9 to 12.0.0 and it work fine don't install node v 12.11.0 it have a bug username_0: i am going to closed that issue it's due to node version i upgraded from 8.0.9 to 12.0.0 and it work fine don't install node v 12.11.0 it have a bug Status: Issue closed
JuliaLang/julia
258026361
Title: RFC: Get rid of #undef and replace it with null in Array{Union{T, Null} Question: username_0: I mentioned this idea at https://github.com/JuliaLang/julia/issues/16029#issuecomment-327420330 but I thought this proposal entails large enough changes that it deserves its own issue. The idea is to get rid of `#undef` entries in non-`isbits` arrays, and replace it with `null`(from #23642). This would have several advantages: - Use a single representation for missing values everywhere: currently `#undef` is kind of weird since you can get this value, but never set it again, and having multiple notions of missing/uninitialized value increases the complexity of the language. - Guarantee that an `Array{T}` always contains valid values of type `T`: currently there's always the possibility that indexing into a non-`isbits` array throws an `UndefRefError`, which is akin to the "billion dollar mistake" (though in a less severe form since it only affects arrays). Concretely, in order to be able to set uninitialized values to `null`, we must force all uninitialized array constructors to create `Array{Union{Null, T}}` rather than `Array{T}` objects. For non-`isbits` arrays, this is not a problem since there is room to store the type tag already, or a `NULL` pointer can be translated to `null` as a built-in special case. For `isbits` arrays, with @username_1's recent work to optimize `Union`s (https://github.com/JuliaLang/julia/pull/22441), the performance impact is limited: it is equivalent to allocating an `Array{T}` plus an `Array{UInt8}` (the latter storing the type, i.e. whether the value is initialized or not). Moreover, with `Union{Null, T}` the type tag for `Null` is equal to `0`, which means the `Array{UInt8}` part can be allocated directly and efficiently using `calloc`. But of course for `isbits` arrays it would be more common to create an initialized `Array{T}` directly using `zeros` or `fill` (which would be the recommended constructors). Once all entries in an unitialized `Array{Union{Null, T}}` have been replaced with valid objects, the array can be turned into an `Array{T}` via a simple call to `convert`. This never requires making a copy, so it's very efficient. Then the `Array{T}` can safely be passed to functions which expect it to contain only valid `T` objects, which is now always guaranteed. If it turns out that creating the `Array{UInt8}` type tag part is too costly in some very particular cases, an `unsafe_array` function could be provided, which would allow creating uninitialized `Array{T}` objects for `isbits` types. This should probably not be allowed for non-`isbits` types (though if it was the current behavior of `#undef` could possibly be retained, with the expectation that most users would never see it). Since it forces filling arrays with valid values on construction, this proposal effectively amounts to forbidding uninitialized arrays (https://github.com/JuliaLang/julia/issues/9147). Answers: username_1: @username_7 and I discussed this at JuliaCon (getting rid of #undef). I'm not sure about the `Union{T, Null}` approach, as it seems to imply every array would potentially cost an extra byte per element. Jameson and I talked about just getting rid of the uninitialized Array constructors (i.e. `Vector{T}(n)`), so you'd be left w/ `fill(val)` to initialize arrays. My impression is that it actually wouldn't be that crazy of a deprecation/removal, as you almost always know a value to initialize with. While it would certainly be nice to reduce language complexity w/ one less form of "undef"/"missinginess", we mainly discussed this in the context of https://github.com/JuliaLang/julia/pull/18632, which has been a long-desired optimization. While that PR has been deferred due to being "an optimization", we should still consider any deprecations that might be needed in order to facilitate those optimizations. username_0: I think "getting rid of the uninitialized Array constructors" is essentially equivalent to this proposal, since then you have no way to create an "empty" `Array{T}` except by creating an `Array{Union{Null, T}}` and convert it once it's fully initialized. So (as I say in the description) in most cases you'd rather use `zeros` or `fill` if you know a reasonable initial value and it's not too costly to create. username_1: Oh, one other thing I'd mention is that with nice syntax for `Union{T, Null}`, it would be a very trivial change to require of users to go from `Vector{T}(n)` to `Vector{T?}(n)`. username_2: Since `#undef` values are entirely poisonous, would this change be breaking? I think this could be a feature in 1.x. username_0: What would be breaking would be removing uninitialized array constructors, which is mostly what this proposal is about. So I don't think the milestone is appropriate. username_2: You're absolutely right - and the representation of the eltype would need to be carefully considered, too. username_3: If we remove `#undef`, we also need to talk about undefined variables. Some variables (namely captured mutable variables) are implemented as struct fields, so variables and struct fields need to support the same behaviors with respect to undefined-ness. My interpretation of the "billion dollar mistake" is the introduction of first-class null references --- i.e. null references you can pass to a function. The problem is that null references propagate, so that a problem can appear far from the source. Immediately raising an error on reading a null reference helps localize the problem. So whatever the merit of replacing `#undef` with first-class Null values might be, it moves closer to the "billion dollar mistake" rather than farther IMO. Also, `Any` includes `Null`, so would `Vector{Any}(n)` yield an array full of `null`s? I understand wanting to encourage constructors like `fill`, but replacing an array of undefined references with an array of objects that are technically "defined" but not what you want is not necessarily an improvement. username_4: The implementation could be changed to implicitly use `T?` or `Some{T}?` for values that can be uninitialized (depending on whether `nothing isa T` or not). My suspicion is that the really problematic part of the billion dollar mistake is not that `null` is first-class in C/C++/Java, but rather that its _exclusion_ cannot be easily expressed in their type systems. In other words, for any reference type in those languages (pointers in C/C++, objects in Java), you cannot express that you always have to be worried that you might have a `null` instead of a valid reference to a value of that type. In this approach that would still be a problem for `Any` but assuming `Any` is the parent type of `Void`, it wouldn't be a problem for any other abstract types: if you have a `Vector{<:Number}` you can be sure that there are no `nothing` values; you could only have `nothing` values in a `Vector{<:Number?}`. username_4: I should add that I think we're getting a bit late in the 1.0 timeline to be considering such a fundamental change. However, since it's hard to write code that relies on `#undef` precisely because it's not a first-class value, we may want to consider if it might be possible to leave ourselves the option of changing this behavior in a non-breaking or minimally breaking way. username_3: Yes, being able to express the exclusion of null mitigates the issue to a large extent, but the point is that raising an error on reading an undefined location is not the billion dollar mistake. In fact it's almost as far as you can get from it --- not only *can* you express that something is non-null, you're forced to express that in many contexts (i.e. it implicitly puts a `::NonNull` on certain operations). username_0: Deprecating uninitialized array constructors would clearly be breaking; fortunately it's not terribly hard to do (marking them as `unsafe_*`). Subtler issues about undefined variables and fields could be less breaking and may require more work. username_3: I'm definitely interested in changing the array constructors for #16029. Keeping our current semantics but making it "harder" to get uninitialized arrays is something I could be on board with. username_5: I agree with everything Jeff has said. The typical "legitimate" use case for `#undef` is when constructing (in multiple steps) a mutable struct or array (eg make an array and populate the elements in a loop). One way to accommodate this might be to have a "construction" phase followed by a "publishing" step where all refs are validated to be defined (and perhaps that any user defined invariants are followed)? It's a bit of an odd idea, I admit. I'm username_0: That's more or less what this proposal suggests, by requiring a call to `convert(Array{T}, x)` once initialization is done, which would guarantee early failures if there are nulls. username_5: @username_0 I wonder if I'd prefer something... completely orthogonal to the type system, I suppose. For example, I want my `Vector{Any}` to only contain nulls where I put them, so if I forget to populate each element explicitly it throws. username_0: Yeah, but that system doesn't work for `isbits` types, where an uninitialized value cannot be distinguished from a valid one. It's annoying to have inconsistent safety features depending on the kind of type you are using (or which the caller may have passed to you). username_6: I recall reading early threads which wanted the uninitialized array constructors because `zeros` and `ones` was far less performant. It seems that's still the case? ```julia @benchmark zeros(1000) BenchmarkTools.Trial: memory estimate: 7.94 KiB allocs estimate: 1 -------------- minimum time: 455.375 ns (0.00% GC) median time: 1.205 μs (0.00% GC) mean time: 1.926 μs (25.12% GC) maximum time: 24.584 μs (90.79% GC) -------------- samples: 10000 evals/sample: 144 @benchmark Vector{Float64}(1000) BenchmarkTools.Trial: memory estimate: 7.94 KiB allocs estimate: 1 -------------- minimum time: 101.663 ns (0.00% GC) median time: 183.116 ns (0.00% GC) mean time: 381.337 ns (44.79% GC) maximum time: 2.417 μs (68.70% GC) -------------- samples: 10000 evals/sample: 956 ``` username_3: Yes, we'd still need a way to get uninitialized arrays. But the syntax for it can change. username_3: Conclusion from triage: radical changes to `#undef` are too much to bite off for 1.0; we should focus on the array constructors. Status: Issue closed
madisongh/digsigserver
715904677
Title: stderr: ERR: flash variable set not defined - kernel-bup-payload Question: username_0: past the signing issue, last error i believe digiserver in debug output [2020-10-06 14:01:57 -0400] [22555] [INFO] Goin' Fast @ http://0.0.0.0:9999 [2020-10-06 14:01:57 -0400] [22558] [INFO] Starting worker [22558] [2020-10-06 14:02:21 -0400] [22558] [INFO] manifest line: DTBFILE=tegra186-quill-p3310-1000-c03-00-base.dtb [2020-10-06 14:02:21 -0400] [22558] [INFO] manifest line: ODMDATA=0x1090000 [2020-10-06 14:02:21 -0400] [22558] [INFO] manifest line: LNXFILE=boot.img [2020-10-06 14:02:21 -0400] [22558] [INFO] manifest line: BOARDID=3310 [2020-10-06 14:02:21 -0400] [22558] [INFO] manifest line: FAB=C04 [2020-10-06 14:02:21 -0400] [22558] [INFO] manifest line: fuselevel=fuselevel_production [2020-10-06 14:02:21 -0400] [22558] [INFO] manifest line: localbootfile=boot.img [2020-10-06 14:02:21 -0400] [22558] [INFO] manifest line: boardcfg= [2020-10-06 14:02:21 -0400] [22558] [INFO] manifest line: CHIPREV=0 [2020-10-06 14:02:21 -0400] [22558] [INFO] manifest line: BOARDSKU= [2020-10-06 14:02:21 -0400] [22558] [INFO] manifest line: BOARDREV= [2020-10-06 14:02:21 -0400] [22558] [INFO] manifest line: BUPGENSPECS=fab=B00 fab=B02 fab=C04 fab=D00 fab=D01 [2020-10-06 14:02:21 -0400] [22558] [DEBUG] Setting: FAB=B00 [2020-10-06 14:02:21 -0400] [22558] [INFO] Running: ['tegra186-flash-helper', '--bup', '-u', '/tmp/tmpbbotn6a8/rsa_priv.pem', '-v', '/tmp/tmpbbotn6a8/sbk.txt', 'flash.xml.in', 'tegra186-quill-p3310-1000-c03-00-base.dtb', 'jetson-xavier-nx-devkit-sb.cfg', '0x1090000', 'boot.img'] [2020-10-06 14:02:21 -0400] [22558] [WARNING] signing error, stdout: stderr: ERR: flash variable set not defined [2020-10-06 14:02:21 -0400] - (sanic.access)[INFO][127.0.0.1:42768]: POST http://127.0.0.1:9999/sign/tegra 500 13 and logfile of error cat /home/dingo/tegra-test-distro/build/tmp/work/jetson_xavier_nx_devkit_sb-oe4t-linux/kernel-bup-payload/1.0-r0/temp/log.do_deploy.22863 DEBUG: Executing python function sstate_task_prefunc DEBUG: Python function sstate_task_prefunc finished DEBUG: Executing python function extend_recipe_sysroot NOTE: Direct dependencies are ['/home/dingo/tegra-test-distro/layers/meta-tegra/recipes-bsp/tegra-binaries/tegra186-flashtools-native_32.4.3.bb:do_populate_sysroot', 'virtual:native:/home/dingo/tegra-test-distro/layers/meta/recipes-core/coreutils/coreutils_8.32.bb:do_populate_sysroot', '/home/dingo/tegra-test-distro/layers/meta-tegra/recipes-bsp/tegra-binaries/tegra-bootfiles_32.4.3.bb:do_populate_sysroot', 'virtual:native:/home/dingo/tegra-test-distro/layers/meta/recipes-kernel/dtc/dtc_1.6.0.bb:do_populate_sysroot', '/home/dingo/tegra-test-distro/layers/meta-tegra/recipes-bsp/tegra-binaries/tegra-redundant-boot-base_32.4.3.bb:do_populate_sysroot'] NOTE: Installed into sysroot: [] NOTE: Skipping as already exists in sysroot: ['tegra186-flashtools-native', 'coreutils-native', 'tegra-bootfiles', 'dtc-native', 'tegra-redundant-boot-base', 'quilt-native', 'tegra-helper-scripts-native', 'autoconf-native', 'gnu-config-native', 'attr-native', 'automake-native', 'xz-native', 'libtool-native', 'texinfo-dummy-native', 'gettext-minimal-native', 'gcc-cross-aarch64', 'gcc-runtime', 'patch-native', 'python3-native', 'pseudo-native', 'glibc', 'custom-flash-layout', 'mender-custom-flash-layout', 'tegra-flashvars', 'bootfiles', 'pkgconfig-native', 'flex-native', 'tegra-binaries-patches', 'kern-tools-native', 'qemuwrapper-cross', 'bc-native', 'update-rc.d-native', 'cpio-native', 'rpm-native', 'openssl-native', 'kern-tools-tegra-native', 'binutils-cross-aarch64', 'mklibs-native', 'createrepo-c-native', 'opkg-utils-native', 'prelink-native', 'libgcc', 'dnf-native', 'kmod-native', 'pigz-native', 'linux-tegra', 'cross-localedef-native', 'depmodwrapper-cross', 'opkg-native', 'perl-native', 'ldconfig-native', 'makedevs-native', 'bison-native', 'keystore', 'arm-trusted-firmware', 'm4-native', 'libmpc-native', 'zlib-native', 'linux-libc-headers', 'gmp-native', 'mpfr-native', 'libtirpc-native', 'ncurses-native', 'readline-native', 'sqlite3-native', 'util-linux-native', 'libffi-native', 'libnsl2-native', 'gdbm-native', 'bzip2-native', 'shared-mime-info-native', 'shadow-native', 'qemu-native', 'systemd-systemctl-native', 'popt-native', 'db-native', 'elfutils-native', 'file-native', 'curl-native', 'libxml2-native', 'ninja-native', 'cmake-native', 'expat-native', 'glib-2.0-native', 'binutils-native', 'python3-iniparse-native', 'librepo-native', 'libcomps-native', 'libdnf-native', 'gtk-doc-native', 'debianutils-native', 'libsolv-native', 'libarchive-native', 'libcap-ng-native', 'libpcre2-native', 'meson-native', 'xmlto-native', 'itstool-native', 're2c-native', 'libpcre-native', 'gettext-native', 'python3-six-native', 'gpgme-native', 'libcheck-native', 'libmodulemd-native', 'gobject-introspection-native', 'json-c-native', 'swig-native', 'e2fsprogs-native', 'lzo-native', 'python3-setuptools-native', 'docbook-xsl-stylesheets-native', 'docbook-xml-dtd4-native', 'libxslt-native', 'libassuan-native', 'libgpg-error-native', 'libyaml-native', 'unzip-native'] DEBUG: Python function extend_recipe_sysroot finished DEBUG: Executing shell function do_deploy 1+0 records in 1+0 records out 4096 bytes (4.1 kB, 4.0 KiB) copied, 5.3751e-05 s, 76.2 MB/s WARNING: exit code 22 from a shell command. ERROR: Execution of '/home/dingo/tegra-test-distro/build/tmp/work/jetson_xavier_nx_devkit_sb-oe4t-linux/kernel-bup-payload/1.0-r0/temp/run.do_deploy.22863' failed with exit code 22: 1+0 records in 1+0 records out 4096 bytes (4.1 kB, 4.0 KiB) copied, 5.3751e-05 s, 76.2 MB/s WARNING: exit code 22 from a shell command. Answers: username_1: That message (`ERR: flash variable set not defined`) is reported by the flash helper script after sourcing the `flashvars` file that should have been passed along from the build but finding that the `FLASHVARS` environment variable isn't defined. It looks like it's trying to sign for a TX2-based (tegra186) machine, rather than a Xavier-based machine (tegra194). That might be part of the problem. Status: Issue closed
reazen/relude
535911570
Title: Consider switching to GitHub actions for CI stuff Question: username_0: We are currently using CircleCI for builds, but it might make sense to switch everything to GitHub actions at this point. If we do that, I'd like to do it for all of our projects for consistency. Answers: username_0: I setup a GH action to build relude, and the initial cut should work fine, so I think I'm going to update all the reazen projects to use GH actions, and then kill all of our CircleCI stuff. username_0: I think I have all the reazen projects switched over to GH actions. Status: Issue closed
bgruening/galaxytools
103335157
Title: DotKnot Question: username_0: DotKnot is a heuristic method for pseudoknot prediction in a given RNA sequence. DotKnot extracts stem regions from the secondary structure probability dot plot calculated by RNAfold. Recursive H-type pseudoknots and intramolecular kissing hairpins are constructed and their presence in the sequence is verified. http://dotknot.csse.uwa.edu.au/ Answers: username_1: I am working on that tool username_1: done Status: Issue closed
grails/grails-core
185980158
Title: Grails 3.2.2: Interceptor not executed Question: username_0: ### Steps to Reproduce 1. install grails 3.2.2 2. grails -v to confirm i am running 3.2.2 3. grails create-app myapp 4. add an interceptor myapp/grails-app/controllers/myapp/AuthenticationInterceptor.groovy 5. grails clean && grails war 6. java -jar -Xdebug -Xrunjdwp:transport=dt_socket,server=y,suspend=n,address=5005 build/libs/myapp.war 7. From my IntelliJ IDEA, set a breakpoint in the before() method of AuthenticationInterceptor 8. Start debug from IDE 9. Access http://localhost:8080/ ### Expected Behaviour 1. Execution suspends at the breakpoint ### Actual Behaviour 2. Grails default index.html welcome page is shown; interceptor breakpoint is not hit. ### Environment Information - **Operating System**: macOS Sierra Version 10.12.1 - **Grails Version:** 3.2.2 - **JDK Version:** java version "1.8.0_102", Java(TM) SE Runtime Environment (build 1.8.0_102-b14), Java HotSpot(TM) 64-Bit Server VM (build 25.102-b14, mixed mode) ### Example Application https://github.com/username_0/myapp.git ### Note The breakpoint is hit when I do: grails run-app --debug-jvm Answers: username_1: I can't reproduce this problem. How are you launching the project from within the IDE? username_1: I am using IntelliJ IDEA 2016.2.4 username_0: IntelliJ IDEA 2016.2.5 Build #IU-162.2228.15, built on October 14, 2016 username_0: <img width="1066" alt="screen shot 2016-10-28 at 10 58 08 am" src="https://cloud.githubusercontent.com/assets/9458969/19816767/7accea3c-9cfd-11e6-8562-14f9b688b65a.png"> username_1: My mistake. I see now that you are running with `java -jar...`. I had overlooked that. Sorry for the noise. username_2: I can confirm this as I just ran into this problem. Interceptors will not run when running with java -jar Happens in Grails 3.2.0 and Grails 3.2.2 I've have not tried to deploy the war file on a Tomcat server yet. username_3: `grails-app/conf/BootStrap.groovy` seems to suffer the same problem, not executed when running with `java -jar` since Grails 3.2.0. username_4: I looked at the `grails.factories` for the interceptor plugin that was bundled with the war and it contained: ``` grails.compiler.traits.TraitInjector=grails.compiler.traits.InterceptorTraitInjector grails.core.ArtefactHandler=org.grails.plugins.web.interceptors.InterceptorArtefactHandler ``` When the classloader gets the factories, the result of `url.text` for the interceptor `grails.factories` is: ``` #Grails Factories File #Thu Oct 27 15:24:17 UTC 2016 grails.compiler.traits.TraitInjector=grails.compiler.traits.InterceptorTraitInjector ``` So something is not being read correctly. I verified the URL path was the same path of the file I inspected. username_5: Workaround is to use `grails-plugin-interceptor` from 3.2.1 for the moment: ```groovy compile "org.grails:grails-plugin-interceptors:3.2.1" ``` Status: Issue closed
pionl/laravel-chunk-upload
711770903
Title: Laravel 8 support Question: username_0: Laravel 8 released in September 8. However this package does not support the latest version by dependencies reason, so it should be support. Answers: username_1: +1 username_2: same... username_3: +1 username_4: There is a working fork at https://github.com/drjdr/laravel-chunk-upload There a PR to merge this repo. username_5: thanks for the update rico. Hopefully this will get merged soon. Status: Issue closed username_6: 🚀 Laravel 8 support - all acceptance tests passed
fullcalendar/fullcalendar
691280393
Title: Event property doesnt work Question: username_0: ### Reduced Test Case https://codesandbox.io/s/rough-river-hf04l ### Bug Description Events don't work based on the full calendar documentation. I return an array of objects from my api like so `{"data":[ {"id":1,"title":"Test","start":"2020-09-10 13:00:00","end":"2020-09-24 14:00:00","color":"purple"} ]}` Now I have also tried event sources without any luck. The only way I can get this to work is to do my own axios call and then updating the events array with my response/data. I can do it this way but then I am not able to hook into the next and prev buttons/dates without the need to create my own buttons/functionality which is plain crazy. Am I missing something?<issue_closed> Status: Issue closed
vernemq/vernemq
1025061172
Title: VMQ-Diversity can't authenticate client Question: username_0: ### Environment - VerneMQ Version: Latest - OS: Linux, Docker image - Docker image: tried vernemq/vernemq and erlio/docker-vernemq - docker-compose file `version: "3.8"` `services:` `vernemq1:` `image: erlio/docker-vernemq` `ports:` ` - "1883:1883"` `- "8888:8888"` `environment:` `DOCKER_VERNEMQ_ACCEPT_EULA: "yes"` `DOCKER_VERNEMQ_ALLOW_ANONYMOUS: "off"` `DOCKER_VERNEMQ_plugins.vmq_diversity: "on"` `DOCKER_VERNEMQ_plugins.vmq_passwd: "off"` `DOCKER_VERNEMQ_plugins.vmq_acl: "off"` `DOCKER_VERNEMQ_vmq_diversity.auth_mysql.enabled: "on"` `DOCKER_VERNEMQ_vmq_diversity.mysql.host: "db"` `DOCKER_VERNEMQ_vmq_diversity.mysql.port: "3306"` `DOCKER_VERNEMQ_vmq_diversity.mysql.user: "user"` `DOCKER_VERNEMQ_vmq_diversity.mysql.password: "<PASSWORD>"` `DOCKER_VERNEMQ_vmq_diversity.mysql.database: "system"` `DOCKER_VERNEMQ_vmq_diversity.mysql.password_hash_method: "sha256"` `depends_on:` `- db` `db:` `image: mysql` `environment:` ` MYSQL_DATABASE: "system"` `# So you don't have to use root, but you can if you like` `MYSQL_USER: "user"` ` # You can use whatever password you like` ` MYSQL_PASSWORD: "<PASSWORD>"` ` # Password for root access` ` MYSQL_ROOT_PASSWORD: "<PASSWORD>"` `volumes:` ` - ./mysqldata:/var/lib/mysql` `ports:` ` - "3306:3306"` - Cluster size/standalone: standalone ### Expected behavior I tried this issue with Mysql and PostgreSQL. When I followed the steps in the vernemq documentation for authentication with database. #### I created a table named vmq_auth_acl: for MySql: ` CREATE TABLE vmq_auth_acl ( mountpoint character varying(10) NOT NULL, client_id character varying(128) NOT NULL, username character varying(128) NOT NULL, password character varying(128), publish_acl TEXT, subscribe_acl TEXT, CONSTRAINT vmq_auth_acl_primary_key PRIMARY KEY (mountpoint, client_id, username) );` For Postgresql I added as well the `CREATE EXTENSION pgcrypto;` and changed TEXT to JSON #### Inserted a user using the sql statements in de vernemq documentation. postrgeSQL: `WITH x AS ( SELECT ''::text AS mountpoint, 'test-client'::text AS "client_id", 'test-user'::text AS username, '123'::text AS password, gen_salt('bf')::text AS salt, '[{"pattern": "a/b/c"}, {"pattern": "c/b/#"}]'::json AS publish_acl, '[{"pattern": "a/b/c"}, {"pattern": "c/b/#"}]'::json AS subscribe_acl) INSERT INTO vmq_auth_acl (mountpoint, "client_id", username, password, publish_acl, subscribe_acl) SELECT x.mountpoint, x."client_id", x.username, crypt(x.password, x.salt), publish_acl, subscribe_acl FROM x;` MySQL: `INSERT INTO vmq_auth_acl (mountpoint, client_id, username, password, publish_acl, subscribe_acl) VALUES ('', 'test-client', 'test-user', SHA2('123', 256), '[{"pattern":"a/b/c"},{"pattern":"c/b/#"}]', '[{"pattern":"a/b/c"},{"pattern":"c/b/#"}]');` I tried Crypt for postgreSQL and tried SHA256 for sql since it was stated that password() was depreciated. So inserted the password using `SHA2('123', 256)` for mysql and `crypt(x.password, x.salt)` for postrgessql. #### Tried to subscribe to a topic `mosquitto_sub -h localhost -p 1883 -t "c/b/#" -u "test-user" -P "123"` I received `[warning] can't authenticate client {[],<<"anon-2eJxgD5vEx+S+w9xT5LRlPi1YJc=">>} from 172.21.0.1:56634 due to invalid_credentials` ### Actual behavior Expected a success in authentication since Vernemq was able to connect to the database and I checked the database, there is indeed a table vmq_auth_acl with the user that I inserted. So I have no idea why it is failing the authentication, I thought I followed all the steps from the documentation. Answers: username_0: Oh damn, so stupid, now it is working indeed, Thank you very much!
music-encoding/music-encoding
302727195
Title: Dead link in documentation Question: username_0: At http://music-encoding.org/support/guidelines and the other documentation pages such as http://music-encoding.org/documentation/3.0.0/chapters , the “Elements” link (http://music-encoding.org/documentation/3.0.0/elements) is dead. Answers: username_1: What do you mean? When I look at this link: http://music-encoding.org/documentation/3.0.0/elements I see this page: <img width="1340" alt="screen shot 2018-03-06 at 6 48 54 am" src="https://user-images.githubusercontent.com/3487289/37038599-12db1c22-210a-11e8-8777-8d8ac4080978.png"> username_0: That's peculiar. I get this: ![404](https://user-images.githubusercontent.com/3779074/37039403-afdf2890-2157-11e8-88d0-d58280cb15c0.png) username_1: It must be the Danish censors... username_2: The link works for me too. Maybe you can try a "shift + reload" ? username_0: OK. I tested the link in various environments. Here are the results: Desktop PC at home: Firefox 43.0.4 OK Chrome 64.0.xxx OK MSIE 11 OK Desktop PC at office: Firefox 52.0.1 404 MSIE 11 404 Chrome 64.0.xxx 404 Laptop on WiFi at home or Work (location makes no difference): Firefox 56.0.1 404 MSIE 11 OK Chrome 64.0.xxx OK It doesn't seem to be a cache thing since I haven't used most of these combinations before. Only the 'Elements' link seems to be affected. But if I am the only one experiencing this problem, it's probably not worth investigating further. But it seems to have started recently when the pages were slightly redesigned. username_2: It looks on the screenshot that the page requested is `http://music-encoding.org/documentation/3.0.0/elements/` (with a slash at the end) instead of `http://music-encoding.org/documentation/3.0.0/elements`. I wonder how this can be different. username_1: Yes, I was just noticing that. It seems to be a problem with the slash at the end of the address. This works for me: http://music-encoding.org/documentation/3.0.0/elements This does not work for me: http://music-encoding.org/documentation/3.0.0/elements/ The actual URL is: http://music-encoding.org/documentation/3.0.0/elements.html Github is making an alias from elements.html to elements (or it is setup this way somehow in the configuration file for the website). To fix the problem, the contents should be moved to: http://music-encoding.org/documentation/3.0.0/elements/index.html This should allow these three URLs to work: http://music-encoding.org/documentation/3.0.0/elements http://music-encoding.org/documentation/3.0.0/elements/ http://music-encoding.org/documentation/3.0.0/elements/index.html To fix the problem, a permalink in the Jekyll header should be added placing the page in `/documentation/3.0.0/elements/index.html` in this Markdown file: https://raw.githubusercontent.com/music-encoding/music-encoding.github.io/master/documentation/3.0.0/elements.md See: https://jekyllrb.com/docs/permalinks Or the contents of `elements.md` should be moved from `/documentation/3.0.0/elements.md` to `documentation/3.0.0/elements/index.md`, which would be the easiest thing to do since there is no need to add the permalink to the Jekyll header in this case. Alternatively all links to the elements page should be changed to: http://music-encoding.org/documentation/3.0.0/elements.html username_1: (And this is probably true for other pages as well, where it might be best to force the contents to `/index.html` allowing for slashes to be optional rather than having `.html` be optional.) username_0: I notice that Firefox (at least the newer versions, it appears) adds the slash if I omit it and therefore fails. I did use the URL without the trailing slash: I just followed the link. Strangely enough http://music-encoding.org/documentation/3.0.0/atts has the same faulty behaviour on my laptop (Firefox adds a slash and fails), while http://music-encoding.org/documentation/3.0.0/models works on both my laptop and office PC (no slash added). username_2: This should not be fixed because the links have `.html`. See http://music-encoding.org/guidelines/v3/elements.html Status: Issue closed username_1: One possibility is to use redirections: https://github.com/jekyll/jekyll/issues/3368 http://www.marran.com/tech/creating-redirects-with-jekyll https://github.com/jekyll/jekyll-redirect-from username_1: Note that this URL also works: http://music-encoding.org/guidelines/v3/elements With some sort of automatic redirect to http://music-encoding.org/guidelines/v3/elements.html So additionally an alias could be made to: http://music-encoding.org/guidelines/v3/elements/ Since that typically the mapping in most webservers for a URL like http://music-encoding.org/guidelines/v3/elements (hence the original confusion)
ioBroker/AdapterRequests
312248671
Title: MiPow Playbulb Question: username_0: https://www.mipow.com/collections/playbulb Hier ein Node.js Projekt https://github.com/athombv/com.mipow.playbulb Migrated from Trello with 0 votes Answers: username_0: Homebridge modules exist: maxclaey/homebridge-playbulb as example. Please check if the new ioBroker.ham Adapter can help here: see https://forum.iobroker.net/viewtopic.php?f=36&t=14972
electron/electron
267329155
Title: Old zlib(1.2.8 with security issue) used in electron-v1.7.9-win32-ia32.zip Question: username_0: The electron.exe in the electron-v1.7.9-win32-ia32.zip was be scaned and used zlib_v1.2.8. I check the electron.exe.sym in electron-v1.7.9-win32-x64-symbols.zip found that the zlib_v1.2.8 is included in pdfium.lib a static lib in libchromiumcontent. Answers: username_1: Electron has far more security vulnerabilities than this: the latest version is still based on old Chromium (58 & 59) so it inherits many of the security vulnerabilities published in Chromium [60](https://chromereleases.googleblog.com/2017/07/stable-channel-update-for-desktop.html), [61](https://chromereleases.googleblog.com/2017/09/stable-channel-update-for-desktop.html) and [62](https://chromereleases.googleblog.com/2017/10/stable-channel-update-for-desktop.html) username_2: **CVE:s:** - https://www.cvedetails.com/cve/CVE-2016-9842/ - https://www.cvedetails.com/cve/CVE-2016-9843/ **zlib changelog:** - https://zlib.net/ChangeLog.txt In any case the step to resolve this is to upgrade the Chromium dependency since zlib has been updated there (https://bugs.chromium.org/p/chromium/issues/detail?id=691075), NodeJS who also uses zlib is using the latest version since 7.6.0: https://github.com/nodejs/node/blob/master/doc/changelogs/CHANGELOG_V7.md#2017-02-21-version-760-current-italoacasas username_3: As per our [SECURITY](https://github.com/electron/electron/blob/master/SECURITY.md) document if Electron is exposed to third part vulnerabilities can you please report to the indicated email address and not to the public issue tracker. Status: Issue closed username_0: I have try 1.8.2-beta1, it seems fixed the zlib1.2.8 security issue, but there comes a new security issue for openjpeg, I have already send security issue email by the SECURITY document.
LordOfDarknessAlexander/Auto_ObsessionsX
59139527
Title: AuctionCars selling same car multiple times and other issues Question: username_0: When in carView, the user can click sell, then go back then sell the same car. The 'sell' button should check the AuctionSell state to see if an auctions currently exist with that car id before creating the sale. Auctions which have expired should still be displayed but greyed out (setting css opacity to 0.45) and functionality disabled, so they user has visual record of what they've sold and are selling. Sales should also persist, saving when: the user leaves the state, an auction expires, an auction is started, at regular intervals in the background(if there are active auctions) and loading when the user enters the state<issue_closed> Status: Issue closed
DataDog/documentation
95444767
Title: Incorrect paths in Mac OS X doc Question: username_0: Based on an install I did today the instructions in http://docs.datadoghq.com/guides/basic_agent_usage/osx/ are full of inconsistencies. ~/.datadog-agent/bin/agent is actually found at /usr/local/bin/datadog-agent The info command is /usr/local/bin/datadog-agent info and not ~/.datadog-agent/bin/info The config files are ~/.datadog-agent/datadog.conf ~/.datadog-agent/datadog.conf and not ~/.datadog-agent/agent/datadog.conf ~/.datadog-agent/agent/conf.d All logs are in /var/log/datadog and not ~/.datadog-agent/supervisord/logs/ Answers: username_1: Hi Tom, Thanks for leaving this comment. We recently changed the paths for the mac agent. Looks like we missed updating the documentation. Thanks for catching that. Fixed in #462 Status: Issue closed
Msybl/Shuttle_v1
114910991
Title: TO DO LIST Question: username_0: __Elements of TO DO LIST__ When user enters app first time, app needs to know this is driver or passanger. - [ ] Register user for driver / passanger - [ ] Create register fragment for driver and passanger After registration completed, app should remember his information for next usages. - [ ] Reload user information - [ ] Skip registration activity - [ ] Create an activity that welcomes user
pennmem/ram_utils
310500674
Title: save a single .h5 file with all summary objects Question: username_0: Generating a report currently produces a handful of .h5 files containing data about the session that can be used for analysis or generating reports. Since a session is the basic unit of analysis, it would make sense to save all of these as a single file so that analysts/etc. do not need to manually load multiple files and potentially do merging on the analysis end. Answers: username_1: I think you mean you want to use the "bundling" in `traitschema` 1.2. This just uses whatever format for individual `Schema` subclass instances and then zips them together (which was a lot easier than figuring out how to combine things into a single HDF5 or whatever file). username_0: After looking at it more, the way to do this is to do away with separate math, session, and classifier summary objects. Instead, there should just be a single session summary object that contains all of the necessary information for producing the report plots and doing post-hoc analysis username_0: From talking with other end-users, it's not such a problem to keep classifier summary separate from session summary. However, it would make sense to fold the math summary into the session summary
Factom-Asset-Tokens/fat-js
512066622
Title: 0 amount transactions are rejected Question: username_0: fatd 1.0 allows for fat 0 and fat 1 transaction with 0/empty input/output but fat-js reject them with errors like `Input amount must be a positive nonzero integer`. Answers: username_1: Merged to master in 96228ea50a100a240c8672eebf40111f1ce49e97 Status: Issue closed
GoogleChrome/lighthouse
205717049
Title: Test for offline-first navigations Question: username_0: While the test for a "200 OK" response to an offline navigation does a good job of confirming that a web app works offline, it doesn't check that navigations are handled via an [offline-first](https://twitter.com/jaffathecake/status/788289680735436800) strategy. Currently, a web app might use a [network-falling-back-to-cache](https://developers.google.com/web/fundamentals/instant-and-offline/offline-cookbook/#network-falling-back-to-cache) strategy for handling navigations and still get a perfect Lighthouse score, even though the network request will behave poorly given lie-fi or a high-latency network. Would it be possible to create a new test which checked whether the HTTP request used for a navigation was fulfilled via a network request, and fail that test if it was? One wrinkle is that a web app which used a [stale-while-revalidate](https://developers.google.com/web/fundamentals/instant-and-offline/offline-cookbook/#stale-while-revalidate) strategy for navigations should pass the test, since the HTTP request was fulfilled from the cache. While there's a network request made, the network doesn't block the page from getting the cached result. Based on my [last attempt](https://github.com/GoogleChrome/lighthouse/pull/578) at a related PR, I'm not sure that I have the familiarity with Chrome's debugging protocol to properly implement this, but I'm happy to review PRs, and provide some passing and failing test cases. CC: @addyosmani @slightlyoff Answers: username_1: It seems like this should be identical to testing a site that's serving from the cache, you just don't want to get distracted by a separate request that never gets returned to the page username_0: `¯\_(ツ)_/¯` regarding as to whether it would need debugger protocol. I'm really ignorant here. The question to answer is: did the response used to fulfill the navigation request come from the network?
oceanprotocol/aquarius
959047951
Title: Aquarius is not indexing Moonbase Alpha Question: username_0: Aquarius is currently not indexing the moonbase alpha network. This means that the market is not working on Moonbase Alpha. The chain IDs that Aquarius is indexing are shown here: https://aquarius.oceanprotocol.com/api/v1/aquarius/chains/list The market is getting a 500 error returned from aquarius: ![500 error - 2021-08-03_13-58](https://user-images.githubusercontent.com/20739535/128012490-124c74ec-7e91-4a8a-82f6-32bdd1760536.png) Answers: username_1: @username_0 , could you check now ? username_0: Still getting the same 500 error when loading the market on Moonbase Alpha. It only seems to give the error if Moonbase Alpha is the only network that you are on username_1: these are the related errors in Aquarius: ![image](https://user-images.githubusercontent.com/54084524/128085172-05d19533-e503-417b-b944-e77d92202731.png) ![image](https://user-images.githubusercontent.com/54084524/128085373-e91ebcc0-9140-4a53-ad0d-083b1fb31da5.png) username_2: This is still an issue. And this also happens on polygon mumbai. username_3: With Aquarius 3.0.2 running on a remote machine, the indexing works and I can query the new ES endpoint. @username_0 I assume this is the old Aquarius. Can you confirm? Can you also send me a full response? 500 doesn't automatically mean the assets are not indexed, it could be something else. @username_2 @username_1 what is the status of the v3.0.2 deployment? Maybe this is a wild goose chase and we are trying to fix something that is already fixed. Worst case scenario, if v3.0.2 is not ready for deployment, maybe we can make a test with aquarius2.oceanprotocol.com on Moonbeam and Mumbai? username_1: Aquarius deployed on Moonbeam and Mumbai and should be accessible through aquarius2.oceanprotocol.com endpoint (v3.0.2). username_2: You can easily test this [here](https://market-13bh6vl0j-oceanprotocol.vercel.app/search?sort=created&sortOrder=desc) and select the desired networks. Moonbeam seems to be working in 3.0.2 but mumbai is still not indexing anything username_3: I selected Mumbai, there is no 500 error. There are simply no results returned. When I try indexing Mumbai, I get no metadata created events. Can someone publish an asset on Mumbai, to make sure we have results? username_3: @username_2 any updates on this? username_2: No, please try and publish an asset using the api to test this. username_3: Marian published an asset today during a call with me https://market-13bh6vl0j-oceanprotocol.vercel.app/asset/did:op:f23E410e48621839Ba374f81a2882801D0334B74. it seems to be correctly indexed and shown on the marketplace. The issues we had were with our broken Mumbai faucet and the broken ETH faucet in Mumbai itself. It was a rough ride but it seems we're in the clear now. Status: Issue closed
tapos-datta/ExoPlayer-With_CustomVisualizer
739608403
Title: used buildAudioRenderers method is deprecated in DefaultRenderersFactory Question: username_0: Hi! In DefaultRenderersFactory the 'buildAudioRenderers' method now accepts 8 params instead of 10. So, audioProcessors is missing which plays the key role in customizing it. Please update the class CustomRendererFactory accordingly. Thanks Answers: username_1: Yeah sure, Thank you for your concern. I have updated the sources accordingly the version of ExoPlayer-2.12.0. Status: Issue closed username_0: Thanks but isn't just overriding buildAudioSink method enough? I used this and it is working. ``` public class CustomRendererFactory extends DefaultRenderersFactory implements MediaSourceEventListener { private AudioProcessor[] aMProcessors; private TeeAudioProcessor teeAudioProcessor; public CustomRendererFactory(Context context, AudioBufferSink listener) { super(context); teeAudioProcessor = new TeeAudioProcessor(listener); } @Nullable @Override protected AudioSink buildAudioSink(Context context, boolean enableFloatOutput, boolean enableAudioTrackPlaybackParams, boolean enableOffload) { aMProcessors = new AudioProcessor[]{teeAudioProcessor}; return new DefaultAudioSink( AudioCapabilities.getCapabilities(context), new DefaultAudioSink.DefaultAudioProcessorChain(aMProcessors), enableFloatOutput, enableAudioTrackPlaybackParams, enableOffload); } } ``` username_1: Hi! In DefaultRenderersFactory the 'buildAudioRenderers' method now accepts 8 params instead of 10. So, audioProcessors is missing which plays the key role in customizing it. Please update the class CustomRendererFactory accordingly. Thanks username_1: Welcome, and yes, you can. The main difference is the way of passing a custom audio processor to DefaultRenderFactory. In the newer version, the audio processor is needed to pass through an audioProcessorChain. Status: Issue closed
department-of-veterans-affairs/va.gov-team
839329306
Title: [VAMC-Scaffolds] Create or Clone all required pages for VA Dublin health care in VISN 7 [BA] Question: username_0: ## User Story As a Veteran Web user, I need a easier and more efficient site, so I can quickly get to my objective and complete my journey on the local VA.Gov/MC website. ## Goal _Implement required VAMC-upgraded va.gov webpages_ ## Objectives or Key Results this is meant to further - _Increase overall quality experience for veteran web user_ - _Reduce Click through count by_Highlighting top stories_ ## Resources - Tools - Documentation [Menu and Page checklist](https://drive.google.com/drive/folders/1QliZFSkQHEruGgjOGRihG8tohYxswjm1) [Clonable Website](Salt-lake-city-health-care) ## Tasks _Creation of the following Drupal types/menus and pages_ * All pages are cloned or created, the System name, Related office or Health Care system, Owner, Meta title, Meta Description, and Menu items are updated. **1st Step: Go into Manage, Content, and click on Menus** _Find VA [SYSTEM NAME] (listed alphabetically) that you are cloning in the menu list and select "Edit" _using the blue button "+add link" add the following "nolink" (See instructions about "nolink" menu link titles: - [ ] SERVICES AND LOCATIONS (Link: "nolink" / uncheck "enabled checkbox") - [ ] NEWS AND EVENTS (Link: "nolink" / uncheck "enabled checkbox") - [ ] ABOUT VA "Short System Name" (e.g.:ALTOONA) (Link: "nolink" / uncheck "enabled checkbox") **CLONING Clone from Salt Lake City of Sheridan Communicate on VAMC-upgrade channel** ** Clone or create these Listing pages and Add to Menu just below SERVICES AND LOCATIONS - [ ] Health Services Listing Page - [ ] Locations Listing Page _ * Find the following facilities and make sure they are associated under the correct system and owner - [ ] Validate these sites are matched in VAST DB and Playlist, then move under [System] ** Add facility pages to Menu under "Locations" (Listing page) - [ ] vha_557 Carl Vinson Veterans' Administration Medical Center - [ ] vha_557GA Macon VA Clinic - [ ] vha_557GB Albany VA Clinic - [ ] vha_557GC Milledgeville VA Clinic - [ ] vha_557GE Brunswick VA Clinic - [ ] vha_557GF Tifton VA Clinic - [ ] vha_557HA Perry VA Clinic _ **CLONING Clone from Salt Lake City or Sheridan Communicate on VAMC-upgrade channel** _ ** Top task items listed below require "DISABLED" menu items associated to them** Place disabled menu items below --[System Name] "DISABLED" - [ ] Make an appointment - [ ] Register for care - [ ] Billing and insurance - [ ] Dav-vans transportation for veterans - [ ] Medical records - [ ] Pharmacy - [ ] [System Name] - Campus map - [ ] Chaplain services ** Items below require "DISABLED" menu items associated to them** Place disabled menu items below "Health Services" (listing page) - [ ] Caregiver support [Truncated] - [ ] Performance - [ ] Leadership (Listing Page) * Add these pages to Menu under Work with Us - [ ] Jobs and Careers - [ ] Internship and Fellowship - [ ] Volunteer and Donate **NEW SECTION** - more listing pages** * Add these pages to Menu under NEWS AND EVENTS - [ ] Event Listing Page - [ ] News Releases Listing Page - [ ] Stories Listing Page ## Acceptance Criteria - [ ] Validation that all pages created and assigned to the correct SYSTEM. (review in SYSTEM lookup page and take inventory. - [ ] Validate that all menu pages show up in Menu and are visible in the correct order on the menu - [ ] Validate that pages are ready to take data load information - [ ] All system pages listed above have been created<issue_closed> Status: Issue closed
jlippold/tweakCompatible
555888564
Title: `PreferenceLoader` working on iOS 13.3.1 Question: username_0: ``` { "packageId": "preferenceloader", "action": "working", "userInfo": { "arch32": false, "packageId": "preferenceloader", "deviceId": "iPhone8,4", "url": "http://cydia.saurik.com/package/preferenceloader/", "iOSVersion": "13.3.1", "packageVersionIndexed": true, "packageName": "PreferenceLoader", "category": "System", "repository": "rpetrich repo", "name": "PreferenceLoader", "installed": "2.2.4~beta1", "packageIndexed": true, "packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.", "id": "preferenceloader", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.1.5", "shortDescription": "load preferences in style", "latest": "2.2.4~beta1", "author": "<NAME>", "packageStatus": "Unknown" }, "base64": "<KEY> "chosenStatus": "working", "notes": "" } ```<issue_closed> Status: Issue closed