repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
bazelbuild/rules_rust
351479687
Title: rules_rust broken by bazel@HEAD Question: username_0: Hello dear rules owners, I see that bazel@HEAD breaks rules_rust: https://buildkite.com/bazel/bazel-with-downstream-projects-bazel/builds/386#183d2d56-2b16-4010-9373-ef1b3b66b6eb Would you mind taking a look? Thank you very much! This is the announcement email: We are very close to remove in-memory //tools/default* package in favor of using @bazel_tools//tools/jdk: and @bazel_tools//tools/cpp: instead. The changes will be ready with next release. Motivation: //tools/default was initially created as virtual in-memory package. It generates content dynamically based on current configuration. There is no need of having //tools/defaults any more as LateBoundAlias can generate dynamic configuration-based label resolving. Also, having //tools/default makes negative impact on performance, and introduces unnecessary code complexity. All references to //tools/defaults:* targets should be removed or replaced to corresponding target in @bazel_tools//tools/jdk: and @bazel_tools//tools/cpp: packages. Scope of changes and impact: All targets in //tools/default will not exist any more. If you have any references inside your BUILD or *.bzl files to any of its, then bazel will fail to resolve. Migration plan: Please replace all occurrences: //tools/defaults:jdk -- by @bazel_tools//tools/jdk:current_java_runtime -- or/and @bazel_tools//tools/jdk:current_host_java_runtime //tools/defaults:java_toolchain -- by @bazel_tools//tools/jdk:current_java_toolchain //tools/defaults:crosstool -- by @bazel_tools//tools/cpp:current_cc_toolchain -- or/and @bazel_tools//tools/cpp:current_cc_host_toolchain -- if you need reference to libc_top, then @bazel_tools//tools/cpp:current_libc_top These targets will not be supported any more: //tools/defaults:coverage_report_generator //tools/defaults:coverage_support
prestodb/presto
271401118
Title: Create reusable type mapping tests for JDBC connectors Question: username_0: Type mapping between Presto and external database accessed by JDBC connector is done in `BaseJdbcClient` (`presto-base-jdbc`). While implementation is shared, correctness should be integration-tested against each database we have connector for (currently: postgresql, sqlserver, mysql). This should be implemented as reusable test (like `AbstractTestQueries`) with possibility to disable selected tests by a concrete connector e.g. for unsupported types. Context https://github.com/prestodb/presto/pull/9285#issuecomment-341991161 Answers: username_1: Also relevant for the Redshift connector. username_2: [This and the 4 following tests](https://github.com/prestodb/presto/blob/master/presto-mysql/src/test/java/com/facebook/presto/plugin/mysql/TestMySqlDistributedQueries.java#L97) might be of help/inspiration :) username_0: @username_2 thanks! Do you recall why the data mapping tests are distributed? (That's probably why I didn't find them 🙂 ) username_2: No idea, probably because the original tests that were refactored into that DSL were distributed. As far as I see, there's no dependency on the tests being distributed. But maybe @username_3 will know? He started the CHAR/VARCHAR work. username_3: I don't thing there is any particular need for them to be distributed. username_3: Also if the test don't need to be distributed then we could possibly extend `LocalQueryRunner` to have method similar to: `Pair<MaterializedResult, Page> LocalQueryRunner::execute`. Then we could additionally test if connector writes correct representation of data types. We could for instance validate that char blocks returned by connector don't have trailing spaces. CC: @username_2 username_2: Don't the current tests check that? username_3: I don't think so, they only check that `MaterializedResult` is equal to inserted value? I don't think this implies that Slice representation is correct for instance. username_0: Surely the current tests do not do this. Hence #9282 (the bug was not about incorrect semantics, but about incorrect cast implementation). Status: Issue closed
harfbuzz/harfbuzz
761149264
Title: Compilation issue with ghs arm compilator Question: username_0: Here is the error message : "..\..\..\..\..\..\..\..\..\..\ptf_gfi_core\System\harfbuzz\Dev\LibSources\hb-open-type.hh", line 524: error #304: no instance of overloaded function "OT::SortedUnsizedArrayOf<Type>::as_array" matches the argument list operator hb_sorted_array_t<const Type> () const { return as_array (); } ^ The c++ compilator (compatible with c++11) seems to be strict not finding a method without argument. As far i understand the cast operator shall return &arrayZ[0] ? So one fix could be to call_array (0) ?<issue_closed> Status: Issue closed
awslabs/aws-crt-nodejs
876792669
Title: electron error Error: node_modules/aws-crt/dist/bin/darwin-x64/aws-crt-nodejs Question: username_0: Hi, I am trying to use `aws-iot-device-sdk-v2` in an electron app. When I run `electron-forge start` I get this error: ``` App threw an error during load Error: Cannot find module '/Users/jake/Code/mir-kiosk-electron/electron-kiosk/node_modules/aws-crt/dist/bin/darwin-x64/aws-crt-nodejs' at webpackEmptyContext (/Users/jake/Code/mir-kiosk-electron/electron-kiosk/.webpack/main/index.js:111944:10) at Object.<anonymous> (/Users/jake/Code/mir-kiosk-electron/electron-kiosk/.webpack/main/index.js:112436:91) at Object../node_modules/aws-crt/dist/native/binding.js (/Users/jake/Code/mir-kiosk-electron/electron-kiosk/.webpack/main/index.js:112445:30) ```` After some googling I fixed this by adding this to `webpack.main.config.js` It will be good I don't have to use this. I think the problem is `binding.js` uses `__dirname` ``` node: { __dirname: true, // I }, target: "node", // in order to ignore built-in modules like path, fs, etc. externals: [nodeExternals()], ```` The problem is I get a missing module error if I run the packaged app. I guess node_module is not packaged. ``` Uncaught Exception: Error: Cannot find module 'macaddress' ```` Do you have any idea as to how to fix it? https://github.com/username_0/mir-kiosk-electron/tree/master/electron-kiosk Status: Issue closed Answers: username_1: @username_0 How did you solve this? username_0: Hi, I made a patch. https://github.com/username_0/mir-kiosk-electron/blob/node-ipc/electronKiosk/patches/aws-crt%2B1.8.0.patch https://github.com/awslabs/aws-crt-nodejs/pull/220/commits/44faba53400fb0e993665b0fad79eb4a435fce72 By the way, this does not work in windows. `aws-crt` is not supported in electron on Windows.
linebender/norad
830401812
Title: Python bindings and mutation patterns Question: username_0: This is a quick sketch of what would be involved in adapting norad so that it would be useable from languages like python, *as a drop-in replacement for existing libraries*; the particular challenge here is getting norad objects to have the same semantics ('reference semantics') as objects in those languages, and making that work with Rust's ownership model. This is follow-up from the discussion in https://github.com/fonttools/fonttools/issues/1095. I am going to make a bunch of assumptions about how some of the python tools work, so please correct me wherever I'm wrong! My understanding of what we would want in a python API is basically: everything is an 'object', which reference semantics. If I get a glyph from a layer, and I change its outline, and then I go and get another copy of that glyph from the layer, those glyphs will be identical; they point to the same underlying data. This is not how things currently work in norad; in norad mutation works through one of two mechanisms, which i'll call "borrowing" and "check-out/check-in": - *borrowing*: This is how layers currently work in norad. To mutate a layer, you call a method that returns a reference to a layer object, and you can mutate this layer; however you *cannot* hold on to a copy of this layer, or really pass it anywhere else. - *check-in/check-out*: This how glyphs currently work: *norad doesn't even let you get a mutable reference to a glyph, at all*. In norad, if you want to mutate a glyph, you get a glyph from a layer, you modify it, and then you add it back to that layer. You can think of this as being like a check-out/check-in mechanism. ### A design to support bindings I think that trying to make norad *as it is currently written* fit into the python model will be tricky, but I think there's a reasonably straight-forward answer, which is that we have a separate set of types and interfaces explicitly designed to work with python. This should let us continue to share all of the base types and parsing/validation/serialization logic, while letting us build two separate APIs that will respect the two distinct use-cases. So basically: we add a python-specific wrapper, in rust, for each type, like: ```rust pub struct PyUfo { meta: PyMetaInfo, font_info: PyFontInfo, ...etc } pub struct PyLayer { glyphs: Rc<RefCell<Map<String, PyGlyph>>, ...etc } pub struct PyGlyph { inner: Rc<RefCell<Glyph>> } ``` etcetera. Note: This assumes that the glyph is a 'leaf' type, that is it is the finest granularity object that you're allowed to mutate and expect those mutations to show up elsewhere. This might not be the case; for instance you might expect to be able to get a `Contour` out of a glyph and change its properties and have those be reflected everywhere; in this case we could also need a `PyContour` type, and `PyGlyph` would look more like the `Glyph` that's already in norad. You can mostly ignore the `Rc<RefCell<_>>` bit. The `Rc` means is that we're using a `R`eference `c`ounted pointer, and the `RefCell` means basically that the internal data is not subject to rust's borrowing rules at compile-time. (`Rc` + `RefCell` is assuming that this object will not be shared between OS threads, which seems like a reasonable assumption for python; if we *do* want that behaviour then we would instead use `Arc` + `Mutex`, which ensures that our reference counts and data access are thread-safe). ### Borrowing problems One possible concern with this approach involves borrowing expectations; the rule with `RefCell` is that when you actually want to *mutate* the data, you acquire a kind of 'lock'. If this object is already borrowed, you can't get that lock. In practice I think we can avoid this completely by ensuring all of that acquire/release happens on the rust side; I'd have to look into this a bit, though, to make sure. It might mean we have to write something to generate the python bindings ourselves, to ensure that things like setters and getters are doing that borrowing under the covers. If we *do* have to expose this somehow, what we would do is to just throw a python exception if something was already borrowed. I was initially thinking this would be a larger part of the design, as a sort of safety valve; when folks migrated existing python code to this library they might hit some new exceptions, but I actually think we can probably avoid this altogether? ### other thoughts *an alternative design based on proxy objects*: I think if we want a drop-in replacement for an existing tool written in python, something like what I describe here will be the best route. There are options, like having 'proxy objects' that just hold a reference to the font or layer as well as a method for mapping mutations on themselves to mutations on the shared object. This honestly has a certain nerdy appeal, especially since we could do cool stuff like having a `def delete:` on a `Glyph` object that removes it from the layer and updates the layer_contents, but I think it's probably a bit more complicated and it's a bit less clear to me how well it would work, although I'm more curious as I end this paragraph than I was when I began? ### next steps This is intended as a sketch, and an actual design will require a bit more thought and research. I'm going to hold off on doing that work until I have a better sense of how much of a priority is this, and whether it's my priority or someone else's. If @username_1 is interested in doing the work then I'm happy to offer whatever advice and guidance I can. Otherwise if @davelab6 thinks that this is worth a week or two of my time then I'm confident we can get something working pretty quickly; the only part I'm unsure of is how to generate the python bindings in a way that would play nicely with with this interior-mutability pattern. Answers: username_1: Another suggestion that Raph had was the only object you expose to Python is the font, and you pass around with a path-key to access or mutate deeper structures. So changing the X position of a point is actually done by the moral equivalent of `font.set_value(“public.glyphs/a/2/1/x”, -5) ` That may help to put all the locking in the same place. username_0: @username_1 that sounds like approximately what I was thinking about with 'proxy objects', as an alternative design, although I probably could have expressed it more clearly. :) username_1: Incidentally we've since discovered that UFO loading is not the bottleneck we thought it was (yay profiling!) so I would not suggest this was a very high priority... What I have with iondrive (creating ufoLib2 objects in Rust) is fast enough for my needs. username_0: okay, sounds good!
Icinga/ipl-html
816261461
Title: SelectElement default value is ignored Question: username_0: When creating/adding a form element of type `select`, the order of options is significant. The order should **not** be significant, i.e. it shouldn't matter if `value` is passed in first. The following will ignore the default value: ```php $this->addElement('select', 'name', [ 'value' => 'bar', 'options' => [ 'foo' => 'foo', 'bar' => 'bar' ] ); ``` The following will pre-select `bar` as expected: (Note the different placement of the `value` attribute) ```php $this->addElement('select', 'name', [ 'options' => [ 'foo' => 'foo', 'bar' => 'bar' ], 'value' => 'bar' ); ```<issue_closed> Status: Issue closed
CollaboratingPlatypus/PetaPoco
139416988
Title: Problem when using bracket on order by clause Question: username_0: I have a column named like user in my table and I need using it in my order by clause, but user in sql server is a reserve word, so I need use brackets in my sql query. My query looks like that: ```SELECT [Level], [Description], [CreatedAt], [User] FROM MyTable ORDER BY [User]``` When I try execute a query like the above query in the PetaPoco I get a exception because I'm using brackets in my order by clause and whether I trying execute without brackets the result is not like expected. Answers: username_1: until fixed, maybe just use a non-reserved alias like SELECT [Level], [Description], [CreatedAt], [User] as MyUser FROM MyTable ORDER BY MyUser Status: Issue closed username_2: Added integration test. Confirms this has already been fixed
dart-lang/language
888921366
Title: How does static metaprogramming handle incomplete code for auto-complete? Question: username_0: We discussed this in a meeting, but I wanted to capture it in an issue so we don't forget. Currently, analyzer is happy to analyze incomplete or incorrect programs in order to make auto-complete work while the user is in the middle of typing code. For example: <img width="543" alt="Screen Shot 2021-05-11 at 3 23 55 PM" src="https://user-images.githubusercontent.com/46275/117892096-f154a300-b26c-11eb-9d80-1cf5e9bea219.png"> Here, even though the class is syntactically incorrect, it can still type check the class enough to show `bar()` as an auto-complete option. If something like macros are applied to that class, though, this can break down. If the analyzer/compiler only applies macros when the code is in a valid state, then auto-complete will stop working while they are typing in a class that uses macros. If analysis can apply macros to incomplete code, then we need to define how that works or what that means. One idea is that the language basically ignores the issue. It only specifies the behavior of applying macros to correct programs. When the user is in the middle of typing code, analyzer takes their current incomplete program and discards whatever incomplete stuff it needs to get to a valid source file. Then it runs the macros and type checks that. Maybe something like that would work. Whatever we do, it's critical that static metaprogramming doesn't degrade the IDE user experience. @username_1 @username_4 Answers: username_1: @devoncarew, @username_5 and I recently had the first in a series of meetings about some of the UX issues around static metaprogramming and I'm in the process of writing up a doc with what we had thought about. Our intent is to share it with you as soon as it's reasonably coherent as a way of furthering discussions like this. Code completion is one of the topic areas, and I'll be sure to add thoughts about incomplete code. username_2: Definitely looking forward to @username_1's doc. As someone who's been hacking on the analyzer for 7 years, I have a bunch of thoughts on this topic too. I'll try to give my personal perspective here briefly; I'm happy to follow up off-line if that would be useful. The issue IMHO is better characterized as how to handle *code with static errors* rather than *incomplete* code. (Some incomplete code is free of static errors and that code can be analyzed normally; some code with static errors isn't incomplete but actually has extra junk, or stuff that's simply wrong (like code that fails to type check) and we need to handle that too). So really what we are talking about here is error recovery. I like the simplicity of Bob's suggestion that we discard code until we reach a valid program, but I think in order to reduce the amount of discarding we need to do, we will need some escape hatches to allow invalid code to be represented. As a trivial example, if a method's return type is misspelled, we shouldn't just discard that method from the model; the code generator should still be able to see that the method exists. But if it queries its return type, it should either get `dynamic`, or a special "invalid type" with some well-defined behaviors. Personally I favor the "invalid type" approach because it makes it easier to avoid cascading errors, but that may be more implementation effort because currently the analyzer uses `dynamic`. Similarly, if we allow code generators to peek at the expressions and statements inside a method declaration, it would be nice to have "invalid statement" and "invalid expression" nodes so that we can show the code generator as much of the method body as possible without forcing it to deal directly with the invalid code. A particularly tricky area is what to do with loops in class hierarchies. The language spec forbids the class hierarchy from looping in order to avoid non-termination of algorithms like type inference, subtype checking, etc. The analyzer's implementation of these algorithms has been carefully written to avoid non-termination even in the presence of class hierarchy loops. I don't think we can expect writers of code generators to exercise the same care, and if we're going to be running user-supplied code generators during analysis, it's really important to avoid non-termination issues. So I think we should ensure that when their code generator queries the code structure, it sees a class hierarchy without loops. We *could* handle this by discarding "extends" and "implements" clauses until no loops remain, but I think this would lead to a poor user experience because code generators often have expectations about the class structure of the code they're processing, and will issue errors if they don't see the class structure they expect. It would be really confusing if our efforts to "fix" class hierarchy loops caused code generators to output nonsense error messages. I think a better approach here is the "invalid type" approach: if there's a loop in the class hierarchy, the analyzer should synthetically change one of the "extends" or "implements" clasuses to e.g. "extends <invalid-type>"; that way either the code generator can choose to suppress its error message, or even if it doesn't and the error message makes it to the user, the name "invalid-type" should provide the user with better clues about what went wrong. Side question: in order to fix loops in the class hierarchy, the analyzer will need to make arbitrary choices about which edge to cut. How consistent should its choices be? IMHO it would be best for it to be deterministic and independent of the order of queries from the code generator. One final note: thus far, we have left the design of error recovery pretty much entirely up to the analyzer team, and kept any discussion of error recovery out of the spec. I think that's a really good general principle because it allows the analyzer team more creative freedom and lets them pivot faster when they learn things about how error recovery works in real-world scenarios. But I think it would be good to classify our errors into some broad categories and specify what categories of invalid code (if any) user-provided code should be expected to deal with. I think it should be something like: user-provided code generators won't ever see syntax errors, class hierarchy loops, or self-referential typedefs. But they may see special "invalid type" nodes. They may also see: unimplemented abstract methods, type mismatches between a class method and the method it overrides, and naming conflicts (e.g. classes with duplicate names). Also, when looking inside method bodies, they may see "invalid expression" and "invalid statement" nodes, unterminated switch cases, mismatched return types, use of uninitialized variables, and failure to return a value from a method with a non-nullable return type. (That's just my first impression though. I'm fairly open to discussion, especially on that last sentence about method bodies) username_1: That's an interesting recovery strategy, but it's not consistent with the approach used by the analyzer for other kinds of errors. I'm not saying that it has to be consistent, just making an observation. I think it would be informative to try to write a couple of macros as a concrete way of thinking about which recovery approach would provide the best UX for both end users and for macro authors. username_0: This is a great discussion. Trying to synthesize your comments, how does something like this sound: * Tools may run macros even on code that contains some kind of errors. It's entirely up to tools to choose whether and in which circumstances that they do. The language doesn't specify the boundary. * But, if the code isn't valid, it will never end up being run. The macros are only invoked in order to provide a better IDE experience. * The API that macros use to work with code has a notion of erroneous pieces of code. Introspecting over code in a macro may yield "invalid type", "invalid statement", etc. Macro authors can choose to handle those gracefully if they choose in order to provide a better experience when the macro is invoked on invalid code. If they don't the macro may fail, which essentially means it generates some cascaded compile errors. It's up to tools to decide if they want to hide those or not. In other words tools *may* eagerly run macros on ill-defined code. The API gives macro authors affordances to handle that gracefully if they so choose. But the resulting code never gets actually executed, so the language doesn't have to get in the business of specifying its behavior. Does that seem like a reasonable set of trade-offs? username_3: Why will it stop working? Consider a macro that generates a data class. Let's assume the input for a macro is declared as "template class", which is easily recognized by the analyzer as a template (as opposed to a normal code). ```dart @dataClass template class Foo { int x; String y; // etc. } ``` Further, suppose we already have a state with a valid template. The macro has been called already, and the resulting code is known. What happens when we move the cursor into the definition of the template and start editing? The straightforward answer is: nothing. The macro doesn't get called when the code of the template is incomplete. But the previous results of a macro are not discarded - at least until the cursor leaves the template block. So all suggestions work as if nothing had happened. If the user exits the block leaving the template in a bad state - I'm not sure even in this case the results of prior macro invocation should be invalidated (thus causing an avalanche of other errors), but maybe. However, if the macro gets called with incomplete code (as proposed) and has to decide how to handle the errors "gracefully" - it's painful even to think of how the macro has to be coded to behave predictably. The only possibility I see for the macro is to memorize the input and output of the last correct invocation and somehow figure out what has changed and how this affects the result - but why bother? There's no guarantee that this behavior will be more intuitive than the straightforward algorithm outlined above. username_1: Yep, that's part of the proposal I'm writing up. username_4: There is actually one case which I think might fall into this same bucket as what we are discussing here. The case is where a macro introspects on an api, but the api relies on type inference to be filled in, and the macro itself needs to be ran in order for that type inference to work. Consider this class: ```dart class Thing { final name = someMacroGeneratedField; } ``` If the macro generating `someMacroGeneratedField` introspects on `Thing` and asks for the type of `name`, what does it get? The issue is actually a lot more subtle/interesting than even this contrived example, because even if `someMacroGeneratedField` can be resolved to some identifier, some other macro could run and change the identifier that should have been resolved to by shadowing it. This is a general problem that we are working through right now but it seems likely that macros will need to at least have to deal with introspecting on apis where the type is _not yet known_, and never will be whenever that macro runs, because it hasn't been confidently inferred yet. My assumption at this point is some macros would want to fail and require a type on the left hand side (although this is not ideal, I don't see a good workaround at this time). Other macros might be ok with just treating it as dynamic. I think we need to give them some special notion of an "unknown type" to choose how to handle this. username_1: I think there's an interesting requirement implied by that statement: while applying the macro M, the introspection API must reveal a state that's consistent with the state the code would be in if the macros were being run for the first time. In other words, if the analyzer caches any macro results and then has to re-apply M, the introspection API can't reveal any code in the cache that would have been generated after M was run the first time. @devoncarew @username_5 username_5: The requirement that every time is like the first time looks reasonable to me.
BragaAndrei/testing
548562966
Title: The font in the news bar does not work properly Question: username_0: Description Defect nr.ID: 001 Title: Font Text Issues macOS Mojave version 10.14.6, Safari Version 12.1.2 (14607.3.9) Pre-condition: The font is shown to the customer in a glitchy way Steps to reproduce: 1. Go to www.airport.md 2. Scroll down to the news bar Expected result: The news text font is shown as glitchy Severity: Low Priority: 1 Environment macOS Mojave version 10.14.6 , Safari Version 12.1.2 (14607.3.9) <img width="980" alt="Screen Shot 2020-01-11 at 13 00 17" src="https://user-images.githubusercontent.com/59793996/72218719-c0644100-3546-11ea-9324-60a477fc1e48.png"><issue_closed> Status: Issue closed
Fatal1ty/mashumaro
815895565
Title: Inconsistent checks for invalid value type for str type Question: username_0: If a class based on DataClassDictMixin has a field with type str it will construct instances from data that contains data of other types for that field, including numbers, lists, and dicts. However fields of other types, eg int, do not accept other non-compatible types. Not sure if this is intentional and I'm missing something here, but it kinda seems like unexpected/undesirable behaviour when you want the input data to be validated. The following example only throws an error on the very last line: ```python from dataclasses import dataclass from mashumaro import DataClassDictMixin @dataclass class StrType(DataClassDictMixin): a: str StrType.from_dict({'a': 1}) StrType.from_dict({'a': [1, 2]}) StrType.from_dict({'a': {'b': 1}}) @dataclass class IntType(DataClassDictMixin): a: int IntType.from_dict({'a': 'blah'}) ``` Answers: username_1: There is no strict validation at the moment for the sake of performance. It's not needed in many cases but I'm going to add optional validation. It will be turned on in the field or config options.
Vulnerator/Vulnerator
373638609
Title: Multiple CKL Files do not show up in the Vulnerator Excel Report Question: username_0: The vulnerator tool does not import all finding details into the vulnerator spreadsheet which should show multiple entries from a several CKL files. I attempted to import 3 separate CKL files generated from STIG VIEWER into the Vulnerator to consolidate my findings in to one report. The vunerator spreadsheet only logged one of three entries into the mitigation column leaving the other two entries out. For example if I had three CKL files, and all three files had three separate comments and finding entries pertaining to one CCI, the vulnerator tool will randomly pick one of the three files and import the data it sees into the spreadsheet rather than importing all three entries into the report. Answers: username_1: @username_0 I've updated how vulnerabilities are parsed for v6.2.0, and am currently verifying the accuracy of all data now - I will ensure that all relevant information is captured for every item in the checklist. Thank you for the bug report!
godotengine/godot
403148414
Title: Inspector dock changes size when selected Question: username_0: <!-- Please search existing issues for potential duplicates before filing yours: https://github.com/godotengine/godot/issues?q=is%3Aissue --> **Godot version:** 5b5db08 **Issue description:** When clicking the Inspector dock, it will change its size. It only affects this dock alone, no matter its current location: ![peek 2019-01-25 11-00](https://user-images.githubusercontent.com/30739239/51747391-428ef780-20a1-11e9-9abe-7adf98e921ba.gif) Answers: username_1: Might be related to #25281. username_2: Im trying to fix [#25281](https://github.com/godotengine/godot/issues/25281) and i found that the problem might be related to tab_container resizeing.. in tab_container.cpp there are these lines: ``` for (int i = 0; i < tabs.size(); i++) { Control *c = tabs[i]; if (!c->is_visible_in_tree()) continue; Size2 cms = c->get_combined_minimum_size(); ms.x = MAX(ms.x, cms.x); ms.y = MAX(ms.y, cms.y); } ``` Not skipping invisible tabs solve this for me: ``` for (int i = 0; i < tabs.size(); i++) { Control *c = tabs[i]; Size2 cms = c->get_combined_minimum_size(); ms.x = MAX(ms.x, cms.x); ms.y = MAX(ms.y, cms.y); } ``` However, i think this only makes the symptoms go away.. i think i need some feedback to go on. username_3: Related, if not duplicate, to #24572 username_4: Still happens in Godot 3.1 RC1 ![QCsLlevr0V](https://user-images.githubusercontent.com/1311555/54076148-fa6a1400-429f-11e9-9a7d-f621605ce1b1.gif) username_2: Yes, i made a pull request with a possible fix, but i think it was kicked to 3.2 username_5: @username_2 where is your pull, I can't find it? username_2: Here! https://github.com/godotengine/godot/pull/25353 Status: Issue closed
angular/angular
1165882638
Title: Tour of Heroes Docs: Missing CSS Styles in src/styles.css Question: username_0: ### Description Tour of Heroes (TOH) Part 0 is missing button styles in `src/styles.css`. As a result, when developers go through the TOH, button CSS styles that are present in the live examples in src/styles.css are not present in the code snippets at the end of the lesson. This is most apparent in [TOH part-2](https://angular.io/tutorial/toh-pt2), where a list of buttons are shown without the expected styles. ### What is the affected URL? https://stackblitz.com/edit/angular-ivy-3vttek?file=src%2Fstyles.css ### Please provide the steps to reproduce the issue 1. Create new project in TOH Part 0 and apply code snippets to project 2. Follow steps in TOH part 1 and apply code snippets 3. Follow steps in TOH part 2 and apply code snippets 4. View output project styling for heroes.component.html without the expected button styling Minimal Reproduction https://stackblitz.com/edit/angular-ivy-3vttek?file=src%2Fstyles.css ### Please provide the expected behavior vs the actual behavior you encountered **Expected**: heroes.component.html will have expected button styles applied as seen in the [live example ](https://stackblitz.com/run?file=src/app/heroes/heroes.component.html)for TOH part 2 **Actual**: heroes.component.html does not have button styles applied. ### Please provide a screenshot if possible Actual <img width="331" alt="TOH-p2-button-styles-actual" src="https://user-images.githubusercontent.com/20045743/157781153-7268ae58-6ac6-4110-a9a8-a2420ce76340.png"> Expected <img width="331" alt="TOH-p2-button-styles-expected" src="https://user-images.githubusercontent.com/20045743/157781181-0bdf8ea3-80aa-45a2-a935-1bcaf9f953f2.png"> ### Please provide the exception or error you saw ```true N/A ``` ### Is this a browser-specific issue? If so, please specify the device, browser, and version. ```true N/A ``` Answers: username_1: There is already a dedicated section in TOH part 2 explaining how to add styles after adding buttons: https://angular.io/tutorial/toh-pt2#style-the-heroes username_0: @username_1 It explains how to add, but not what to add. Giving a code snippet provides an example for this section and reinforces differences between component-styles and global-styles. I'd be happy to update this issue to instead suggest a code snippet addition to the Final code review in TOH part 2, rather than TOH part 0, since that flows better with the rest of the documentation. username_1: The `what to add ` is in the last part of the section (with related explanation about component vs global styles): ``` Open the heroes.component.css file and paste in the private CSS styles for the HeroesComponent. You'll find them in the [final code review](https://angular.io/tutorial/toh-pt2#final-code-review) at the bottom of this guide. ``` About: ``` I'd be happy to update this issue to instead suggest a code snippet addition to the Final code review in TOH part 2 ``` The code snippet is already in the final code review part 2 username_1: Oh, i see sorry, I thought it was about button styles from the component not being featured in part 2. Then I would be in favor of fixing it in part 0 as such styles are details beginners shouldn't have to care about to focus on Angular framework itself for this tutorial. username_0: No worries friend, I could have explained it better. And agreed 👍🏻 Status: Issue closed
amagovpt/kit-selo
533357570
Title: Refererência desnecessária a software específico Question: username_0: O texto leva a crer que o uso específico dessa aplicação é necessário para cumprir aquele requisito. Sendo o PDF uma norma aberta, o teste para a sua acessibilidade, por um lado, não depende do uso de um software em específico, e, por outro lado, não se cinge ao funcionamento da função de cópia nesse mesmo software.
quasarframework/quasar
243213913
Title: validation for form in the popup Dialog Question: username_0: feature suggestion/request when I create a Dialog with form components,such as : ``` import { Dialog, Toast } from 'quasar' Dialog.create({ title: 'Prompt', form: { name: { type: 'text', label: 'Textbox', model: '' }, mobilephone: { type: 'text', label: 'Textbox', model: '' }, }, buttons: [ 'Cancel', { label: 'Ok', handler (data) { Toast.create('Returned ' + JSON.stringify(data)) } } ] }) ``` can I do some validation for the mobilephone number? I need some feature to support vuelidate. or can I pass a @blur event handler to mobilephone ? thx very much Status: Issue closed Answers: username_1: Hi, Check http://beta.quasar-framework.org/components/dialog.html#Prevent-Closing-the-Dialog This will allow you to use validations and prevent closing the dialog until all data is filled in correctly. If you need something more specific, create your own Modal. username_2: @username_1 , can you show some code snippet? I can not know how validate from dialogs.
Azure/AKS
287428773
Title: Cannot get logs of running pod via kubectl and kube-svc, kubernetes-dashboard restarted many times Question: username_0: In azure kubernetes cluster, to get logs or exec to interactive of any pods giving error timed out and after `kubectl get pods --all-namespaces -o wide` . The namespace kube-system's pods kube-svc, kubernetes-dashboard restarted many times. logs of `kubectl cluster-info dump` as `==== START logs for container heapster of pod kube-system/heapster-75667786bb-m74tg ==== Request log error: an error on the server ("unknown") has prevented the request from succeeding (get pods heapster-75667786bb-m74tg) ==== END logs for container heapster of pod kube-system/heapster-75667786bb-m74tg ==== ==== START logs for container heapster-nanny of pod kube-system/heapster-75667786bb-m74tg ==== I0110 11:57:54.533538 1 pod_nanny.go:56] Invoked by [/pod_nanny --cpu=80m --extra-cpu=0.5m --memory=140Mi --extra-memory=4Mi --threshold=5 --deployment=heapster --container=heapster --poll-period=300000 --estimator=exponential] I0110 11:57:54.626367 1 pod_nanny.go:68] Watching namespace: kube-system, pod: heapster-75667786bb-m74tg, container: heapster. I0110 11:57:54.626427 1 pod_nanny.go:69] cpu: 80m, extra_cpu: 0.5m, memory: 140Mi, extra_memory: 4Mi, storage: MISSING, extra_storage: 0Gi I0110 11:57:54.627246 1 pod_nanny.go:110] Resources: [{Base:{i:{value:80 scale:-3} d:{Dec:<nil>} s:80m Format:DecimalSI} ExtraPerNode:{i:{value:5 scale:-4} d:{Dec:<nil>} s: Format:DecimalSI} Name:cpu} {Base:{i:{value:146800640 scale:0} d:{Dec:<nil>} s:140Mi Format:BinarySI} ExtraPerNode:{i:{value:4194304 scale:0} d:{Dec:<nil>} s:4Mi Format:BinarySI} Name:memory}] E0110 11:58:24.628329 1 reflector.go:205] k8s.io/contrib/addon-resizer/nanny/kubernetes_client.go:108: Failed to list *v1.Node: Get https://10.0.0.1:443/api/v1/nodes?resourceVersion=0: dial tcp 10.0.0.1:443: i/o timeout E0110 11:58:54.627514 1 nanny_lib.go:87] timed out waiting for the condition E0110 11:58:55.629076 1 reflector.go:205] k8s.io/contrib/addon-resizer/nanny/kubernetes_client.go:108: Failed to list *v1.Node: Get https://10.0.0.1:443/api/v1/nodes?resourceVersion=0: dial tcp 10.0.0.1:443: i/o timeout E0110 11:59:26.629759 1 reflector.go:205] k8s.io/contrib/addon-resizer/nanny/kubernetes_client.go:108: Failed to list *v1.Node: Get https://10.0.0.1:443/api/v1/nodes?resourceVersion=0: dial tcp 10.0.0.1:443: i/o timeout E0110 11:59:57.630452 1 reflector.go:205] k8s.io/contrib/addon-resizer/nanny/kubernetes_client.go:108: Failed to list *v1.Node: Get https://10.0.0.1:443/api/v1/nodes?resourceVersion=0: dial tcp 10.0.0.1:443: i/o timeout E0110 12:00:04.628441 1 nanny_lib.go:87] timed out waiting for the condition E0110 12:00:28.631175 1 reflector.go:205] k8s.io/contrib/addon-resizer/nanny/kubernetes_client.go:108: Failed to list *v1.Node: Get https://10.0.0.1:443/api/v1/nodes?resourceVersion=0: dial tcp 10.0.0.1:443: i/o timeout E0110 12:00:59.632082 1 reflector.go:205] k8s.io/contrib/addon-resizer/nanny/kubernetes_client.go:108: Failed to list *v1.Node: Get https://10.0.0.1:443/api/v1/nodes?resourceVersion=0: dial tcp 10.0.0.1:443: i/o timeout E0110 12:01:14.726064 1 nanny_lib.go:87] timed out waiting for the condition E0110 12:01:30.632856 1 reflector.go:205] k8s.io/contrib/addon-resizer/nanny/kubernetes_client.go:108: Failed to list *v1.Node: Get https://10.0.0.1:443/api/v1/nodes?resourceVersion=0: dial tcp 10.0.0.1:443: i/o timeout E0110 12:02:01.726960 1 reflector.go:205] k8s.io/contrib/addon-resizer/nanny/kubernetes_client.go:108: Failed to list *v1.Node: Get https://10.0.0.1:443/api/v1/nodes?resourceVersion=0: dial tcp 10.0.0.1:443: i/o timeout E0110 12:02:24.726631 1 nanny_lib.go:87] timed out waiting for the condition E0110 12:02:32.727699 1 reflector.go:205] k8s.io/contrib/addon-resizer/nanny/kubernetes_client.go:108: Failed to list *v1.Node: Get https://10.0.0.1:443/api/v1/nodes?resourceVersion=0: dial tcp 10.0.0.1:443: i/o timeout ==== END logs for container heapster-nanny of pod kube-system/heapster-75667786bb-m74tg ==== ==== START logs for container kubedns of pod kube-system/kube-dns-v20-6c8f7f988b-t8p97 ==== I0110 11:40:58.678123 1 dns.go:48] version: 1.14.4-2-g5584e04 I0110 11:40:58.703856 1 server.go:70] Using configuration read from directory: /kube-dns-config with period 10s I0110 11:40:58.703908 1 server.go:113] FLAG: --alsologtostderr="false" I0110 11:40:58.703919 1 server.go:113] FLAG: --config-dir="/kube-dns-config" I0110 11:40:58.703926 1 server.go:113] FLAG: --config-map="" I0110 11:40:58.703931 1 server.go:113] FLAG: --config-map-namespace="kube-system" I0110 11:40:58.703936 1 server.go:113] FLAG: --config-period="10s" I0110 11:40:58.703942 1 server.go:113] FLAG: --dns-bind-address="0.0.0.0" I0110 11:40:58.703947 1 server.go:113] FLAG: --dns-port="10053" I0110 11:40:58.703954 1 server.go:113] FLAG: --domain="cluster.local." I0110 11:40:58.703961 1 server.go:113] FLAG: --federations="" I0110 11:40:58.703968 1 server.go:113] FLAG: --healthz-port="8081" I0110 11:40:58.703973 1 server.go:113] FLAG: --initial-sync-timeout="1m0s" I0110 11:40:58.703977 1 server.go:113] FLAG: --kube-master-url="" I0110 11:40:58.703984 1 server.go:113] FLAG: --kubecfg-file="/config/kubeconfig" I0110 11:40:58.703988 1 server.go:113] FLAG: --log-backtrace-at=":0" I0110 11:40:58.704003 1 server.go:113] FLAG: --log-dir="" I0110 11:40:58.704008 1 server.go:113] FLAG: --log-flush-frequency="5s" I0110 11:40:58.704012 1 server.go:113] FLAG: --logtostderr="true" I0110 11:40:58.704017 1 server.go:113] FLAG: --nameservers="" I0110 11:40:58.704021 1 server.go:113] FLAG: --stderrthreshold="2" I0110 11:40:58.704026 1 server.go:113] FLAG: --v="2" I0110 11:40:58.704030 1 server.go:113] FLAG: --version="false" I0110 11:40:58.704038 1 server.go:113] FLAG: --vmodule="" I0110 11:40:58.704080 1 server.go:176] Starting SkyDNS server (0.0.0.0:10053) I0110 11:40:58.704134 1 server.go:200] Skydns metrics not enabled I0110 11:40:58.704142 1 dns.go:147] Starting endpointsController I0110 11:40:58.704147 1 dns.go:150] Starting serviceController I0110 11:40:58.704261 1 logs.go:41] skydns: ready for queries on cluster.local. for tcp://0.0.0.0:10053 [rcache 0] I0110 11:40:58.704272 1 logs.go:41] skydns: ready for queries on cluster.local. for udp://0.0.0.0:10053 [rcache 0] I0110 11:40:59.204376 1 dns.go:171] Initialized services and endpoints from apiserver I0110 11:40:59.204403 1 server.go:129] Setting up Healthz Handler (/readiness) I0110 11:40:59.204412 1 server.go:134] Setting up cache handler (/cache) I0110 11:40:59.204438 1 server.go:120] Status HTTP port 8081 [Truncated] W0110 11:40:15.322895 1 server.go:191] WARNING: all flags other than --config, --write-config-to, and --cleanup are deprecated. Please begin using a config file ASAP. time="2018-01-10T11:40:15Z" level=warning msg="Running modprobe ip_vs failed with message: ``, error: exec: \"modprobe\": executable file not found in $PATH" time="2018-01-10T11:40:15Z" level=error msg="Could not get ipvs family information from the kernel. It is possible that ipvs is not enabled in your kernel. Native loadbalancing will not work until this is fixed." W0110 11:40:15.385440 1 server_others.go:263] Flag proxy-mode="" unknown, assuming iptables proxy I0110 11:40:15.387256 1 server_others.go:117] Using iptables Proxier. I0110 11:40:15.554170 1 server_others.go:152] Tearing down inactive rules. I0110 11:40:15.706742 1 conntrack.go:98] Set sysctl 'net/netfilter/nf_conntrack_max' to 131072 I0110 11:40:15.711494 1 conntrack.go:52] Setting nf_conntrack_max to 131072 I0110 11:40:15.712763 1 conntrack.go:83] Setting conntrack hashsize to 32768 I0110 11:40:15.712986 1 conntrack.go:98] Set sysctl 'net/netfilter/nf_conntrack_tcp_timeout_established' to 86400 I0110 11:40:15.713096 1 conntrack.go:98] Set sysctl 'net/netfilter/nf_conntrack_tcp_timeout_close_wait' to 3600 I0110 11:40:15.731308 1 config.go:202] Starting service config controller I0110 11:40:15.731336 1 controller_utils.go:1041] Waiting for caches to sync for service config controller I0110 11:40:15.731544 1 config.go:102] Starting endpoints config controller I0110 11:40:15.731555 1 controller_utils.go:1041] Waiting for caches to sync for endpoints config controller I0110 11:40:15.832032 1 controller_utils.go:1048] Caches are synced for endpoints config controller I0110 11:40:15.832141 1 controller_utils.go:1048] Caches are synced for service config controller ==== END logs for container kube-proxy of pod kube-system/kube-proxy-qj987 ==== ==== START logs for container redirector of pod kube-system/kube-svc-redirect-b8hd6 ==== ` Answers: username_1: I am experiencing the same issue - timeout when retrieving logs. username_2: I also get timeouts and the request times are very slow :( username_3: Same issue here in US East. ``` root@aks-default-15610861-0:~# kubectl logs -f tiller-deploy-7b87468d9-p4sv4 -n kube-system Error from server: Get https://aks-default-15610861-1:10250/containerLogs/kube-system/tiller-deploy-7b87468d9-p4sv4/tiller?fo llow=true: dial tcp 10.125.8.9:10250: getsockopt: connection timed out root@aks-default-15610861-0:~# telnet 10.125.8.9 10250 Trying 10.125.8.9... Connected to 10.125.8.9. Escape character is '^]'. ^] telnet> quit Connection closed. root@aks-default-15610861-0:~# ``` In addition my dns pods are crashing. ``` root@aks-default-15610861-0:~# kubectl get pods --all-namespaces NAMESPACE NAME READY STATUS RESTARTS AGE kube-system heapster-2574232661-lxmn7 2/2 Running 0 1h kube-system kube-dns-v20-2253765213-djdt9 1/3 CrashLoopBackOff 34 1h kube-system kube-dns-v20-2253765213-j9g5r 2/3 CrashLoopBackOff 43 1h kube-system kube-proxy-9lkbm 1/1 Running 0 1h kube-system kube-proxy-qnl26 1/1 Running 0 1h kube-system kube-proxy-x56f4 1/1 Running 0 1h kube-system kube-svc-redirect-2kr8h 1/1 Running 0 1h kube-system kube-svc-redirect-wdg8f 1/1 Running 0 1h kube-system kube-svc-redirect-z6h6l 1/1 Running 0 1h kube-system kubernetes-dashboard-6d9c57c89c-xldlg 1/1 Running 0 1h kube-system tiller-deploy-7b87468d9-p4sv4 1/1 Running 0 38m kube-system tunnelfront-5ff8ddff6d-87t86 1/1 Running 0 1h ``` username_4: Same issue here in eastus. username_5: Same issue, West Europe username_6: I was able to connect my AKS cluster with kubectl but now it stopped responding. What ever kubectl commands I try, it gives "Unable to connect to the server: dial tcp: i/o timeout". Any hints how to get over this? There is no way any more to monitor my pods :-( username_7: I am seeing similar issues. I am able to retrieve pods, deployments, but I can't get logs from any pod, can't connect to the dashboard, can't run busybox inside the cluster. username_8: Ran into this issue recently in the East U.S. region. I'm using the cluster for testing and am turning the VMs off and back on as part of this testing (to work around other suspected AKS issues). This last time it came back up `kubectl logs` and `kubectl exec` were both not working, with the same timeout message as in the original report. After turning VMs off and back on yet again, `kubectl logs` and `kubectl exec` are now working again. Status: Issue closed username_9: Closing due inactivity. Feel free to re-open if still an issue. username_10: I had this problem and it turned out that I had a resource that used the subnet dedicated to AKS. you have to check this and if so, remove the resource. username_11: I ran into the same problem, I am running on bare-metal using kubeadm & ubuntu 16.x. In my case, the master was unable access the kubelet because firewall ports were not open on one of my nodes, opening them up resolved the issue. Command ufw status --to list open ports ufw allow 10250/tcp ufw allow 30000:32767/tcp
dlang/dub
162050070
Title: single file package, random failure Question: username_0: Program exited with code 1 error: the process (/home/basile/bin/dub) has returned the signal 512 Answers: username_1: Could it be that two DUB instances are running in parallel? It appears that the compiled executable gets modified while it is still being executed. A possible fix within DUB in that case would be to `rm` the file before invoking the linker. username_0: This wouldn't be logical. On linux, executable files are not locked when executed and we can recompile them even if the previous version is being run. I will invistigate more but I keep this issue opened as a reminder. Status: Issue closed
sockeqwe/fragmentargs
242352665
Title: New release that allows to obtain a bundle from builder Question: username_0: Hello @username_1, our team really likes fragmentargs and we use it in a multiple projects already. Recently we where looking for a way to use fragmentargs to generate a Bundle that we can use where we can not instantiate a Fragment directly. If have seen that this functionality is available in the snapshot version since may 2016. Is there any release planned within the next few days/weeks that will provide this feature in a non-snapshot version? Best Regards, Andreas Answers: username_1: Oh boy, I'm really sorry ... I just have focused my open source time on other project as they have had higher priority for me since we use FragmentArgs version 3 at work and it fits our needs. There are still some tasks / issues open for a stable 4.0 version: https://github.com/username_1/fragmentargs/milestone/4 The most important ones are #46 #44 #73 #83 #82 and #17 . The remaining ones could also be "fixed" in a minor version like `4.0.1` I can't give any serious estimation when I will have time to work on this, but "within few days" is not possible for me. I think first week of August is realistic. If someone wants to contribute, pull requests are very welcome! username_0: Thanks for the quick reply! I will try to support you getting tasks/issues fixed for version 4.0 👍 Status: Issue closed
pytorch/pytorch
399614338
Title: I run the correct official demo TypeError: 'NoneType' object is not iterable Question: username_0: I run the correct official demo code: ![image](https://user-images.githubusercontent.com/12770958/51221607-2d191f80-1975-11e9-8f25-0e61647cd68b.png) I don't know why will quote this built-in function error Answers: username_0: I'm sorry, but I know where I wrong Status: Issue closed
don/cordova-plugin-ble-central
127625798
Title: ios not discover phone with android Question: username_0: Hello, since few days I was trying to figure it out and I failed. When I use scan command it nor discover my android phone, but in my iPad bluetooth settings it is discovered (see screen bellow). I need only to get uuid, I username_2't want to connect/write etc - generally I only want to discover devices, which are in bluetooth range. Please help. Kind regards. ![1453276126373](https://cloud.githubusercontent.com/assets/4047675/12442775/1178ce76-bf53-11e5-9e22-a7fdfaaabcb8.jpg) Answers: username_1: I encountered the same problem as well. ble.scan is unable to detect an android or iOS phone. Can anyone help? username_2: In order for this plugin to discover another phone, that phone needs to be *advertising*. The easiest way to do this is to have the target phone advertising as a Bluetooth Low Energy peripheral. It's also possible to advertise without actually offering services. username_1: Forgive me for my ignorance. Is it just a phone setting to get the phone (both Android and iOS) to advertise as a peripheral, or an app is necessary to achieve that? If an app need to be written, are there Cordova plugins available, particularly for Android phones? I think we can do this with native Android code, but I am using Cordova. Thanks. username_2: For now I think you need native code to make iOS and Android advertise Bluetooth services. I username_2't know of any Cordova plugins to do this. I've been working on cordova-plugin-ble-peripheral but it's not complete. username_1: Thank you. Hope to see your plugin soon. Status: Issue closed
CatServer/CatServer
427356593
Title: 加入夸克(Quark)无法启动 Question: username_0: 服务端加入Quark之后启动失败,官方服务端+Forge启动没问题 `[12:22:31] [main/INFO] [LaunchWrapper]: Loading tweak class name catserver.server.launcher.username_2Tweaker [12:22:31] [main/INFO] [LaunchWrapper]: Using primary tweak class name catserver.server.launcher.username_2Tweaker [12:22:31] [main/INFO] [LaunchWrapper]: Calling tweak class catserver.server.launcher.username_2Tweaker [12:22:32] [main/INFO] [FML]: Forge Mod Loader version 14.23.5.2815 for Minecraft 1.12.2 loading [12:22:32] [main/INFO] [FML]: Java is Eclipse OpenJ9 VM, version 1.8.0_202, running on Windows 10:amd64:10.0, installed at C:\Program Files\AdoptOpenJDK\jdk-8.0.202.08\jre [12:22:32] [main/ERROR] [FML]: Apache Maven library folder was not in the format expected. Using default libraries directory. [12:22:32] [main/ERROR] [FML]: Full: C:\Users\lxu36\Desktop\username_2\libraries\maven-artifact-3.5.3.jar [12:22:32] [main/ERROR] [FML]: Trimmed: c:/users/lxu36/desktop/catserver/ [12:22:33] [main/INFO] [FML]: Searching C:\Users\lxu36\Desktop\username_2\.\mods for mods [12:22:33] [main/INFO] [FML]: Searching C:\Users\lxu36\Desktop\username_2\.\mods\1.12.2 for mods [12:22:33] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in AppleCore-mc1.12.2-3.2.0.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [12:22:33] [main/WARN] [FML]: The coremod AppleCore (squeek.applecore.AppleCore) is not signed! [12:22:33] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in Aroma1997Core-1.12.2-2.0.0.2.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [12:22:33] [main/WARN] [FML]: The coremod aroma1997.core.coremod.CoreMod does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:33] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in astralsorcery-1.12.2-1.10.11.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [12:22:33] [main/WARN] [FML]: The coremod hellfirepvp.astralsorcery.core.AstralCore does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:33] [main/INFO] [Astral Core]: [AstralCore] Initialized. [12:22:33] [main/INFO] [FML]: Loading tweaker guichaguri.betterfps.tweaker.BetterFpsTweaker from BetterFps-1.4.8.jar [12:22:33] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in BetterWithLib-1.12-1.5.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [12:22:33] [main/WARN] [FML]: The coremod LoadingPlugin (betterwithmods.library.core.LoadingPlugin) is not signed! [12:22:33] [main/WARN] [FML]: The coremod com.bloodnbonesgaming.bnbgamingcore.core.BNBGamingCorePlugin does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:33] [main/WARN] [FML]: The coremod BNBGamingCore (com.bloodnbonesgaming.bnbgamingcore.core.BNBGamingCorePlugin) is not signed! [12:22:33] [main/INFO] [FML]: Loading tweaker codechicken.asm.internal.Tweaker from ChickenASM-1.12-1.0.2.7.jar [12:22:33] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in CTM-MC1.12.2-0.3.3.22.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [12:22:33] [main/WARN] [FML]: The coremod team.chisel.ctm.client.asm.CTMCorePlugin does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:33] [main/WARN] [FML]: The coremod CTMCorePlugin (team.chisel.ctm.client.asm.CTMCorePlugin) is not signed! [12:22:33] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in Farseek-1.12-2.3.1.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [12:22:33] [main/WARN] [FML]: The coremod farseek.core.FarseekCoreMod does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:33] [main/WARN] [FML]: The coremod FarseekCoreMod (farseek.core.FarseekCoreMod) is not signed! [12:22:33] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in foamfix-0.10.3-1.12.2.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [12:22:33] [main/WARN] [FML]: The coremod pl.asie.foamfix.coremod.FoamFixCore does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:33] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in Forgelin-1.8.2.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [12:22:33] [main/WARN] [FML]: The coremod net.shadowfacts.forgelin.preloader.ForgelinPlugin does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:33] [main/WARN] [FML]: The coremod ForgelinPlugin (net.shadowfacts.forgelin.preloader.ForgelinPlugin) is not signed! [12:22:34] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in InventoryTweaks-1.64+dev.146.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [12:22:34] [main/WARN] [FML]: The coremod invtweaks.forge.asm.FMLPlugin does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:34] [main/WARN] [FML]: The coremod ivorius.ivtoolkit.IvToolkitLoadingPlugin does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:34] [main/WARN] [FML]: The coremod IvToolkit (ivorius.ivtoolkit.IvToolkitLoadingPlugin) is not signed! [12:22:34] [main/INFO] [FML]: Loading tweaker org.spongepowered.asm.launch.MixinTweaker from JustEnoughIDs-1.0.2-26.jar [12:22:34] [main/WARN] [FML]: The coremod micdoodle8.mods.miccore.MicdoodlePlugin does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:34] [main/WARN] [FML]: The coremod MicdoodlePlugin (micdoodle8.mods.miccore.MicdoodlePlugin) is not signed! [12:22:34] [main/ERROR] [FML]: The coremod api.player.forge.PlayerAPIPlugin is requesting minecraft version 1.12.1 and minecraft is 1.12.2. It will be ignored. [12:22:34] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in Quark-r1.5-146.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [12:22:34] [main/WARN] [FML]: The coremod vazkii.quark.base.asm.LoadingPlugin does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:34] [main/WARN] [FML]: The coremod LoadingPlugin (vazkii.quark.base.asm.LoadingPlugin) is not signed! [12:22:34] [main/WARN] [FML]: Found FMLCorePluginContainsFMLMod marker in ResourceLoader-MC1.12.1-1.5.3.jar. This is not recommended, @Mods should be in a separate jar from the coremod. [12:22:34] [main/WARN] [FML]: The coremod lumien.resourceloader.asm.LoadingPlugin does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:34] [main/WARN] [FML]: The coremod blusunrize.immersiveengineering.common.asm.IELoadingPlugin does not have a MCVersion annotation, it may cause issues with this version of Minecraft [12:22:34] [main/INFO] [LaunchWrapper]: Loading tweak class name net.minecraftforge.fml.common.launcher.FMLInjectionAndSortingTweaker [12:22:34] [main/INFO] [LaunchWrapper]: Loading tweak class name guichaguri.betterfps.tweaker.BetterFpsTweaker [12:22:34] [main/INFO] [LaunchWrapper]: Loading tweak class name codechicken.asm.internal.Tweaker [12:22:34] [main/INFO] [LaunchWrapper]: Loading tweak class name org.spongepowered.asm.launch.MixinTweaker [12:22:34] [main/INFO] [mixin]: SpongePowered MIXIN Subsystem Version=0.7.11 Source=file:/C:/Users/lxu36/Desktop/username_2/./mods/JustEnoughIDs-1.0.2-26.jar Service=LaunchWrapper Env=SERVER [12:22:34] [main/WARN] [FML]: The coremod JEIDLoadingPlugin (org.dimdev.jeid.JEIDLoadingPlugin) is not signed! [12:22:34] [main/INFO] [mixin]: Compatibility level set to JAVA_8 [12:22:34] [main/INFO] [LaunchWrapper]: Loading tweak class name net.minecraftforge.fml.common.launcher.FMLDeobfTweaker [12:22:34] [main/INFO] [LaunchWrapper]: Calling tweak class net.minecraftforge.fml.common.launcher.FMLInjectionAndSortingTweaker [Truncated] [12:22:38] [main/INFO] [BetterFps]: Patching net.minecraft.block.Block... (aow) [12:22:39] [main/INFO] [mixin]: A re-entrant transformer 'guichaguri.betterfps.transformers.PatcherTransformer' was detected and will no longer process meta class data [12:22:39] [main/INFO] [STDOUT]: [team.chisel.ctm.client.asm.CTMTransformer:preTransform:230]: Transforming Class [net.minecraft.block.Block], Method [getExtendedState] [12:22:39] [main/INFO] [STDOUT]: [team.chisel.ctm.client.asm.CTMTransformer:finishTransform:242]: Transforming net.minecraft.block.Block Finished. [12:22:39] [main/INFO] [STDOUT]: [pl.asie.patchy.helpers.ConstructorReplacingTransformer$FFMethodVisitor:visitTypeInsn:75]: Replaced NEW for net/minecraft/util/ClassInheritanceMultiMap to pl/asie/foamfix/coremod/common/FoamyClassInheritanceMultiMap [12:22:39] [main/INFO] [STDOUT]: [pl.asie.patchy.helpers.ConstructorReplacingTransformer$FFMethodVisitor:visitMethodInsn:87]: Replaced INVOKESPECIAL for net/minecraft/util/ClassInheritanceMultiMap to pl/asie/foamfix/coremod/common/FoamyClassInheritanceMultiMap [12:22:39] [main/INFO] [mixin]: A re-entrant transformer '$wrapper.pl.asie.foamfix.coremod.FoamFixTransformer' was detected and will no longer process meta class data [12:22:39] [main/INFO] [Astral Core]: [AstralTransformer] Transforming alm : net.minecraft.enchantment.EnchantmentHelper with 1 patches! [12:22:39] [main/INFO] [Astral Core]: [AstralTransformer] Applied patch PATCHMODIFYENCHANTMENTLEVELS [12:22:39] [main/INFO] [Astral Core]: [AstralTransformer] Transforming amu : net.minecraft.world.World with 2 patches! [12:22:39] [main/INFO] [Astral Core]: [AstralTransformer] Skipping PATCHSUNBRIGHTNESSWORLDCLIENT as it can't be applied for side SERVER [12:22:39] [main/INFO] [Astral Core]: [AstralTransformer] Applied patch PATCHSUNBRIGHTNESSWORLDCOMMON [12:22:39] [main/INFO] [STDOUT]: [pl.asie.foamfix.coremod.FoamFixTransformer:spliceClasses:112]: Added INTERFACE: pl/asie/foamfix/coremod/patches/IFoamFixWorldRemovable [12:22:39] [main/INFO] [STDOUT]: [pl.asie.foamfix.coremod.FoamFixTransformer:spliceClasses:149]: Added METHOD: net.minecraft.world.World.foamfix_removeUnloadedEntities [12:22:39] [main/INFO] [FML]: [Quark ASM] Transforming WorldServer [12:22:39] [main/INFO] [FML]: [Quark ASM] Applying Transformation to method (Names [areAllPlayersAsleep, func_73056_e, g] Descriptor ()Z / ()Z) [12:22:39] [main/INFO] [FML]: [Quark ASM] Located Method, patching... [12:22:39] [main/INFO] [FML]: [Quark ASM] Located patch target node L0 [12:22:39] [main/INFO] [FML]: [Quark ASM] Patch result: true ` Answers: username_1: paste.ubuntu.com是个好东西 username_0: 抱歉,头一次提交BUG,没考虑到那么多,之前顺手给删了,有时间我在试试吧 Status: Issue closed
lampepfl/dotty
108887623
Title: AbstractMethodError with outer links Question: username_0: Dotty compiled by dotty has compiled hello world, so I am looking now into more complicated phases: ``` Exception in thread "main" java.lang.AbstractMethodError: dotty.tools.dotc.transform.PatternMatcher$Translator$OptimizingMatchTranslator.dotty$tools$dotc$transform$PatternMatcher$Translator$MatchTranslator$$$outer()Ldotty/tools/dotc/transform/PatternMatcher$Translator; ``` Answers: username_0: How to reproduce: compile dotty by dotty(is part of test suite). Than compile such file: ```scala object A { def main(args: Array[String]): Unit = { args match { case Array() => println("empty"); case _ => println(args)} } } ``` ``` Exception in thread "main" java.lang.AbstractMethodError: dotty.tools.dotc.transform.PatternMatcher$Translator$OptimizingMatchTranslator.dotty$tools$dotc$transform$PatternMatcher$Translator$MatchTranslator$$$outer()Ldotty/tools/dotc/transform/PatternMatcher$Translator; at dotty.tools.dotc.transform.PatternMatcher$Translator$MatchTranslator.translateMatch(PatternMatcher.scala:1166) at dotty.tools.dotc.transform.PatternMatcher$Translator$OptimizingMatchTranslator.translateMatch(PatternMatcher.scala:75) at dotty.tools.dotc.transform.PatternMatcher.transformMatch(PatternMatcher.scala:51) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.goMatch(TreeTransform.scala:724) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transformUnnamed(TreeTransform.scala:1114) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.liftedTree7$1(TreeTransform.scala:1213) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.$anonfun$transform$10(TreeTransform.scala:1205) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer$$Lambda$689/746074699.apply(Unknown Source) at dotty.tools.dotc.reporting.Reporting.traceIndented(Reporter.scala:148) at dotty.tools.dotc.core.Contexts$Context.traceIndented(Contexts.scala:53) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transform(TreeTransform.scala:1204) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transformUnnamed(TreeTransform.scala:1087) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.liftedTree7$1(TreeTransform.scala:1213) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.$anonfun$transform$10(TreeTransform.scala:1205) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer$$Lambda$689/746074699.apply(Unknown Source) at dotty.tools.dotc.reporting.Reporting.traceIndented(Reporter.scala:148) at dotty.tools.dotc.core.Contexts$Context.traceIndented(Contexts.scala:53) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transform(TreeTransform.scala:1204) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transformNamed(TreeTransform.scala:1004) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.liftedTree7$1(TreeTransform.scala:1212) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.$anonfun$transform$10(TreeTransform.scala:1205) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer$$Lambda$689/746074699.apply(Unknown Source) at dotty.tools.dotc.reporting.Reporting.traceIndented(Reporter.scala:148) at dotty.tools.dotc.core.Contexts$Context.traceIndented(Contexts.scala:53) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transform(TreeTransform.scala:1204) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transformStat$2(TreeTransform.scala:1238) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.$anonfun$662(TreeTransform.scala:1242) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer$$Lambda$690/596470015.apply(Unknown Source) at dotty.tools.dotc.core.Decorators$ListDecorator$.loop$4(Decorators.scala:51) at dotty.tools.dotc.core.Decorators$ListDecorator$.mapconserve$extension(Decorators.scala:67) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transformStats(TreeTransform.scala:1242) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transformUnnamed(TreeTransform.scala:1183) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.liftedTree7$1(TreeTransform.scala:1213) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.$anonfun$transform$10(TreeTransform.scala:1205) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer$$Lambda$689/746074699.apply(Unknown Source) at dotty.tools.dotc.reporting.Reporting.traceIndented(Reporter.scala:148) at dotty.tools.dotc.core.Contexts$Context.traceIndented(Contexts.scala:53) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transform(TreeTransform.scala:1204) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transformNamed(TreeTransform.scala:1011) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.liftedTree7$1(TreeTransform.scala:1212) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.$anonfun$transform$10(TreeTransform.scala:1205) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer$$Lambda$689/746074699.apply(Unknown Source) at dotty.tools.dotc.reporting.Reporting.traceIndented(Reporter.scala:148) at dotty.tools.dotc.core.Contexts$Context.traceIndented(Contexts.scala:53) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transform(TreeTransform.scala:1204) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.transformStat$2(TreeTransform.scala:1238) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.$anonfun$662(TreeTransform.scala:1242) [Truncated] at dotty.tools.dotc.core.Phases$Phase.runOn(Phases.scala:273) at dotty.tools.dotc.transform.TreeTransforms$TreeTransformer.runOn(TreeTransform.scala:474) at dotty.tools.dotc.Run.$anonfun$$anonfun$compileUnits$1$1(Run.scala:59) at dotty.tools.dotc.Run$$Lambda$194/388357135.applyVoid(Unknown Source) at scala.compat.java8.JProcedure1.apply(JProcedure1.java:18) at scala.compat.java8.JProcedure1.apply(JProcedure1.java:10) at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186) at dotty.tools.dotc.Run.$anonfun$compileUnits$1(Run.scala:57) at dotty.tools.dotc.Run$$Lambda$185/501187768.apply$mcV$sp(Unknown Source) at scala.compat.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:12) at dotty.tools.dotc.util.Stats$.monitorHeartBeat(Stats.scala:58) at dotty.tools.dotc.Run.compileUnits(Run.scala:65) at dotty.tools.dotc.Run.compileSources(Run.scala:49) at dotty.tools.dotc.Run.compile(Run.scala:33) at dotty.tools.dotc.Driver.doCompile(Driver.scala:21) at dotty.tools.dotc.Driver.process(Driver.scala:44) at dotty.tools.dotc.Driver.main(Driver.scala:48) at dotty.tools.dotc.Main.main(Main.scala) ``` Status: Issue closed
shankarpandala/lazypredict
1005312125
Title: Custom metric Question: username_0: Hi there, Thank you for this code. It would be nice if you could provide an example of how to use custom metrics. Answers: username_1: I second this. All examples use `custom_metrics=None`. I tried assigning a series of strings looking at the names in the source code but no luck.
surbhit21/GSoC-2019
471522833
Title: File not found error Question: username_0: Hi, I tried running the optimization and it seems that there is some issue with the unzipping. I am adding the error trace. `Internal Server Error: /tasks/Optimization/ Traceback (most recent call last): File "/home/surbhit/Work/GSoC-2019/venv/lib/python3.6/site-packages/django/core/handlers/exception.py", line 34, in inner response = get_response(request) File "/home/surbhit/Work/GSoC-2019/venv/lib/python3.6/site-packages/django/core/handlers/base.py", line 115, in _get_response response = self.process_exception_by_middleware(e, request) File "/home/surbhit/Work/GSoC-2019/venv/lib/python3.6/site-packages/django/core/handlers/base.py", line 113, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/home/surbhit/Work/GSoC-2019/venv/lib/python3.6/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view return view_func(*args, **kwargs) File "/home/surbhit/Work/GSoC-2019/venv/lib/python3.6/site-packages/rest_framework/viewsets.py", line 116, in view return self.dispatch(request, *args, **kwargs) File "/home/surbhit/Work/GSoC-2019/venv/lib/python3.6/site-packages/rest_framework/views.py", line 495, in dispatch response = self.handle_exception(exc) File "/home/surbhit/Work/GSoC-2019/venv/lib/python3.6/site-packages/rest_framework/views.py", line 455, in handle_exception self.raise_uncaught_exception(exc) File "/home/surbhit/Work/GSoC-2019/venv/lib/python3.6/site-packages/rest_framework/views.py", line 492, in dispatch response = handler(request, *args, **kwargs) File "/home/surbhit/Work/GSoC-2019/GSoC-2019/REST_FindSim/tasks/views.py", line 68, in create os.mkdir(os.path.join(BASE_FILE_PATH, 'tsv/')+file_label) FileNotFoundError: [Errno 2] No such file or directory: 'media/files/tsv/surbhit11563865698.6517096'` Answers: username_1: Hi, I think this happens when using os.mkdir() and pyrhon fails to find the directory"media/files/tsv/". I will add this empty directory into repo. Hope this can fix the problem. username_0: I am having the "media/files/tsv/" directory and the zip is also getting stored there but the specific folder is not created in this case the `surbhit11563865698.6517096` folder username_1: Maybe it's because the path here is a relative path rather than a absolute path? username_1: I knew the reason. It's because os.mkdir() can not create dir in a complicated path. I changed this into os.makedirs(). I think this will solve the problem. username_0: The os.makedirs(), did not solve the error on my machine. Is there anything else I have to do besides pulling the last commit and running it? Also, the folders are created but they are empty. `FileNotFoundError at /tasks/Optimization/ [Errno 2] No such file or directory: 'media/files/tsv/surbhit11563881005.749884/testop.zip'` username_1: I think you need to migrate. Because the models are changed and the tsv files and model files are uploading into media/files/tsv/ and media/files/model/ respectively. Status: Issue closed username_0: Thanks Chen, migrating worked like a charm. It is running on my machine now. :)
Zrips/CMI
426963307
Title: itemname and itemlore can't find the specified player Question: username_0: **Description of issue or feature request:** I want to use these commands from the console: /cmi itemlore [playerName] [linenumber] [new lore line] /cmi itemname [playerName] [new name] I receive the message that the player not found. The player is online, the nickname is correct. --- **Cmi Version (using`/cmi version`):** 192.168.3.11 **Server Type (Spigot/Paperspigot/etc):** Paper **Server Version (using `/ver`):** Paper(571) 1.13.2 Answers: username_1: [14:04:32 INFO]: Can't find player with this name! ``` Which resulted in player not found showing twice. https://i.imgur.com/dXVy980.png Status: Issue closed username_2: Format was changed to -p:[playerName] when indicating player, just to simplify player name detection
openstates/openstates.org
302391978
Title: Bill search does not "remember" the search term used in main search Question: username_0: Example flow: 1. On the homepage, search a term in the header's search box 2. On the search results page, click the `View all bill results` button 3. The bill-search page will come up, but it won't remember what the search term was (ie, it'll display all bills for the country, unfiltered, and the `Search text` field on the left side panel will not contain the original search term) Answers: username_1: dupe of #56 Status: Issue closed
electron/electron
1058947734
Title: [Bug]: Electron helper high cpu usage Question: username_0: ### Preflight Checklist - [X] I have read the [Contributing Guidelines](https://github.com/electron/electron/blob/main/CONTRIBUTING.md) for this project. - [X] I agree to follow the [Code of Conduct](https://github.com/electron/electron/blob/main/CODE_OF_CONDUCT.md) that this project adheres to. - [X] I have searched the [issue tracker](https://www.github.com/electron/electron/issues) for a feature request that matches the one I want to file, without success. ### Electron Version 16.0.1 ### What operating system are you using? macOS ### Operating System Version macOS big sur 11.5.2 ### What arch are you using? x64 ### Last Known Working Electron version _No response_ ### Expected Behavior Observing system load should be fine I think. https://systeminformation.io/statsfunctions.html the module works fine itself, but when i run it with electron or as a fork process it impact cpu usage ### Actual Behavior ![image](https://user-images.githubusercontent.com/1147704/142694927-da78bf5a-4c17-4c4d-b356-c26706f4d352.png) the electron helper is about 99% usage ### Testcase Gist URL https://github.com/username_0/revealer ### Additional Information ![image](https://user-images.githubusercontent.com/1147704/142695082-6956a398-43d3-4652-9581-9d81739bc4c3.png) ```bash git clone https://github.com/username_0/revealer cd revealer npm i npm run dev ``` Answers: username_0: ![image](https://user-images.githubusercontent.com/1147704/142720867-bcaf51bb-d932-42fb-9ddd-2f03219f1bcf.png) macOS High Sierra 10.13.6 is okay username_1: Also starting from @electron v16 facing an abnormally high CPU load issue, see https://github.com/username_1/ElectronMail/issues/461. So had to downgrade to v15. Could not yet narrow down the scope. username_2: Thanks for reporting this and helping to make Electron better! Because of time constraints, triaging code with third-party dependencies is usually not feasible for a small team like Electron's. Would it be possible for you to make a standalone testcase with only the code necessary to reproduce the issue? For example, [Electron Fiddle](https://www.electronjs.org/fiddle) is a great tool for making small test cases and makes it easy to publish your test case to a [gist](https://gist.github.com) that Electron maintainers can use. Stand-alone test cases make fixing issues go more smoothly: it ensure everyone's looking at the same issue, it removes all unnecessary variables from the equation, and it can also provide the basis for automated regression tests. I'm adding the `blocked/need-repro` label for this reason. After you make a test case, please link to it in a followup comment. Thanks in advance! Your help is appreciated.
Opentrons/opentrons
1071079222
Title: bug: If analysis fails outside a Protocol Engine command, app allows proceeding with the run anyway Question: username_0: # Overview The v5.0.0-beta.0 app doesn't block you from trying to run a protocol that's known to have fatal errors. # Steps to reproduce Upload this protocol through the v5.0.0-beta.0 app: ```python metadata = { "apiLevel": "2.11" } def run(protocol): tip_rack = protocol.load_labware("opentrons_96_tiprack_300ul", 1) pipette = protocol.load_instrument("p300_single_gen2", tip_racks=[tip_rack]) raise RuntimeError() well_plate = protocol.load_labware("biorad_96_wellplate_200ul_pcr", 2) pipette.pick_up_tip() pipette.return_tip() ``` # Current behavior * The server responds with an analysis of `"not-ok"` (that nonetheless has some succeeded commands). * The app shows a deck map with all the labware that had successfully loaded before the error. In the example above, the tip rack, but not the well plate. * The app allows proceeding with the run. * The run eventually hits the error and the app shows a red "run failed" banner, with no further details. # Expected behavior * The app should not allow proceeding with the run. * The app should probably not even show a deck layout? Answers: username_0: For this part, we'll probably have to come up with more structured error schemas. Status: Issue closed
gugod/App-perlbrew
92562944
Title: fish shell perlbrew use error Question: username_0: fish shell execute `perlbrew use 5.23.0` error need modify $HOME/perl5/perlbrew/etc/perlbrew.fish function __perlbrew_set_env to set -l code (eval $perlbrew_command env $argv | perl -pe 's/^(export|setenv)\s+(\w+)=(.*)/set -xg $2 $3/; s/^unset[env]*/set -eug/; s/$/;/; y/:/ /') Answers: username_1: Can some fish users verify if this is sttill an issue with perlbrew 0.74 ? username_2: I am still having the issue with perlbrew 0.76, and @username_0 fix works for me.
leo-project/leofs
378528981
Title: [test] Confirm if the rebalance with multiple nodes works as intended Question: username_0: After we got https://github.com/leo-project/leofs/issues/1156 on github, we've started looking into whether or not the rebalance with multiple nodes works as intended. In case of the #1156, a user tried to add 5 nodes into 5 nodes cluster at once and do the rebalance. as a result, many objects were not accessible. I also tried to add 2 nodes into 2 nodes cluster at once and do the rebalance and the result was almost same with the user's one. The reason why some objects were not accessible has not identified yet however there seems to be something wrong in the way to handle OLD/NEW RING. During the rebalance - GET/HEAD should go to storage nodes based on the OLD RING. - PUT/DELETE should go to storage nodes based on the NEW RING This logic seems not to work as intended so we will look into further around this.
aws/aws-iot-device-sdk-js
404432050
Title: Support for non-ssl (ws://) Question: username_0: In our development environment with serverless-offline, IoT connections are being rejected by the browser for invalid certificate authority. This makes sense as we're using self-signed certs and there doesn't appear to be any way to disable checks for self-signed certs when running on a browser and native WebSocket. Side note.. I've seen `rejectUnauthorized: false` and I'm fairly sure it used to work but I can no longer make it work in a browser. Is there something I'm missing here? I've since made several small code changes to aws-iot-device-sdk-js to support ws:// URLs, though I can't say for sure it covers all of the use-cases. Would that be a welcome PR? Status: Issue closed Answers: username_1: Hey @username_0, Thanks for your interest! Since we only support connecting to the AWS IoT endpoints (which REQUIRE SSL), I don't really think this is something we want to support or encourage. Cheers! Colden
cashapp/stagehand
729187521
Title: Improve consistency of terminology in documentation Question: username_0: We should be more consistent with the terminology we use in the project - both in the long form documentation and headerdocs, as well as some of the (private) variable names. For example, when talking about durations, the "duration" can apply to different parts of the animation, so I've started differentiating by saying: * The time from the start of the execution phase to when the animation completes is the `end-to-end duration`. This is usually a calculated value based on the cycle duration and the repeat style (when available). * The time for the animation to go from a relative timestamp of 0 to 1 (or 1 to 0) is the `cycle duration`. This is typically how consumers specify the duration. * The time for the animation to go between two arbitrary relative timestamps is the `segment duration`. This is used in a few places like snapshot testing and the upcoming interactive animations. _See original suggestion in https://github.com/cashapp/stagehand/pull/49#r499112281_
Kleak/flouter
775531851
Title: Use with BottomNavigationBar() Question: username_0: How to use the library with BottomNavigationBar()? I tried these but not sure what the best solution is: https://stackoverflow.com/questions/55888715/how-to-persist-bottomnavigationbar-when-using-flutter-navigation-and-routes https://stackoverflow.com/questions/49681415/flutter-persistent-navigation-bar-with-named-routes Answers: username_1: I didn't tested nested flouter router as of now but it's in the roadmap #5 as soon as this is well tested i will add an example and how to do it in the readme. For now you should be able to follow what is in the link you put above but i think it will not update the URL in your browser. Status: Issue closed
whiteout-io/imap-client
93357979
Title: Timeout Question: username_0: Hi, I write a imap sync application with your lib. I get some errors in upload: [2015-07-06T19:19:23.368Z][imap-client] uploading a message of 5878160 bytes to INBOX [2015-07-06T19:19:26.069Z][browserbox IMAP] [Error: [2] Socket timed out!] [2015-07-06T19:19:26.071Z][browserbox] [2] closing connection [2015-07-06T19:19:26.071Z][browserbox] [2] entering state: 4 [2015-07-06T19:19:26.071Z][imap-client] [Error: IMAP connection encountered an error! Error: [2] Socket timed out!] /home/berci/nodejs/imapmigra/node_modules/imap-client/src/imap-client.js:103 this.onError(new Error(msg)); // report the error ^ TypeError: undefined is not a function at ImapClient._onError (/home/berci/nodejs/imapmigra/node_modules/imap-client/src/imap-client.js:103:18) at BrowserBox.<anonymous> (/home/berci/nodejs/imapmigra/node_modules/imap-client/node_modules/browserbox/src/browserbox.js:130:18) at ImapClient._onError (/home/berci/nodejs/imapmigra/node_modules/imap-client/node_modules/browserbox/src/browserbox-imap.js:362:18) at ImapClient._onTimeout (/home/berci/nodejs/imapmigra/node_modules/imap-client/node_modules/browserbox/src/browserbox-imap.js:407:14) at Timer.listOnTimeout (timers.js:119:15) Can you help me? David<issue_closed> Status: Issue closed
vuejs/vuepress
908344162
Title: Use double brackets as markdown link Question: username_0: Hello, I don't know if this has been referenced before, but I couldn't find it here. I have a lot of files that are referenced each other with `[[name of file]]`. I would like to make these double brackets clickable in order to follow the links. I searched for a possible config file, but I couldn't find it.
phpcfdi/credentials
1154458499
Title: ¿Se puede saber si un CFDI está cancelado? Question: username_0: Hola, Tengo una duda con el Webservice del SAT ¿se puede saber si un CFDI está cancelado?, hasta ahorita la única solución que le he encontrado es una combinación: - Webservice del SAT para la extracción masiva de los CFDI's vigentes - Web Scraping para obtener las canceladas (semi-automatizada ya que requiere la captura del captcha) ¿Habrá alguna otra forma automatizada de saber si los CFDI's están cancelados? Status: Issue closed Answers: username_1: Mira el proyecto [phpcfdi/sat-estado-cfdi](https://github.com/phpcfdi/sat-estado-cfdi) Lee bien la documentación ;)
reapit/foundations
599797719
Title: The what's new link from the help docs is broken Question: username_0: **Describe the bug** Clicking on the help > what's new section takes me to the homepage of the docs, should redirect to the correct page in foundations documentation. Directs to: https://marketplace.reapit.cloud/developer/api-docs#whats-new **Expected behavior** Should direct to: https://marketplace.reapit.cloud/developer/api-docs/whats-new<issue_closed> Status: Issue closed
w3ctag/security-questionnaire
341245575
Title: gh-pages branch out of sync with master Question: username_0: Is the gh-pages branch serving a purpose here? What's displayed at https://w3ctag.github.io/security-questionnaire/ appears to be older than master (i.e. https://rawgit.com/w3ctag/security-questionnaire/master/index.html ) Thanks, working with PING folks to coalesce the various documents here! Answers: username_1: I changed the settings to display the `master` branch (currently 61520b7d65ebf0404097a82164add05e7fdf83f1) rather than the `gh-pages` branch (currently 1f854b0a6f57d70596e1a3e0c1c81a5c086d5eab). Status: Issue closed username_1: And I've now deleted the `gh-pages` branch to avoid future confusion. (Everything on `gh-pages` was on `master`... `master` was just 7 commits ahead.)
maglub/rpi-sous-vide
241436343
Title: Add generic "remote data logging" (influx, etc) to input/output Question: username_0: Currently the remote logging is hard coded into the input/output/control scripts. * Create a generic logging script, which calls whatever plugin you have created for it. Answers: username_0: Added generic bin/logging, which symlinks to logging-available/none per default. An example script for influxdb is available too. Status: Issue closed
gryffon/ringteki
278567497
Title: Before the Throne under Stronghold Question: username_0: Did it just to see if you could. You can. Answers: username_1: Same thing happened to me with the Public Forum province when I first tested it. There's currently no restriction to the player from choosing these cards as your stronghold province. username_2: Thanks for the report - fix submitted #1274 Status: Issue closed
inveniosoftware/invenio-communities
1067328100
Title: Use Api client pattern for requests axios calls Question: username_0: `curate.js`, `members.js`, `records.js`, `requests.js`... should also use the [API front-end client](https://github.com/inveniosoftware/invenio-communities/pull/368/files#diff-ca2b00d4bd11dc8fe60a830144b5347057b2708919baa86ef51417f8d84c6fdeR38) to make their calls when their related functionalities are implemented.
pytorch/pytorch
264794978
Title: how to use pytorch in c++? Question: username_0: i train models in python,but i need c++ to deploy models in fact. how to operate? Status: Issue closed Answers: username_1: You should take a look at ONNX http://onnx.ai/ which is supported by PyTorch, see http://pytorch.org/docs/master/onnx.html
earthlab/hub-ops
345826401
Title: Collect all hub charts in a directory Question: username_0: Create a `hub-config` directory that contains all the charts for the various hubs. Make sure to update the documentation, deploy script, etc Answers: username_1: i have a stupid question @username_0 related to this - all data management stuff. why is `Chart.yml` always capitalized but the others are lower case? Does that matter? username_0: The structure of the directories that configure each hub come from how [helm charts](https://docs.helm.sh/developing_charts/#charts) are organised. I've never wondered why it is `Chart.yml` and not `chart.yml` :) To my shame I've also never experimented with lower vs uppercase. It might well work because I think OSX or Windows filesystems can't actually tell the difference(??). username_1: ok! this is something to explore. but i would like the hub-config dir so we can manage all of these hubs in one place! it's confusing now to have them all floating around :) so would love to sort this feature out! @username_0 Status: Issue closed
cmangos/issues
642470043
Title: 🐛 [Bug Report] Question: username_0: ## 🐛 Bugreport The issue is that when you attempt to zone into Karazhan - the game client freezes and eventually crashes after a few minutes. ### Expected behavior Any level 70 character, in a raid group with the Master's Key (attunement) should be able to zone into the Karazhan raid instance. ### Version & Environment Client Version: "2.4.3" (TBC) CMaNGOS Repo & Commit Hash: https://github.com/cmangos/mangos-tbc/tree/aa971d7 Database Repo & Commit Hash: https://github.com/cmangos/tbc-db/tree/f6d1e Operating System: Ubuntu 18.04 LTS ### Steps to reproduce 1. Create a level 70 character. 2. Assign that character the Master's Key (.add 24490) 3. Join a group with another character & convert it to a raid. 4. Teleport to Karazhan (.tele group karazhan) 5. Attempt to zone in after unlocking the front door with The Master's Key. NOTE: One thing I noted was that a second character on a second test account was able to zone into Karazhan without issue. But when the first character crashed, and the second character re-zoned - his client crashed as well. It seems that this may only be affecting the first character to zone in? ### Crashlog ============================================================================== World of WarCraft (build 8606) Exe: <REMOVED> Time: Jun 20, 2020 4:24:27.959 PM User: <REMOVED> Computer: <REMOVED> ------------------------------------------------------------------------------ This application has encountered a critical error: ERROR #132 (0x85100084) Fatal Exception Program: C:\Games\WoW TBC\Wow.exe Exception: 0xC0000005 (ACCESS_VIOLATION) at 0023:59434901 The instruction at "0x59434901" referenced memory at "0x59434901". The memory could not be "written". WoWBuild: 8606 Realm: Sunwell [<REMOVED>:8085] Local Zone: Karazhan Local Player: Carneasada, 0000000000000001, (-11101.8,-1998.31,49.8927) ------------------------------------------------------------------------------ ---------------------------------------- x86 Registers ---------------------------------------- EAX=59434901 EBX=0019FDC0 ECX=33FD0F00 EDX=ABA9ABA8 ESI=0E05C008 EDI=00006801 EBP=0019FDAC ESP=0019FD8C EIP=59434901 FLG=00010202 CS =0023 DS =002B ES =002B SS =002B FS =0053 GS =002B ---------------------------------------- Stack Trace (Manual) [Truncated] --- Thread ID: 8836 --- 74C46359 02BDFF80 0001:00006359 C:\Windows\System32\KERNEL32.DLL 770E7C24 02BDFFDC 0001:00066C24 C:\Windows\SYSTEM32\ntdll.dll 770E7BF4 02BDFFEC 0001:00066BF4 C:\Windows\SYSTEM32\ntdll.dll --- Thread ID: 8716 --- 74C46359 02DBFF80 0001:00006359 C:\Windows\System32\KERNEL32.DLL 770E7C24 02DBFFDC 0001:00066C24 C:\Windows\SYSTEM32\ntdll.dll 770E7BF4 02DBFFEC 0001:00066BF4 C:\Windows\SYSTEM32\ntdll.dll --- Thread ID: 10772 --- 74C46359 02EBFF80 0001:00006359 C:\Windows\System32\KERNEL32.DLL 770E7C24 02EBFFDC 0001:00066C24 C:\Windows\SYSTEM32\ntdll.dll 770E7BF4 02EBFFEC 0001:00066BF4 C:\Windows\SYSTEM32\ntdll.dll --- Thread ID: 7572 --- 7511F59F 0407FF0C 0001:0010E59F C:\Windows\System32\KERNELBASE.dll 0065EF34 0407FF68 0001:0025DF34 C:\Games\WoW TBC\Wow.exe 0075FDD4 0407FF80 0001:0035EDD4 C:\Games\WoW TBC\Wow.exe Answers: username_1: Please, try again with latest master. username_0: @username_1 - the latest commit to master seems to have resolved the problem. Tested on commit tag `1331cff` Status: Issue closed
mlozadad/mlozadad.github.io
258067161
Title: [Portafolio] Actualizar imágenes con algo relacionado al servicio Question: username_0: -[ ] Tomar como referencia al momento de editar contenido: https://www.xatakafoto.com/guias/crea-un-portfolio-de-fotografia-con-exito -[ ] Imagen referencia http://gallery.debdesk.com/portfolio-4-columns/ -[ ] Imagen referencia https://www.google.com.pe/search?q=photos+for+portfolio&tbm=isch&tbs=rimg:CegA3qLiOKxlIji_1J9DewHwFXJsZG3hpFxZtiM0-HcB69HpVeZuJWs_15Ms25QW3dTiE3H8lp7hhHQN7uDZl3G5dxGCoSCb8n0N7AfAVcEdqqZqSjvEg2KhIJmxkbeGkXFm0Ro7OJnM-pYgYqEgmIzT4dwHr0ehHOxU2gOUmRsCoSCVV5m4laz_1kyEYpWgREOn7PsKhIJzblBbd1OITcRFKETV_1aSKW8qEgkfyWnuGEdA3hHuZ0NkM1SQPCoSCe4NmXcbl3EYEf-E9JyqWQuq&tbo=u&sa=X&ved=0ahUKEwiWisCntafWAhUMEZAKHXpKBnwQ9C8IHA&biw=1366&bih=638&dpr=1
pholser/junit-quickcheck
314727525
Title: Mapped Generation does not seem to do what I think it should do Question: username_0: Here's my property: ``` @Property() public void shouldShrinkTo105( @From(DivisibleBy5.class) // @When(seed = -4386629332000517955L) // int i // ) { System.out.println(i); } public static class DivisibleBy5 extends Generator<Integer> { public DivisibleBy5() { super(Integer.class); } @Override public Integer generate(SourceOfRandomness random, GenerationStatus status) { return gen().type(int.class) // .map(i -> Math.abs(i * 5)) // .filter(i -> i > 100) // .filter(i -> i < 1000) // .generate(random, status); } } ``` What I think it should produce is multiples of 5 between 100 and 1000. However the output looks like the map function is not being used: ``` 447 882 250 842 826 920 357 861 117 135 958 867 638 252 143 ... ``` I'm not sure I'm understanding the purpose of map correctly. Answers: username_1: @username_0 Thanks for this -- will investigate. username_1: @username_0 I think the integer multiplication is overflowing in some cases. For example, using the seed you provided, the first integer to be handed to the `map()` function is `-1233545336`. That multiplied by `5` ordinarily would be `-6167726680`, but that has magnitude too great for a 32-bit `int`. Instead, you get `-1872759384`. Status: Issue closed username_0: Many thanks. Should have been obvious - but wasn't! username_1: @username_0 Side note: TIL you can use (JDK >= 8) [methods on java.lang.Math](https://docs.oracle.com/javase/8/docs/api/java/lang/Math.html#multiplyExact-int-int-) to perform the usual arithmetic operations on `int` and `long` that raise exceptions on overflow.
mdlayher/ndp
796474395
Title: s/Dial/Listen/g? Question: username_0: func Dial(ifi *net.Interface, addr Addr) (*Conn, net.IP, error) {...} Answers: username_1: Probably could change that yeah. But at this point I'm not sure it's necessary. username_0: totally not an urgent thing at all during an internal discussion, it was suggested that language like "Dial" is more for client side convention wise also within `Dial(ifi *net.Interface, addr Addr)` there is not a deeper call to any Dialing function Status: Issue closed username_1: I finally decided to make this change, thanks!
cityenergyproject/seattle
245570359
Title: Add Building Type dropdown to the top of the side bar Question: username_0: ![screen shot 2017-07-25 at 5 06 29 pm](https://user-images.githubusercontent.com/12550451/28598993-0fde26f8-715c-11e7-87bc-2d4aa79e1cb7.png) Selecting a building type from the dropdown menu on the left should redraw the histograms. If a building is added to the building comparison _before_ changing the building type dropdown, it should fade to half opacity and become non-interactive with the exception of the "x" to remove it. If the right building type is selected the building in question should become active again.<issue_closed> Status: Issue closed
microsoft/AttackSurfaceAnalyzer
815595704
Title: Results UNKNOWN Question: username_0: **Describe the bug** I'm attempting to use this for the first time, and I am not getting any results back. After taking a before-and-after snapshot and running an analysis, the UI does not let me select a result type to view, and if I export, it returns: ``` { "results": { "UNKNOWN": 0 }, "metadata": { "compare-version": "2.2.76+68c0dfa092", "compare-os": "WINDOWS", "compare-osversion": "Microsoft Windows NT 10.0.17763.0" } } ``` **To Reproduce** I used the following steps: 1. Download and install ASA from github 2. Run asa.exe gui 3. Run a static scan 4. Add an unsigned EXE to the /Downloads folder 5. Run a static scan 6. Run an analysis on the two results If I export, I get a file with the contents above. The shell also lists this: ``` [07:03:01 INF] Begin ?02?/?24?/?21? ?07?:?02?:?58? ?AM. [07:03:01 INF] Starting 9 Collectors. [07:03:01 INF] Starting FileSystemCollector. [07:03:01 INF] Scanning root C:\ [07:04:54 INF] Completed FileSystemCollector in 00h:01m:52s:817ms. [07:04:54 INF] Starting OpenPortCollector. [07:04:54 INF] Completed OpenPortCollector in 00h:00m:00s:086ms. [07:04:55 INF] Starting ServiceCollector. [07:04:56 INF] Completed ServiceCollector in 00h:00m:01s:458ms. [07:04:56 INF] Starting UserAccountCollector. [07:05:00 INF] Completed UserAccountCollector in 00h:00m:03s:999ms. [07:05:00 INF] Starting RegistryCollector. [07:06:23 INF] Completed RegistryCollector in 00h:01m:23s:342ms. [07:06:24 INF] Starting CertificateCollector. [07:06:24 INF] Completed CertificateCollector in 00h:00m:00s:051ms. [07:06:24 INF] Starting FirewallCollector. [07:06:24 INF] Completed FirewallCollector in 00h:00m:00s:093ms. [07:06:24 INF] Starting ComObjectCollector. [07:06:34 INF] Completed ComObjectCollector in 00h:00m:09s:786ms. [07:06:34 INF] Starting EventLogCollector. [07:06:39 INF] Completed EventLogCollector in 00h:00m:04s:736ms. [07:08:15 INF] Exporting UNKNOWN [07:08:34 INF] Completed Comparing in 00h:00m:25s:753ms. [07:08:34 INF] Loaded Analyses from Embedded. [07:08:34 INF] Completed Analysis in 00h:00m:00s:054ms. [07:08:36 INF] Exporting UNKNOWN [07:08:42 INF] Completed Flushing in 00h:00m:07s:858ms. ``` **Expected behavior** I'd expect some results from my scan **System Configuration (please complete the following information):** - OS: Microsoft Windows Service 2019 Standard. On a VM, if that makes any difference. - OS Version: 10.0.17763 - Application Version: 2.2.76 - CLI or GUI: GUI Answers: username_1: Thanks for the report Nathan. I'll start taking a look today. This is indeed unexpected. username_1: I can't reproduce this using the CLI the version published to nuget. Testing the GUI now. The workaround to do what you were doing in the Cli is 1. `asa collect -a` 2. Run your targeted operation 3. `asa collect -a` 4. `asa export-collect` ``` PS C:\Users\gstocco> asa collect -f --directories Downloads [13:30:41 INF] AttackSurfaceAnalyzer v.2.2.76+68c0dfa092 [13:30:42 WRN] Attack Surface Analyzer should be run as Administrator. Results will not be complete. [13:30:42 INF] Begin 2021-02-24T13:30:42.2567714-08:00. [13:30:42 INF] Starting 1 Collectors. [13:30:42 INF] Starting FileSystemCollector. [13:30:42 INF] Scanning root Downloads [13:31:06 INF] Completed FileSystemCollector in 00h:00m:23s:715ms. PS C:\Users\gstocco> copy .\Documents\GitHub\RecursiveExtractor\RecursiveExtractor\bin\Debug\net5.0\RecursiveExtractor.dll . PS C:\Users\gstocco> move .\RecursiveExtractor.dll .\Downloads\RecursiveExtractor.dll PS C:\Users\gstocco> asa collect -f --directories Downloads [13:32:13 INF] AttackSurfaceAnalyzer v.2.2.76+68c0dfa092 [13:32:14 WRN] Attack Surface Analyzer should be run as Administrator. Results will not be complete. [13:32:14 INF] Begin 2021-02-24T13:32:14.1957657-08:00. [13:32:14 INF] Starting 1 Collectors. [13:32:14 INF] Starting FileSystemCollector. [13:32:14 INF] Scanning root Downloads [13:32:23 INF] Completed FileSystemCollector in 00h:00m:08s:958ms. PS C:\Users\gstocco> asa export-collect [13:32:32 INF] AttackSurfaceAnalyzer v.2.2.76+68c0dfa092 [13:32:32 INF] Provided null run Ids using latest two runs. [13:32:32 INF] Comparing 2021-02-24T13:30:42.2567714-08:00 vs 2021-02-24T13:32:14.1957657-08:00. [13:32:33 INF] Completed Comparing in 00h:00m:00s:388ms. [13:32:33 INF] Loaded Analyses from Embedded. [13:32:33 INF] Completed Analysis in 00h:00m:00s:388ms. [13:32:33 INF] Completed Flushing in 00h:00m:00s:000ms. [13:32:33 INF] Output written to: C:\Users\gstocco\2021-02-24T13_30_42.2567714-08_00_vs_2021-02-24T13_32_14.1957657-08_00_summary.json.txt ``` username_1: Can't reproduce this on the GUI either on Windows 10. This might be a compatibility issue with Server 2019. username_2: Even I'm facing the same issue. In GUI, we are unable to export results for a specific type. ![export not working](https://user-images.githubusercontent.com/55632430/111114773-bba37e80-8589-11eb-84d2-fe5f08a5e0c0.png) username_1: @username_2 are you also on server 2019? username_1: I've tested on server and get results successfully on the command line but something does seem strange in the GUI on server. The GUI was completely rewritten for 2.3 in Blazor so any issues you're seeing in the old GUI wouldn't carry over. 2.3 is currently in Beta. Status: Issue closed
scp-fs2open/fs2open.github.com
788327202
Title: Warn in debug if beams have discarded damage factors Question: username_0: Aside from the wiki and in-game testing, it can be difficult to catch that Armor, Shield, and Subsystem factors have no effect for beams by default. It would be useful for debug builds to display a warning if a beam has a tabled damage factor not equal to 1 while $Beams Use Damage Factors: in game_settings.tbl is not true. However. the retail tables include the targeting laser, an unused fighter beam with 0 damage and non-1 damage factors. All other retail beams have all three damage factors set to 1. Perhaps the warning could be suppressed when using retail data, which could still add a warning to every mod that doesn't disable the targeting laser. Alternatively, the warning could be suppressed for beams with 0 damage, since those damage factors have no effect anyway. Answers: username_1: For whatever it's worth, I think that's the best solution. It's the cleanest check, the easiest to explain (`// Avoid a warning for retail's Targeting Laser` or something to that effect), and _always_ avoids issuing a superfluous warning.
coenjacobs/mozart
753493935
Title: Mozart vs PHP Scoper, and Autoloader behavior question Question: username_0: Howdy! I need to prefix the Composer dependencies of a WordPress plugin, I've been studying the differences between Mozart and PHP Scoper and trying to find out what's best for me. I see that WooCommerce [will be using](https://github.com/woocommerce/woocommerce/pull/28147) Mozart on their next releases, while [Yoast](https://developer.yoast.com/blog/safely-using-php-dependencies-in-the-wordpress-ecosystem/) and [Google Site Kit](https://github.com/google/site-kit-wp/pull/696) are using PHP Scoper. PHP Scoper is a tool focused on Phars, therefore it doesn't seem to be able to handle Composer Autoloader gracefully. **Yoast implementation of PHP-Scoper** Prefixes the `vendor` libraries, remove them from `composer.json` and run `composer dump-autoload`. Then it implements their own `spl_autoload_register` to autoload the prefixed libraries. **Google Site Kit implementation of PHP-Scoper** Seems to drop Composer Autoloader entirely, implementing their own `spl_autoload_register`. I don't like to add `spl_autoload_register` because it adds a file read for every class lookup, and disks are the slowest thing in computing. I understand that Mozart is designed to work alongside Composer, using the Composer Autoloader itself. Is that correct? Would you recommend Mozart or PHP Scoper for prefixing a WordPress plugin? Why? I saw that you (<NAME>) had written an article about this subject on your blog, but the link is dead. Thanks for your time, for this library, and for sharing your knowledge. Answers: username_0: Found the blog article with the dead link: https://username_1.me/2019/07/05/why-i-felt-there-was-a-need-for-mozart/ username_1: Obviously, I prefer Mozart for WordPress projects. I know about the other implementations and in #96 (including some more information, which was once published on my now dead blog) I have explained some of the philosophy and choices behind Mozart. Fun fact, PHP-Scoper didn't even exist back when I started Mozart. That obviously doesn't mean that you should or shouldn't use it, but that's one of the main reasons Mozart now exists. On the topic of autoloading, yes I do believe that Mozart has a unique approach here. Mozart relies on doing as _little as possible_. In practice, this means that Mozart doesn't change the way Composer is autoloading anything and thus the default autoloader, provided by Composer is used in your project, after Mozart is done transforming the files.
gautema/CQRSlite
187489347
Title: Autofac integration Question: username_0: Hi @username_1, Thanks for the wonderful project. I have more of a question than an issue...I'm trying to get the sample project to work with Autofac, I've gone about doing that by converting the .net core DI code to Autofac, here is my startup.cs: public class Startup { public IContainer ApplicationContainer { get; private set; } public IServiceProvider ConfigureServices(IServiceCollection services) { services.AddMemoryCache(); var builder = new ContainerBuilder(); services.AddMvc(); // services.AddSingleton<InProcessBus>(new InProcessBus()); builder.Register<InProcessBus>(x => new InProcessBus()).AsSelf().SingleInstance(); builder.Register<ICommandSender>(c => c.Resolve<InProcessBus>()).SingleInstance(); builder.Register<IEventPublisher>(c => c.Resolve<InProcessBus>()).SingleInstance(); builder.RegisterType<Session>().As<ISession>().InstancePerLifetimeScope(); builder.RegisterType<InMemoryEventStore>().As<IEventStore>().SingleInstance(); builder.RegisterType<MemoryCache>().As<ICache>().InstancePerLifetimeScope(); builder.Register<IRepository>(x => new CacheRepository(new Repository(x.Resolve<IEventStore>()), x.Resolve<IEventStore>(), x.Resolve<ICache>())).InstancePerLifetimeScope(); builder.RegisterType<ReadModelFacade>().As<IReadModelFacade>().InstancePerDependency(); var targetAssembly = typeof(InventoryCommandHandlers).GetTypeInfo().Assembly; builder.RegisterAssemblyTypes(targetAssembly) .Where(type => type.GetInterfaces().Any(y => y.IsAssignableFrom(typeof(CQRSlite.Commands.ICommandHandler<>)))) .AsImplementedInterfaces() .InstancePerDependency(); builder.Populate(services); this.ApplicationContainer = builder.Build(); var sp= new AutofacServiceProvider(ApplicationContainer); var registrar = new BusRegistrar(new DependencyResolver(sp)); registrar.Register(typeof(InventoryCommandHandlers)); return sp; } } However when i step through this, the GetService Method in DependecyResolver.cs can not resolve IHandlerRegistrar. Any guidance would be appreciated. Status: Issue closed Answers: username_0: Duh...I forgot to register IHandlerRegistrar, it was late at night ;-) Here is the autofac equivalent https://github.com/username_0/CQRSlite/commit/7b9802d0a485bb5dc624941ee01bed002eb763da username_1: Great 😊
pulibrary/pulmap
109867499
Title: Expand filter options on search results page Question: username_0: Place filter options/facets below searchbar. Include most common facets (e.g. subject as select) next to search button outside of toggle area. See: http://www.trulia.com/for_rent/Princeton,NJ/map_v Answers: username_1: @username_0 Is there a way to break apart the facets? We want the subject facet to be next to search bar and not in the expanded toggled view. Also, can we change the facets to be select elements instead of lists of links? Status: Issue closed
TCArknight/Savage-Rifts
329843883
Title: Master of Magic Question: username_0: Master of Magic doesn't appear to be seeing Knowledge Arcane. Not a major issue, just validation one. I've tried both the Knowledge Arcane skill and using base Knowledge and putting Arcane in the field. Answers: username_1: Fixed - 07/06/18 Status: Issue closed
danielgindi/Charts
159773492
Title: Failing to build the project Question: username_0: Hello all, I have a problem in building a project using Charts, i tried to add it manually and using coacoapods, both failed. I get 27 errors <img width="257" alt="screen shot 2016-06-11 at 14 24 23" src="https://cloud.githubusercontent.com/assets/18231944/15985310/9e7908ec-2fe0-11e6-8b5a-312cf016d9e5.png"> Status: Issue closed Answers: username_1: update your Xcode seems a solution to me
home-assistant/core
651106095
Title: SpeedTest Integration Failing Question: username_0: <!-- READ THIS FIRST: - If you need additional help with this template, please refer to https://www.home-assistant.io/help/reporting_issues/ - Make sure you are running the latest version of Home Assistant before reporting an issue: https://github.com/home-assistant/core/releases - Do not report issues for integrations if you are using custom components or integrations. - Provide as many details as possible. Paste logs, configuration samples and code into the backticks. DO NOT DELETE ANY TEXT from this template! Otherwise, your issue may be closed without comment. --> ## The problem <!-- Describe the issue you are experiencing here to communicate to the maintainers. Tell us what you were trying to do and what happened. --> The SpeedTest component does not run or provide results. ## Environment <!-- Provide details about the versions you are using, which helps us to reproduce and find the issue quicker. Version information is found in the Home Assistant frontend: Developer tools -> Info. --> Home Assistant Core 0.112.2 - Home Assistant Core release with the issue: 0.112.2 - Last working Home Assistant Core release (if known): - Operating environment (OS/Container/Supervised/Core): FreeNAS (FreeBSD Jail) - Integration causing this issue: SpeedTest - Link to integration documentation on our website: https://www.home-assistant.io/integrations/speedtestdotnet/ ## Problem-relevant `configuration.yaml` <!-- An example configuration that caused the problem for you. Fill this out even if it seems unimportant to you. Please be sure to remove personal information like passwords, private URLs and other credentials. --> Configured via UI, no .yaml involved. ```yaml ``` ## Traceback/Error logs <!-- If you come across any trace or error logs, please provide them. --> No errors occur in the logs, it's as though nothing happens. I've manually called the service, and no logs generate. ```txt ``` ## Additional information This has been an ongoing issue, but completely stopped around version 0.100 (not positive). I had it working previously by installing the ca_root_nss package in the jail. It seems to have a certificate issue, but now that no longer works even though the certificates are installed in the base environment. Answers: username_1: @username_0 Does the issue still exist? The integration works fine for me in 0.115, but I am also running HA Core which might make the important difference here.
jupyterlab/jupyterlab
413823909
Title: Please add Settings for Terminal Font Size and styles Question: username_0: Kindly add settings to change the Terminal font sizing and other settings like line gaps, right now each line on terminal is too close to each other and some text is hidden due to that, also can not change font size separately, whole screen zooming messes up other stuff If you are having issues with installation or configuration, you may ask for help on the [JupyterLab Gitter channel](https://gitter.im/jupyterlab/jupyterlab) or file an issue here. Answers: username_1: Here are the terminal settings currently on the 1.0a1 prerelease, which I think covers everything you mentioned. You can get these by trying out the 1.0a1 prerelease, or you can wait (maybe a few months?) for the 1.0 release. ```javascript { // Terminal // @jupyterlab/terminal-extension:plugin // Terminal settings. // ************************************* // Font family // The font family used to render text. "fontFamily": "monospace", // Font size // The font size used to render text. "fontSize": 13, // Line height // The line height used to render text. "lineHeight": 1, // Scrollback Buffer // The amount of scrollback beyond initial viewport "scrollback": 1000, // Theme // The theme for the terminal. "theme": "dark" } ``` username_1: Closing as resolved. Status: Issue closed
celery/kombu
238332844
Title: SQS Messages not being deleted Question: username_0: I have a custom consumer reading off of an SQS queue. It is written as follows: ```python import logging from django.conf import settings from django.core.management.base import BaseCommand from kombu import Queue from kombu.async import Hub, set_event_loop from kombu.mixins import ConsumerMixin from tasks.celery import app logger = logging.getLogger(__name__) class Worker(ConsumerMixin): def __init__(self, connection, queues): self.connection = connection self.queues = queues super(Worker, self).__init__() logger.info("Started worker %r for queues %r", self, self.queues) def get_consumers(self, Consumer, channel): return [Consumer(queues=self.queues, accept=['pickle', 'json'], callbacks=[self.process_task])] def process_task(self, body, message): logger.info('Processing message: %r', body) try: # Do things pass finally: logger.info("ACKing message %r", message) message.ack() print('ack\'d') class Command(BaseCommand): # This is a Django command help = "Sync task and result messages with database." def add_arguments(self, parser): parser.add_argument('--queue', '-q', default=settings.PLATFORM_QUEUE) def handle(self, queue, *args, **options): set_event_loop(Hub()) with app.connection() as conn: try: logger.info("Launching worker") worker = Worker(conn, queues=[Queue(queue)]) worker.run() except KeyboardInterrupt: print('bye bye') ``` A root logger sends all logs to stdout. Reviewing the stdout, I can see that the requests to delete a message is generated, however it does not appear to ever be sent to AWS: ``` 2017-06-24 10:13:56,611 tasks.consumer: INFO ACKing message <Message object at 0x7fb07fc96168 with details {'properties': {'correlation_id': 'b62c944f-7811-438b-949e-7f9e598a8c44'}, 'body_length': 77, 'content_type': 'application/json', 'delivery_info': {'routing_key': 'export', 'exchange': 'task_exchange'}, 'state': 'RECEIVED', 'delivery_tag': 'AQEBqrM3jZ2n1CUKEmGiXms9Ro3efS+CgZ/KzAC1qRXwWbOiZQTXVXP1eyod6xzitfYE8OrcsmwVnJwfzMNOWsqn09iSIbvfK3WvkX0YN+pH81rSOOvx0RyKGLPwTzardlbqkQJb4LaNj15Q2OeRF9BlpQJ3gpVeO2feW23ZXaJ7+fzmduOXutW44IxFg8Sx4mXBZ0ieR84G01lDp3ReFl9nVpumfPGQvRqDDp+wVe6gN8NIYER3LV5PD8u+eUIbULwhNh6qKmLsxy4F7cxDkap1+6ueAoytE3fkvHD+eUdj7Lg='}> [Truncated] x-amz-date:20170624T101357Z host;x-amz-date e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 2017-06-24 10:13:57,875 botocore.auth: DEBUG StringToSign: AWS4-HMAC-SHA256 20170624T101357Z 20170624/us-west-2/sqs/aws4_request c674918d8890a427b39cf31b211c7e089d1de7a2c077825768f7c4625200aeb1 2017-06-24 10:13:57,877 botocore.auth: DEBUG Signature: 9fc700844cac343ce92bd476249de09c6bef79e94ec5b2b9d880e825701d710c ack'd 2017-06-24 10:13:59,786 botocore.endpoint: DEBUG Making request for OperationModel(name=ReceiveMessage) (verify_ssl=True) with params: {'method': 'POST', 'body': {'Version': '2012-11-05', 'QueueUrl': 'https://us-west-2.queue.amazonaws.com/12345/dev-platform.fifo', 'WaitTimeSeconds': 20, 'Action': 'ReceiveMessage', 'MaxNumberOfMessages': 10}, 'query_string': '', 'headers': {'User-Agent': 'Boto3/1.4.1 Python/3.5.2 Linux/3.13.0-112-generic Botocore/1.4.80'}, 'url_path': '/', 'context': {'has_streaming_input': False, 'client_config': <botocore.config.Config object at 0x7fb082846518>, 'client_region': 'us-west-2'}, 'url': 'https://us-west-2.queue.amazonaws.com/'} ``` Inserting a breakpoint above `message.ack()`, I can see that [`getresponse()` method of the `AsyncHTTPSConnection` class](https://github.com/celery/kombu/blob/09bd23bbd83344b09cbf38b7257107e560db9f25/kombu/async/aws/connection.py#L112-L115) creates a request and adds that request the the PyCurl `CurlClient` instance. As said above, it appears that a web request request is never actually made to AWS. Any ideas as to why this would be? I don't really understand how the `Hub` object works, I only added it to get around the issue described in #746. Perhaps I need to give the hub instance a push to have it process the PyCurl requests? As mentioned in #737, it feels like it would be nice to drop the PyCurl requirement in favor of Boto3 or Requests, however implementing that seems a bit beyond me at the moment. Answers: username_1: @username_0 a few questions here: - Why did you need a custom worker implementation? To integrate it as a Django command with `manage.py` ? - How do you know that the messages are not deleted? Are you seeing the queue size increase in AWS? - On your question about the `Hub` usage, does this also happen if you use the `celery worker` command directly and not your custom implementation? username_0: No, `celery worker` works as expected, this is simply an issue with the custom consumer (which is what leads me to assume that I'm doing something wrong rather than this being an actual bug). username_0: Really, the crux of my question is how async hub is actually _async_. I'm not manually doing anything with multiprocessing, gevent, threading, or anything else so I'm a bit unsure how this could actually be asynchronous and it doesn't appear that creating a `Hub` instance sets any of this up either. This makes me think that I'm missing an implementation detail. username_1: @username_0 have you tried perhaps sub-classing the [WorkController class](https://github.com/celery/celery/blob/master/celery/worker/worker.py#L64) instead of `ConsumerMixin`? If you notice it does use [Hub](https://github.com/celery/celery/blob/master/celery/worker/worker.py#L82) in the list of bootsteps so I am guessing it will behave exactly as the command-line worker invocation. username_0: @username_1 I have not yet tried that. Are you suggesting that I subclass that class, override the [`WorkController` property of the `Celery` class](https://github.com/celery/celery/blob/de2d075fb07850eee247766ebeb11d5530a43734/celery/app/base.py#L1090-L1096), and use that to spin up a `celery worker` process? If so, any pointers on where to actually put the message processing logic? It seems like [`WorkerController. _process_task ()`](https://github.com/celery/celery/blob/de2d075fb07850eee247766ebeb11d5530a43734/celery/worker/worker.py#L223-L231) may be a candidate but that currently only send the work to a pool so there may be a better place. Ultimately, I felt that because I'm changing _how_ these messages should be processed, it felt like I was breaking Celery's design and this was best done with Kombu (which did indeed seem to offer a much simpler implementation). It's just this one small bug/question about how the async `Hub` works that seems to be getting in the way... username_0: I took the simple route and used `boto3` to manually ack the message: ```python class Worker(ConsumerMixin): def __init__(self, connection, queues): self.connection = connection self.queues = queues super(Worker, self).__init__() logger.info("Started worker %r for queues %r", self, self.queues) def get_consumers(self, Consumer, channel): return [Consumer(queues=self.queues, accept=['pickle', 'json'], callbacks=[self.process_task])] def process_task(self, body, message): logger.info('Processing message: %r', message) try: # do stuff pass finally: logger.info("ACKing message %r", message) if self.connection.as_uri().startswith('sqs://'): # HACK: Can't seem to get message.ack() to work for SQS # backend. Without this hack, messages will keep # re-appearing after each See https://github.com/celery/kombu/issues/758 return self._sqs_ack(message) return message.ack() def _sqs_ack(self, message): logger.debug("Manually ACKing SQS message %r", message) client = boto3.client('sqs', 'us-west-2') client.delete_message( QueueUrl=message.delivery_info['sqs_queue'], ReceiptHandle=message.delivery_info['sqs_message']['ReceiptHandle'] ) message._state = 'ACK' message.channel.qos.ack(message.delivery_tag) ``` However, I ultimately think the async curl tooling is problematic. For example, I'm interested in using SQS with the RPC backend. This appears to not work for same reasons mentioned in #746. I'm going to append my bug to that ticket. username_2: This still happening in 2018, this makes the use of Celery SQS broker causing tasks to be executed repeatedly forever, once they reach the visibility timeout. username_3: Is there any progress on this? username_4: No but PRs are welcome. username_5: Is this issue still being actively looked into? username_5: @username_0 Which version of kombu were you using when you created this issue? My team is currently operating a service which is on a rather outdated version, so it could be possible that this issue is affecting us too. username_0: @username_5 I believe it was `4.2.1`. Like I wrote [above](https://github.com/celery/kombu/issues/758#issuecomment-311702699), I used a hack and moved on. username_6: I think this is still happening? I have tasks that get repeatedly executed over and over again and I think it's because of this username_5: Yes, @username_6, I think this issue is still occurring. At my company also, we repeatedly see inexplicable issues with messages being stuck on the celery daemon without being ack'ed to sqs and I think the issue is a result of this.
ialbert/biostar-central
35145522
Title: Mobile site doesn't display votes Question: username_0: The mobile site only displays vote counts on comments, not answers. Answers: username_1: Yes, please fix, this prevents me a lot from voting since I almost exclusively read on my mobile. username_2: @ialbert username_3: +1 It would be nice to see the votes and cast votes on mobile. All that is displayed is the question text with the 'similar posts' section below (at least on Android with Firefox and Chrome, though this is probably not OS-specific). username_4: Yep, that's still true :( ![img_4159](https://cloud.githubusercontent.com/assets/64738/20021042/682d3e9c-a291-11e6-8781-e4dbd3c4e656.PNG) ![img_4160](https://cloud.githubusercontent.com/assets/64738/20021104/a20db376-a291-11e6-9b98-7506d3035e87.PNG) username_5: This issue has been fixed in the most current release of biostars. Status: Issue closed
dart-lang/site-www
447728578
Title: 'Effective Dart' page issue - harder to navigate Question: username_0: Page URL: https://www.dartlang.org/guides/language/effective-dart Page source: https://github.com/dart-lang/site-www/tree/master/src/_guides/language/effective-dart/index.md Found a typo? You can fix it yourself by going to the page source and clicking the pencil icon. Or finish creating this issue. Description of issue: I love a lot of the new design for the site, it looks very snazzy, but I'm having trouble navigating Effective Dart. There are a couple of issues in particular: 1. I'm finding it difficult to navigate between the major areas of Effective Dart. Specifically I'm missing the sub items that used to exist in the left nav panel. Without being too prescriptive, it seems like it would be great to give Effective Dart the same treatment as some other nav subitems like "Development > Command-line & server apps." 2. It's harder to see overview of issues in each major section (e.g. Style). Previously there was a table of contents on the right of the page as there is now, but there was _also_ a table of contents at the top of the page. The top table of contents was IMO much easier to read because of: * Larger indents of rules underneath their subcategories (e.g. Identifiers > DO name types using UpperCamelCase). * Less word wrap because of the greater width of the main section. Because the RHS table of contents is so narrow, a lot more scrolling needs to happen to see all of the rules. Answers: username_1: Thanks for your feedback, @username_0. I'd wondered about these things when we implemented the new site, so it's good to get your take on it. We removed the in-page TOCs (#1602) due to a CSS issue, but we should be able to fix that. username_1: I think that in the sidenav we want to have the following: Effective Dart * Overview (https://dart.dev/guides/language/effective-dart) * Style * Documentation * ... username_2: +1 I like the site, but personally I do find the TOC at the top to be extremely useful when reviewing code to follow the style guide. I hope it can be added back to the top! username_0: That sounds great! Thanks @username_1! Status: Issue closed username_2: Thanks for adding this! I was trying it, and noticed that there's some weird rendering/behavior. The bullet points are not aligned with the text (but interestingly they also work as part of the link). Possibly related to the issue mentioned above? In the attached screenshot, my mouse was where the red dot is at the level of the bullet point that for some reason is right above "Libraries", but (I hope) you can see that the link that it "corresponds" to (a couple lines below) is highlighted as if the mouse was hovering over it. ![effective_dart_toc_issue](https://user-images.githubusercontent.com/1861094/58359291-861f2580-7e37-11e9-9c93-1d82c81efb40.png) Can we re-open this issue? Or should I file a new one? What's the usual procedure? username_1: Page URL: https://www.dartlang.org/guides/language/effective-dart Page source: https://github.com/dart-lang/site-www/tree/master/src/_guides/language/effective-dart/index.md Found a typo? You can fix it yourself by going to the page source and clicking the pencil icon. Or finish creating this issue. Description of issue: I love a lot of the new design for the site, it looks very snazzy, but I'm having trouble navigating Effective Dart. There are a couple of issues in particular: 1. I'm finding it difficult to navigate between the major areas of Effective Dart. Specifically I'm missing the sub items that used to exist in the left nav panel. Without being too prescriptive, it seems like it would be great to give Effective Dart the same treatment as some other nav subitems like "Development > Command-line & server apps." 2. It's harder to see overview of issues in each major section (e.g. Style). Previously there was a table of contents on the right of the page as there is now, but there was _also_ a table of contents at the top of the page. The top table of contents was IMO much easier to read because of: * Larger indents of rules underneath their subcategories (e.g. Identifiers > DO name types using UpperCamelCase). * Less word wrap because of the greater width of the main section. Because the RHS table of contents is so narrow, a lot more scrolling needs to happen to see all of the rules. username_1: I forgot that the formatting issue showed up on stable Chrome (which I don't normally use) but not on Canary. I'll take a look into this and see if it's an easy fix. (Or anyone else is welcome to take a look... CSS is not my strong point.) username_1: I wonder if this is an interaction with the code that makes links not be hidden by the always-visible black bar at the top of the page. I'm going to revert the TOC part of this change, for now, and leave this bug open so we'll remember to take a look later. (Or if someone can figure out a fix that works in both Chrome stable and Chrome canary, you're welcome to take a shot at it!) Btw, the styles look different (not great, but readable) on Safari: ![image](https://user-images.githubusercontent.com/2164483/58359901-94bb0c00-7e3a-11e9-86a4-89e0d38c799b.png) username_1: Removing the default TOC also removed the link anchors from the page. username_1: I'm going to close this issue because always having a top-of-page TOC isn't easy for us to implement, there are workarounds (make your window smaller, use the main page), and I suspect that people are now more used to the new navigation. Feel free to open a new issue if the lack of top-of-page TOC still really bugs you. Status: Issue closed
ray-project/ray
552475438
Title: Does multiagent envs support partial step() returns? Question: username_0: Hello, I am using rllib for its support of multi-agent environments. I am dealing with a use-case where the step() function will be returning transitions for only a random fraction of the agents present in the environment each time. I have the feeling that this might already work seamlessly with the way rllib samples in multi-agent environment, but I am not sure. Could you confirm or infirm, please? Thanks! Yann. Answers: username_1: Yes, this should work -- you can just return a subset of the agent ids in the dict returned from step. For example, {"agent1": obs} instead of {"agent1": obs, "agent2": obs} if only agent1 is ready.
bancodobrasil/stop-analyzing
781476229
Title: I would like to help Question: username_0: Hi, everyone! I would like to contribute to this project! I'm a fullstack developer, also I have some experience contributing to OpenSource, so maybe I can help other users that want to contribute as well. Looking forward to contributing! Thanks! Answers: username_1: Hi @username_0 ! Welcome to stop analyzing 👋🏻 Thanks for oferring some help here. As you have some experience contributing to Open Source, you could take a look and review this PR (https://github.com/bancodobrasil/stop-analyzing-embed/pull/111). What do you think? Have you check the https://github.com/bancodobrasil/stop-analyzing/issues/2 that gives more details on the main goals of the project? username_1: @username_0, also there is this issue you could also take a look at, if you may: https://github.com/bancodobrasil/stop-analyzing-api/issues/27 username_0: Hi @username_1! Yes, I checked #2, very interesting project. I'll take a look at the issue and the PR. Thank you! Status: Issue closed
AntonGepting/tmux-interface-rs
960676622
Title: Incorrect deserialization of session creation time Question: username_0: https://github.com/username_1/tmux-interface-rs/blob/1f8316a6355fca9f84b2c0bf9cd9e904d04ca014/src/variables/session/session.rs#L135 When deserializing session create time via `SESSION_CREATED`, it's treating the value as milliseconds, when in fact the value is in seconds — at least when used with tmux 3.2a. Answers: username_0: The same would be true for `activity` and `last_attached` fields of `Session` (this I tested on my machine) as well as (probably, untested) for `activity` in `Window`. username_1: Hello, thank you for using the library, for reporting this bug and fixing it in your PR. It was changed during refactoring in developing version, you can temporary try to build with `dev` branch to test and see if it is acceptable for you, has no other problems or side effects for your use case: ``` [dependencies] tmux_interface = { git = "https://github.com/username_1/tmux-interface-rs.git", branch = "dev", features = ["tmux_latest"] } ``` In `dev` it is returning back just an `usize` integer number (was not sure what is better just an integer or some `Duration` type of the value). Tomorrow I’ll check it closely and think about either publishing `dev` as the new crate version `0.1.1` or better accept your fix first and publish it as `0.1.1`. username_0: This is an API-breaking change, shouldn't it be `0.2.0`? username_0: Also I just noticed that there's a `NOTE: u64`. It does indeed sound like a good idea to make it a `u64` or even `i64` (because negative timestamps exist and time libraries like chrono would use signed values). Status: Issue closed username_1: https://github.com/username_1/tmux-interface-rs/blob/1f8316a6355fca9f84b2c0bf9cd9e904d04ca014/src/variables/session/session.rs#L135 When deserializing session create time via `SESSION_CREATED`, it's treating the value as milliseconds, when in fact the value is in seconds — at least when used with tmux 3.2a. username_1: Just a note, basically for myself. I’ve looked into tmux sources, it seems like it uses `timeval` structure internally for storing the time information and uses only `time_t tv_sec` part of it (only seconds) for an output as a string (`long long signed integer` -> `i64`): https://github.com/tmux/tmux/blob/bb4bc8caf4a7fa1680333a42679ca72390b60001/format.c#L3102 Sometimes it uses the second (2.) part of `timeval` `long int tv_usec` (microseconds) for logging purposes too, but it is not relevant in this case: https://github.com/tmux/tmux/blob/6c2bf0e22119804022c8038563b9865999f5a026/log.c#L120 For future versions I trend to use basic `i64` type instead of `Duration` at this point: * end user can decide how to work with it, what type or crate to use for it * files i/o maybe will be a bit nicer without converting it back and forth * `Duration` contains seconds `u64` field and nanos field `u32` unused in this case. A bit oversized. At least I think so today. Thank you for your contributions and explanations. I mentioned you in `README.md`, hope you don't mind? username_0: You're welcome! :) username_0: Version 0.2 works for me, so I'm closing the issue. Thanks for prompt reaction! Status: Issue closed
nymag/clay-space-edit
186831286
Title: Components inside of a clay-space do not refresh after saving Question: username_0: For example, updating the package-nav with a new tag on http://qa.www.vulture.com/pages/new.html?edit=true# does not refresh the package-nav with new DOM after save, even though the correct rendered component is returned from the PUT request. Likely something around https://github.com/nymag/clay-space-edit/blob/master/src/controllers/space-controller.js#L29 according to @username_1 Answers: username_1: Resolved by #47, thanks @cperryk! Status: Issue closed
phoenixframework/phoenix_live_view
494587261
Title: redirect/2 is not working from childs Question: username_0: <!-- *Note:* Currently Live View is under active development and we are focused on getting a stable and solid initial version out. For this reason, we will be accepting only bug reports in the issues tracker for now. We will open the issues tracker for features after the current milestone is ironed out. And remember – be nice and have fun! --> ### Environment * Elixir version (elixir -v): 1.9.1 * Phoenix version (mix deps): 1.4.9 * NodeJS version (node -v): v12.4.0 * NPM version (npm -v): 6.11.3 * Operating system: Mac ### Actual behavior At the moment `redirect/2` function is not working from child LiveView component, for example https://github.com/coingaming/bennu/blob/5eff5ce5259b611a1742ccca15b0ad3ed3071f1c/lib/bennu/live_form.ex#L262-L264 If this expression is called - exception is raised. It worked just fine before. ```elixir 14:38:32.572 [error] GenServer #PID<0.964.0> terminating ** (ArgumentError) cannot invoke redirect/2 from a child LiveView (phoenix_live_view) lib/phoenix_live_view.ex:1171: Phoenix.LiveView.assert_root_live_view!/2 (phoenix_live_view) lib/phoenix_live_view.ex:1073: Phoenix.LiveView.redirect/2 (site_builder_flask) lib/bennu/live_form.ex:368: SiteBuilderFlask.Component.BOSiteDetails.DefaultCoreuiLive.handle_event/3 (phoenix_live_view) lib/phoenix_live_view/channel.ex:85: Phoenix.LiveView.Channel.handle_info/2 (stdlib) gen_server.erl:637: :gen_server.try_dispatch/4 (stdlib) gen_server.erl:711: :gen_server.handle_msg/6 (stdlib) proc_lib.erl:249: :proc_lib.init_p_do_apply/3 ``` Maybe this exception make sense for `live_redirect`, but why it is raised with `redirect`? ### Expected behavior `redirect/2` works from a child LiveView like it was before Answers: username_0: With this workaround redirects from child are working again https://github.com/coingaming/bennu/commit/b73eb52779a31abae833a81ecf14bad6bae2361c But I still think what this behaviour is a bug username_1: It was intentional. Send a message to the `root_pid` if you need to redirect. Thanks! Status: Issue closed username_2: @username_1 Where is the `root_pid` documented, can't find it on hexdocs: https://hexdocs.pm/phoenix_live_view/search.html?q=root_pid username_1: It’s on the %Socket{} >
jlippold/tweakCompatible
573686244
Title: `Gesturesi11` working on iOS 13.3 Question: username_0: ``` { "packageId": "com.hius.gesturesi11", "action": "working", "userInfo": { "arch32": false, "packageId": "com.hius.gesturesi11", "deviceId": "iPhone9,3", "url": "http://cydia.saurik.com/package/com.hius.gesturesi11/", "iOSVersion": "13.3", "packageVersionIndexed": true, "packageName": "Gesturesi11", "category": "Tweaks", "repository": "BigBoss", "name": "Gesturesi11", "installed": "1.7", "packageIndexed": true, "packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 100% with 1 working reports.", "id": "com.hius.gesturesi11", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.1.5", "shortDescription": "iPhone X and 11 features for all devices", "latest": "1.7", "author": "Hius", "packageStatus": "Working" }, "base64": "<KEY>", "chosenStatus": "working", "notes": "" } ```<issue_closed> Status: Issue closed
apache/airflow
674310266
Title: pool_slots seems to be ignored Question: username_0: <!-- Welcome to Apache Airflow! For a smooth issue process, try to answer the following questions. Don't worry if they're not all applicable; just try to include what you can :-) If you need to include code snippets or logs, please put them in fenced code blocks. If they're super-long, please use the details tag like <details><summary>super-long log</summary> lots of stuff </details> Please delete these comment blocks before submitting the issue. --> <!-- IMPORTANT!!! PLEASE CHECK "SIMILAR TO X EXISTING ISSUES" OPTION IF VISIBLE NEXT TO "SUBMIT NEW ISSUE" BUTTON!!! PLEASE CHECK IF THIS ISSUE HAS BEEN REPORTED PREVIOUSLY USING SEARCH!!! Please complete the next sections or the issue will be closed. This questions are the first thing we need to know to understand the context. --> **Apache Airflow version**: 1.10.9 (apache-airflow[crypto,celery,postgres,hive,jdbc,mysql,ssh,slack]==1.10.9) **Python version**: 3.7.6 (docker: python:3.7.6-slim-buster) **Kubernetes version (if you are using kubernetes)** (use `kubectl version`): **Environment**: - **Cloud provider or hardware configuration**: Tested locally and on AWS - **OS** (e.g. from /etc/os-release): Debian 10 (buster) and Ubuntu 18.04.4 LTS (Bionic Beaver) - **Kernel** (e.g. `uname -a`): 5.7.8 and 4.15.0 - **Install tools**: - **Others**: **What happened**: pool_slots appears to be fixed at 1. If I set a higher slot count for tasks that are in a 10-slot pool, Airflow still runs 10 of them at once. **What you expected to happen**: Sum(pool_slots) for all tasks currently executing in some pool should not exceed the total number of slots in that pool <!-- What do you think went wrong? --> **How to reproduce it**: 1. Start Airflow 1.10.9 (this may occur on other versions too, but was only tested on 1.10.9). 2. Create pool: ``` airflow pool -s pool_slots_test 10 'pool_slots test pool' [Truncated] If this is a UI bug, please provide a screenshot of the bug or a link to a youtube video of the bug in action You can include images using the .md sytle of ![alt text](http://url/to/img.png) To record a screencast, mac users can use QuickTime and then create an unlisted youtube video with the resulting .mov file. ---> **Anything else we need to know**: <!-- How often does this problem occur? Once? Every time etc? Any relevant logs to include? Put them here in side a detail tag: <details><summary>x.log</summary> lots of stuff </details> --> Answers: username_0: Solved, `pool_slots` arrived in 1.10.10, according to Git. I'd misread the changelog somenow. Status: Issue closed
Fluorescence-Tools/tttrlib
591555326
Title: Travis fails to build the daily build on macOS Question: username_0: Travis on macOS fails to build the daily builds. The daily builds create a distributable app that is embedded in a dmg file. Unmounting the dmg file fails in the build process with the message ``` Fixing permissions... Done fixing permissions. Blessing started Blessing finished Unmounting disk image... hdiutil: detach: timeout for DiskArbitration expired hdiutil: detach: drive not detached fileicon: ERROR: Target not found or neither file nor folder: '/Users/travis/build/Fluorescence-Tools/chisurf/dist/osx/ChiSurf-Installer.dmg' python: can't open file 'upload_sftp': [Errno 2] No such file or directory ``` Others have similar problems (see https://github.com/al45tair/dmgbuild/issues/13)
thellmund/Android-Week-View
486711927
Title: Accessibility Support Question: username_0: Calendar events are not recognized by accessibility focus. Clicking on calendar events doesn't respond while Accessibility is on Answers: username_1: I’ll make sure to add this in an upcoming release. username_1: I started working on accessibility support in the `th/accessibility` branch. You can tap on events and on empty days. In the latter case, TalkBack will simply read the corresponding date. I still need to change it so that it reads the exact time that the user tapped when clicking on an empty date. Besides that, I think it’s pretty much complete. You can try it out if you switch to `implementation 'com.github.username_1:Android-Week-View:th~accessibility-SNAPSHOT'` in your `build.gradle`. If you think there’s anything missing, let me know. username_1: Added in #127. Status: Issue closed
apache/trafficserver
878509963
Title: Intermittent assertion failure `captive_action.cancelled == 0` Question: username_0: On the current master branch (ed89b7b694908b12f21452195e2bbfd4e0fbb156), I observed below assertion failure intermittently. ``` 2021-05-07 15:17:51.924268+0900 traffic_server[20019:1779140] Fatal: HttpCacheSM.cc:161: failed assertion `captive_action.cancelled == 0` Process 20019 stopped * thread #2, name = '[ET_NET 0]', stop reason = signal SIGABRT frame #0: 0x00007fff20394946 libsystem_kernel.dylib`__pthread_kill + 10 libsystem_kernel.dylib`__pthread_kill: -> 0x7fff20394946 <+10>: jae 0x7fff20394950 ; <+20> 0x7fff20394948 <+12>: movq %rax, %rdi 0x7fff2039494b <+15>: jmp 0x7fff2038eb49 ; cerror_nocancel 0x7fff20394950 <+20>: retq Target 0: (traffic_server) stopped. (lldb) bt * thread #2, name = '[ET_NET 0]', stop reason = signal SIGABRT * frame #0: 0x00007fff20394946 libsystem_kernel.dylib`__pthread_kill + 10 frame #1: 0x00007fff203c3615 libsystem_pthread.dylib`pthread_kill + 263 frame #2: 0x00007fff20318411 libsystem_c.dylib`abort + 120 frame #3: 0x00000001012202ae libtscore.10.dylib`ink_abort(message_format="%s:%d: failed assertion `%s`") at ink_error.cc:99:3 frame #4: 0x000000010121c3cc libtscore.10.dylib`::_ink_assert(expression="captive_action.cancelled == 0", file="HttpCacheSM.cc", line=161) at ink_assert.cc:37:3 frame #5: 0x00000001000a69c0 traffic_server`HttpCacheSM::state_cache_open_write(this=0x0000000109001990, event=1108, data=0x000000013004e840) at HttpCacheSM.cc:161:3 frame #6: 0x0000000100005aff traffic_server`Continuation::handleEvent(this=0x0000000109001990, event=1108, data=0x000000013004e840) at I_Continuation.h:219:12 frame #7: 0x0000000100338a33 traffic_server`CacheVC::callcont(this=0x000000013004e840, event=1108) at P_CacheInternal.h:640:25 frame #8: 0x0000000100339633 traffic_server`CacheVC::openWriteStartDone(this=0x000000013004e840, event=3900, e=0x0000000000000000) at CacheWrite.cc:1627:10 frame #9: 0x0000000100005aff traffic_server`Continuation::handleEvent(this=0x000000013004e840, event=3900, data=0x0000000000000000) at I_Continuation.h:219:12 frame #10: 0x000000010033acef traffic_server`Cache::open_write(this=0x0000000101c415e0, cont=0x0000000109001990, key=0x00000001082c5c38, info=0x000000013004ed88, apin_in_cache=0, (null)=0x0000000000000000, type=CACHE_FRAG_TYPE_HTTP, hostname="localhost8001http100kblocalhost:8001http://localhost:8001/100kb", host_len=9) at CacheWrite.cc:1868:10 frame #11: 0x00000001002f9ca9 traffic_server`CacheProcessor::open_write(this=0x000000010069da00, cont=0x0000000109001990, expected_size=0, key=0x00000001082c5c28, request=0x0000000109000370, old_info=0x000000013004ed88, pin_in_cache=0, type=CACHE_FRAG_TYPE_HTTP) at Cache.cc:3254:24 frame #12: 0x00000001000a768f traffic_server`HttpCacheSM::open_write(this=0x0000000109001990, key=0x00000001082c5c28, url=0x0000000109000388, request=0x0000000109000370, old_info=0x000000013004ed88, pin_in_cache=0, retry=true, allow_multiple=false) at HttpCacheSM.cc:363:20 frame #13: 0x000000010010257f traffic_server`HttpSM::do_cache_prepare_action(this=0x0000000108fffbc0, c_sm=0x0000000109001990, object_read_info=0x000000013004ed88, retry=true, allow_multiple=false) at HttpSM.cc:4795:11 frame #14: 0x0000000100109208 traffic_server`HttpSM::do_cache_prepare_write(this=0x0000000108fffbc0) at HttpSM.cc:4724:3 frame #15: 0x000000010010896c traffic_server`HttpSM::set_next_state(this=0x0000000108fffbc0) at HttpSM.cc:7608:5 frame #16: 0x00000001000df33a traffic_server`HttpSM::call_transact_and_set_next_state(this=0x0000000108fffbc0, f=0x0000000000000000)(HttpTransact::State*)) at HttpSM.cc:7339:3 frame #17: 0x00000001000e8094 traffic_server`HttpSM::handle_api_return(this=0x0000000108fffbc0) at HttpSM.cc:1691:5 frame #18: 0x00000001000dad8f traffic_server`HttpSM::do_api_callout(this=0x0000000108fffbc0) at HttpSM.cc:436:5 frame #19: 0x00000001001071b4 traffic_server`HttpSM::set_next_state(this=0x0000000108fffbc0) at HttpSM.cc:7373:5 frame #20: 0x00000001000df33a traffic_server`HttpSM::call_transact_and_set_next_state(this=0x0000000108fffbc0, f=0x0000000000000000)(HttpTransact::State*)) at HttpSM.cc:7339:3 frame #21: 0x00000001000f3aa7 traffic_server`HttpSM::state_hostdb_lookup(this=0x0000000108fffbc0, event=500, data=0x0000000102839f80) at HttpSM.cc:2348:5 frame #22: 0x00000001000da6d8 traffic_server`HttpSM::main_handler(this=0x0000000108fffbc0, event=500, data=0x0000000102839f80) at HttpSM.cc:2701:5 frame #23: 0x0000000100005aff traffic_server`Continuation::handleEvent(this=0x0000000108fffbc0, event=500, data=0x0000000102839f80) at I_Continuation.h:219:12 frame #24: 0x00000001002b39ef traffic_server`reply_to_cont(cont=0x0000000108fffbc0, r=0x0000000102839f80, is_srv=false) at HostDB.cc:503:9 frame #25: 0x00000001002b9e7d traffic_server`HostDBContinuation::dnsEvent(this=0x000000011818f5e0, event=600, e=0x0000000109449a00) at HostDB.cc:1366:11 frame #26: 0x0000000100005aff traffic_server`Continuation::handleEvent(this=0x000000011818f5e0, event=600, data=0x0000000109449a00) at I_Continuation.h:219:12 frame #27: 0x00000001002da3bd traffic_server`DNSEntry::postOneEvent(this=0x0000000120958800, (null)=0, (null)=0x0000000000000000) at DNS.cc:1449:26 frame #28: 0x00000001002da0fc traffic_server`DNSEntry::post(this=0x0000000120958800, h=0x0000000118118000, ent=0x0000000109449a00) at DNS.cc:1435:5 frame #29: 0x00000001002d9d72 traffic_server`DNSEntry::postAllEvent(this=0x0000000120958800, (null)=0, (null)=0x0000000000000000) at DNS.cc:1412:7 frame #30: 0x0000000100005aff traffic_server`Continuation::handleEvent(this=0x0000000120958800, event=0, data=0x0000000000000000) at I_Continuation.h:219:12 frame #31: 0x00000001002d959a traffic_server`dns_result(h=0x0000000118118000, e=0x0000000120958800, ent=0x0000000109449a00, retry=false, tcp_retry=false) at DNS.cc:1379:6 frame #32: 0x00000001002d74d6 traffic_server`dns_process(handler=0x0000000118118000, buf=0x0000000109449a00, len=43) at DNS.cc:1792:7 frame #33: 0x00000001002d434c traffic_server`DNSHandler::recv_dns(this=0x0000000118118000, (null)=0, (null)=0x0000000000000000) at DNS.cc:927:11 frame #34: 0x00000001002d138a traffic_server`DNSHandler::mainEvent(this=0x0000000118118000, event=0, e=0x0000000000000000) at DNS.cc:940:3 frame #35: 0x0000000100005aff traffic_server`Continuation::handleEvent(this=0x0000000118118000, event=0, data=0x0000000000000000) at I_Continuation.h:219:12 frame #36: 0x00000001002df6c4 traffic_server`DNSConnection::trigger(this=0x000000011812d678) at DNSConnection.cc:87:12 frame #37: 0x000000010043da46 traffic_server`NetHandler::waitForActivity(this=0x00000001080bc0f0, timeout=60000000) at UnixNet.cc:534:27 frame #38: 0x0000000100484f01 traffic_server`EThread::execute_regular(this=0x00000001080b8000) at UnixEThread.cc:303:14 frame #39: 0x000000010048537d traffic_server`EThread::execute(this=0x00000001080b8000) at UnixEThread.cc:364:11 frame #40: 0x000000010048353b traffic_server`spawn_thread_internal(a=0x0000000101c3d770) at Thread.cc:92:12 frame #41: 0x00007fff203c3954 libsystem_pthread.dylib`_pthread_start + 224 frame #42: 0x00007fff203bf4a7 libsystem_pthread.dylib`thread_start + 15 ``` Answers: username_0: ``` (lldb) frame select 13 (lldb) p this->history (History<65>) $0 = { history = { [0] = { location = (file = "HttpSM.cc", func = "state_read_client_request_header", line = 697) event = 100 reentrancy = 2 } [1] = { location = (file = "HttpSM.cc", func = "set_next_state", line = 7497) event = 34463 reentrancy = 2 } [2] = { location = (file = "HttpSM.cc", func = "do_cache_lookup_and_read", line = 4699) event = 6656 reentrancy = 2 } [3] = { location = (file = "HttpCacheSM.cc", func = "state_cache_open_read", line = 100) event = 1102 reentrancy = -31073 } [4] = { location = (file = "HttpSM.cc", func = "state_cache_open_read", line = 2612) event = 1102 reentrancy = 1 } [5] = { location = (file = "HttpSM.cc", func = "set_next_state", line = 7457) event = 34463 reentrancy = 1 } [6] = { location = (file = "HttpSM.cc", func = "state_hostdb_lookup", line = 2337) event = 500 reentrancy = 1 } [7] = { location = (file = "HttpSM.cc", func = "set_next_state", line = 7606) event = 34463 reentrancy = 1 } [8] = { location = (file = "HttpCacheSM.cc", func = "state_cache_open_write", line = 160) event = 1108 reentrancy = -31073 } [9] = { location = (file = 0x0000000000000000, func = 0x0000000000000000, line = 0) event = 0 reentrancy = 0 } ... ``` username_1: Probably #7705 username_0: Duplicated. username_2: The commit without this fail is a8b6746d303c4fafdc2e609941750fc4c2ac9b14
backstage/backstage
1056309079
Title: None Question: username_0: @Rugvip , will we want to create a `.snyk` file for `@techdocs/cli`, like was done for `@backstage/cli`? https://github.com/backstage/backstage/blob/master/packages/cli/.snyk#L23-L27 Other reported vulnerabilities affecting `@techdocs/cli`: | Issue | Vulnerability ID | Ignored for @backstage/cli | | ----- | --------------- | --------------------------- | | #8112 | SNYK-JS-BROWSERSLIST-1090194 | True | | #8092 | SNYK-JS-IMMER-1540542 | True | | #7967 | SNYK-JS-SHELLQUOTE-1766506 | False | | #7966 | SNYK-JS-PROMPTS-1729737 | False | Answers: username_1: @username_0 That sounds good to me!
ant-design/ant-design-pro
891751492
Title: ant design pro v4 怎么自定义动态主题🧐[问题] Question: username_0: ### 🧐 问题描述 ant design pro v4 看文档是需要安装umi-plugin-antd-theme,然后config/theme.config.json自动识别,配合SettingDrawer就可以动态切换主题颜色和其他,这个没问题。但是我现在不想用SettingDrawer,想自定义一个按钮,就切换两三个主题色就行了,该怎么实现,官网的文档最后自定义有点看不明白,请问有详细案例吗? ### 💻 示例代码 <!-- 如果你有解决方案,在这里清晰地阐述 --> ### 🚑 其他信息 这两步具体应该怎么配,有点模糊,求大神帮助。 ![image](https://user-images.githubusercontent.com/23291027/118247770-26176300-b4d6-11eb-9de8-d226a22dfd2f.png) ![image](https://user-images.githubusercontent.com/23291027/118247795-2dd70780-b4d6-11eb-85bc-d3042f546229.png) Status: Issue closed Answers: username_1: `(window as any).umi_plugin_ant_themeVar`, 里面有当前的列表和要加载的 css,你可以用下面的方法来加载 css ``` const updateTheme = ( dark: boolean, color?: string, hideMessageLoading = false, publicPath = "/theme" ) => { // ssr if ( typeof window === "undefined" || !(window as any).umi_plugin_ant_themeVar ) { return; } const formatMessage = getFormatMessage(); let hide: any = () => null; if (!hideMessageLoading) { hide = message.loading( formatMessage({ id: "app.setting.loading", defaultMessage: "正在加载主题", }) ); } const href = dark ? `${publicPath}/dark` : `${publicPath}/`; // 如果是 dark,并且是 color=daybreak,无需进行拼接 let colorFileName = dark && color ? `-${encodeURIComponent(color)}` : encodeURIComponent(color || ""); if (color === "daybreak" && dark) { colorFileName = ""; } const dom = document.getElementById("theme-style") as HTMLLinkElement; // 如果这两个都是空 if (!href && !colorFileName) { if (dom) { dom.remove(); localStorage.removeItem("site-theme"); } return; } const url = `${href}${colorFileName || ""}.css`; if (dom) { dom.onload = () => { window.setTimeout(() => { hide(); }); }; dom.href = url; } else { const style = document.createElement("link"); style.type = "text/css"; style.rel = "stylesheet"; style.id = "theme-style"; style.onload = () => { window.setTimeout(() => { hide(); }); }; style.href = url; if (document.body.append) { document.body.append(style); } else { document.body.appendChild(style); } } localStorage.setItem("site-theme", dark ? "dark" : "light"); }; ```
cheahengsoon/Salus
202468460
Title: Fix ArgumentNullException in Convert.ToBase64String Question: username_0: ### Version: 1 | SalusHockeyApp ### ### Stacktrace ### <pre>0 System Convert ToBase64String 1 MyEvents.Views Profile+<Dashboard_Clicked>d__7 MoveNext 2 System.Runtime.ExceptionServices ExceptionDispatchInfo Throw 3 System.Runtime.CompilerServices AsyncMethodBuilderCore <ThrowAsync>m__0 4 Android.App SyncContext+<Post>c__AnonStorey0 <>m__0 5 Java.Lang Thread+RunnableImplementor Run 6 Java.Lang IRunnableInvoker n_Run 7 mono.java.lang RunnableImplementor n_run 8 mono.java.lang RunnableImplementor run 9 android.os Handler handleCallback 10 android.os Handler dispatchMessage 11 android.os Looper loop 12 android.app ActivityThread main 13 java.lang.reflect Method invokeNative 14 java.lang.reflect Method invoke 15 com.android.internal.os ZygoteInit$MethodAndArgsCaller run 16 com.android.internal.os ZygoteInit main 17 dalvik.system NativeStart main Java.Lang.Thread 2 </pre> ### Reason ### android.runtime.JavaProxyThrowable: System.ArgumentNullException: Value cannot be null. ### Link to HockeyApp ### * [https://rink.hockeyapp.net/manage/apps/353113/crash_reasons/154435326](https://rink.hockeyapp.net/manage/apps/353113/crash_reasons/154435326)<issue_closed> Status: Issue closed
aristanetworks/ansible-cvp
899542055
Title: Feature Request - Add support for absent/factory-reset in cv_device_v3 module Question: username_0: ## Request Type - [x] New feature (non-breaking change which adds functionality) - [ ] Breaking change (fix or feature that would cause existing functionality to change) - [ ] Other (please describe): ## Detailed Description In the current version (3.0), we do not support the absent mode for the state field (only present is supported): ``` tasks: - name: "Absent device" arista.cvp.cv_device_v3: devices: "{{devices}}" state: absent ``` Result: ``` TASK [Absent device] *********************************************************************************************************************************************************************************************************************** Monday 24 May 2021 11:34:30 +0100 (0:00:00.033) 0:00:00.033 ************ Monday 24 May 2021 11:34:30 +0100 (0:00:00.032) 0:00:00.032 ************ fatal: [cv_server]: FAILED! => changed=false msg: State==absent is not yet supported ! PLAY RECAP ********************************************************************************************************************************************************************************************************************************* cv_server : ok=0 changed=0 unreachable=0 failed=1 skipped=0 rescued=0 ignored=0 ``` ## Context Should give the ability to remove devices in 2 different ways: * Remove the device from the provisioning page (same action as in the UI: right-click on device > Remove). Trial syntax : ``` - name: "Absent device" arista.cvp.cv_device_v3: devices: "{{devices}}" state: absent ``` * Factory reset the device (in the UI: right-click on device > Factory Reset) Trial syntax: ``` - name: "Factory reset device" arista.cvp.cv_device_v3: devices: "{{devices}}" state: factory-reset ```<issue_closed> Status: Issue closed
ONCdb/ONCdbWeb
267055100
Title: installation instructions Question: username_0: betelgeuse:2017.ONCdb robberto$ python ONCdbWeb/onc_app/app_onc.py Traceback (most recent call last): File "ONCdbWeb/onc_app/app_onc.py", line 6, in <module> from SEDkit import sed ModuleNotFoundError: No module named 'SEDkit' I have found and downloaded SEDkit from GitHub, but I am not sure about where/how install it to make it visible to app_onc.py
ECLK/IncidentManagement
586646618
Title: CSV Export - Garbage characters displayed for 'Sinhala' , 'Tamil' Contents Question: username_0: **Describe the bug** Garbage characters displayed for 'Sinhala' , 'Tamil' Contents **To Reproduce** Steps to reproduce the behavior: 1. login as any user 2. export any CVS report with Sinhalese and Tamil content 3. Now check whether Sinhalese and Tamil content available **Expected behavior** Should display Sinhalese and Tamil Unicode properly **Actual Output** Garbage characters displayed for 'Sinhala' , 'Tamil' Contents **Screenshots** Answers: username_1: this is fixed even without the pdf service. at the moment this pdf genertor is on the same backend side on review page. Status: Issue closed
SynBioHub/synbiohub
625360499
Title: Deleting a public collection does not appear to delete all triples Question: username_0: After deleting a public collection, I noticed that some triples persisted. The existing query may not be able to clear everything. Answers: username_0: So the problem here is with deleteStaggered. If the "sbh:member" triples get deleted, then in the next staggered delete query, the triples for that object will not longer be targeted for deletion. The easiest solution is to stop doing staggered deletions, since I think they are not limited normally anyway by virtuoso (need to check). However, the reason this was introduced was very big deletes took a very long time and sometimes would crash (I think). Another solution would be to make sure sbh:member triples are deleted last. username_0: Decided to remove the staggering for now. Status: Issue closed
jlippold/tweakCompatible
314379355
Title: `CloudToButt` working on iOS 7.1.2 Question: username_0: ``` { "packageId": "com.cpdigitaldarkroom.cloud2butt", "action": "working", "userInfo": { "packageStatus": "Unknown", "packageStatusExplaination": "This tweak has not been reviewed. Please submit a review if you choose to install.", "author": "CP Digital Darkroom", "iOSVersion": "7.1.2", "url": "http://cydia.saurik.com/package/com.cpdigitaldarkroom.cloud2butt/", "latest": "1.0-1", "name": "CloudToButt", "category": "Tweaks", "arch32": true, "packageIndexed": false, "commercial": false, "shortDescription": "Change \"cloud\" to \"butt\" across iOS", "packageInstalled": true, "id": "com.cpdigitaldarkroom.cloud2butt", "packageName": "CloudToButt", "repository": "BigBoss", "deviceId": "iPhone3,3", "packageVersionIndexed": false, "tweakCompatVersion": "0.0.7", "packageId": "com.cpdigitaldarkroom.cloud2butt" }, "base64": "<KEY>", "chosenStatus": "working", "notes": "" } ```
httpwg/http2-spec
1176222794
Title: Title of Appendix B (comment 32) Question: username_0: Title of Appendix B: The first bullet in this section mentions RFC 8740. Should "Changes from RFC 7540" be "Changes from RFC 7540 and RFC 8740"? Also, should the first bullet be moved to the end of the bullet list? Original: Appendix B. Changes from RFC 7540 Answers: username_1: I would prefer not to change anything here. username_2: I agree.
jlippold/tweakCompatible
445184358
Title: `DHelper` working on iOS 12.1.2 Question: username_0: ``` { "packageId": "com.synnyg.dhelper", "action": "working", "userInfo": { "arch32": false, "packageId": "com.synnyg.dhelper", "deviceId": "iPhone9,2", "url": "http://cydia.saurik.com/package/com.synnyg.dhelper/", "iOSVersion": "12.1.2", "packageVersionIndexed": true, "packageName": "DHelper", "category": "Application", "repository": "Packix", "name": "DHelper", "installed": "1.1.2", "packageIndexed": true, "packageStatusExplaination": "This package version has been marked as Working based on feedback from users in the community. The current positive rating is 100% with 1 working reports.", "id": "com.synnyg.dhelper", "commercial": true, "packageInstalled": true, "tweakCompatVersion": "0.1.5", "shortDescription": "A complete download manager", "latest": "1.1.2", "author": "SynnyG", "packageStatus": "Working" }, "base64": "<KEY> "chosenStatus": "working", "notes": "" } ```
zold-io/zold
377050959
Title: NameError: uninitialized constant Zold::Front::Usagewatch on requesting host.ip.address:4096 Question: username_0: NameError: uninitialized constant Zold::Front::Usagewatch /usr/local/lib/ruby/gems/2.4/gems/zold-0.16.9/lib/zold/node/front.rb:193:in `block (2 levels) in <class:Front>' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:39:in `block (2 levels) in get' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:53:in `calc' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:39:in `block in get' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:38:in `synchronize' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:38:in `get' /usr/local/lib/ruby/gems/2.4/gems/zold-0.16.9/lib/zold/node/front.rb:193:in `block in <class:Front>' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1635:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1635:in `block in compile!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:992:in `block (3 levels) in route!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1011:in `route_eval' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:992:in `block (2 levels) in route!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1040:in `block in process_route' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1038:in `catch' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1038:in `process_route' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:990:in `block in route!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:989:in `each' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:989:in `route!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1097:in `block in dispatch!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `block in invoke' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `catch' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `invoke' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1094:in `dispatch!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:924:in `block in call!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `block in invoke' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `catch' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `invoke' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:924:in `call!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:913:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-2.0.5/lib/rack/deflater.rb:34:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/xss_header.rb:18:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/path_traversal.rb:16:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/json_csrf.rb:26:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/base.rb:50:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/base.rb:50:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/frame_options.rb:31:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-2.0.5/lib/rack/null_logger.rb:9:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-2.0.5/lib/rack/head.rb:12:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:194:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1957:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1502:in `block in call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1729:in `synchronize' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1502:in `call' /usr/local/lib/ruby/gems/2.4/gems/thin-1.7.2/lib/thin/connection.rb:86:in `block in pre_process' /usr/local/lib/ruby/gems/2.4/gems/thin-1.7.2/lib/thin/connection.rb:84:in `catch' /usr/local/lib/ruby/gems/2.4/gems/thin-1.7.2/lib/thin/connection.rb:84:in `pre_process' /usr/local/lib/ruby/gems/2.4/gems/thin-1.7.2/lib/thin/connection.rb:50:in `block in process' /usr/local/lib/ruby/gems/2.4/gems/eventmachine-1.2.7/lib/eventmachine.rb:1077:in `block in spawn_threadpool' Please contact me if more informaton is needed. Answers: username_1: @username_2[/z](https://www.username_1.com/u/username_2) please, pay attention to this issue username_1: @username_0[/z](https://www.username_1.com/u/username_0) this project will fix the problem faster if you donate a few dollars to it; just [click here](https://www.username_1.com/contrib/CAZPZR9FS) and pay via Stripe, it's very fast, convenient and appreciated; thanks a lot! username_2: @username_3 release, tag is `0.16.11` username_3: @username_2 OK, I will release it now. Please check the progress [here](http://www.username_3.com/t/16447-435685334) username_3: @username_2 Done! FYI, the full log is [here](http://www.username_3.com/t/16447-435685334) (took me 11min) username_2: @username_0 I believe it's fixed now Status: Issue closed username_0: @username_2 , many thanks for addressing the issue! username_0: @username_2 it looks like the issue is still reproducible. ``` NameError: uninitialized constant Zold::Front::Usagewatch /usr/local/lib/ruby/gems/2.4/gems/zold-0.16.12/lib/zold/node/front.rb:199:in `block (2 levels) in <class:Front>' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:39:in `block (2 levels) in get' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:53:in `calc' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:39:in `block in get' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:38:in `synchronize' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:38:in `get' /usr/local/lib/ruby/gems/2.4/gems/zold-0.16.12/lib/zold/node/front.rb:197:in `block in <class:Front>' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1635:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1635:in `block in compile!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:992:in `block (3 levels) in route!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1011:in `route_eval' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:992:in `block (2 levels) in route!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1040:in `block in process_route' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1038:in `catch' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1038:in `process_route' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:990:in `block in route!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:989:in `each' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:989:in `route!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1097:in `block in dispatch!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `block in invoke' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `catch' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `invoke' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1094:in `dispatch!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:924:in `block in call!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `block in invoke' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `catch' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `invoke' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:924:in `call!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:913:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-2.0.5/lib/rack/deflater.rb:34:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/xss_header.rb:18:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/path_traversal.rb:16:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/json_csrf.rb:26:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/base.rb:50:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/base.rb:50:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/frame_options.rb:31:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-2.0.5/lib/rack/logger.rb:15:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-2.0.5/lib/rack/common_logger.rb:33:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:231:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:224:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-2.0.5/lib/rack/head.rb:12:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:194:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1957:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1502:in `block in call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1729:in `synchronize' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1502:in `call' /usr/local/lib/ruby/gems/2.4/gems/thin-1.7.2/lib/thin/connection.rb:86:in `block in pre_process' /usr/local/lib/ruby/gems/2.4/gems/thin-1.7.2/lib/thin/connection.rb:84:in `catch' /usr/local/lib/ruby/gems/2.4/gems/thin-1.7.2/lib/thin/connection.rb:84:in `pre_process' /usr/local/lib/ruby/gems/2.4/gems/thin-1.7.2/lib/thin/connection.rb:50:in `block in process' /usr/local/lib/ruby/gems/2.4/gems/eventmachine-1.2.7/lib/eventmachine.rb:1077:in `block in spawn_threadpool' ``` Feel free to see this in action on [this node](http://172.16.58.3:4096). username_0: @username_2 , by the way, do I have to create a new ticket or it is eligible to reopen this one? username_2: @username_3 release, tag is `0.16.15` username_3: @username_2 OK, I will release it now. Please check the progress [here](http://www.username_3.com/t/16447-436341145) username_3: @username_0 @username_2 Oops, I failed. You can see the full log [here](http://www.username_3.com/t/16447-436341145) (spent 8min) ``` Stopping ... \u001b[32m.\u001b[0m#<Thread:0x0000000004336630@/home/r/repo/test/node/fake_node.rb:45 run> terminated with exception (report_on_exception is true): \u001b[1mTraceback\u001b[m (most recent call last): 3: from /home/r/repo/test/node/fake_node.rb:46:in `block (3 levels) in run' 2: from /home/r/repo/lib/zold/verbose_thread.rb:38:in `run' 1: from /home/r/repo/test/node/fake_node.rb:49:in `block (4 levels) in run' /home/r/repo/lib/zold/commands/node.rb:212:in `run': \u001b[1mAlias should be a 4 to 16 char long alphanumeric string: invalid-alias (\u001b[1;4mRuntimeError\u001b[m\u001b[1m) \u001b[m\u001b[32m.\u001b[0mThin web server (v1.7.2 codename Bachmanity) Maximum connections set to 1024 Listening on 0.0.0.0:42522, CTRL+C to stop Stopping ... \u001b[32m.\u001b[0mThin web server (v1.7.2 codename Bachmanity) Maximum connections set to 1024 Listening on 0.0.0.0:33477, CTRL+C to stop Stopping ... \u001b[32m.\u001b[0mThin web server (v1.7.2 codename Bachmanity) Maximum connections set to 1024 Listening on 0.0.0.0:35513, CTRL+C to stop Stopping ... \u001b[32m.\u001b[0mThin web server (v1.7.2 codename Bachmanity) Maximum connections set to 1024 Listening on 0.0.0.0:46726, CTRL+C to stop Stopping ... \u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[31mE\u001b[0m \u001b[31mError: TestZold#test_script_redeploy_on_upgrade_sh: Errno::ENOTEMPTY: Directory not empty @ dir_s_rmdir - /tmp/d20181106-22501-yvw5js /usr/local/rvm/rubies/ruby-2.5.1/lib/ruby/2.5.0/fileutils.rb:1337:in `rmdir' /usr/local/rvm/rubies/ruby-2.5.1/lib/ruby/2.5.0/fileutils.rb:1337:in `block in remove_dir1' /usr/local/rvm/rubies/ruby-2.5.1/lib/ruby/2.5.0/fileutils.rb:1348:in `platform_support' /usr/local/rvm/rubies/ruby-2.5.1/lib/ruby/2.5.0/fileutils.rb:1336:in `remove_dir1' /usr/local/rvm/rubies/ruby-2.5.1/lib/ruby/2.5.0/fileutils.rb:1329:in `remove' /usr/local/rvm/rubies/ruby-2.5.1/lib/ruby/2.5.0/fileutils.rb:691:in `block in remove_entry' /usr/local/rvm/rubies/ruby-2.5.1/lib/ruby/2.5.0/fileutils.rb:1386:in `ensure in postorder_traverse' /usr/local/rvm/rubies/ruby-2.5.1/lib/ruby/2.5.0/fileutils.rb:1386:in `postorder_traverse' /usr/local/rvm/rubies/ruby-2.5.1/lib/ruby/2.5.0/fileutils.rb:689:in `remove_entry' /usr/local/rvm/rubies/ruby-2.5.1/lib/ruby/2.5.0/tmpdir.rb:95:in `mktmpdir' /home/r/repo/test/test_zold.rb:41:in `block (2 levels) in <class:TestZold>' \u001b[0m bin/rails test home/r/repo/test/test_zold.rb:38 \u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[33mS\u001b[0m\u001b[32m.\u001b[0m\u001b[33mS\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[33mS\u001b[0m\u001b[33mS\u001b[0m\u001b[33mS\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[33mS\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0mThin web server (v1.7.2 codename Bachmanity) Maximum connections set to 1024 Listening on 0.0.0.0:35386, CTRL+C to stop Stopping ... \u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[33mS\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[33mS\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m\u001b[32m.\u001b[0m Finished in 269.730417s, 1.0010 runs/s, 11.6820 assertions/s. 270 runs, 3151 assertions, 0 failures, 1 errors, 8 skips You have skipped tests. Run with --verbose for details. rake aborted! Command failed with status (1): [ruby -I"lib:lib:test" -I"/home/r/.ruby/gems/rake-12.3.1/lib" "/home/r/.ruby/gems/rake-12.3.1/lib/rake/rake_test_loader.rb" "test/commands/routines/test_reconnect.rb" "test/commands/routines/test_spread.rb" "test/commands/test_alias.rb" "test/commands/test_calculate.rb" "test/commands/test_clean.rb" "test/commands/test_create.rb" "test/commands/test_diff.rb" "test/commands/test_fetch.rb" "test/commands/test_invoice.rb" "test/commands/test_list.rb" "test/commands/test_merge.rb" "test/commands/test_node.rb" "test/commands/test_pay.rb" "test/commands/test_propagate.rb" "test/commands/test_pull.rb" "test/commands/test_push.rb" "test/commands/test_remote.rb" "test/commands/test_remove.rb" "test/commands/test_show.rb" "test/commands/test_taxes.rb" "test/node/test_async_entrance.rb" "test/node/test_emission.rb" "test/node/test_entrance.rb" "test/node/test_farm.rb" "test/node/test_farmers.rb" "test/node/test_front.rb" "test/node/test_nodup_entrance.rb" "test/node/test_safe_entrance.rb" "test/node/test_spread_entrance.rb" "test/node/test_sync_entrance.rb" "test/node/test_trace.rb" "test/test__helper.rb" "test/test_age.rb" "test/test_amount.rb" "test/test_cached_wallets.rb" "test/test_copies.rb" "test/test_dir_items.rb" "test/test_gem.rb" "test/test_hexnum.rb" "test/test_http.rb" "test/test_id.rb" "test/test_key.rb" "test/test_log.rb" "test/test_metronome.rb" "test/test_patch.rb" "test/test_prefixes.rb" "test/test_remotes.rb" "test/test_signature.rb" "test/test_size.rb" "test/test_sync_wallets.rb" "test/test_tax.rb" "test/test_tree_wallets.rb" "test/test_txn.rb" "test/test_upgrades.rb" "test/test_verbose_thread.rb" "test/test_version.rb" "test/test_wallet.rb" "test/test_wallets.rb" "test/test_zold.rb" "test/upgrades/test_protocol_up.rb" ] /home/r/.ruby/gems/rake-12.3.1/exe/rake:27:in `<top (required)>' Tasks: TOP => default => test (See full trace by running task with --trace) container 59928a6c1bafcb0805fe18068b77dad4dea6587846ba900d28dddcd5643a742a is dead Tue Nov 6 18:47:48 CET 2018 ``` username_2: @username_3 release, tag is `0.16.15` username_3: @username_2 OK, I will release it now. Please check the progress [here](http://www.username_3.com/t/16447-436344765) username_2: @username_0 please, try again, with version 0.16.15 username_3: @username_2 Done! FYI, the full log is [here](http://www.username_3.com/t/16447-436344765) (took me 10min) username_0: @username_2 , the node has been updated to 0.16.15, but the issue is still persisting. Do I need to provide more information on that? username_2: @username_0 this is a mystery... let me investigate username_2: @username_0 what is your operating system? username_2: @username_3 release, tag is `0.16.16` username_0: @username_2 my system is FreeBSD: ``` FreeBSD dmz1 11.2-RELEASE-p4 FreeBSD 11.2-RELEASE-p4 #0: Thu Sep 27 07:46:01 UTC 2018 <EMAIL>:/usr/obj/usr/src/sys/GENERIC i386 ``` username_3: @username_2 OK, I will release it now. Please check the progress [here](http://www.username_3.com/t/16447-436374011) username_3: @username_2 Done! FYI, the full log is [here](http://www.username_3.com/t/16447-436374011) (took me 11min) username_2: @username_0 I believe it's fixed in 0.16.16, please try username_2: ``` NameError: uninitialized constant Zold::Front::Usagewatch /usr/local/lib/ruby/gems/2.4/gems/zold-0.16.9/lib/zold/node/front.rb:193:in `block (2 levels) in <class:Front>' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:39:in `block (2 levels) in get' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:53:in `calc' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:39:in `block in get' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:38:in `synchronize' /usr/local/lib/ruby/gems/2.4/gems/zache-0.2.0/lib/zache.rb:38:in `get' /usr/local/lib/ruby/gems/2.4/gems/zold-0.16.9/lib/zold/node/front.rb:193:in `block in <class:Front>' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1635:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1635:in `block in compile!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:992:in `block (3 levels) in route!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1011:in `route_eval' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:992:in `block (2 levels) in route!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1040:in `block in process_route' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1038:in `catch' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1038:in `process_route' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:990:in `block in route!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:989:in `each' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:989:in `route!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1097:in `block in dispatch!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `block in invoke' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `catch' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `invoke' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1094:in `dispatch!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:924:in `block in call!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `block in invoke' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `catch' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1076:in `invoke' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:924:in `call!' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:913:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-2.0.5/lib/rack/deflater.rb:34:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/xss_header.rb:18:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/path_traversal.rb:16:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/json_csrf.rb:26:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/base.rb:50:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/base.rb:50:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-protection-2.0.4/lib/rack/protection/frame_options.rb:31:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-2.0.5/lib/rack/null_logger.rb:9:in `call' /usr/local/lib/ruby/gems/2.4/gems/rack-2.0.5/lib/rack/head.rb:12:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:194:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1957:in `call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1502:in `block in call' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1729:in `synchronize' /usr/local/lib/ruby/gems/2.4/gems/sinatra-2.0.4/lib/sinatra/base.rb:1502:in `call' /usr/local/lib/ruby/gems/2.4/gems/thin-1.7.2/lib/thin/connection.rb:86:in `block in pre_process' /usr/local/lib/ruby/gems/2.4/gems/thin-1.7.2/lib/thin/connection.rb:84:in `catch' /usr/local/lib/ruby/gems/2.4/gems/thin-1.7.2/lib/thin/connection.rb:84:in `pre_process' /usr/local/lib/ruby/gems/2.4/gems/thin-1.7.2/lib/thin/connection.rb:50:in `block in process' /usr/local/lib/ruby/gems/2.4/gems/eventmachine-1.2.7/lib/eventmachine.rb:1077:in `block in spawn_threadpool' ``` Please contact me if more informaton is needed. username_1: @username_2[/z](https://www.username_1.com/u/username_2) please, pay attention to this issue username_1: @username_0[/z](https://www.username_1.com/u/username_0) this project will fix the problem faster if you donate a few dollars to it; just [click here](https://www.username_1.com/contrib/CAZPZR9FS) and pay via Stripe, it's very fast, convenient and appreciated; thanks a lot! username_0: @username_2 , it is working so far. Will see how it goes. Status: Issue closed username_2: @username_0 it's stable, won't appear again, thanks for reporting! username_1: Job `gh:zold-io/zold#523` is not assigned, can't get performer <!-- https://www.username_1.com/footprint/CAZPZR9FS/6949e34d-36fd-4325-ac32-d113fa277657, version: 0.37.5, hash: ${buildNumber} --> username_1: This job is not in scope <!-- https://www.username_1.com/footprint/CAZPZR9FS/6e304f3b-7dd0-4c8e-abf2-011b9de55739, version: 0.37.5, hash: ${buildNumber} -->
garronej/denoify
673073097
Title: deno package registry api Question: username_0: You can also use the page parameter to specify the page, default is 1 I hope this information is helpful 🙂 Answers: username_1: It helps a great deal, thank you! username_2: Are there docs open source docs for this? username_1: Hi @username_2! Not that I know off beside: https://github.com/denoland/deno_registry2/blob/main/API.md username_1: @username_2 You can also check [this](https://github.com/denoland/deno_registry2/issues/40#issuecomment-668774286) out, very useful. Status: Issue closed
CMPUT301W15T13/TravelPlanner
57558802
Title: 08.07.01 Question: username_0: As an approver, I want to return a submitted expense claim that was not approved, denoting the claim status as returned and setting my name as the approver for the expense claim. Answers: username_0: As an approver, I want to return a submitted expense claim that was not approved, denoting the claim status as returned and setting my name as the approver for the expense claim. username_0: Previously done Status: Issue closed username_1: As an approver, I want to return a submitted expense claim that was not approved, denoting the claim status as returned and setting my name as the approver for the expense claim. Status: Issue closed
LonamiWebs/Telethon
959219495
Title: Answer on non owner request Question: username_0: In bots i have problem - lib says, than "this not active command" etc. Plz fix it, is very old problem Answers: username_1: Please reopen once you're able to provide more information (describe the situation, problem, what is happening, and what should happen, in more detail). Status: Issue closed username_0: if you send from non owner telethon says "Incoming massage!" He send msg to queue... and skip it.
gzaspa/make-object-alive-mobile
568808339
Title: Зробити навчальну програму, яка детектить та розпізнає обличчя на android Question: username_0: Зробити програму на основі статті: https://medium.com/devnibbles/facial-recognition-with-android-1-4-5e043c264edc В цій статті розглядається такі питання: 1. Facial Detection on Android 2. Facial Recognition using Google AutoML (off-device) 3. Facial Recognition using TensorFlow (off-device) 4. Facial Recognition using TensorFlow Lite (on-device)
xBelladonna/LemmeSmash
476672309
Title: Handle conflicting proxy tags for keysmashes and owospeak Question: username_0: As of right now there is no check to ensure that proxy tags for both functions are unique. This leads to problems and general unpredictability when either the prefix and/or suffix for keysmash tags is the same as either the prefix and/or suffix for the owospeak tags. A hasty fix was pushed but this didn't account for null tags and as such will not let a user perform initial registration of any tags. The fix has since been reverted, and whatever new fix is implemented needs to allow users to register initially, account for null tags and also intelligently check to ensure the same sets of symbols are not reused anywhere. Answers: username_0: Fixed in #cc53de9 Status: Issue closed
kreait/firebase-php
175124159
Title: how to authenticate user using username and password Question: username_0: Hello i am new in using firebase and interested in using this library i try to create user using username and password but find no thing about this in your document i currently using google api to signup new user and login in this reference Registration Method: POST `URL: https://www.googleapis.com/identitytoolkit/v3/relyingparty/signupNewUser?key=<my-firebase-api-key>` `Login Method: POST URL: https://www.googleapis.com/identitytoolkit/v3/relyingparty/verifyPassword?key=<my-firebase-api-key>` but can not work with rules to check if authenticated ``` "users" : { "$uid":{ ".read":"auth.uid != null", ".write":"auth.uid != null", } } ``` please help to get the best way to do auth Regards Answers: username_1: It's not possible to authenticate a user by username and password through a server application - but it is also not needed: because you are already authenticated, you can impersonate any user with an Authentication override: http://firebase-php.readthedocs.io/en/1.x/authentication.html username_0: i updated issue i mean using email and password i see at firebase doc something like CreatUser(email,password) using node is there is something like this here ? username_1: Not at the moment - but could you provide me the links to the documentation in question? Also to the documentation where you found the URLs to create the users via the identity toolkit? I will then see if I can adapt this in the PHP SDK. Thank you! username_0: of course this is API reference [https://developers.google.com/identity/toolkit/web/reference/](url) and here where in find full answer about auth using rest api [http://stackoverflow.com/questions/37322747/using-mail-and-password-to-authenticate-via-the-rest-api-firebase](url) the point now how to verify if user authenticated using rules `"users" : { "$uid":{ ".read":"auth.uid != null", ".write":"auth.uid != null", } } ` username_0: and that is the document for creating user using email and password with there build in function in nodejs (JavaScript) https://firebase.google.com/docs/auth/web/password-auth ``` firebase.auth().createUserWithEmailAndPassword(email, password).catch(function(error) { // Handle Errors here. var errorCode = error.code; var errorMessage = error.message; // ... }); ``` username_1: 3rd party SDKs can only make use of the REST API, so I am not sure I will be able to replicate this functionality, but I will look what I can do. username_0: I build an api using laravel that handle login using email and password and signup and getUserInfo code : ``` public function Login(Request $request) { // FIREBASE CONFIGRATION $config = new Configuration(); $config->setAuthConfigFile(base_path() . 'GOOGLE-SERVICE-FILE.json'); $firebase = new Firebase('https://YOUR-PROJECT.firebaseio.com', $config); // USER INPUT $email = $request->input('email'); $password = $request->input('<PASSWORD>'); $client = new GuzzleHttp\Client(); // Create a POST request using google api $key = 'YOUR KEY'; $responsee = $client->request( 'POST', 'https://www.googleapis.com/identitytoolkit/v3/relyingparty/verifyPassword?key=' . $key, [ 'headers' => [ 'content-type' => 'application/json', 'Accept' => 'application/json' ], 'body' => json_encode([ 'email' => $email, 'password' => $password, 'returnSecureToken' => true ]), 'exceptions' => false ] ); $body = $responsee->getBody(); $js = json_decode($body); if (isset($js->error)) { return response()->json([ 'success' => false, 'message' => $js->error->message ]); } else { return response()->json([ 'success' => true, 'localId' => $js->localId, 'idToken' => $js->idToken, 'email' => $js->email, 'refreshToken' => $js->refreshToken, 'expiresIn' => $js->expiresIn, ]); } } ``` [Truncated] ]); } else { return response()->json([ 'success' => true, 'localId' => $js->users[0]->localId, 'email' => $js->users[0]->email, ]); } } ``` I hope this will be helpful Regards username_1: Progress will be logged at #51 Status: Issue closed
falconry/falcon
493096353
Title: app hanging on request.stream.read() Question: username_0: I must be doing something wrong but can't for the life of me figure out what it is. I've created this simple test program that (at least for me) shows the behavior I'm seeing: ``` import falcon from wsgiref import simple_server app = falcon.API() class res(object): def on_get(self, req, resp): body = req.stream.read() resp.status = falcon.HTTP_200 def on_post(self, req, resp): body = req.stream.read() resp.status = falcon.HTTP_200 app.add_route('/', res()) server = simple_server.make_server('127.0.0.1', 9000, app) server.serve_forever() ``` Making GET or POST requests don't return until the client hangs up (ctrl+c on curl, X on browser). Debugging the application shows `req.stream.closed = False`, not sure if that means anything. Please let me know if I can provide any more info. Answers: username_0: Thanks @vytas7! Both for the quick reply and the solution. `bounded_stream` worked perfectly. Status: Issue closed
Puzzlepart/prosjektportalen365
1122257362
Title: Weird project template selector alignment/UX Question: username_0: **Describe the bug** The template selector doesn't look too good. **To Reproduce** Steps to reproduce the behavior: 1. Go to a new project 2. Click on project template selector **Expected behavior** The selector should look aligned and not ugly **Screenshots** ![image](https://user-images.githubusercontent.com/1837390/152215237-b5489928-bda7-437b-a57a-c43c6dcca8f5.png) Answers: username_0: Resolved ![image](https://user-images.githubusercontent.com/1837390/152218067-2fc9088f-573c-4d99-96ae-86d43ea3e5e9.png) Status: Issue closed
jorymorrison/MLD
510988186
Title: Incorrect slash when saving results (linux) Question: username_0: When saving the results of the program, the program uses a forward slash after attempting to save to the MLD folder resulting in the saving of the file in the home directory as MLD\{tittle}. Answers: username_1: The desired behavior, in my mind, would be to have the results saved to a "results" directory located inside the directory of execution. username_1: Looks like you addressed this in 1d083c7. Will Close pending PR. Status: Issue closed username_1: Let's leave open until the PR goes through to have better trace-ability. username_1: When saving the results of the program, the program uses a forward slash after attempting to save to the MLD folder resulting in the saving of the file in the home directory as MLD\{tittle}. username_1: Fixed in #8 Status: Issue closed
nwjs/nw.js
187881527
Title: How do I distribute my app on linux? Question: username_0: I'm using 0.14.7, and ubuntu 32bit. I recently updated from 0.12.2, and now the build process seems to be different or doesn't work. I've added my files to package.nw and then appended to nw via: cat nw app.nw > app && chmod +x app Then in the docs it says "On Linux, you need to create proper .desktop file." But there aren't any directions specific for this case, If I try running nw, it says "There is no application installe for "shared library" files. Do you want to search for an application to open this file?" So I'm assuming the .desktop file would fix this? It says you can create it via a text editor, but I have no idea what I should put inside it. Can't there be some more explanation for this? in 0.12.2 the .desktop step wasn't necessary. Answers: username_1: Please ask questions in our mailing list: https://groups.google.com/forum/#!forum/nwjs-general Status: Issue closed
facebookresearch/Detectron
347314776
Title: Can Detectron model be converted to Caffe model? Question: username_0: I'm trying to convert Keypoint-R-CNN example model based on Detectron/Caffe2 to Caffe. By default, Detectron divides the model into "base" part and "head" part, but I made a new Caffe prototxt including both keypoint net and net(base net) after extracting detectron_net.pbtxt and detectron_keypoint_net_pbtxt together and manually changed each names of items. Also, I found it's possible to convert all of Caffe2's weight/bias to numpy arrays and then those to Caffe using FetchBlob("weight name"). But now, I feel confused that it seems like, there would be no matching CUDA/cuDNN implementation in Caffe2 for Caffe's "Deconvolution(gpu)" code( Does "ConvTranspose" work same as "Deconvolution"?) Plus, some of Caffe's hyper parameters in prototxt are gone in Caffe2's pbtxt such as " num_output" or "lr_mult" in convolution_param. Not only that, Detectron's convert_pkl_to_pb.py converts the example weight file(.pkl, arond 480MB) to so small 2 .pb files(less than 200KB in total) with fusing AffineTransform option. If I apply this fusing option, it merge some blobs, so the original network structure is changed and reduced. As there is no AffineTransform layer in Caffe, I need to turn on this option. So, I would like someone to tell me that Caffe2 can be translated to Caffe or not. Is there anybody who succeed in converting Caffe2 model to Caffe? I'm also considering ONNX or MMdnn. Thank you in advance! ### System information * Operating system: ubuntu 16.04 * Compiler version: gcc 5.4.0 * CUDA version: 9.2 * cuDNN version: 7.1.4 * NVIDIA driver version: 396.45 * GPU models (for all devices if they are not all the same): YES * `PYTHONPATH` environment variable: YES * `python --version` output: python3.5 Answers: username_0: Yes. ONNX is such a awesome tool, but when I tried to convert a tensorflow model to caffe before, it run slower almost 10 times(both running based on cuDNN layers). So, conceptually, it's possible to convert Caffe2 to Caffe, however, actually there are too many general interfaces inserted. That's why I want to use numpy to directly convert as "convert_pkl_to_pb.py" does. **3. Hyper paramerters** **num_output**: After loading the model file and I checked the each blob size of total net(keypoint_net, body net) and filled each num_output value in the .prototxt one by one manually. **lr_mult, decay_mult**: As I just want to run inference step, I don't need those. **4. Newly added layers** IMO, for running Detectron on Caffe, the newly added operators also need to be ported. ResizeNearst(upsample_nearest_op.cu) GenerateProposals(generate_proposals_op.cc) CollectAndDistributeFpnRpnProposals(collect_and_distribute_fpn_rpn_proposals_op.cc) BatchPermutation(batch_permutation_op.cu) BBoxTransform(bbox_transform_op.cc) BoxWithNMSLimit(box_with_nms_limit_op.cc) ConvTranspose(conv_transpose_op_cudnn.cc) -> thanksfully, someone did it (https://github.com/BVLC/caffe/pull/5924). username_1: I just tried it and it worked for me. I used a fresh install from current version of the repository + #449 (+ #110 Python3 compatibility fix). The output looks like this for me ``` 89 214 992 detectron_keypoint_net_init.pb 3 829 detectron_keypoint_net.pb 9 298 detectron_keypoint_net.pbtxt 430 333 detectron_keypoint_net.png 206 638 747 detectron_net_init.pb 26 954 detectron_net.pb 62 998 detectron_net.pbtxt 3 082 557 detectron_net.png ``` username_0: ok ;) thank you for your help!! I will try Status: Issue closed username_2: @username_0 Have to converted Caffe 2 model to Caffe model? I am working on Faster RCNN and want to convert it to caffe model? any suggestion? Thanks!
babel/babel
226603665
Title: transformFromAst with a Node instead of the AST root Question: username_0: ### Input Code I'm the developer of a mutation testing framework for JavaScript which involves parsing, modifying and generating code. In order to stay up to date with the latest JavaScript features, I've decided to switch to Babel. Unfortunately, I can't find any API that allows me to generate code based on a Node instead of the entire AST. I created the code below to change `1 + 2` into `1 - 2`: ```TypeScript import * as _ from 'lodash'; import * as babel from 'babel-core'; class Mutator { mutate() { const code = `1 / 2; 1 + 2;`; let input = babel.transform(code); console.log('input:', input.code); const ast = input.ast as babel.types.File; let mutatedAst = _.cloneDeep(ast) as babel.types.File; //Create a backup that we can use to generate code if (ast) { let statement = ast.program.body[1]; if (babel.types.isExpressionStatement(statement)) { let expression = statement.expression; if (babel.types.isBinaryExpression(expression)) { let mutatedNode = _.cloneDeep(statement) as babel.types.ExpressionStatement; // Clone the Node so we don't modify the original object ((mutatedNode.expression) as babel.types.BinaryExpression).operator = '-'; // Modify the operator mutatedAst.program.body = [mutatedNode]; // Create a Program with only our statement let output = babel.transformFromAst(mutatedAst); // Generate the code! console.log('output:', output.code); } } } } } new Mutator().mutate(); ``` Would it be possible to do this without emptying the body of the `program`? If would be great if I could call something like `babel.transformFromAst(mutatedNode)`. Answers: username_1: You could create a new `File` node with your single node in it, if I'm understanding you properly, e.g. ``` var t = babel.types; var ast = t.file(t.program([mutatedAst)); ``` That said, I don't quite understand what you're trying to do. For instance you're usage of `transform` and `transformFromAst` are a bit strange. I'd expect something more like ``` const code = `1 / 2; 1 + 2;`; const output = babel.transform(code, { plugins: [ function() { return { visitor: { BinaryExpression(path) { if (path.node.operator === "+") path.node.operator = "-"; } }, } }, ], }); console.log('input:', input.code); console.log('output:', output.code); ``` This may be an easier discussion to have on out Slack if you're interested. username_1: Gonna close this since there aren't any clear next steps. Feel free to hop on slack if you have more questions. Status: Issue closed
anzwdev/al-code-outline
618815725
Title: New Table Wizard: missing length Question: username_0: Witam Andrzej, when I enter this into the New Table Wizard: ![image](https://user-images.githubusercontent.com/34504100/82032181-2995c580-969b-11ea-94d8-a7c8ac379360.png) then I get this: ![image](https://user-images.githubusercontent.com/34504100/82032308-4d590b80-969b-11ea-985c-43fa91ff72fc.png) Answers: username_1: Czesc :) Thank you for letting me know about it. Looks like a bug, I'll fix it as soon as possible. username_1: I've fixed the issue and uploaded new version of the extension to the Visual Studio Marketplace. username_0: Confirmend; thank you :-) Status: Issue closed
sminez/penrose
695415503
Title: focus bouncing when playiing video in monocle mode Question: username_0: **Describe the bug** A clear and concise description of what the bug is. when playing a video using mpv in monocle mode the focus will bounce between the video and the other window. This may be related to #52 If the mouse pointer is moved to the status bar this stops **To Reproduce** Steps to reproduce the behavior: switch to monocle layout open a terminal run mpv -vo xv video,mp4 switch to another workspace and back **Desktop (please complete the following information):** - OS: linux - Distribution debian - Version sid Answers: username_1: @username_0 as requested in #47 when you raised an issue previously, please can you fill in all of the bug template including providing a minimal main.rs that reproduces the bug. The [minimal](https://github.com/username_1/penrose/blob/develop/examples/minimal/main.rs) example in the `examples` directory is a good starting point for this. In particular: I need to know if this is related to `mpv` specifically (I assume this issue does not happen for other programs given that you are mentioning it but I have no way of knowing). I also need to know if this only happens on the `monocle` layout, and if so how you have configured that layout. Does this only happen when you toggle between workspaces? As stated in the README of this repo: this is a hobby project. I don't have time or capacity to try and work this sort of thing out from the the level of detail you are providing when raising bugs. Sorry. username_0: Hi I have so far worked out that this is an interaction of wm.set_client_insert_point(InsertPoint::AfterFocused); which I have after let mut wm = WindowManager::init(config, &conn); I am not yet sure what it is that it is interacting with but I am working on tracking this down but I have run out of time at the moment. I have seen the effect in a number of thing not just mpv I just listed mpv as it was an easy example. username_1: @username_0 I am closing this for now as the details provided so far indicate that this is a configuration issue specific to your use of `penrose` as a crate. If you manage to narrow things down and can provide full details in the future, please feel free to open a new issue using the issue template. Status: Issue closed username_0: I am currently thinking that it may be an issue with the focus variable that is cached in manager getting out if sync with the value at the workspace level but I have not found evidence of this yet. This could explain why it shows up with monocle and paper and not the other layout which ignore focus I will send through more info when I have had timn to look further at this .
dansanti/l10n_cl_stock_picking
194405591
Title: Variable price no definida Question: username_0: Al crear un picking y validarlo, sale que la variable price no esta definida, esto en el archivo [stock_picking.py](https://github.com/username_1/l10n_cl_stock_picking/blob/9.0/models/stock_picking.py#L402) Que variable deberia ir ahi???. **rec.price_uni**t talvez Answers: username_1: hay algunos cambios que tengo que subir Status: Issue closed
AdamZink/frequency-finder
548483143
Title: Only mutate the child candidates Question: username_0: When creating each child, immediately recalculate the wav_signal, bucketed_signal, and fitness scores. This will allow parents to stay intact like a more traditional genetic algorithm approach. Also, assess if performance is better (expecting more generations due to less mutations, but could be faster because of less total recalculations)<issue_closed> Status: Issue closed
Landry333/Big-Owl
742010889
Title: REPO-159: Add templates for issue creations Question: username_0: When creating an issue, there should be templates for different types of issue: - User Story - Acceptance Test - Bug Report - Testing - Epic - Pull request - Refactoring - Technical Story - Documentation That way, issues will be written consistently.<issue_closed> Status: Issue closed
pnbruckner/homeassistant-config
372578952
Title: Composite tracker: Handle watched entities that can temporarily have an invalid source_type Question: username_0: At startup some types of device_trackers seem to be entered into the State Machine with an invalid source_type attribute (specifically, one that is set to None.) This causes the corresponding tracker to be removed from the state change watch list. Ignore these invalid states so future updates will still be used.<issue_closed> Status: Issue closed
ProjectSidewalk/SidewalkWebpage
386515120
Title: Landing Page choropleth color really dark Question: username_0: The new version on sidewalk-test.cs shows the following choropleth: ![image](https://user-images.githubusercontent.com/2873216/49334488-22a0f300-f58c-11e8-8494-f1166fd6629c.png) This is the original on sidewalk.cs: ![image](https://user-images.githubusercontent.com/2873216/49334487-1b79e500-f58c-11e8-9043-5211bfea6d9b.png) The color scale increased from 10 to 11 colors. Was this intentional or a side-effect of some change? Answers: username_1: Purposeful to highlight difference between completely done and 99% done—a huge distinction when trying to get people to select incomplete neighborhoods. Sent from my iPhone > username_0: I see, gotcha! Status: Issue closed