repo_name
stringlengths
4
136
issue_id
stringlengths
5
10
text
stringlengths
37
4.84M
ptitSeb/box86
728342702
Title: Unimplemented Opcode Question: username_0: 18626|0xae06130: Unimplemented Opcode 66 0F 3A 21 C8 10 66 0F Illegal Opcode FF FF FF FF FF CE 18626|SIGSEGV @0xf7613b64 (???) (x86pc=0x84d4e60/./DevilDaybi_i386:"???", esp=0xf626ead8), for accessing 0x5c7393d8 (code=1), db=(nil)((nil):(nil)/(nil):(nil)/???) Full log. Answers: username_1: This is an SSE4 opcode. Box86 is supposed to handle up to SSSE3 only. What game/app is that? username_0: Thanks. Its a game a few friends and I wrote and we are trying to make it work on a Raspberry PI 4. username_1: if you can compile with `-march=core2` it should works with box86. username_0: Thank you for your help. I have tried compiling with the following line - `cmake .. -DARM_DYNAREC=1 -DRPI4=1 -DCMAKE_BUILD_TYPE=RelWithDebInfo -DCMAKE_C_FLAGS:STRING=-march=core2;` And I got this error: ``` -- The C compiler identification is GNU 8.3.0 -- The CXX compiler identification is GNU 8.3.0 -- Check for working C compiler: /usr/bin/cc -- Check for working C compiler: /usr/bin/cc -- broken CMake Error at /usr/share/cmake-3.13/Modules/CMakeTestCCompiler.cmake:52 (message): The C compiler "/usr/bin/cc" is not able to compile a simple test program. It fails with the following output: Change Dir: /home/ludi/sources/box86/build/CMakeFiles/CMakeTmp Run Build Command:"/usr/bin/make" "cmTC_2a545/fast" /usr/bin/make -f CMakeFiles/cmTC_2a545.dir/build.make CMakeFiles/cmTC_2a545.dir/build make[1]: Entering directory '/home/ludi/sources/box86/build/CMakeFiles/CMakeTmp' Building C object CMakeFiles/cmTC_2a545.dir/testCCompiler.c.o /usr/bin/cc -march=core2 -o CMakeFiles/cmTC_2a545.dir/testCCompiler.c.o -c /home/ludi/sources/box86/build/CMakeFiles/CMakeTmp/testCCompiler.c cc: error: unrecognized -march target: core2 cc: note: valid arguments are: armv2 armv2a armv3 armv3m armv4 armv4t armv5 armv5t armv5e armv5te armv5tej armv6 armv6j armv6k armv6z armv6kz armv6zk armv6t2 armv6-m armv6s-m armv7 armv7-a armv7ve armv7-r armv7-m armv7e-m armv8-a armv8.1-a armv8.2-a armv8.3-a armv8.4-a armv8-m.base armv8-m.main armv8-r iwmmxt iwmmxt2 native cc: error: missing argument to ‘-march=’ make[1]: *** [CMakeFiles/cmTC_2a545.dir/build.make:66: CMakeFiles/cmTC_2a545.dir/testCCompiler.c.o] Error 1 make[1]: Leaving directory '/home/ludi/sources/box86/build/CMakeFiles/CMakeTmp' make: *** [Makefile:121: cmTC_2a545/fast] Error 2 CMake will not be able to correctly generate this project. Call Stack (most recent call first): CMakeLists.txt:3 (project) -- Configuring incomplete, errors occurred! ``` Have I done it incorrectly? username_1: Oh, now, I was speaking about your games on the x86 side. box86 is not compatible with SSE4 for now, no mater what compile flag you use. username_0: Thanks! username_1: I renamed the ticket and will probably close it soon. I'm still unsure if I'll support SSE4 on box86, I still haven't encounter a single app or game were it's mendatory to have SSE4.
vsanthanam/VSAlert
312949252
Title: Request for NSTextaligment Question: username_0: Would it be possible to include customized text-alignment per alert? By default it aligns to centre, but would be nice if user can control this. Answers: username_1: If you'd like to try this functionality now, you can clone the repo and checkout the branch 'alert-messsage-text-alignment'. I've edited the sample app so the first alert has its text aligned to the left. note: this is an early implementation -- it doesn't support UIAppearance nor is it documented. It'll make its way into the master branch later today. username_1: This feature is now available in release 1.2.1 Status: Issue closed username_0: Thanks for the quick response!
rundeck/docs
494825562
Title: Execution fails on new installation project with script resource model source Question: username_0: However, it seems that even though this execution fails, the nodes are loaded from the script source properly, and the second execution works as expected. **How to reproduce Behavior** * Install Rundeck * Setup a project with a `script` Resource Model Source that defines some nodes * Create a job that executes on one of the defined nodes * Run the job Status: Issue closed Answers: username_1: Somehow this bug got buried in our Docs project. If this is still an issue on the current version of Rundeck please feel free to open a new issue in the rundeck/rundeck project.
ThirteenAG/WidescreenFixesPack
341098769
Title: [SH2] Using multi-core processor. Question: username_0: Hi! Maybe it's not possible, but I want to ask anyway. Is it possible modify the .exe code or some file to make Silent Hill 2 working on multi-core processor without the need of changing it to single-core affinity? Answers: username_1: That would probably require rewriting the whole engine, AFAIK. username_2: Just make a .BAT file with this in it username_2: Just make a `.BAT` file with this in it `start /affinity 1 /high sh2pc.exe` username_0: Thank you, but I already know that. In fact, the Widescreen Fix has that option activated automatically. But that metod it doesn't work well on Windows 7. I was asking cause I wanted to know if there's a chance to enhance the game itself, making it work with multi-core. username_3: I wish I could pin-point how I'm able to use the WSF's built-in single core affinity with this game on my Windows 7 machine because I can get it to run with that without issue... username_4: I think one of the ADX threads deadlocks at some point (likely due to a data race). Unfortunately, ADX is linked statically both into sh2 and sh3 binaries (and probably sh4 as well), so it's not trivially upgradeable by replacing the dll. username_5: Some information posted here might be relevant: https://github.com/elishacloud/Silent-Hill-2-Enhancements/issues/22#issuecomment-418599617 Status: Issue closed
zixpo/candybar
829223248
Title: can't find symbol variable version name after updating gradle Question: username_0: ![Screenshot 2021-03-11 201539](https://user-images.githubusercontent.com/7301670/110805178-dd111b80-82a6-11eb-99ae-27b7f0d73090.jpg) Answers: username_1: Nossa assim fica difícil em. Enviado do meu Samsung Mobile da Claro Obter o Outlook para Android<https://aka.ms/ghei36> username_2: Add this line to `library/build.gradle` file - https://github.com/zixpo/candybar/blob/main/library/build.gradle#L39 Status: Issue closed username_1: Rest Enviado do meu Samsung Mobile da Claro Obter o Outlook para Android<https://aka.ms/ghei36> username_0: Please update Google Play Billing Library version to 3 username_2: Would do it before the deadline username_1: would you pay? Enviado do meu Samsung Mobile da Claro Obter o Outlook para Android<https://aka.ms/ghei36>
philliplab/quasispecies_sim_reports
361257234
Title: Redo the analysis and plotting of the simulated results - waaaaaay too slow Question: username_0: <!--- @huboard:{"order":17.992801799640066,"milestone_order":0.9994002099440126,"custom_state":""} --> Answers: username_0: So actually it is not slow. Something weird must have happened in that one run. Memory usage is another story... username_0: Sigh - it is because the results are cached - that is way it was so much faster now - review this issue again username_0: The distance matrix calculations are very "slow" - But I think it is not the calculation, but the creation and population of the data structure, so caching will have little benefit? explore this - also how effective is digest on such a monster dataset? Status: Issue closed
renrizzolo/react-native-sectioned-multi-select
362206558
Title: Custom chips render Question: username_0: Hi @renrizzolo, great component and the infinite customization options. I just want one more :) : a `custom chips` renderer, I want to render the chips to be uniform with my app: I use `react-native-paper`. Thanks 👍
gbif/data-mobilization
987454189
Title: database_fish_Dingqu_Jinshajiang_China Question: username_0: ## database_fish_Dingqu_Jinshajiang_China Region: China Taxon: Fish Type: undefined Why is this important: Dingqu Basin is the primary tributary left bank of the upper Yangtze River and the largest tributary of the Jinshajiang River, China. Fish composition of this basin presents typical Qinghai-Tibet Plateau fauna, which is crucial for evolutionary and conservation research of Qinghai-Tibet Plateau fish. Priority: medium License: Unspecified Users contact info: <EMAIL>
banctilrobitaille/torch-vectorized
952979204
Title: Getting negative eigenvalues for PDS matrices Question: username_0: Always getting negative eigenvalues for randomly generated PDS matrices when one eigenvalue is very small and close to 0 or very big. Any way to deal with this situation? The following is one specific example that gives a negative eigenvalue and the eigenvalue is not really close to 0. ``` T = torch.tensor([[ 1.3999e+00, 1.5765e+00, -5.5419e+03], [ 1.5765e+00, 2.1994e+00, -7.3147e+03], [-5.5419e+03, -7.3147e+03, 2.4693e+07]], dtype=torch.float64) vlinalg.vSymEig(T.reshape((1,9,1,1,1)), eigenvectors=False)[0] ``` The result is `tensor([[[[[-7.9096e-02]]], [[[ 2.6781e-01]]], [[[ 2.4693e+07]]]]], dtype=torch.float64)` But if we use the build-in function in Pytorch: ``` torch.symeig(T, eigenvectors=True)[0] ``` Then the eigenvalues are all positive: `tensor([4.5855e-03, 1.8413e-01, 2.4693e+07], dtype=torch.float64)` Answers: username_1: Hello @username_0, thank you for reporting this issue. I played a bit with your example and it seems that its related to numerical approximation imprecision when computing the second eigenvalue (which also impact the third): `eig_vals[:, 0, :, :, :] = q + 2 * p * torch.cos(phi)` `eig_vals[:, 1, :, :, :] = q + 2 * p * torch.cos(phi + pi * (2.0 / 3.0))` `eig_vals[:, 2, :, :, :] = 3 * q - eig_vals[:, 0, :, :, :] - eig_vals[:, 2, :, :, :]` From what I've seen, the amount of digits used in the approximation of pi can influence the result quite a bit when dealing with huge number (as in your case). I'll investigate how I can improve the computation accuracy.
seek4science/seek
1005569089
Title: Sample / Custom Metadata behaviour with booleans seems to be incorrect Question: username_0: when setting, ("a boolean is not a valid Boolean"). Unchecking a boolean doesn't seem to work either. Seems to be something to do with posting a value 'on' or 'off' rather than true or false Answers: username_0: should also be possible to set to nil if non-mandatory, so a combobox or dropdown looks necessary Status: Issue closed
jlippold/tweakCompatible
418132632
Title: `Spodinhancer` working on iOS 12.1.1 Question: username_0: ``` { "packageId": "com.hackyouriphone.spodinhancer", "action": "working", "userInfo": { "arch32": false, "packageId": "com.hackyouriphone.spodinhancer", "deviceId": "iPod7,1", "url": "http://cydia.saurik.com/package/com.hackyouriphone.spodinhancer/", "iOSVersion": "12.1.1", "packageVersionIndexed": false, "packageName": "Spodinhancer", "category": "HYI - Tweaks", "repository": "HackYouriPhone", "name": "Spodinhancer", "installed": "1.1", "packageIndexed": false, "packageStatusExplaination": "This tweak has not been reviewed. Please submit a review if you choose to install.", "id": "com.hackyouriphone.spodinhancer", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.1.3", "shortDescription": "An ultimate suite of Spotify tweaks", "latest": "1.1", "author": "<NAME>", "packageStatus": "Unknown" }, "base64": "<KEY> "chosenStatus": "working", "notes": "" } ```
vercel/next.js
694029596
Title: Prefetching not working for optional catch all routes in Firefox Question: username_0: # Bug report ## Describe the bug When using `next/link` for optional catch all routes, it is prefetched in Chrome but not in Firefox ## To Reproduce The code can be found in https://github.com/username_0/nextjs-bug-mwe, it is hosted at https://nextjs-bug-mwe.vercel.app/ ## Expected behavior The file `working.json` should be prefetched. In general, links with `next/link` should be prefetched ## Screenshots Firefox network logs: ![image](https://user-images.githubusercontent.com/29740136/92300823-eca95380-ef55-11ea-8a4a-7a87a4324930.png) Chrome network logs: ![image](https://user-images.githubusercontent.com/29740136/92300826-f8951580-ef55-11ea-8075-2b2200cc6b45.png) Note that `working.json` is in the Chrome log but not the Firefox one, it is only loaded when actually clicking the link ## System information - OS: Linux - Browser (if applies): Firefox 80.0.1 - Version of Next.js: 9.5.3 - Version of Node.js: not sure, hosted on Vercel Answers: username_1: what is your`network.prefetch-next` setting in firefox? username_0: Turns out this was it, thanks very much! Is this disabled by default in Firefox? I don't remember disabling this username_1: i have run into a similar issue and for me it was caused by "ublock origin" addon default settings. see my comment [here](https://github.com/vercel/next.js/discussions/11578#discussioncomment-51258) username_0: Oh that's annoying, thanks very much for your help though! Status: Issue closed username_0: # Bug report ## Describe the bug When using `next/link` for optional catch all routes, it is prefetched in Chrome but not in Firefox ## To Reproduce The code can be found in https://github.com/username_0/nextjs-bug-mwe, it is hosted at https://nextjs-bug-mwe.vercel.app/ This bug can also be seen in the Next.js docs ## Expected behavior The file `working.json` should be prefetched. In general, links with `next/link` should be prefetched ## Screenshots Firefox network logs: ![image](https://user-images.githubusercontent.com/29740136/92300823-eca95380-ef55-11ea-8a4a-7a87a4324930.png) Chrome network logs: ![image](https://user-images.githubusercontent.com/29740136/92300826-f8951580-ef55-11ea-8075-2b2200cc6b45.png) Note that `working.json` is in the Chrome log but not the Firefox one, it is only loaded when actually clicking the link ## System information - OS: Linux - Browser (if applies): Firefox 80.0.1 - Version of Next.js: 9.5.3 - Version of Node.js: not sure, hosted on Vercel username_0: Sorry for reopening this. On Firefox I'm often getting a prefetch of around 500 bytes, not showing any response, and not having the effect of prefetching, but on Chrome they're in the order of kilobytes, with the response showing the data from the page. Is this a uBlock Origin thing too? username_2: Hi, I'm going to close this as data prefetching has been changed in the latest version and it appears to be working correctly in Firefox as shown in the below screenshot. To ensure issues you are experiencing aren't related to extensions, you can try disabling them and see if they can be reproduced without in the future. <img width="1396" alt="Screen Shot 2020-12-01 at 1 40 22 PM" src="https://user-images.githubusercontent.com/22380829/100788499-c5d6ae00-33da-11eb-82fa-dc3d652bfd31.png"> Status: Issue closed username_3: This issue has been automatically locked due to no recent activity. If you are running into a similar issue, please create a new issue with the steps to reproduce. Thank you.
kubernetes/minikube
611345446
Title: fail to install minikube on windows with virtualbox Question: username_0: minikube start --cpus 2 --memory 2048 --insecure-registry=['192.168.99.1:5000'] --vm-driver=virtualbox * minikube v1.5.2 on Microsoft Windows 10 Pro 10.0.18362 Build 18362 * Creating virtualbox VM (CPUs=2, Memory=2048MB, Disk=20000MB) ... * Found network options: - NO_PROXY=192.168.99.100 - no_proxy=192.168.99.100 * X Failed to enable container runtime: enable docker.: Process exited with status 1 * * Sorry that minikube crashed. If this was unexpected, we would love to hear from you: - https://github.com/kubernetes/minikube/issues/new/chooseì Status: Issue closed Answers: username_0: check docker's env variables
concourse/concourse
178699440
Title: Display how long an individual task takes Question: username_0: Right now concourse displays the duration of an entire build, but it would be cool to display that for individual tasks as well. That way I can get even more surgical information about my build history. Answers: username_1: Dupe of https://github.com/concourse/concourse/issues/428 Status: Issue closed
GrantQuick/BlackbaudSkyApi-FENXT
1161902893
Title: Budget Dollar Amounts Question: username_0: I can see where the connector pulls in the table of the budget list of accounts.... but the database has one more step deeper than each department that lists the actual budget amounts for those departments. Inside Financial Edge's Database view, you have to double-click a specific department to get to view the actual budget amounts for that fiscal year. Is there a way to get to the actual budget amounts using this connector?
HectorPeeters/fontoxpath
891924026
Title: Assignment 9: Assessment of own report (individual) Question: username_0: Individually, assess your draft report. Use this [rubric](https://brightspace.tudelft.nl/d2l/common/dialogs/quickLink/quickLink.d2l?ou=281758&type=content&rcode=tudelft-660804) to evaluate your own report. As a group, write down three questions you have for your lecturer. You can ask these questions during Session 4.<issue_closed> Status: Issue closed
Nordstrom/kubelogin
283957368
Title: kubelogin fails to properly create it's config file on windows Question: username_0: `kubelogin config --alias=xxx --server-url=https://xxx --kubectl-user=xxx 2017/12/20 17:04:17 Couldn't find config file in root directory. Creating config file... 2017/12/20 17:04:17 failed to create file in root directory: exec: "touch": executable file not found in %PATH%` windows doesn't have `touch` Answers: username_0: https://github.com/Nordstrom/kubelogin/blob/master/cmd/cli/main.go#L211 Status: Issue closed
olifolkerd/tabulator
539954244
Title: xlsx download with multiple sheet not working properly in version 4.5.2 Question: username_0: Download with multiple sheets works fine in 4.4.3 as described in instructions but having trouble in 4.5.2. Download proceeds but only the sheet from the active table shows data on its tab. Looking into code, the problem seems related to download function accepting 5 arguments now but only being provided 4 internally. This results in slight malfunction and data is lost. Example where 4 parameters are provided to 5-argument download function: Download.prototype.commsReceived = function (table, action, data) { switch (action) { case "intercept": this.download(data.type, "", data.options, data.intercept); break; } }; The intercept function gets lost since the download function expects it to be 5th arg.<issue_closed> Status: Issue closed
nooitaf/nurdbar
105685110
Title: No multiline IRC Question: username_0: The IRC client doesn't handle multiple line messages both receiving and sending. Answers: username_1: https://github.com/username_1/nurdbar/issues/1 username_2: ah dat is jammer username_1: well it does, but its not shown on the bar-screen username_1: multi line messages are now properly word-wrapped and should work on any screen size Status: Issue closed
NativeScript/NativeScript
178269115
Title: Nativescript with Angular 2: proper i18n approach? Question: username_0: Angular 2 has its own i18n functionality: http://www.savethecode.com/angular2-i18n-native-support/ Will NativeScript adapt Angular's i18n for its own use? Or the nativescript-i18n plugin is to be officially supported part of NS going forward?
swcarpentry/python-novice-inflammation
417271197
Title: Episode 1 - BODMAS - Knowledge Question: username_0: If we are introducing new syntax/programming concepts do we need to get people to think about Mathematical operator order of precedence as well? and in a way knowledge of the python operator order of precedence. Answers: username_1: Python's operator precedence rules mirror those in mathematics. I don't know if adding this information would be helpful or add possible confusion. username_0: true adding the extra could create extra confusion so would the example be better as say `3 + 5` Avoiding any unintended inference to operator precedence or would this remove the idea of combining operations? username_2: Hi @username_0 and welcome to our repository! Using `3 + 5 * 4` as an example in the introduction serves two purposes: 1) convey the fact that Python interpreter can be used as a "calculator" 2) It follows standard mathematical rules. As with any lesson out there, we have to draw a borderline somewhere and this is the case here: if the order of mathematical operations is a problem, then this lesson is not the right lesson to start with (SWC lessons are for beginning scientists). We also don't mention other operations (such as exponentiation, which is ** and not ^, which is a bitwise XOR) for the same reason: this is not the goal here. Here we just want to say that you can do math in Python interpreter and quickly transition to variables. username_3: I think it is important to introduce the concept of BODMAS, but not in a confusing way. They should tell the participants that when python does mathematical operations, it follow the law of BODMAS.Therefore, they do not have to write an extra command to specify the order of the mathematical operation to get the correct answer.
JayNewstrom/DockerGradleAndroid
578618356
Title: Add "Remote Build" support Question: username_0: There are a few tools that use ssh and rsync to do builds remotely. https://github.com/Instamotor-Labs/mirakle is the one I've been using recently, and want to support. Ideally this would probably just be built upon the base docker image, and use a different name or tag. Status: Issue closed Answers: username_0: 6fa2107
flowable/flowable-engine
224737391
Title: Problem with cancelling boundary events Question: username_0: In our application we have a sub process with a signal boundary event using cancelActivity="true". Within this sub process we create a child execution using a service task with a Java delegate. During this child execution the signal for the signal boundary event is thrown. This leads to a database error because the execution of the sub process is deleted in method _deleteChildExecutions_ of class _BoundaryEventActivityBehavior_ but not the child execution that references the execution of the sub process. If the statement `if (parentExecution.getCurrentFlowElement() instanceof CallActivity)` would be ommited in method _deleteChildExecutions_ all would be fine, because then the child execution (that was not created by a call activity) would be deleted too. Thanx in advance Ralf Status: Issue closed Answers: username_1: Hi Ralf, Ok, that check was there to prevent an unnecessary sql query, but your case shows that it's needed anyways. Changed the logic according to your suggestion. Thanks
shortcutme/Sublime-IntelliDocs
76744167
Title: Python: lack of support for some functions Question: username_0: It seems don't work correctly for Python, as good example I can give iteritems(). Move cursor there, press F2 - nothing happens. Next example is match(), if it's invoked as a method on variable build with re.compile() then no documentation pop-up is displayed.
dart-lang/sdk
206193225
Title: Strong mode should infer `Null` for generic types inside constant expressions Question: username_0: If a constant expression refers to a generic type, strong mode should replace the uses with `Null` (or later `Nothing`). For example: ``` dart List<T> foo<T>() { return const <T>[]; } ``` Answers: username_1: I think @username_3 was going to follow up on this, but just to mention: My guess is that the right approach (ignoring type bounds) is to infer a type as usual, and then replace all covariant occurrences with Null and contravariant occurrences with dynamic or Object. This yields a type which is a subtype of the precise inferred type. I haven't thought through whether bounds add anything interesting to the problem or not. username_2: That makes sense for cases where the `T` isn't the actual type. E.g.: ```dart List<T Function(T Function(T))> foo<T>() => const []; ``` In this case, the inference wants a `List<Something>` that is a subtype (and therefore assignable) to `List<T Function(T Function(T))>` for any `T` (and T isn't even bounded here, so it's "easy"). Obviously `List<Null>` is a solution, but less precise than possible since `List<Null Function(Object Function(Null))>` is also one. I guess we want the greatest lower bound that can be reasonably found. Or, is that really important? You can't use the class at runtime anyway, and you can assign it to any static type you care about, so does it matter that the runtime type isn't precise? If we do infer a preciser type, with a bound on `T`, we'd use the bound instead of `Object`. With an F-bound, we'll ... do whatever we do for F-bounds (fail, default to `Object` or introduce the anonymous infinite type) instead of `Object`. The contravariant positions are still satisfied by `Null`. username_3: I believe that would be the way to go. Just double checking, it is possible to create a constant expression where we would like to have a type argument inferred, where we need to match a formal type parameter of an enclosing declaration (such that we must find a constant value that fits with all possible actual type arguments), and where the inferred type argument occurs contravariantly: ```dart typedef void Foo<T>(T t); void foo<T>(T t) {} void bar<T>([Foo<T> arg = foo]) {} main() { bar<int>(); } ``` For the default value `foo` of `arg` in the declaration of `bar`, we need to infer `foo<Object>` or `foo<dynamic>` rather than `foo<Null>` (at some point, `foo<Nothing>`), ignoring that we don't support these generic instantiation tear-offs yet. Bounds _do_ create the usual problems. We would have a rule like this: 'for a formal type parameter which is used in a contravariant position, replace it by the least upper bound of all possible actual type arguments', as well as the (unproblematic) 'for a formal type parameter which is used in a covariant position, replace it by the Null (or Nothing)': ```dart class C<S extends C<S>> {} class C1 extends C<C1> {} typedef void Foo<T>(T t); void foo<U extends C<U>>(U t) {} void bar<V extends C<V>>([Foo<V> arg = foo]) {} main() { bar<C1>(); } ``` We cannot use `foo<Object>` nor `foo<dynamic>` because the given type arguments do not satisfy the bounds (and even though we may contemplate using malbounded type annotations, this is a malbounded runtime entity, so that's still just plain wrong ;-), and we cannot express the least upper bound of all the possible type arguments `U` satisfying `U extends C<U>`. So I'd give the same recommendation that I usually do for instantiate-to-bound: Refuse to solve this problem, i.e., reject that default value at compile time. Adjusting the title to catch up with the developments so far. username_3: Ah, I forgot to reload so I didn't see your response first, Lasse. Just for completeness, though, I could add that we do have cases where inference con't provide a sufficiently "universal" type: ```dart typedef T Foo<T>(T t); T foo<T>(T t) => null; void bar<T>([Foo<T> arg = foo]) {} main() { bar<int>(); } ``` In this case neither a top nor a bottom type will work (because we have both covariance and contravariance), so there will of course be some cases where we must declare that the inference failed and the program is rejected. username_1: I don't think we've committed to treating instantiated tearoffs of generic functions (`foo<T>` in @username_3's example above) as const objects? Or did we decide that was fine? If so, then yes, there are some examples we can't find a sound approximation for and will have to error out. I don't expect that to be a common issue. The point that @username_2 makes about simply using `Null` as the type argument to the outermost type constructor and not worrying about getting the most precise runtime type seems like a good one. I think there's probably a reasonable definition that can be made around this. I don't immediately see any benefit to having a more precise type offhand. username_3: Right, const objects are specified syntactically, and the supported syntactic forms do not include generic tear-offs like `foo<T>`. We could (1) add syntactic support for generic tear-offs, and add this form to the list of constant expressions, with constraints as needed (e.g., can't use formal type parameters in scope); and we could (2) add support for generic tear-offs via inferred type arguments (such that we can have them semantically for expressions like `foo`), without adding syntactic support (`foo<T>`). The things I said were intended to be compatible with (2), but I needed a hypothetical (1) in order to explain what I meant; finally, we could of course have both (1) and (2). username_2: /L -- <NAME> - <EMAIL> 'Faith without judgement merely degrades the spirit divine' Google Denmark ApS - Frederiksborggade 20B, 1 sal - 1360 København K - Denmark - CVR nr. 28 86 69 84 username_1: See comment on this bug here: https://github.com/dart-lang/dart-lang-evolution/issues/111 for the last state of the discussion around partial instantiations, and the inference thereof. There is an open issue against strong mode analyzer to restrict inference to only inferring partial instantiations on tearoffs (not on function expressions).
Cr0wTom/Mi-Notebook-Pro-Mods
654690879
Title: Usb C Hub Power error Question: username_0: I found a problem with the USB C Hubs, I state that I use both the original cable and the power supply, when I connect them to the computer all the peripherals work quietly, but if I try to load the PC through them the charge stops and starts again in 3 seconds cycles . I have tried all versions of bios up to the latest version, only in versions 0502 and 0300 this problem does not occur, but I would like to modify my DVMT then update the bios. Any help?
Princeton-CDH/winthrop-django
333354534
Title: migrate manifest and canvas URIs from plum to figgy Question: username_0: See https://github.com/Princeton-CDH/derrida-django/issues/163 for equivalent work on derrida Status: Issue closed Answers: username_0: I've reviewed the migrated migration test - updated manifest and canvas data looks good in the admin site, and the thumbnail images are coming from figgy.
Princeton-CDH/derrida-django
254431417
Title: As a user, I want to see an indication of which library books in the list are annotated so I can easily identify books with annotations. Question: username_0: @username_1 can you take this one on? The Solr index should provide the flag you need to key on (there's a comment the template), and there is a preliminary design for how this should look (with a placeholder graphic) in Zeplin. There's a similar visual indicator planned for annotated pages, so would be great if the styles were reusable for that. Answers: username_0: @username_1 can you take this one on? The Solr index should provide the flag you need to key on (there's a comment the template), and there is a preliminary design for how this should look (with a placeholder graphic) in Zeplin. There's a similar visual indicator planned for annotated pages, so would be great if the styles were reusable for that. username_0: @username_1 what do you think about adding alt text/title text for the icon as a way of providing a cue for what it means? I know that won't help on mobile, but could be useful on desktop. username_1: We can definitely do that. We could even style a tooltip to show the purpose of the icon. What do you think the test would say for these? username_0: @username_1 perhaps "contains annotations" ? That would be fairly consistent with the filter language username_0: @username_1 I like the tooltip, thanks for adding it. This one looks good to me, I'll mark it as ready for the project team to look at. username_2: @username_0 @username_1 Looks good! Status: Issue closed
USGS-R/delaware-model-prep
588750245
Title: Order of seg_id's for distance matrix Question: username_0: I'm wondering what order the seg_id's are in when the distance matrix is calculated. In other words, which row in the distance matrix corresponds to which seg_id. I don't think the order itself will matter, but I think it will make a difference if the order is *different* in the input datasets to the RGCN model compared to the order in the distance matrix.<issue_closed> Status: Issue closed
MichaCo/DnsClient.NET
406319066
Title: Https http://dnsclient.michaco.net/ Question: username_0: Hi, It would be nice if dnsclient.michaco.net is available on https. Now we get: ![image](https://user-images.githubusercontent.com/5808377/52209770-36671f00-2885-11e9-91f7-35c0faaba00b.png) It's correct that you are using Azure? Maybe this could be helpful: https://gooroo.io/GoorooTHINK/Article/16420/Lets-Encrypt-Azure-Web-Apps-the-Free-and-Easy-Way/20047 Answers: username_1: Hey, Yeah I know, I could use lets encrypt. Just haven't had time to setup the deployments and get everything running and tested... Will happen eventually Status: Issue closed username_0: 🎉
infor-design/enterprise-ng
1050291436
Title: Datagrid: Expose the maskOptions and filterMaskOptions function Question: username_0: **Describe the bug** Need to expose the maskOptions and filterMaskOptions functions in enterprise-ng. **Expected behavior** Need something like the following to be added to soho-datagrid.d.ts... ``` maskOptions?: SohoMaskOptions | SohoDataGridColumnMaskOptionsFunction; filterMaskOptions?: SohoMaskOptions | SohoDataGridColumnMaskOptionsFunction; type SohoDataGridColumnMaskOptionsFunction = ( row: number, cell: any, fieldValue: any, columnDef: SohoDataGridColumn, rowData: Object ) => SohoMaskOptions; ``` **Version** - ids-enterprise-ng: v10.10.0 **Additional context** Is related to https://github.com/infor-design/enterprise/issues/5847 Status: Issue closed Answers: username_2: This issue is now resolved.
tensorflow/tensorflow
947491007
Title: There should be No expansion ReLU on MobileNet v2 bottleneck Question: username_0: I argue that this line should be removed: https://github.com/tensorflow/tensorflow/blob/a4d25192f0340b7786760d57e6b5859e561d9b49/tensorflow/python/keras/applications/mobilenet_v2.py#L450 The code seems wrong vs. the [MobileNet v2 paper](https://arxiv.org/pdf/1801.04381.pdf), **which this code is supposed to faithfully implement**, for the following reasons: The paper uses all of Figure 1 (read the caption), part D of Figure 2 (also read the caption Note), and part B of Figure 3 (also read the caption and check the parts that are annotated ReLU vs. those that aren't), as well as well as the last two paragraphs of Section 3.2 "Linear Bottlenecks" to argue _against_ having ReLU on the bottlenecks. One can specifically see that the _expansion_ layer in part D of Figure 2 is shown without ReLU, and the same setup is shown with part B of Figure 3, all of which matches perfectly with the description given in the last two paragraphs of the section. Currently, it is like the entire point of a central section of the paper was missed or ignored. Answers: username_1: @username_0 This issue is more suitable for TensorFlow Models repo. Please post it on [Tensorflow/Models](https://github.com/tensorflow/models/issues?q=is%3Aopen+is%3Aissue) repo from here. Thanks! username_0: @username_1 can you direct me to the location of this file within that repository? username_0: @username_1 Just to clarify, I don't think this file exists in that repository. I did look. username_2: Looking at the original paper, the Table 1 (Bottleneck residual block) shows that the (1x1 conv2d , ReLU6) is the operation used for the block, which I think is matching with the current implementation in Keras. I have also check the pytorch implementation in https://github.com/pytorch/vision/blob/6db1569c89094cf23f3bc41f79275c45e9fcb3f3/torchvision/models/mobilenet.py#L45, which also use Relu6 as activation. Also adding @marksandler2 and @andrewghoward who are the original author of the paper to confirm the correctness here.
phylotastic/phylotastic-portal
185218960
Title: pulldowns do not close as expected Question: username_0: Select item from Public Trees, My trees pulldowns. The pulldowns do not close as expected (OSX, Chrome). Close pulldown when user selects an item. Answers: username_1: @username_0 I made the dropdown closed right after user select a tree. username_0: @username_1 , I can't test this until #214 is fixed, because there is no way to put anything into public trees, apparently. username_1: After creating a tree, you can publish it by opening 'Tree Settings' and tick on Public: true. <img width="334" alt="screen shot 2016-10-26 at 10 11 41 am" src="https://cloud.githubusercontent.com/assets/1614884/19734405/a4a5c7f6-9b64-11e6-8e26-c9ee7f3f723a.png"> username_0: I'm not seeing that. I think you might need to push it to dev. username_0: However, someone added a public tree named "Fox" and I verified that the pull-down is working as expected. So this is closed. Status: Issue closed username_1: @username_0 You can only see the "Tree settings" button when you are the owner of the tree. Would you please create a new tree and see that whether the button is in tree viewer?
amsbr/OctoPrint-EEPROM-Marlin
264236746
Title: Not loading eeprom Question: username_0: I've installed the plugin and see it listed with other plugins. with printer connected I click the plugin from the left navbar then click on Load EEPROM but it doesn't list any of the eeprom settings. Just a blank page apart from a bit of text at the top and the two buttons ![marlin editor](https://user-images.githubusercontent.com/31435380/31390023-af8fefce-adca-11e7-839e-ecf3e086280c.JPG) Answers: username_1: Hi @username_0 The setting #define EEPROM_SETTINGS is enabled on your firmware? username_0: Duh! My apologies, M503 was disabled in the firmware. I recently uploaded Marlin 1.1.6 using a predefined configuration.h for the CR-10 but this had been disabled to allow other functions to fit in the program memory. I only discovered this last night and changed it. I'd rather have this feature than a boot screen so the boot screen went to give the room required for this feature. My newbie mistake. username_1: Hi @username_0 All right! Status: Issue closed
jlippold/tweakCompatible
625324106
Title: `Audicy` working on iOS 13.5 Question: username_0: ``` { "packageId": "com.ouraigua.xcode.audicy", "action": "working", "userInfo": { "arch32": false, "packageId": "com.ouraigua.xcode.audicy", "deviceId": "iPhone10,5", "url": "http://cydia.saurik.com/package/com.ouraigua.xcode.audicy/", "iOSVersion": "13.5", "packageVersionIndexed": true, "packageName": "Audicy", "category": "Utilities", "repository": "Packix", "name": "Audicy", "installed": "2.0.3", "packageIndexed": true, "packageStatusExplaination": "A matching version of this tweak for this iOS version could not be found. Please submit a review if you choose to install.", "id": "com.ouraigua.xcode.audicy", "commercial": true, "packageInstalled": true, "tweakCompatVersion": "0.1.5", "shortDescription": "Import audio/video to Music app, Videos app, Ringtones, Podcasts, Camera Roll...", "latest": "2.0.3", "author": "<NAME>", "packageStatus": "Unknown" }, "base64": "<KEY>", "chosenStatus": "working", "notes": "" } ```<issue_closed> Status: Issue closed
void-linux/void-packages
806420476
Title: Packages can't depend on package-$version Question: username_0: See 0961351b5226020b44bc8f1426310112c9151055 `util-linux` shouldn't have to be revbumped every time `util-linux-libs` is, or vice versa. Changes that only affect one of the packages should allow us to revbump only one of them, but that currently isn't possible, because `util-linux` has `util-linux-libs-${version}_${revision}` in `depends`. The ABI is unlikely to change between revisions (we even depend on it not changing between versions). Therefore, it should be possible to depend on `util-linux-libs-${version}`, which currently results in the package not being found: `=> ERROR: util-linux-2.36.1_2: target dependency 'util-linux-libs-2.36.1' does not exist!`. It would be nice to define a mechanism for depending on a specific *version* of a package, without caring about specific revision. I believe other templates could also be improved if such a mechanism becomes available. Answers: username_1: `util-linux-libs>=${version}_1<${version}_999` is close enough.
RQuispeC/pytorch-ACSCP
507773769
Title: Running the code on new images Question: username_0: Hi there! Thank you for sharing the code. Can you please guide me over how can I use this code on my own dataset? I want to give in a few images as input, and want the pre-trained model of fold3 to evaluate those images and output a crowd count. Looking forward to your help. TIA! Status: Issue closed Answers: username_0: Making changes in the evaluate_model.py script only visually removes the ground truth images on the plot. It seems that in some other script the program is still comparing it with the ground truths. I'm trying to understand the flow of the program, but one thing is confusing me. In the actual training and testing with the UCF dataset, the ground truth density maps were provided. When testing on new images, will the program still be requiring ground truth density maps and json files for each new image for comparing, or will the predicted density maps made at run-time suffice to give a crowd count? How do I populate the 'gt_density' blob in data_loader.py since there will not be any ground truth densities available in my case. Looking forward to your kind assistance. username_1: Hi username_0, I reviewed your comment and realize that my previous answer was not complete! sorry for that. To test on your own dataset, you also need to update `data_loader.py` . `data_loader.py` loads the data for train and test, in your case you may have to edit function `__iter__` in this section: ```python if self.test_loader: #loader is for testing, then we divide the image in chunks of size (img_height, img_width) _, h, w = img.shape orig_shape = (h, w) # compute padding if self.test_overlap: padding_h = self.img_height - max(h % self.img_height, (h - self.img_height//2) % self.img_height) padding_w = self.img_width - max(w % self.img_width, (w - self.img_width//2) % self.img_width) else: padding_h = self.img_height - (h % self.img_height) padding_w = self.img_width - (w % self.img_width) # add padding img = np.concatenate((img, np.zeros((img.shape[0], padding_h, img.shape[2]))), axis =1) den = np.concatenate((den, np.zeros((img.shape[0], padding_h, img.shape[2]))), axis =1) img = np.concatenate((img, np.zeros((img.shape[0], img.shape[1], padding_w))), axis =2) den = np.concatenate((den, np.zeros((img.shape[0], img.shape[1], padding_w))), axis =2) assert img.shape[1] % 2 == 0 and img.shape[2] % 2 == 0, "Inputs images must have even dimensions, found {}".format(img.shape) # create batch for test _, h, w = img.shape new_shape = (h, w) disp_height = self.img_height // 2 if self.test_overlap else self.img_height disp_width = self.img_width // 2 if self.test_overlap else self.img_width for i in range(0, h - self.img_height + 1, disp_height): for j in range(0, w - self.img_width + 1, disp_width): chunk_img = img[0, i:i + self.img_height, j:j + self.img_width] chunk_den = den[0, i:i + self.img_height, j:j + self.img_width] chunk_img = chunk_img.reshape((1, chunk_img.shape[0], chunk_img.shape[1])) chunk_den = chunk_den.reshape((1, chunk_den.shape[0], chunk_den.shape[1])) imgs.append(chunk_img) dens.append(chunk_den) ``` basically you need to remove any reference to variables `den`, `chunk_den` or `dens`, and populate variable `blob['gt_density']` with empty value. Finally, you also need to update function `recontruct_test` similarly to `__iter__`. username_1: Hi there! Thank you for sharing the code. Can you please guide me over how can I use this code on my own dataset? I want to give in a few images as input, and want the pre-trained model of fold3 to evaluate those images and output a crowd count. Looking forward to your help. TIA! username_0: I was able to grasp the program flow yesterday, and i'm happy to see that the changes suggested by you align with the ones I had made after my understanding. For plotting the results however, `architecture/utils.py` also has to be edited, and every reference to `gt_density_map` has to be removed. Furthermore, I had completely removed `blob['gt_density']` from `data_loader.py` and then also removed the corresponding variable `gt_data` from `evaluate_model.py`. The code seems to be working perfectly now to give my desired output! Thank you for your time and assistance. Cheers! Status: Issue closed
box/spout
83743307
Title: ODS output Question: username_0: Hi I have a repo for fast-efficient writing to ODS,and I was wondering if it is a file-format you are interested to have as part of spout? Answers: username_1: Hi @username_0 ! Adding a new format is not straight forward, as it needs to support both reading and writing, and doing so in a scalable way. Spout also has an extensive test suite that covers most of the use cases. I am not familiar with ODS but I'll be happy to check your repo to see how easy it is to integrate it. Is ODS a popular format? username_0: Are you joking when you ask "Is ODS a popular format?". It's the open specification that libreoffice / openoffice use as a spreadsheet format. In some countries public sector departments have to save in open formats, and most of Europe is transitioning towards them for vendor agnostic IT files. Even newer MS Office installs offer users the chance to save by default in the open formats (not just MS open doc). It is far superior to Excel in terms of scalability and performance (I can open larger documents in OpenOffice / LibreOffice than Excel). As a more personal example, my own business CRM & ERP software packages currently use ODS as an alternative for writing, because there were scalability issues with PHPExcel. For reading, I could do some work to integrate, to be honest I wanted to know if it was worth me doing the work and submitting a pull request ;). username_1: My question was more in terms of global usage. I agree that the format is really important for some people but my point was, if it represents less than 0.1% of the global usage, does it worth adding this format? Anyway, I did my research and it seems like it represents a non negligible usage. So I think it would be a great addition to Spout and I'll be happy to help you get a pull request through. username_2: :+1: username_1: Just added an ODS writer: #99. Reader is coming with updated documentation but you can start playing with it if you want :) @username_0, feel free to submit pull requests or comments on this since you have experience with an ODS writer before. The current implementation can most likely be improved :) username_0: Do you know what, it stalled as I have a major work project ongoing atm and not much free time at PC, thanks ;) Status: Issue closed
facebook/fresco
102510884
Title: PNG issue Question: username_0: I get A TooManyBitmapsException when I load several PNG format images by use GridView on screen. After that, I can't display one of these images on a SimpleDraweeView full screen. Use a Sumsung device make many screenshots and run the comparison sample, The exception can reproduce every time. I upload a [video](https://youtu.be/5cFpkJPIWNM) about this on youtube. I searched and saw all issues about this, but I have no idea to resolve it. #213 #214 #256 #307 #308 #469 #483 #539 #556 Answers: username_0: Now, The only way is rolling back to UIL before application release :( username_1: Do you specify ResizeOptions? If not, each of these images requires a bitmap of the screen size, even though they are 9 times smaller (I.e. your app is using 9x more memory on images than it should). Without some more context it is hard to tell. username_2: For PNGs you need to [enable downsampling](http://frescolib.org/docs/resizing-rotating.html#downsampling) when specifying ResizeOptions. Status: Issue closed username_0: Awesome, thank you. username_3: That's Right,thank you @username_2 username_4: i wanna load gifs in my gallery.But when i get the first frame of the gif as a bitmap, i don't know how to load it with fresco
binary-com/binary-bot
362186245
Title: Open trade after one candle Question: username_0: if sma 7 crosses ema 21 and l want to buy but l want to buy after one complete candle what must l do Answers: username_1: in "in candle list read (close)" block, change the "close" to "open" username_0: can you please show me how to set it up username_1: Attach the zip file here username_2: We're closing this issue due to inactivity, you're welcome to open it again if you still haven't got your answer. Status: Issue closed
zkSNACKs/WalletWasabi
1157295475
Title: Value cannot be null. (Parameter 'right') Question: username_0: ### General Description When transactions were confirmed I got this. It seems like the history is not updated after this error, so txs stay unconfirmed. ### Logs ``` 2022-03-02 14:44:22.490 [1] WARNING NotificationHelpers.TryGetNotificationInputs (111) System.ArgumentNullException: Value cannot be null. (Parameter 'right') at NBitcoin.Money.op_Subtraction(Money left, Money right) at WalletWasabi.Fluent.Helpers.NotificationHelpers.TryGetNotificationInputs(ProcessedResult result, String& message) in WalletWasabi.Fluent\Helpers\NotificationHelpers.cs:line 82 2022-03-02 14:44:22.896 [1] WARNING NotificationHelpers.TryGetNotificationInputs (111) System.ArgumentNullException: Value cannot be null. (Parameter 'right') at NBitcoin.Money.op_Subtraction(Money left, Money right) at WalletWasabi.Fluent.Helpers.NotificationHelpers.TryGetNotificationInputs(ProcessedResult result, String& message) in WalletWasabi.Fluent\Helpers\NotificationHelpers.cs:line 82 2022-03-02 14:44:22.897 [1] WARNING NotificationHelpers.TryGetNotificationInputs (111) System.ArgumentNullException: Value cannot be null. (Parameter 'right') at NBitcoin.Money.op_Subtraction(Money left, Money right) at WalletWasabi.Fluent.Helpers.NotificationHelpers.TryGetNotificationInputs(ProcessedResult result, String& message) in WalletWasabi.Fluent\Helpers\NotificationHelpers.cs:line 82 2022-03-02 14:44:22.898 [1] WARNING NotificationHelpers.TryGetNotificationInputs (111) System.ArgumentNullException: Value cannot be null. (Parameter 'right') at NBitcoin.Money.op_Subtraction(Money left, Money right) at WalletWasabi.Fluent.Helpers.NotificationHelpers.TryGetNotificationInputs(ProcessedResult result, String& message) in WalletWasabi.Fluent\Helpers\NotificationHelpers.cs:line 82 2022-03-02 14:44:22.899 [1] WARNING NotificationHelpers.TryGetNotificationInputs (111) System.ArgumentNullException: Value cannot be null. (Parameter 'right') at NBitcoin.Money.op_Subtraction(Money left, Money right) at WalletWasabi.Fluent.Helpers.NotificationHelpers.TryGetNotificationInputs(ProcessedResult result, String& message) in WalletWasabi.Fluent\Helpers\NotificationHelpers.cs:line 82 ``` Status: Issue closed Answers: username_2: #7429 cannot possible fix this problem. username_2: ### General Description When transactions were confirmed I got this. It seems like the history is not updated after this error, so txs stay unconfirmed. ### Logs ``` 2022-03-02 14:44:22.490 [1] WARNING NotificationHelpers.TryGetNotificationInputs (111) System.ArgumentNullException: Value cannot be null. (Parameter 'right') at NBitcoin.Money.op_Subtraction(Money left, Money right) at WalletWasabi.Fluent.Helpers.NotificationHelpers.TryGetNotificationInputs(ProcessedResult result, String& message) in WalletWasabi.Fluent\Helpers\NotificationHelpers.cs:line 82 2022-03-02 14:44:22.896 [1] WARNING NotificationHelpers.TryGetNotificationInputs (111) System.ArgumentNullException: Value cannot be null. (Parameter 'right') at NBitcoin.Money.op_Subtraction(Money left, Money right) at WalletWasabi.Fluent.Helpers.NotificationHelpers.TryGetNotificationInputs(ProcessedResult result, String& message) in WalletWasabi.Fluent\Helpers\NotificationHelpers.cs:line 82 2022-03-02 14:44:22.897 [1] WARNING NotificationHelpers.TryGetNotificationInputs (111) System.ArgumentNullException: Value cannot be null. (Parameter 'right') at NBitcoin.Money.op_Subtraction(Money left, Money right) at WalletWasabi.Fluent.Helpers.NotificationHelpers.TryGetNotificationInputs(ProcessedResult result, String& message) in WalletWasabi.Fluent\Helpers\NotificationHelpers.cs:line 82 2022-03-02 14:44:22.898 [1] WARNING NotificationHelpers.TryGetNotificationInputs (111) System.ArgumentNullException: Value cannot be null. (Parameter 'right') at NBitcoin.Money.op_Subtraction(Money left, Money right) at WalletWasabi.Fluent.Helpers.NotificationHelpers.TryGetNotificationInputs(ProcessedResult result, String& message) in WalletWasabi.Fluent\Helpers\NotificationHelpers.cs:line 82 2022-03-02 14:44:22.899 [1] WARNING NotificationHelpers.TryGetNotificationInputs (111) System.ArgumentNullException: Value cannot be null. (Parameter 'right') at NBitcoin.Money.op_Subtraction(Money left, Money right) at WalletWasabi.Fluent.Helpers.NotificationHelpers.TryGetNotificationInputs(ProcessedResult result, String& message) in WalletWasabi.Fluent\Helpers\NotificationHelpers.cs:line 82 ```
symfony/symfony
211032659
Title: [SecurityBundle] something is broken after recent commits Question: username_0: | Q | A | ---------------- | ----- | Bug report? | yes | Feature request? | no | BC Break report? | no | RFC? | no | Symfony version | 3.3.0-dev Today we've got build error in our system: Restored to last worked version on 'd69bb304001e57e00678b86e8e499067889db538' Hope this info will help to spot a problem. ``` [04:39:32][Ant output] [exec] [exec] > Sensio\Bundle\DistributionBundle\Composer\ScriptHandler::clearCache [04:39:32][Ant output] [exec] [exec] [04:39:32][Ant output] [exec] [exec] [04:39:32][Ant output] [exec] [exec] [Symfony\Component\DependencyInjection\Exception\ServiceNotFoundException] [04:39:32][Ant output] [exec] [exec] The service "security.authentication.provider.simple_form.default" has a de [04:39:32][Ant output] [exec] [exec] pendency on a non-existent service "security.user.provider.concrete.carrier [04:39:32][Ant output] [exec] [exec] User". ``` Our security.yml: ```yml security: providers: carrierUser: entity: class: System\Entity\CarrierUser property: email erase_credentials: false firewalls: dev: pattern: ^/(_(profiler|wdt)|css|images|js)/ security: false default: anonymous: ~ provider: carrierUser simple-form: check_path: /login_check require_previous_session: false authenticator: tms.security.authentication.provider failure_handler: auth.failure_handler failure_forward: true failure_path: false encoders: System\Entity\CarrierUser: bcrypt ``` Answers: username_0: Little findings: on working version there line in container build logs: ```Symfony\Component\DependencyInjection\Compiler\InlineServiceDefinitionsPass: Inlined service "security.user.provider.concrete.carrieruser" to "security.firewall.map.context.default".``` but in not working version there is line ``Symfony\Component\DependencyInjection\Compiler\InlineServiceDefinitionsPass: Inlined service "security.user.provider.concrete.carrierUser" to "security.firewall.map.context.default".``` Previous version automatically lowercased every provider name, but new version looks like not converting to lower cases. Maybe this is a case. username_1: Which commit exactly did you use when experiencing this issue? username_0: today latest i think, just checked out from teamcity work dir: here is: 28a00dac0c396f2880b84bc33d7abae698d4d94a username_0: But I don't know exactly which commit fails. username_1: Can you check if the issue still persists when you only use lowercase letters for the user provider name? username_2: If only the case sensitivity is the problem AFAIR this was done on purpose. And it should be in Changelog/Upgrade document. username_0: Yes, works then lowercasered. If it's by purpose then should throw exception about saying what is wrong exactly during configuration validation I think. I'll probably tracked which commit broken this and will write details soon username_0: found a problem, it was during inconsistent getUserProviderId() usage, sometimes it was called like: getUserProviderId(strtolower($name)); sometimes like getUserProviderId($name); Fixed by putting strtolower() inside getUserProviderId https://github.com/symfony/symfony/commit/fbd9f88e312d03614b7e96ed3a1c8c993eb8c0b5#diff-2be909961a57bf75fbb600c1f5fc46e3R320 username_1: @username_0 That commit was reverted later on. That's why I wonder if you still experience the issue when using the latest `master` branch. username_1: Never mind, I see that f6637dd9008deb786823aba02102870e4bd26e25 wasn't merged up to master yet. Status: Issue closed
adafruit/Adafruit_MQTT_Library
941910144
Title: How to use to set callback to a class member function ? Question: username_0: Is there a way to use the library inside of a class and then set a callback to be one of the class functions ? I can't figure out a way to do this ? It seems to work without a problem with the PubSubClient but I'd rather use this library because it it more robust when it comes to reconnecting. I'm using esp8266 with Arduino. <img width="1022" alt="Screen Shot 2021-07-11 at 11 Sun 11 55 58 PM" src="https://user-images.githubusercontent.com/8324295/125272802-f69f8f80-e2a7-11eb-86cc-2c36917f4284.png"> Status: Issue closed Answers: username_1: Nope, you'd need to make `data_callback` and `command_callback` globally scoped functions or `static`
mpv-android/mpv-android
512616764
Title: Interpolation flickers on S9+ Question: username_0: Hi! I recently tried to use interpolation on my Galaxy s9+, but found out it is flickering like crazy. I tried with hwdec and swdec, but the issue remains the same. I will include everything needed to fix the issue, aslong as i get help there as i dont know much in that area. Here is an example how it looks: https://streamable.com/c40oe Answers: username_1: version? This bug should have been fixed in 2019-09-02. username_0: @username_1 the Last release which is on this github. 2019-09-02 username_1: Please provide a [`logcat`](https://developer.android.com/studio/command-line/logcat) log. username_0: @username_1 as soon as i am at home, ill Provide a logcat log username_0: wasnt at home yet. currently in a bad family situation. i try to get it till the end of the week., username_0: Alrighty - i wont even be able to send a logcat for a few more days as im not home for a pretty long time. if there is a way to logcat whilst only using a phone that would be awesome, then i could do it faster. Sorry to disturb you while i cant provide more information. i saw you released a new version - i will test it with that one when i have access to a pc or when i know how to logcat on mobile only. Status: Issue closed
USGS-WiM/LIQWIDS
467109205
Title: Will only print one page Question: username_0: Cooperator reports only one page prints. @username_0 to test this on Monday. Answers: username_0: Testing: https://test.wim.usgs.gov/liqwids/?characteristic=Nitrate&site=USGS-403653073252901,11NPSWRD_WQX-FIIS_AIMS_20,USGS-405135073235501 Not seeing this behavior in Chrome. This seems to be a bug in Edge. For me, Edge shows 3 pages, but everything is pushed into the first page and it all overlaps in a mess. ![image](https://user-images.githubusercontent.com/12058013/61242053-6820c380-a70a-11e9-8355-15dafaba0cd8.png) I'm unable to open this in IE on my work machine. Firefox only will print one page, cutting off the rest. [testprint.pdf](https://github.com/USGS-WiM/LIQWIDS/files/3393821/testprint.pdf)
gjoseph92/soundDB
193215210
Title: Accept Endpoints, directories, and files in Accessors, not just Datasets Question: username_0: Not everyone might want to use iyore; a folder full of NVSPL files comes to mind here. For the `ds` argument, Accessors should also accept: - iyore Endpoint - string path to a directory - In this case, just create an in-memory Dataset rooted in that directory. - Should the structure be `endpointName: *`, or should each Accessor have a built-in default pattern for its file type? - string path to a file - Just read and return the data using `Accessor.parse`
onevcat/Kingfisher
361622491
Title: How to use NSURLProtocol in Kingfisher Question: username_0: I know a protocol =====>AuthenticationChallengeResponsable , can solve self sined SSL problem, but there is no demo , I do not know how to write it . Answers: username_1: Maybe you could check this https://github.com/username_1/Kingfisher/wiki/Cheat-Sheet#authentication-with-nsurlcredential Basically it is just forwarding [urlSession(_:task:didReceive:completionHandler:)](https://developer.apple.com/documentation/foundation/urlsessiontaskdelegate/1411595-urlsession) and [urlSession(_:didReceive:completionHandler:)](https://developer.apple.com/documentation/foundation/urlsessiondelegate/1409308-urlsession), so you could do the same thing to verify or ignore your self-signed cert.
codeforyouth/vs-covid19
582524482
Title: パフォーマンスの改善 Question: username_0: #### 参考情報 https://googlechrome.github.io/lighthouse/viewer/?psiurl=https%3A%2F%2Fcovid19-support.info%2F&strategy=mobile&category=performance&category=accessibility&category=best-practices&category=seo&category=pwa&utm_source=lh-chrome-ext
dask/dask
626527267
Title: v2.17.0 unit test failure "KeyError: 'optimization'" Question: username_0: **What happened**: Running the unit tests for 2.17.0 on Alpine Linux with Python 3.8 fails in several places with `KeyError: 'optimization'`. **What you expected to happen**: All the unit tests to succeed **Minimal Complete Verifiable Example**: Well, just run the unit tests like normal on an Alpine Linux edge system. ``` (I don't know what a MCVE code is) ``` **Anything else we need to know?**: Log: ``` ERROR collecting dask/dataframe/tests/test_categorical.py dask/dataframe/tests/test_categorical.py:28: in <module> cat_series.append((ds.compute().index, ds.index)) dask/base.py:166: in compute (result,) = compute(self, traverse=False, **kwargs) dask/base.py:438: in compute dsk = collections_to_dsk(collections, optimize_graph, **kwargs) dask/base.py:217: in collections_to_dsk _opt_list.append(opt(_graph_and_keys[0], _graph_and_keys[1], **kwargs)) dask/dataframe/optimize.py:28: in optimize fuse_subgraphs = config.get("optimization.fuse.subgraphs") dask/config.py:459: in get result = result[k] E KeyError: 'optimization' ERROR collecting dask/dataframe/tests/test_dataframe.py dask/dataframe/tests/test_dataframe.py:38: in <module> full = d.compute() dask/base.py:166: in compute (result,) = compute(self, traverse=False, **kwargs) dask/base.py:438: in compute dsk = collections_to_dsk(collections, optimize_graph, **kwargs) dask/base.py:217: in collections_to_dsk _opt_list.append(opt(_graph_and_keys[0], _graph_and_keys[1], **kwargs)) dask/dataframe/optimize.py:28: in optimize fuse_subgraphs = config.get("optimization.fuse.subgraphs") dask/config.py:459: in get result = result[k] E KeyError: 'optimization' ERROR collecting dask/dataframe/tests/test_indexing.py dask/dataframe/tests/test_indexing.py:20: in <module> full = d.compute() dask/base.py:166: in compute (result,) = compute(self, traverse=False, **kwargs) dask/base.py:438: in compute dsk = collections_to_dsk(collections, optimize_graph, **kwargs) dask/base.py:217: in collections_to_dsk _opt_list.append(opt(_graph_and_keys[0], _graph_and_keys[1], **kwargs)) dask/dataframe/optimize.py:28: in optimize fuse_subgraphs = config.get("optimization.fuse.subgraphs") dask/config.py:459: in get result = result[k] E KeyError: 'optimization' ERROR collecting dask/dataframe/tests/test_shuffle.py [Truncated] full = d.compute() dask/base.py:166: in compute (result,) = compute(self, traverse=False, **kwargs) dask/base.py:438: in compute dsk = collections_to_dsk(collections, optimize_graph, **kwargs) dask/base.py:217: in collections_to_dsk _opt_list.append(opt(_graph_and_keys[0], _graph_and_keys[1], **kwargs)) dask/dataframe/optimize.py:28: in optimize fuse_subgraphs = config.get("optimization.fuse.subgraphs") dask/config.py:459: in get result = result[k] E KeyError: 'optimization' ``` **Environment**: - Dask version: 2.17.0 - Python version: 3.8 - Operating System: Alpine Linux - Install method: source Answers: username_1: Duplicate of https://github.com/dask/dask/issues/6221. We'll probably have a bugfix release later today. Status: Issue closed
backend-br/vagas
937257968
Title: [São Paulo - SP] Desenvolvedor Java - Pleno / Sênior Question: username_0: ## Descrição da vaga Selecionamos para agência de tecnologia da informação de São Paulo - Nível de experiência: Sênior - Setor: Tecnologia da Informação – Serviços - Local de trabalho: Vila Olímpia – SP - Regime: 75% on-site e 25% home office ## Local **São Paulo - SP** ## Requisitos: - Estudos - Idiomas - Conhecimentos em Front End com Javascript. ## Responsabilidades: O Desenvolvedor Java irá atuar no desenvolvimento e manutenção de sistemas, utilizando a linguagem Java, em parceria com demais desenvolvedores da equipe, buscando sempre a excelência e melhora contínua nas entregas das soluções. ## Requisitos necessários: - Formação superior completa em Ciência da Computação, Engenharia da Computação ou áreas correlatas; - Sólidos conhecimentos em Programação com Java EE e JSF; - Padrão MVC; - Integrações – REST, SOAP, WebHook Banco de dados; - Conhecimentos em banco de dados Oracle, SQL Server, My SQL; - Conhecimentos em servidor de aplicação (Weblogic, JBoss, Glassfish); - Conhecimentos em Spring Boot. ## Requisitos Diferenciais: - Instalação/Administração do Weblogic; - Instalação/Administração de Banco de Dados Oracle; ## Ambiente de Trabalho: - Dress code free; - Flexibilidade de horário; - Sala de descompressão com vídeo game; - Máquina Café, bebidas, salgados, doces (livre consumo); ## Contratação **CLT ou PJ a combinar pretensão salarial** ## Como se candidatar **Por favor envie um email informando sua pretensão salarial para <EMAIL> com seu CV anexado - enviar no assunto: Vaga Desenvolvedor Java - Pleno / Sênior** **Ou candidatar-se através do link:** https://jobs.recrutei.com.br/grupo-dream-work/vacancy/11500-dw1110-desenvolvedor-a-mobile-flutter #### Nível - Pleno - Sênior
nltk/nltk
938887533
Title: 3.6.2: sphinx apidoc integration Question: username_0: For more information see: https://www.nltk.org/data.html Attempted to load misc/perluniprops Searched in: - '/home/tkloczko/nltk_data' - '/usr/nltk_data' - '/usr/share/nltk_data' - '/usr/lib/nltk_data' - '/usr/share/nltk_data' - '/usr/local/share/nltk_data' - '/usr/lib/nltk_data' - '/usr/local/lib/nltk_data' ********************************************************************** /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/tokenize/sonority_sequencing.py:docstring of nltk.tokenize.sonority_sequencing:17: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/tokenize/sonority_sequencing.py:docstring of nltk.tokenize.sonority_sequencing:18: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/tokenize/treebank.py:docstring of nltk.tokenize.treebank.TreebankWordDetokenizer:6: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/tokenize/treebank.py:docstring of nltk.tokenize.treebank.TreebankWordDetokenizer:7: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/bleu_score.py:docstring of nltk.translate.bleu_score.SmoothingFunction.method3:6: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/bleu_score.py:docstring of nltk.translate.bleu_score.SmoothingFunction.method3:8: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/bleu_score.py:docstring of nltk.translate.bleu_score.brevity_penalty:74: WARNING: Field list ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm1.py:docstring of nltk.translate.ibm1:25: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm1.py:docstring of nltk.translate.ibm1:34: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm1.py:docstring of nltk.translate.ibm1:35: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm1.py:docstring of nltk.translate.ibm1:37: WARNING: Definition list ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm2.py:docstring of nltk.translate.ibm2:9: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm2.py:docstring of nltk.translate.ibm2:11: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm2.py:docstring of nltk.translate.ibm2:12: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm2.py:docstring of nltk.translate.ibm2:21: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm2.py:docstring of nltk.translate.ibm2:22: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm2.py:docstring of nltk.translate.ibm2:24: WARNING: Definition list ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm3.py:docstring of nltk.translate.ibm3:24: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm3.py:docstring of nltk.translate.ibm3:26: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm3.py:docstring of nltk.translate.ibm3:27: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm3.py:docstring of nltk.translate.ibm3:44: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm3.py:docstring of nltk.translate.ibm3:45: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm3.py:docstring of nltk.translate.ibm3:47: WARNING: Definition list ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm3.py:docstring of nltk.translate.ibm3:53: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm3.py:docstring of nltk.translate.ibm3:54: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm4.py:docstring of nltk.translate.ibm4:18: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm4.py:docstring of nltk.translate.ibm4:20: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm4.py:docstring of nltk.translate.ibm4:42: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm4.py:docstring of nltk.translate.ibm4:45: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm4.py:docstring of nltk.translate.ibm4:50: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm4.py:docstring of nltk.translate.ibm4:52: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm4.py:docstring of nltk.translate.ibm4:53: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm4.py:docstring of nltk.translate.ibm4:71: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm4.py:docstring of nltk.translate.ibm4:72: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm4.py:docstring of nltk.translate.ibm4:74: WARNING: Definition list ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm4.py:docstring of nltk.translate.ibm4:80: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm4.py:docstring of nltk.translate.ibm4:81: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm5.py:docstring of nltk.translate.ibm5:25: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm5.py:docstring of nltk.translate.ibm5:33: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm5.py:docstring of nltk.translate.ibm5:43: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm5.py:docstring of nltk.translate.ibm5:44: WARNING: Block quote ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm5.py:docstring of nltk.translate.ibm5:46: WARNING: Enumerated list ends without a blank line; unexpected unindent. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm5.py:docstring of nltk.translate.ibm5:50: WARNING: Unexpected indentation. /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/ibm5.py:docstring of nltk.translate.ibm5:52: WARNING: Unexpected indentation. [Truncated] looking for now-outdated files... none found pickling environment... done checking consistency... done writing... python-nltk.3 { news install data contribute api/nltk api/nltk.app api/nltk.ccg api/nltk.chat api/nltk.chunk api/nltk.classify api/nltk.cluster api/nltk.corpus api/nltk.corpus.reader api/nltk.draw api/nltk.inference api/nltk.lm api/nltk.metrics api/nltk.misc api/nltk.parse api/nltk.sem api/nltk.sentiment api/nltk.stem api/nltk.tag api/nltk.tbl api/nltk.test api/nltk.test.unit api/nltk.test.unit.lm api/nltk.test.unit.translate api/nltk.tokenize api/nltk.translate api/nltk.twitter } /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/corpus/reader/wordnet.py:docstring of nltk.corpus.reader.wordnet.WordNetCorpusReader.ic:: WARNING: more than one target found for cross-reference 'CorpusReader': nltk.corpus.reader.api.CorpusReader, nltk.corpus.reader.CorpusReader /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/inference/discourse.py:docstring of nltk.inference.discourse.DiscourseTester.multiply:: WARNING: more than one target found for cross-reference 'Expression': nltk.sem.linearlogic.Expression, nltk.sem.logic.Expression /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/inference/discourse.py:docstring of nltk.inference.discourse.ReadingCommand.combine_readings:: WARNING: more than one target found for cross-reference 'Expression': nltk.sem.linearlogic.Expression, nltk.sem.logic.Expression /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/inference/discourse.py:docstring of nltk.inference.discourse.ReadingCommand.combine_readings:: WARNING: more than one target found for cross-reference 'Expression': nltk.sem.linearlogic.Expression, nltk.sem.logic.Expression /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/inference/discourse.py:docstring of nltk.inference.discourse.ReadingCommand.process_thread:: WARNING: more than one target found for cross-reference 'Expression': nltk.sem.linearlogic.Expression, nltk.sem.logic.Expression /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/inference/discourse.py:docstring of nltk.inference.discourse.ReadingCommand.process_thread:: WARNING: more than one target found for cross-reference 'Expression': nltk.sem.linearlogic.Expression, nltk.sem.logic.Expression /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/inference/discourse.py:docstring of nltk.inference.discourse.ReadingCommand.to_fol:: WARNING: more than one target found for cross-reference 'Expression': nltk.sem.linearlogic.Expression, nltk.sem.logic.Expression /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/inference/discourse.py:docstring of nltk.inference.discourse.ReadingCommand.to_fol:: WARNING: more than one target found for cross-reference 'Expression': nltk.sem.linearlogic.Expression, nltk.sem.logic.Expression /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/inference/discourse.py:docstring of nltk.inference.discourse.load_fol:: WARNING: more than one target found for cross-reference 'Expression': nltk.sem.linearlogic.Expression, nltk.sem.logic.Expression /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/sem/evaluate.py:docstring of nltk.sem.evaluate.Model.satisfiers:: WARNING: more than one target found for cross-reference 'Expression': nltk.sem.linearlogic.Expression, nltk.sem.logic.Expression /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/meteor_score.py:docstring of nltk.translate.meteor_score.allign_words:: WARNING: more than one target found for cross-reference 'WordNetCorpusReader': nltk.corpus.reader.wordnet.WordNetCorpusReader, nltk.corpus.reader.WordNetCorpusReader /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/meteor_score.py:docstring of nltk.translate.meteor_score.meteor_score:: WARNING: more than one target found for cross-reference 'WordNetCorpusReader': nltk.corpus.reader.wordnet.WordNetCorpusReader, nltk.corpus.reader.WordNetCorpusReader /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/meteor_score.py:docstring of nltk.translate.meteor_score.single_meteor_score:: WARNING: more than one target found for cross-reference 'WordNetCorpusReader': nltk.corpus.reader.wordnet.WordNetCorpusReader, nltk.corpus.reader.WordNetCorpusReader /home/tkloczko/rpmbuild/BUILD/nltk-3.6.2/nltk/translate/meteor_score.py:docstring of nltk.translate.meteor_score.wordnetsyn_match:: WARNING: more than one target found for cross-reference 'WordNetCorpusReader': nltk.corpus.reader.wordnet.WordNetCorpusReader, nltk.corpus.reader.WordNetCorpusReader done build succeeded, 344 warnings. ``` Answers: username_1: Hi, I have generated the NLTK api docs with [pydoctor](https://github.com/twisted/pydoctor/) here: https://username_1.github.io/apidocs/nltk/nltk.html Pydoctor works by parsing the AST of the code, so it does not require you to actually install the package you want to document to generate the API docs. This might solve your issues here. I find it pretty useful. It does document the whole package automatically, though. In order to remove several low-level packages from the output one should finally fix https://github.com/twisted/pydoctor/issues/36. Maybe pydoctor could be useful for you ?
RunnersHi-Release/RunnersHi_Develop_iOS
715133359
Title: readme image Question: username_0: <img width="446" alt="splashView" src="https://user-images.githubusercontent.com/51286963/95124979-44beba00-078f-11eb-8f1a-d6bf2843a816.png"> ![runnershi](https://user-images.githubusercontent.com/51286963/95125361-d4646880-078f-11eb-9ae1-f55cbc62b208.jpg) <img width="207" alt="running" src="https://user-images.githubusercontent.com/51286963/95125562-1f7e7b80-0790-11eb-8a1b-81284363d3bd.png"> <img width="262" alt="record" src="https://user-images.githubusercontent.com/51286963/95125767-67050780-0790-11eb-8c6c-c090acc50a5d.png"> <img width="560" alt="CoreData" src="https://user-images.githubusercontent.com/51286963/95127024-4dfd5600-0792-11eb-8aed-3830bec840b6.png">
fotosyn/berrycam
765439253
Title: Not accessible from the internet? Question: username_0: Now, in actual version (debugged for iPhone SE and 8) I can access Pi's camera from local network, but not from internet as it was possible with the older version. Answers: username_0: Autostart the script after rebooting the Pi: `sudo nano /etc/rc.local` before line `exit 0` insert: `sudo nohup python3 /home/pi/berrycam/berryCam.py > berryCam.log & tail -f berryCam.log &` press ^X press Y username_1: Thanks @username_0 yep that's pretty much what I've been doing here as well with rc.local Can you give me more detail <EMAIL> in relation to your comments re accessing Pis outside of local network? If anything I would think port forwarding would still be working? username_0: Any news about the internet access problem? username_1: Still to investigate. I don't have a Pi set up with an external IP address at the moment but will try sometime next week. Status: Issue closed username_1: Hi @username_0 I have checked this out using my internet router and Google Wifi to test port forwarding, and all appears to be working as expected. If I access on my 4G network on my iPhone, connecting to the external port it connects to my Pi on the LAN and allows for image capture. I have tested this particular connection with a couple of friends and they have been able to connect and capture also. I am not entirely up to speed with DynDNS - if you can contact me about this at <EMAIL> I can maybe look a bit more into this. It might be better to discuss this 1-1 for obvious reasons. Merry Christmas -- Jim username_0: Hi Jim, did you find my PM? Best regards and Happy New Year, Hans username_1: Got them now Hans thank you. I'll check this connection later. The app doesn't insert www so it may be an issue with the network connection. Let me have a look 👌 username_1: What port number are you using? Is it 20? username_0: 8000
kubernetes/minikube
504802091
Title: Minikube Profile Creation gives the wrong message (Or doesn't work) and listing doesn't work Question: username_0: <!-- Please use this template while reporting an issue, providing as much information as possible. Failure to do so may result in a delayed response. Thank you! --> On a machine which doesn't have any minikube related configuration or data, ran the profile command. **The exact command to reproduce the issue**: ``` PS C:\utilities> .\minikube-windows-amd64.exe profile pranav PS C:\utilities> .\minikube-windows-amd64.exe profile abc PS C:\utilities> .\minikube-windows-amd64.exe profile list ``` **The full output of the command that failed**:<details> ``` PS C:\utilities> .\minikube-windows-amd64.exe profile pranav * Created a new profile : pranav * minikube profile was successfully set to pranav PS C:\utilities> .\minikube-windows-amd64.exe profile abc * Created a new profile : abc * Error while setting kubectl current context : writing kubeconfig: Error writing file C:\Users\bluee/.kube/config: error writing file C:\Users\bluee/.kube/config: open C:\Users\bluee/.kube/config: The system cannot find the path specified. * minikube profile was successfully set to abc PS C:\utilities> .\minikube-windows-amd64.exe profile list * No minikube profile was found. You can create one using `minikube start`. PS C:\utilities> ``` </details> **The operating system version**: Windows 10 Pro Do note that after I used `minikube start`, it did display that it was on the `abc` profile and the cluster started successfully. After the cluster started, I re-ran the `minikube profile list` command and got the following - ``` PS C:\utilities> .\minikube-windows-amd64.exe profile list |---------|-----------|-----------------|-----------|--------------------| | Profile | VM Driver | NodeIP | Node Port | Kubernetes Version | |---------|-----------|-----------------|-----------|--------------------| | abc | hyperv | 192.168.111.228 | 8443 | v1.16.0 | |---------|-----------|-----------------|-----------|--------------------| ! Found 1 invalid profile(s) ! * pranav * You can delete them using the following command(s): $ minikube delete -p pranav ``` This seems to be incorrect. As you can see, despite the fact that there was an error while creating the `abc` profile, the invalid profile listed is `pranav`. -Pranav Answers: username_0: /remove-lifecycle rotten username_1: Hey @username_0 it looks like you're running minikube v1.4.0. I'd recommend upgrading to minikube [v1.10.1](https://github.com/kubernetes/minikube/releases/tag/v1.10.1) which seems to fix this issue on my laptop. Would you mind upgrading and letting us know if that fixes it? username_1: Hopefully it's OK if I close this - there wasn't enough information to make it actionable, and some time has already passed. If you are able to provide additional details, you may reopen it at any point by adding /reopen to your comment. Here is additional information that may be helpful to us: * Whether the issue occurs with the latest minikube release Thank you for sharing your experience! Status: Issue closed
docker/for-mac
212170118
Title: "readlink /var/lib/docker/overlay2/l: invalid argument" occurs after some time Question: username_0: ### Expected behavior `docker -D run 5d35104e2b93` actually runs my built container ### Actual behavior ``` docker -D run 5d35104e2b93 docker: Error response from daemon: ContainerInspect failed. ERRO[0000] error getting events from daemon: net/http: request canceled DEBU[0000] [hijack] End of stdout ``` ### Information ID: 04047E22-3B4B-4EF2-A2F8-458DE6099421 Docker for Mac: version: 17.03.0-ce-mac1 (e5a07a1a6) macOS: version 10.12.3 (build: 16D32) logs: /tmp/04047E22-3B4B-4EF2-A2F8-458DE6099421/20170306-155325.tar.gz [OK] vmnetd [OK] dns [OK] driver.amd64-linux [OK] virtualization VT-X [OK] app [OK] moby [OK] system [OK] moby-syslog [OK] db [OK] env [OK] virtualization kern.hv_support [OK] slirp [OK] osxfs [OK] moby-console [OK] logs [OK] docker-cli [OK] menubar [OK] disk Docker logs: ``` Mar 6 16:01:30 moby root: time="2017-03-06T16:01:30.937928419Z" level=debug msg="Calling GET /containers/84cb59b8c91669702216ef5f7a61f004359b97f97da12072d2496d3dddf07851/json" Mar 6 16:01:30 moby root: time="2017-03-06T16:01:30.938103726Z" level=error msg="Handler for GET /containers/84cb59b8c91669702216ef5f7a61f004359b97f97da12072d2496d3dddf07851/json returned error: readlink /var/lib/docker/overlay2/l: invalid argument" ``` ### Steps to reproduce the behavior ``` version: "3" services: test: tty: true privileged: true build: context: . args: - NPM_TOKEN depends_on: - db - dbseed [Truncated] image: mysql:5.6.23 volumes: - data-volume:/var/lib/db image: sebp/lighttpd ports: - 8080:80 volumes: - reports-volume:/var/www/localhost/htdocs:ro volumes: data-volume: reports-volume: ``` `docker-compose up --build test` To actually fix this I have to reset the whole of docker-for-mac and re-download all my images. Thanks Answers: username_1: I am also experiencing the exact same issues. I can post specs and logs if it will help diagnose. username_2: @username_0 I've looked at your logs and unfortunately it looks like your hyperkit process is frequently crashing with exit_reason 2 and 48. This may have corrupted the D4M file system. I'm closing this as duplicate of https://github.com/docker/for-mac/issues/1176 (exit reason 48, ept fault) and https://github.com/docker/for-mac/issues/1170 (exit reason 2, triple fault). Please subscribe to those issues for updates. @username_1 since this issue may have different causes, please run "Diagnose & Upload" and create a new issue with a description of your problem. That makes it easier for us to keep track of the different issues. Thanks! Status: Issue closed
postmanlabs/postman-app-support
428531137
Title: Documentation Publishing Flow: Required field not throwing error when empty Question: username_0: **Describe the bug** The **API Description** field that's surfaced in the API Network publishing flow is required for publishing but doesn't throw an error when empty. **To Reproduce** Steps to reproduce the behavior: 1. Select a collection to publish 2. Check the **Collection discovery** option 3. Fill out the API Name and API Summary 4. Note that the **Publish** action is disabled 5. Fill out the API Description 6. Notice that the **Publish** action is enabled **Expected behavior** When empty, required fields should throw an error. **Screenshots** Without API Description ![image](https://user-images.githubusercontent.com/4734306/55448129-29686280-558c-11e9-9cee-61386a4dc50a.png) With API Description ![image](https://user-images.githubusercontent.com/4734306/55448142-384f1500-558c-11e9-8274-2a7978fe733a.png) **Additional context** [Filed by Postman] Answers: username_1: @username_0 thanks for reporting this. I want to understand you better. From what I understand, users cannot publish without an API Name, API summary, and API description as the 'publish' option is disabled when at least one of these is empty. When exactly do you want the error to be thrown? Do you mean to say API description widget should have a message like - 'API description cannot be empty' username_0: @username_1 Yes – I guess the screenshot didn't highlight the missing error message that should populate to indicate that the **API Description** field is required. Here's a better one :). ![image](https://user-images.githubusercontent.com/4734306/55454424-89b7ce00-55a5-11e9-981e-14f88cd0155b.png) Status: Issue closed username_1: Got it. Thanks, username_1: **Describe the bug** The **API Description** field that's surfaced in the API Network publishing flow is required for publishing but doesn't throw an error when empty. **To Reproduce** Steps to reproduce the behavior: 1. Select a collection to publish 2. Check the **Collection discovery** option 3. Fill out the API Name and API Summary 4. Note that the **Publish** action is disabled 5. Fill out the API Description 6. Notice that the **Publish** action is enabled **Expected behavior** When empty, required fields should throw an error. **Screenshots** Without API Description ![image](https://user-images.githubusercontent.com/4734306/55448129-29686280-558c-11e9-9cee-61386a4dc50a.png) With API Description ![image](https://user-images.githubusercontent.com/4734306/55448142-384f1500-558c-11e9-8274-2a7978fe733a.png) **Additional context** [Filed by Postman] username_2: Hey @sankalp0o, Do you know if this something that was picked up in some recent changes? Status: Issue closed username_3: Closing this ticket as we have removed the collection discovery option from the publish flow.
facebook/react-native
335810315
Title: Dynamic allocation of textinput Question: username_0: **Environment:** OS: macOS High Sierra 10.13.3 Node: 7.5.0 npm: 4.3.0 Watchman: 4.7.0 Xcode: Xcode 9.4.1 Build version 9F2000 Packages: (wanted => installed) react-native: 0.44.3 **Description:** This is my response array now I have to create a textinput or picker from the input_types value can anyone help me with this. I actually tried to do it with dynamic_data.push() but in onchangetext if we change a single field text it's affecting all the text fields. ``` "form_fields": [ { "id": 12, "created_at": "2017-12-18 11:28:29", "updated_at": "2017-12-18 11:28:29", "name": "children_name_service_7", "label": "Name of the Children?", "info": null, "label_for_search_form": "Name of the Children?", "length": null, "options": null, "class": "Service", "input_type_id": 1, "foreign_id": 7, "form_field_group_id": 5, "is_required": 1, "is_active": 1, "display_order": 1, "depends_on": null, "depends_value": null, "is_enable_this_field_in_search_form": 0, "input_types": { "id": 1, "created_at": "2017-11-23 12:07:45", "updated_at": "2017-11-23 12:07:45", "name": "Single Line of Text", "value": "textInput" } }, { "id": 13, "created_at": "2017-12-18 11:28:29", "updated_at": "2017-12-18 11:28:29", "name": "age_of_the_child_service_7", "label": "Age of the Child", "info": null, "label_for_search_form": "Age", "length": null, "options": null, "class": "Service", "input_type_id": 1, "foreign_id": 7, "form_field_group_id": 5, "is_required": 1, "is_active": 1, [Truncated] "input_type_id": 3, "foreign_id": 7, "form_field_group_id": 5, "is_required": 1, "is_active": 1, "display_order": 3, "depends_on": null, "depends_value": null, "is_enable_this_field_in_search_form": 0, "input_types": { "id": 3, "created_at": "2017-11-23 12:07:45", "updated_at": "2017-11-23 12:07:45", "name": "Select Box", "value": "select" } } ] ```
NicoleKMarino/the_pivot
171282218
Title: Employer Can Create Account Answers: username_1: As a an employer when I visit /login and I click the link "Create Account" I should be redirected to a form and I fill out username and I fill out <PASSWORD> and I fill out email address and I fill out company description and click submit it should alert me that account was succesfully created it should redirect me to employer dashboard I should see "Add a job" in nav bar if I leave out any of these fields it should alert me that account was not successfully created Status: Issue closed
gradle/gradle
808100773
Title: Substitutions configured via eachDependency perform poorly with in parallel Question: username_0: When configurations are resolving in parallel coarse synchronization in `org.gradle.cache.internal.DefaultCrossBuildInMemoryCacheFactory$AbstractCrossBuildInMemoryCache.getIfPresent` on the hot path for `DependencyResolveDetailsWrapperAction` causes excessive blocking: <img width="1433" alt="Screen Shot 2021-02-12 at 4 42 29 pm" src="https://user-images.githubusercontent.com/1479220/107895665-42b00780-6f88-11eb-88aa-2ab94d53db2e.png"> That cache appears unsuitable on that path: ``` org.gradle.cache.internal.DefaultCrossBuildInMemoryCacheFactory$AbstractCrossBuildInMemoryCache.getIfPresent(Object) DefaultCrossBuildInMemoryCacheFactory.java org.gradle.internal.instantiation.generator.AbstractClassGenerator.generate(Class) AbstractClassGenerator.java:163 org.gradle.internal.instantiation.generator.AsmBackedClassGenerator.generate(Class) AsmBackedClassGenerator.java:130 org.gradle.internal.instantiation.generator.ParamsMatchingConstructorSelector.forParams(Class, Object[]) ParamsMatchingConstructorSelector.java:44 org.gradle.internal.instantiation.generator.DependencyInjectingInstantiator.doCreate(Class, Describable, Object[]) DependencyInjectingInstantiator.java:61 org.gradle.internal.instantiation.generator.DependencyInjectingInstantiator.newInstance(Class, Object[]) DependencyInjectingInstantiator.java:55 org.gradle.api.internal.artifacts.ivyservice.dependencysubstitution.DefaultDependencySubstitutions$DependencyResolveDetailsWrapperAction.execute(DependencySubstitution) DefaultDependencySubstitutions.java:475 org.gradle.api.internal.artifacts.ivyservice.dependencysubstitution.DefaultDependencySubstitutions$DependencyResolveDetailsWrapperAction.execute(Object) DefaultDependencySubstitutions.java:459 org.gradle.internal.Actions$CompositeAction.execute(Object) Actions.java:116 org.gradle.api.internal.artifacts.ivyservice.dependencysubstitution.DefaultDependencySubstitutionApplicator.apply(DependencyMetadata) DefaultDependencySubstitutionApplicator.java:43 org.gradle.api.internal.artifacts.ivyservice.dependencysubstitution.CachingDependencySubstitutionApplicator.apply(DependencyMetadata) CachingDependencySubstitutionApplicator.java:44 org.gradle.api.internal.artifacts.ivyservice.resolveengine.graph.builder.NodeState.maybeSubstitute(DependencyState, DependencySubstitutionApplicator) NodeState.java:616 org.gradle.api.internal.artifacts.ivyservice.resolveengine.graph.builder.NodeState.visitDependencies(ExcludeSpec, Collection) NodeState.java:435 ``` ### Context We're trying to switch away from `dependencySubstitutions.all` due to it having to make the assumption that projects might be affected and pushing all configuration dependency resolution for tasks to task graph execution time. ### Steps to Reproduce - Configure substitutions via eachDependency - Resolve two or more configurations in parallel ### Your Environment Gradle 6.8.1 Answers: username_0: Resolving all resolvable configurations on 64 subprojects w/ 8 workers: `dependencySubstitutions.all`: ``` BUILD SUCCESSFUL in 25s ``` `eachDependency`: ``` BUILD SUCCESSFUL in 49s ```
balena-io/balena-cli
396866836
Title: Build log truncation is limited incorrectly Question: username_0: Problem ====== The API has a limit of 512kb when sending requests. The CLI, when updating the image resource after a build, only checks that a single field is less than the 512kb limit; not the whole JSON content. This can cause a `413 PayloadTooLargeError` and a failed deployment. Code in question is here: https://github.com/balena-io/balena-cli/blob/b0da1b4811ebd6664f4149e4c54e52ee846c3a7e/lib/utils/compose.coffee#L282 Suggested fix ========== Before sending the request, `JSON.stringify()` the payload and get the length. If the length is over the maximum, trim the extra amount off **the start** of the build log. The start of the log is less important than the end. This should result in the request being within the size constraints.
pombase/fypo
121317267
Title: PMID:24623719 Question: username_0: spores sensitive to heat spores sensitive to ethanol spores sensitive to cell wall degrading enzymes (glusulase) decreased protein localization to spore wall Answers: username_1: spores sensitive to heat FYPO:0005130 spores sensitive to ethanol FYPO:0005131 spores sensitive to cell wall-degrading enzymes FYPO:0005132 decreased protein localization to ascospore wall FYPO:0005133 username_1: edit file: 09c94cb804bc5cac89400e1cb789c445c78c855c release: ee9e5553f11629ab4d6e12c7b384126cd7b78d5d Status: Issue closed
aws-amplify/amplify-cli
705005250
Title: amplify publish - fails to upload assets to s3 Question: username_0: [Amplify Console GitHub Issue Tracker](https://github.com/aws-amplify/amplify-console/issues) **Describe the bug** After adding hosting by `amplify add hosting` when `amplify publish` was run below error occurred saying ``` Cannot find the distribution folder. Error: Cannot find the distribution folder. ``` **Amplify CLI Version** You can use `amplify -v` to check the amplify cli version on your system subratas-mbp  ~/workspace/react-amplified   master ●  amplify -v 4.29.2 subratas-mbp  ~/workspace/react-amplified   master ●  **To Reproduce** Steps to reproduce the behavior or terminal output if applicable Perform these [steps](https://docs.amplify.aws/start/getting-started/hosting/q/integration/react) **Expected behavior** S3 upload should happen without any issues. **Screenshots** See error details below ``` npm WARN build `npm build` called with no arguments. Did you mean to `npm run-script build`? frontend build command exited with code 0 Publish started for S3AndCloudFront Cannot find the distribution folder. Distribution folder is currently set as: /Users/username_0/workspace/react-amplified/build Cannot find the distribution folder. Error: Cannot find the distribution folder. at Object.scan (/usr/local/lib/node_modules/@aws-amplify/cli/node_modules/amplify-category-hosting/lib/S3AndCloudFront/helpers/file-scanner.js:38:11) ``` **Desktop (please complete the following information):** - OS: [e.g. Mac/Windows/Ubuntu] Mac - Node Version. You can use `node -v` to check the node version on your system subratas-mbp  ~/workspace/react-amplified   master ●  node -v v12.18.0 subratas-mbp  ~/workspace/react-amplified   master ●  **Additional context** **Workaround** As of now the workaround is to run `npm run build` to create this missing `build` directory. And then run `amplify publish` to re-upload all `build\*` to configured s3 bucket. **Probable Root Cause** : `amplify publish` is running `npm build` just before uploading s3 content, should have run `npm run build` instead. Answers: username_1: The build command by default is "npm run-script build" and they are set during `amplify init`. You can still modify it after `amplify init` by running `amplify configure project` Status: Issue closed
Xavalon/XamlStyler
212383056
Title: Ability to locate external configuration automatically Question: username_0: It would be nice, if the location of the external configuration could be done automatically by searching them from the current folder until the drive root (like `stylecop.json`). So it is easily possible to work in different branches and developers don't have to setup anything. So, if possible, then: 1. Project level 2. Solution level 3. Development dircetory level 4. Until root of the hard drive If that could not be done easily, then at least a check for `xamlstyler.json` at the solution level would be great. Answers: username_1: This feature is already available, though the documentation has not been updated (will keep this issue open to track updating documentation). Drop "Settings.XamlStyler" in any directory up through your solution directory and both the plugin and standalone will check for that. username_0: Awesome! Thank you! username_1: Just updated wiki with [a few important details](https://github.com/Xavalon/XamlStyler/wiki/External-Configurations) about this feature. Status: Issue closed
apolukhin/Boost-Cookbook
386536892
Title: Spirit examples using `auto` Question: username_0: Looks like there's an issue with replacing SPIRIT rules with auto. Make a minimized example and report the issue (or fix the examples, if there's no issue). Add CI tests. Answers: username_0: Issue was reported https://github.com/boostorg/spirit/issues/436 Status: Issue closed username_0: Added a correct usage example in https://github.com/username_0/Boost-Cookbook/commit/bfaebcacbaeddb7ae3d7135f39edf03fa574882c
filecoin-project/lotus
944334771
Title: [BUG] docker build error Question: username_0: **Describe the bug** build source code in docker . network: calibnet Dockerfile: ``` FROM golang:1.16.4 AS builder-env RUN rm -rf /etc/apt/sources.list && \ echo "deb http://mirrors.aliyun.com/debian buster main" >>/etc/apt/sources.list && \ echo "deb http://mirrors.aliyun.com/debian-security buster/updates main" >>/etc/apt/sources.list && \ echo "deb http://mirrors.aliyun.com/debian buster-updates main" >>/etc/apt/sources.list && \ apt-get update && apt-get install -y ca-certificates build-essential clang ocl-icd-opencl-dev ocl-icd-libopencl1 jq libhwloc-dev ENV RUSTUP_DIST_SERVER=https://mirrors.ustc.edu.cn/rust-static ENV RUSTUP_UPDATE_ROOT=https://mirrors.ustc.edu.cn/rust-static/rustup ENV IPFS_GATEWAY=https://proof-parameters.s3.cn-south-1.jdcloud-oss.com/ipfs/ ENV GOPROXY=https://goproxy.cn ENV RUSTFLAGS="-C target-cpu=native -g" ENV FFI_BUILD_FROM_SOURCE=1 ADD rust.sh /go RUN chmod +x rust.sh; \ ./rust.sh -y; \ rm rust.sh; \ /bin/bash -c "source $HOME/.cargo/env" ENV PATH="/root/.cargo/bin:${PATH}" WORKDIR / ADD ./lotus / ADD config /root/.cargo/config RUN make calibnet install lotus-bench lotus-miner lotus-worker ``` **Logs** make[1]: Leaving directory '/extern/filecoin-ffi' rm -f lotus go build -ldflags="-X=github.com/filecoin-project/lotus/build.CurrentCommit=+git.dd0966639" -tags=calibnet -o lotus ./cmd/lotus directory cmd/lotus outside available modules make: *** [Makefile:88: lotus] Error 1 The command '/bin/sh -c make calibnet install lotus-bench lotus-miner lotus-worker' returned a non-zero code: 2<issue_closed> Status: Issue closed
Tux/App-ccdiff
701049383
Title: Wishlist: Add support for GNU diff's -r (--recursive) and -p (--show-c-function) options Question: username_0: Hi, It would be great if ccdiff supported more of GNU diff's options (i.e. the ones I use all the time). :-) Particularly -r (--recursive) and -p (--show-c-function). Answers: username_1: `-r / --recursive` was on my own wishlist, but as `ccdiff` might be slow on big files, I just didn't implement that (yet) username_1: For `-p`, I would (for now) say: "Patches welcome"
kubernetes-sigs/cluster-api-provider-azure
875638241
Title: MachinePool has more worker nodes than the desired value Question: username_0: /kind bug **What steps did you take and what happened:** [A clear and concise description of what the bug is.] https://prow.k8s.io/view/gs/kubernetes-jenkins/logs/capz-azure-file-machinepool-1-19/1388742406299455488 ``` Waiting for 1 control plane machine(s) and 2 worker machine(s) to become Ready node/capz-h8u4r3-control-plane-5m8mq condition met node/capz-h8u4r3-mp-0000000 condition met node/capz-h8u4r3-mp-0000002 condition met NAME STATUS ROLES AGE VERSION INTERNAL-IP EXTERNAL-IP OS-IMAGE KERNEL-VERSION CONTAINER-RUNTIME capz-h8u4r3-control-plane-5m8mq Ready master 3m59s v1.19.11-rc.0.23+3ed2eb74c47e2f 10.0.0.4 <none> Ubuntu 18.04.5 LTS 5.3.0-1034-azure containerd://1.3.4 capz-h8u4r3-mp-0000000 Ready <none> 37s v1.19.11-rc.0.23+3ed2eb74c47e2f 10.1.0.4 <none> Ubuntu 18.04.5 LTS 5.3.0-1034-azure containerd://1.3.4 capz-h8u4r3-mp-0000001 NotReady <none> 8s v1.19.11-rc.0.23+3ed2eb74c47e2f 10.1.0.5 <none> Ubuntu 18.04.5 LTS 5.3.0-1034-azure containerd://1.3.4 capz-h8u4r3-mp-0000002 Ready <none> 20s v1.19.11-rc.0.23+3ed2eb74c47e2f 10.1.0.6 <none> Ubuntu 18.04.5 LTS 5.3.0-1034-azure containerd://1.3.4 ``` Eventually, the `NodeReady` node was removed. **What did you expect to happen:** Have no more than the desired MachinePool replicas at any given moment **Anything else you would like to add:** [Miscellaneous information that will assist in solving the issue.] **Environment:** - cluster-api-provider-azure version: master - Kubernetes version: (use `kubectl version`): - OS (e.g. from `/etc/os-release`): Answers: username_1: I believe this is by design because VMSS overprovisioning is enabled by default. The scale set will provision more instances than desired replicas and only keep the first n instances that become ready. cc @username_2 username_2: That's exactly what is happening. How would you like this to behave. It is nice to use over-provisioning to increase overall speed and success rate of VMSS provisioning. username_3: I think it is fine from user standpoint. We found this issue in the e2e tests. In `e2e.test` there is a check to verify all the nodes are ready, which is causing it to hang while waiting to run. One possibility would to validate that we have the expected number of nodes and run `e2e.test` with `--allowed-not-ready-nodes=1` which would allow the tests to proceed.
andres-montanez/Magallanes
63713684
Title: Create repository for documentation Question: username_0: How about to create repo for documentation so people can contribute? I've spend good amount of time by figure how some stuff work and I am happy the little I've learned. Answers: username_1: It's great idea to just **create** some detailed documentation. But why not to do it in this repo? username_0: Sure why not :) What would be the plan? username_1: The thing is everyone's too lazy to start writing :D I'll collect necessary points to acheive and I'll make a new issue with *always-up-to-date* todo list. Then one would do it all or there will be pull requests with chapters (I'd recommend this one, shorter code review). I can't promise but I can prepare the list even today (if I find some spare time). Then you can be first to contribute if you said you found a lot of interesting things in code :) What do you think about such plan? username_1: I've made an issue #213 with list of things to be documented. I think we can close this issue then? Status: Issue closed
FriendsOfSymfony/FOSUserBundle
2921456
Title: Avoid doing the same query on each page Question: username_0: Hi, Currently, using FOSUserBundle, a query is done on each page when the user is logged in, to retrieve its informations. Wouldn't it be possible to get this information in the session? Or can we do that ourselves? Is there already discussions related to this? Because I'd like to have my sessions stored in memory, would be great if this query could be avoided and request the session instead. Thanks!<issue_closed> Status: Issue closed
code4craft/webmagic
236692825
Title: 参考指南: 配置代理是否有过更改 Question: username_0: 请问作者http://webmagic.io/中 **配置代理** 一节是否有过更改,之前看过似乎用的是开源中国某个类库且有过瑕疵,现在是弃用那个库了?能否再发下地址 Answers: username_1: 有,0.7.x版本后代理部分重构了,之前的代理文档可以看文档:0.6.1版文档: [http://webmagic.io/archive/docs/0.6.1](http://webmagic.io/archive/docs/0.6.1) 不过0.6.x版本代理比较繁琐也不太稳定,建议用0.7以后的版本。 username_0: 多谢! Status: Issue closed
adobe/react-spectrum
944532604
Title: useOverlay, useInteractOutside on iOS12 causes modals to disappear instantly Question: username_0: <!--- Thanks for filing an issue 😄! Before you submit, please read the following: Search open/closed issues before submitting since someone might have asked the same thing before! --> # 🐛 Bug Report <!--- Provide a general summary of the issue here --> As title says, on iOS 12 when using `useOverlay` (and `useInteractOutside` under the hood, which I believe might be a source of issue) the modal (or any other dialog basically) disappears instantly as presented on the movie attached. It can be easily reproduced by launching docs of `useDialog` in BrowserStack, on Safari 12 https://user-images.githubusercontent.com/26162660/125646656-35dee1b8-3b24-4603-8700-978fb7e77bee.mp4 ## 🤔 Expected Behavior <!--- Tell us what should happen --> The dialog should not close when appearing. ## 😯 Current Behavior <!--- Tell us what happens instead of the expected behavior. Please include a link to the storybook if possible. --> Dialog closes instantly after appearing. <!--- If you are seeing an error, please include the full error message and stack trace --> ## 💁 Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug --> ## 🔦 Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> Even though the market share of iOS 12 is pretty low, the issue is quite severe - it completely disables the usage of dialog. ## 💻 Code Sample <!-- Please provide a code repository, gist, code snippet or sample files to reproduce the issue --> <!-- You can use this codesandbox template to get started https://codesandbox.io/s/react-spectrum-template-syueo --> ## 🌍 Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> | Software | Version(s) | | ---------------- | ---------- | | react-spectrum | 1.2.2 (current latest) | Browser | iOS Safari 12 | Operating System | iOS 12 ## 🧢 Your Company/Team <!--- Which product team is this bug impacting? (i.e. Adobe/Photoshop) --> [livechat.com](https://www.livechat.com/) ## 🕷 Tracking Issue (optional) <!--- Link to the issue in your system, if publicly available --> Answers: username_1: Just tested in Simulator on iOS 12.0 and wasn't able to reproduce. Tested on both iPhone 5s and iPhone X sims: <img width="459" alt="Screen Shot 2021-07-14 at 11 42 59 AM" src="https://user-images.githubusercontent.com/8961049/125660634-238f03db-b174-4ab4-aa1b-b4e8471415ca.png"> username_2: We'd be happy to accept a fix for this but we won't be picking this up ourselves due to the limited area of effect (only iOS 12 + iPhone Xs) of this bug. Status: Issue closed username_3: Found this issue (or something similar) in the new useSelect on an ipad v 12.5.5 username_4: @username_3 do you have any updates to how to reproduce? I'm not seeing it on my iPhone. though admittedly i am on much different versions Would you be willing to do any debugging? We don't have much to go off unless you have some new repro steps/videos you'd be willing to share. username_3: @username_4 yes ive just finished creating a new issue for this here: https://github.com/adobe/react-spectrum/issues/2686 its different enough, i think, to this issue. username_4: @username_3 o, fabulous, thank you!! username_3: @username_4 glad to help! i think this issue isnt overly specific, since it was the first browserstack version i tried with iOS 12 and stumbled on the bug. can confirm in local with iOS 14 everything was fine
jlippold/tweakCompatible
340856326
Title: `App Admin` not working on iOS 11.3.1 Question: username_0: ``` { "packageId": "com.unlimapps.uaupdatetools", "action": "notworking", "userInfo": { "arch32": false, "packageId": "com.unlimapps.uaupdatetools", "deviceId": "iPhone7,2", "url": "http://cydia.saurik.com/package/com.unlimapps.uaupdatetools/", "iOSVersion": "11.3.1", "packageVersionIndexed": true, "packageName": "App Admin", "category": "Tweaks", "repository": "BigBoss", "name": "App Admin", "packageIndexed": true, "packageStatusExplaination": "This package version has been marked as Not working based on feedback from users in the community. The current positive rating is 0% with 0 working reports.", "id": "com.unlimapps.uaupdatetools", "commercial": false, "packageInstalled": true, "tweakCompatVersion": "0.0.7", "shortDescription": "Downgrade apps to any previous version! No AppSync needed!!!", "latest": "1.0r-61", "author": "UnlimApps Inc.", "packageStatus": "Not working" }, "base64": "<KEY> "chosenStatus": "not working", "notes": "" } ```
atom/language-javascript
182817110
Title: Long lines with many segments screw up highlighting Question: username_0: Example: https://github.com/username_0/Bootstraped-Addon-Backbone/blob/master/resource/modules/utils/Keysets.jsm Open that file in Atom and make sure its Javascript syntax is highlighted. From line 119 onwards, the highlighting is all screwed up. I can "fix" it by either commenting it out or by adding a new line after ``` 'S',``` and another after ```'/',```. Lines 122 and 127 have similar problems. Status: Issue closed Answers: username_1: This is a duplicate of atom/atom#1667 - feel free to subscribe there for updates.
emberjs/ember.js
106057196
Title: Observing array using `array.[]` is not working Question: username_0: Reproduced here: http://ember-twiddle.com/b663e28c60942f4c3d25 Answers: username_1: nothing appears to be consuming `allValues` doing so makes everything work. This appears to be working as expected? username_1: also `'allValues', 'allValues.[]', ` is redundant, observing `'allValues.[]'` is a superset of `'allValues'` username_0: @username_1 `valuesDidChange` is consuming `allValues`. username_0: Is this a case of too aggressive optimization? I have an observer observing a computed property, is that not enough to trigger a consumption? username_0: A simpler example: http://ember-twiddle.com/b663e28c60942f4c3d25 username_1: yes, but `valuesDidChange` only fires, if `allValues` has already been consumed and then changes. username_0: @username_1 So, the takeaway is you *cannot* observe a computed property unless that computed property is consumed elsewhere? username_1: Yes correct, think of them as Schrödingers Computed Properties. Before consumed, the state is undefined, and as such no state change notification is possible. The laziness invariant is actually quite nice, but results in the single caveat we just discussed. username_0: @username_1 So what's the correct way to do this then? Am I going to have to `<div class="hidden">{{allValues}}</div>` to get this to work? username_1: @username_0 or you can call `this.get('values')` in `init` Status: Issue closed
toumorokoshi/transmute-core
230418510
Title: add benchmarking Question: username_0: It would be great to have benchmarking on some of the serialization routes: just to be able to do comparisons of serialization frameworks and technologies. Answers: username_0: this now exists in master, and runs through the full functions that transmute performs, including the validation by schematics: ``` ----------------------------------------------------------------------------------- benchmark: 3 tests ---------------------------------------------------------------------------------- Name (time in us) Min Max Mean StdDev Median IQR Outliers(*) Rounds Iterations ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- test_simple_benchmark 8.1062 (1.0) 12,601.8524 (1.03) 14.6850 (1.0) 51.9610 (1.0) 12.1593 (1.0) 1.1921 (1.0) 333;16469 102301 1 test_complex_benchmark 308.9905 (38.12) 49,899.1013 (4.07) 590.9405 (40.24) 1,980.9461 (38.12) 412.9410 (33.96) 167.6679 (140.65) 9;65 1231 1 test_large_str_benchmark 570.0588 (70.32) 12,270.9274 (1.0) 808.3725 (55.05) 1,007.5184 (19.39) 622.9877 (51.24) 154.9721 (130.00) 29;113 1580 1 ----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- ``` The "simple_benchmark" only does the work that transmute is performing, and that looks pretty good. Status: Issue closed
ant-design/ant-design-mobile
302288533
Title: Add an animationDuration prop to Modal, so we can modify animation duration manually Question: username_0: ### What problem does this feature solve? We use Modal as the wrapper for serveral selection list, which fade in the middle of the window. These selection list Modals are used very frequently. Our users think the fade animation is slow, thus retards their operation. We didn't find a way to quicken the animation. There should be a prop left to developers to customize animation duration. ### What does the proposed API look like? Properties | Description | Type | Default ------------ | ------------- | ------------ | ------------- animationDuration | duration time for animation, which units is second | number | 1 <!-- generated by ant-design-issue-helper. DO NOT REMOVE -->
spotify/ios-sdk
67679601
Title: beta 7: Metadata attribute names are deprecated with no recommended replacements Question: username_0: I find these declarations in SPTAudioStreamingController.h FOUNDATION_EXPORT NSString * const SPTAudioStreamingMetadataTrackName DEPRECATED_ATTRIBUTE; FOUNDATION_EXPORT NSString * const SPTAudioStreamingMetadataTrackURI; FOUNDATION_EXPORT NSString * const SPTAudioStreamingMetadataArtistName DEPRECATED_ATTRIBUTE; FOUNDATION_EXPORT NSString * const SPTAudioStreamingMetadataArtistURI DEPRECATED_ATTRIBUTE; FOUNDATION_EXPORT NSString * const SPTAudioStreamingMetadataAlbumName DEPRECATED_ATTRIBUTE; FOUNDATION_EXPORT NSString * const SPTAudioStreamingMetadataAlbumURI DEPRECATED_ATTRIBUTE; FOUNDATION_EXPORT NSString * const SPTAudioStreamingMetadataTrackDuration DEPRECATED_ATTRIBUTE; All but one of these attribute names are marked "DEPRECATED" with no further explanation. What should we use instead of these? Each corresponds to a string with the same value as the symbol name. Will these key names be preserved in future updates? Answers: username_1: The recommended replacement is to look them up using the track URI. username_2: Can you give a quick example of what method we use to accomplish this. I have been trying to figure this out for the last couple of days. Thanks! username_2: Just a quick example of what I have already tried that still returns nil: SPTRequest.requestItemAtURI(spotifySongToplay.uri, withSession: auth.session) { (error, returnedObject) -> Void in println("TrackURI") println(returnedObject) println(returnedObject.album) //println(spotifySongToplay.uri) //println(returnedObject.imageURL) SPTRequest.requestItemFromPartialObject(returnedObject as SPTPartialObject, withSession: auth.session) { (error, finalObject) -> Void in println("Did get Final object") println(finalObject) println(finalObject.album) println(finalObject.imageURL) println("Artist") println(finalObject.artist) } } Artist, Album, and image URL all come up nil username_0: Per-Olov, It seems that the only non-deprecated way that I can get metadata for a track now is to to rely on this delegate method: -(void)audioStreaming:(SPTAudioStreamingController *)audioStreaming didChangeToTrack:(NSDictionary *)trackMetadata; because all of these are deprecated: -(void)getRelativeTrackMetadata:(int)index callback:(void (^)(NSDictionary *))block DEPRECATED_ATTRIBUTE; -(void)getAbsoluteTrackMetadata:(int)index callback:(void (^)(NSDictionary *))block DEPRECATED_ATTRIBUTE; @property (nonatomic, readonly, copy) NSDictionary *currentTrackMetadata DEPRECATED_ATTRIBUTE; That's fine. But once I have this metadata, the only non-deprecated way to access properties in the metadata dictionary is to use the key names (as strings). This code compiles without warnings with the beta 7 SDK: metadata.title = trackMetadata[@"SPTAudioStreamingMetadataTrackName"]; metadata.trackIdentifier = trackMetadata[SPTAudioStreamingMetadataTrackURI]; metadata.album = trackMetadata[@"SPTAudioStreamingMetadataAlbumName"]; metadata.artist = trackMetadata[@"SPTAudioStreamingMetadataArtistName"]; metadata.trackDuration = [trackMetadata[@"SPTAudioStreamingMetadataTrackDuration"] doubleValue]; It seems weird to me that SPTAudioStreamingMetadataTrackURI is not deprecated and all the other keys are. The corresponding values are all still returned in the metadata dictionary: (lldb) po trackMetadata { SPTAudioStreamingMetadataAlbumName = "An Orchestrated Rise to Fall"; SPTAudioStreamingMetadataAlbumURI = "spotify:album:7IoLXKqgoZSlX3PgwaL8WC"; SPTAudioStreamingMetadataArtistName = "The Album Leaf"; SPTAudioStreamingMetadataArtistURI = "spotify:artist:02uPe16VFxPaiueQsPEDkE"; SPTAudioStreamingMetadataTrackDuration = "353.045"; SPTAudioStreamingMetadataTrackName = Wander; SPTAudioStreamingMetadataTrackURI = "spotify:track:7iXiX53lhIcmYygpzwHlL6"; } username_2: Thanks! this helped me get it working. username_3: Please note that based on feedback on a more recent issue (https://github.com/spotify/ios-sdk/issues/320), the above proposed solution will also be deprecated "very soon". username_4: I'm having trouble working with one of these functions in Swift. Is there any documentation online that discusses alternatives? Not sure what it means by "use full metadata object instead". Here is what I'm trying to accomplish: func tableView(tableView: UITableView, didSelectRowAtIndexPath indexPath: NSIndexPath) { let playListVC = UIStoryboard(name: "Main", bundle: nil).instantiateViewControllerWithIdentifier("playlistView") as! PlaylistDetailController let partial = self.partialPlaylists[indexPath.row] as! SPTPartialPlaylist SPTRequest.requestItemFromPartialObject(partial, withSession: self.session) { (error: NSError!, metadata: AnyObject!) -> Void in playListVC.snapshot = partial as! SPTPlaylistSnapshot playListVC.currentPage = playListVC.snapshot.firstTrackPage playListVC.partialPlaylist = partial playListVC.session = self.session self.navigationController?.pushViewController(playListVC, animated: true) } } Can anyone think of a way to do this without using requestItemFromPartialObject username_5: Playback track metadata in Spotify Audio Playback module is available in SPTPlaybackTrack class. This class has been introduced in beta20. Status: Issue closed
eclipsesource/papyrus-seqd
318427307
Title: [General:008]: Multi selection Question: username_0: The diagram SHALL handle the move, reorder, and deletion of multiple selected elements. Answers: username_1: Currently, only deletion use case should be implemented for multi selection. username_0: Deletion use case works as of now. I'll push this to M4 for reconsidering supporting multi selection for other types of operations.
junegunn/fzf.vim
508507788
Title: Occasional Ag and Rg issues Question: username_0: <!-- Check all that apply [x] --> - Category - [ ] Question - [x] Bug - [ ] Suggestion - OS - [ ] Linux - [x] macOS - [ ] Windows - [ ] Etc. - Vim - [ ] Vim - [x] Neovim nvim version: v0.4.2 From time to time `Ag` and `Rg` commands stop working correctly. I've redefined them a bit to make file grepping a bit better: https://github.com/username_0/dotfiles/blob/master/config/nvim/init.vim#L199-L228 The issue occurs in a rather random fashion whenever I do this: * `Ctrl+G` to invoke customized `Ag` * type something in * press ENTER to open up given file ENTER ends with the following error: ``` Error detected while processing function 30[30]..<SNR>29_callback: line 21: Vim(call):E16: Invalid range: call s:fill_quickfix(list) ``` <img width="1440" alt="nvim" src="https://user-images.githubusercontent.com/6334715/67015999-c093a280-f0f7-11e9-8155-78537b19546c.png"> It gets worse right after that. I can still trigger `Ag` and look for strings, however whenever I press ENTER quicklist disappears, but file is not opened (no explicit error this time): <img width="1440" alt="nvim_2" src="https://user-images.githubusercontent.com/6334715/67016587-b625d880-f0f8-11e9-88b3-d552e0430742.png"> I've been using `Rg` for a couple of weeks, but experienced similar problems, so switched back to `Ag`. The only way to make it work again is to restart nvim. Answers: username_1: I've been struggling with this and I turned off autochdir and so far have had no issues. Not sure why it fixes it though?
StefanScherer/packer-windows
301557572
Title: windows_2016_docker.json produces 0 byte vmdk Question: username_0: I'm trying to get [Octopus Deploy](https://hub.docker.com/r/octopusdeploy/octopusdeploy/) run in Docker (version `18.03.0-ce-rc1, build c160c73`) on my macOS 10.12.6. I've installed VirtualBox (version `5.2.8r121009`), downloaded the Windows 2016 ISO and seemingly successfully executed the following command: ```shell packer build --only=virtualbox-iso -var 'iso_url=/Users/bitbear/Downloads/14393.0.161119-1705.RS1_REFRESH_SERVER_EVAL_X64FRE_EN-US.ISO' windows_2016_docker.json ``` After a few hours, this command finishes with the following: ```shell Build 'virtualbox-iso' finished. ==> Builds finished. The artifacts of successful builds are: --> virtualbox-iso: 'virtualbox' provider box: windows_2016_docker_virtualbox.box ``` However, when I execute the following command, a `.box` file is created with a `WindowsServer2016Docker-disk001.vmdk` file in it that is 0 bytes in size: ```shell vagrant box add windows_2016_docker windows_2016_docker_virtualbox.box ``` After copying `vagrantfile-windows_2016.template` to `Vagrantfile` and modifying it according to work with my setup, `vagrant up` fails with the following: ``` There was an error while executing `VBoxManage`, a CLI used by Vagrant for controlling VirtualBox. The command and stderr is shown below. Command: ["import", "/Users/bitbear/.vagrant.d/boxes/windows_2016_docker/0/virtualbox/box.ovf", "--vsys", "0", "--vmname", "WindowsServer2016Docker_1519934437898_19158", "--vsys", "0", "--unit", "8", "--disk", "/Users/bitbear/VirtualBox VMs/WindowsServer2016Docker_1519934437898_19158/WindowsServer2016Docker-disk001.vmdk"] Stderr: 0%...10%...20%...30%...40%...50%...60%...70%...80%...90%...100% Interpreting /Users/bitbear/.vagrant.d/boxes/windows_2016_docker/0/virtualbox/box.ovf... OK. 0%... Progress state: VBOX_E_FILE_ERROR VBoxManage: error: Appliance import failed VBoxManage: error: Could not open the medium storage unit '/Users/bitbear/.vagrant.d/boxes/windows_2016_docker/0/virtualbox/WindowsServer2016Docker-disk001.vmdk'. VBoxManage: error: VMDK: error reading the magic number in '/Users/bitbear/.vagrant.d/boxes/windows_2016_docker/0/virtualbox/WindowsServer2016Docker-disk001.vmdk' (VERR_EOF). VBoxManage: error: VD: error VERR_VD_VMDK_INVALID_HEADER opening image file '/Users/bitbear/.vagrant.d/boxes/windows_2016_docker/0/virtualbox/WindowsServer2016Docker-disk001.vmdk' (VERR_VD_VMDK_INVALID_HEADER) VBoxManage: error: Details: code VBOX_E_FILE_ERROR (0x80bb0004), component ApplianceWrap, interface IAppliance VBoxManage: error: Context: "RTEXITCODE handleImportAppliance(HandlerArg *)" at line 886 of file VBoxManageAppliance.cpp ``` When I inspect the file `/Users/bitbear/.vagrant.d/boxes/windows_2016_docker/0/virtualbox/WindowsServer2016Docker-disk001.vmdk`, it is indeed an empty file of 0 bytes in size. Any ideas what might be causing this and how I can fix it? Answers: username_1: Which version of Packer do you use? My latest builds are done with Packer 1.1.3. Which size is the box file? It should be about 10 GByte. username_2: Do you have enough free disk space to hold the 10 GByte file? username_0: I'm using version `1.2.1` of packer and I have 160 GB of free disk space. The box is 10.89 GB. username_1: The box size looks good. Can you list the content of the box file? It should look like this: ``` $ .\tar.exe tvf .\windows_2016_docker_virtualbox.box -rw-rw-rw- 0 0 0 2347 Jan 23 01:43 Vagrantfile -rw-rw-rw- 0 0 0 10767669248 Jan 23 01:43 WindowsServer2016Docker-disk001.vmdk -rw-rw-rw- 0 0 0 10416 Jan 23 01:43 box.ovf -rw-rw-rw- 0 0 0 26 Jan 23 01:43 metadata.json ``` username_0: ```shell $ tar tvf windows_2016_docker_virtualbox.box -rw-r--r-- 0 bitbear staff 2286 Mar 1 20:50 Vagrantfile -rw-r--r-- 0 bitbear staff 0 Mar 1 20:50 WindowsServer2016Docker-disk001.vmdk -rw-r--r-- 0 bitbear staff 10412 Mar 1 20:50 box.ovf -rw-r--r-- 0 bitbear staff 26 Mar 1 20:50 metadata.json ``` username_1: Well, I don't know where the error during the packer build happened to produce this corrupt box file. I would suggest to just rebuild the box and see if it works. You could also use the [`username_1/windows_2016_docker`](https://app.vagrantup.com/username_1/boxes/windows_2016_docker) box from Vagrant cloud. username_0: I’ve rebuilt the box four times due to other hints I’ve found around the web of what the problem might be. Nothing has helped, which is why I ended up asking here. I’ll definitely try your box from Vagrant cloud, thanks! username_2: Can duplicate. macOS 10.13.3 $ packer --version 1.2.1 $ vboxmanage --version 5.2.8r121009 ```shell $ ls -lh *.box -rw-r--r-- 1 tanner staff 10G Mar 4 03:30 windows_2016_docker_virtualbox.box ``` The .box is 10G in size ```shell $ tar -tvf windows_2016_docker_virtualbox.box -rw-r--r-- 0 tanner staff 2286 Mar 4 03:25 Vagrantfile -rw-r--r-- 0 tanner staff 0 Mar 4 03:25 WindowsServer2016Docker-disk001.vmdk -rw-r--r-- 0 tanner staff 10412 Mar 4 03:25 box.ovf -rw-r--r-- 0 tanner staff 26 Mar 4 03:25 metadata.json ``` But the .vmdk is zero? Where is the 10G of disk space coming from? username_0: Yes, I'm using HFS+ since I have a Fusion drive. I'm waiting for [Fusion support in APFS](https://www.macrumors.com/2017/09/25/apfs-fusion-drive-high-sierra-update/) before I convert. username_2: HFS+ has Max. file size | 8 exabyte username_1: Interesting, so HFS+ does not have a problem with the box file size and extracting the contents. So it seems a problem while creating the tar file. You can run Packer with debug output ``` export PACKER_LOG=1 packer build ... ``` To see which command is called during the box file creation. I don't know if it's a built-in command in the packer binary or if some external tools from macOS are called. username_2: Not seeing anything obvious in the logs. [typescript.txt](https://github.com/username_1/packer-windows/files/1782300/typescript.txt) username_2: "keep_input_artifact": true HFS+ filesystem [Packer Logs](https://github.com/username_1/packer-windows/files/1783774/typescript.txt) output-virtualbox-iso/ ```shell $ ls -l total 21220296 -rw-r--r-- 1 tanner staff 10864777216 Mar 5 18:46 WindowsServer2016Docker-disk001.vmdk -rwx------ 1 tanner staff 10412 Mar 5 18:23 WindowsServer2016Docker.ovf ``` ```shell $ ls -l *.box -rw-r--r-- 1 tanner staff 10729788371 Mar 5 18:56 windows_2016_docker_virtualbox.box ``` ```shell $ tar -tvf windows_2016_docker_virtualbox.box -rw-r--r-- 0 tanner staff 2286 Mar 5 18:51 Vagrantfile -rw-r--r-- 0 tanner staff 0 Mar 5 18:51 WindowsServer2016Docker-disk001.vmdk -rw-r--r-- 0 tanner staff 10412 Mar 5 18:51 box.ovf -rw-r--r-- 0 tanner staff 26 Mar 5 18:51 metadata.json ``` username_1: Thanks. This really looks like a problem inside packer ``` 2018/03/05 18:51:25 packer: 2018/03/05 18:51:25 Turning dir into box: /<KEY>packer821802642 => windows_2016_docker_virtualbox.box 2018/03/05 18:51:25 packer: 2018/03/05 18:51:25 Compressing with gzip compression level: -1 2018/03/05 18:51:26 packer: 2018/03/05 18:51:26 Skipping directory '/<KEY>packer821802642' for box 'windows_2016_docker_virtualbox.box' 2018/03/05 18:51:26 packer: 2018/03/05 18:51:26 Box add: '/<KEY>packer821802642/Vagrantfile' to 'windows_2016_docker_virtualbox.box' 2018/03/05 18:51:26 ui:  virtualbox-iso (vagrant): Compressing: Vagrantfile  virtualbox-iso (vagrant): Compressing: Vagrantfile 2018/03/05 18:51:26 packer: 2018/03/05 18:51:26 Box add: '/<KEY>WindowsServer2016Docker-disk001.vmdk' to 'windows_2016_docker_virtualbox.box' 2018/03/05 18:51:26 ui:  virtualbox-iso (vagrant): Compressing: WindowsServer2016Docker-disk001.vmdk  virtualbox-iso (vagrant): Compressing: WindowsServer2016Docker-disk001.vmdk  virtualbox-iso (vagrant): Compressing: box.ovf  virtualbox-iso (vagrant): Compressing: metadata.json 2018/03/05 18:56:09 packer: 2018/03/05 18:56:09 Box add: '/<KEY>box.ovf' to 'windows_2016_docker_virtualbox.box' 2018/03/05 18:56:09 ui:  virtualbox-iso (vagrant): Compressing: box.ovf 2018/03/05 18:56:09 packer: 2018/03/05 18:56:09 Box add: '/<KEY>metadata.json' to 'windows_2016_docker_virtualbox.box' 2018/03/05 18:56:09 ui:  virtualbox-iso (vagrant): Compressing: metadata.json 2018/03/05 18:56:09 [INFO] (telemetry) ending vagrant ``` The tar file is created internally: https://github.com/hashicorp/packer/blob/4d3a762e85e7e8050134a4e33e1f87a54029dcb8/post-processor/vagrant/util.go#L71-L155 but this function has no recent changes. username_2: Can duplicate this problem with macOS sierra (10.12.6) and HFS+ filesystem. Time to open a bug in the packer project? username_1: There's one config option `compression_level` which can be set to `0` to turn off compression. Don't know if that makes a difference. Issue with Golang or one of the imported packages?, maybe. Maybe create a small golang program from the DirToBox function, removing the ui parameter and run it with some big files. An issue in packer project may help to get some feedback. But I haven't seen a similar issue there yet, so probably most people work with much smaller vmdk sizes. username_1: @username_0 With `docker-compose` you only can start containers on one Docker host. If you want to mix Linux and Windows workloads you need eg. Docker Swarm running a Linux and Windows node and then you can use a docker-compose.yml to deploy a stack. With later Windows Server like 1709, 1803, ... and latest nightly Docker builds you may use LCOW to start Linux containers directly on the Windows Docker host. But this works only with HyperV enabled in the VM, but VirtualBox does not have nested virtualization feature to do that. username_0: Yeah, I considered that but VirtualBox makes that impossible. So I don't really know what to do. I think I might just manually run Octopus Deploy within a Windows VM and skip Docker for this purpose altogether. username_1: The workaround for now: Use Packer 1.1.3 I just rebuilt a VirtualBox box file (`brew switch packer 1.1.3`) and everything looks fine: ``` $ tar tvf windows_2016_docker_virtualbox.box -rw-r--r-- 0 501 20 2286 Mar 11 14:27 Vagrantfile -rw-r--r-- 0 501 20 10952051712 Mar 11 14:27 WindowsServer2016Docker-disk001.vmdk -rw-r--r-- 0 501 20 10412 Mar 11 14:27 box.ovf -rw-r--r-- 0 501 20 26 Mar 11 14:27 metadata.json ~/code/packer-windows on my* $ ls -l *box -rw-r--r-- 1 stefan staff 10817278351 Mar 11 14:28 windows_2016_docker_virtualbox.box ``` username_1: Created https://github.com/golang/go/issues/24599 for further investigation. username_1: @username_2 Your fix/workaround includes using the homebrew GNU tar and replacing the embedded tar in Vagrant? username_1: Weird. But as long as it works for you. It will be officially fixed with Packer 1.2.3, PR https://github.com/hashicorp/packer/pull/6084 got merged. So using Packer 1.1.3 or 1.2.3 fixes the problem. Status: Issue closed
GotoLink/BetterSand
80502539
Title: Crash Question: username_0: tryng to launch server ---- Minecraft Crash Report ---- // Hey, that tickles! Hehehe! Time: 25/05/15 13.12 Description: Exception in server tick loop cpw.mods.fml.common.LoaderException: java.lang.NoSuchMethodError: net.minecraft.block.Block.func_149691_a(II)Lnet/minecraft/util/IIcon; at cpw.mods.fml.common.LoadController.transition(LoadController.java:162) at cpw.mods.fml.common.Loader.preinitializeMods(Loader.java:515) at cpw.mods.fml.server.FMLServerHandler.beginServerLoading(FMLServerHandler.java:88) at cpw.mods.fml.common.FMLCommonHandler.onServerStart(FMLCommonHandler.java:318) at net.minecraft.server.dedicated.DedicatedServer.func_71197_b(DedicatedServer.java:176) at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:620) at java.lang.Thread.run(Unknown Source) Caused by: java.lang.NoSuchMethodError: net.minecraft.block.Block.func_149691_a(II)Lnet/minecraft/util/IIcon; at bettersand.BetterSand$BlockFluid.<init>(BetterSand.java:173) at bettersand.BetterSand.preInit(BetterSand.java:73) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at cpw.mods.fml.common.FMLModContainer.handleModStateEvent(FMLModContainer.java:530) at sun.reflect.GeneratedMethodAccessor3.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at com.google.common.eventbus.EventSubscriber.handleEvent(EventSubscriber.java:74) at com.google.common.eventbus.SynchronizedEventSubscriber.handleEvent(SynchronizedEventSubscriber.java:47) at com.google.common.eventbus.EventBus.dispatch(EventBus.java:322) at com.google.common.eventbus.EventBus.dispatchQueuedEvents(EventBus.java:304) at com.google.common.eventbus.EventBus.post(EventBus.java:275) at cpw.mods.fml.common.LoadController.sendEventToModContainer(LoadController.java:208) at cpw.mods.fml.common.LoadController.propogateStateMessage(LoadController.java:187) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at com.google.common.eventbus.EventSubscriber.handleEvent(EventSubscriber.java:74) at com.google.common.eventbus.SynchronizedEventSubscriber.handleEvent(SynchronizedEventSubscriber.java:47) at com.google.common.eventbus.EventBus.dispatch(EventBus.java:322) at com.google.common.eventbus.EventBus.dispatchQueuedEvents(EventBus.java:304) at com.google.common.eventbus.EventBus.post(EventBus.java:275) at cpw.mods.fml.common.LoadController.distributeStateMessage(LoadController.java:118) at cpw.mods.fml.common.Loader.preinitializeMods(Loader.java:513) ... 5 more A detailed walkthrough of the error, its code path and all known details is as follows: --------------------------------------------------------------------------------------- -- System Details -- Details: Minecraft Version: 1.7.10 Operating System: Windows 7 (amd64) version 6.1 Java Version: 1.8.0_40, Oracle Corporation Java VM Version: Java HotSpot(TM) 64-Bit Server VM (mixed mode), Oracle Corporation Memory: 465110760 bytes (443 MB) / 1232601088 bytes (1175 MB) up to 1908932608 bytes (1820 MB) JVM Flags: 2 total; -XX:PermSize=2G -XX:MaxPermSize=3G [Truncated] * RailcraftAPI|bore (1.0.0) from Railcraft_1.7.10-9.6.1.0.jar * RailcraftAPI|carts (1.3.1) from Railcraft_1.7.10-9.6.1.0.jar * RailcraftAPI|core (1.3.0) from Railcraft_1.7.10-9.6.1.0.jar * RailcraftAPI|crafting (1.0.0) from Railcraft_1.7.10-9.6.1.0.jar * RailcraftAPI|electricity (1.6.0) from Railcraft_1.7.10-9.6.1.0.jar * RailcraftAPI|events (1.0.0) from Railcraft_1.7.10-9.6.1.0.jar * RailcraftAPI|fuel (1.0.0) from Railcraft_1.7.10-9.6.1.0.jar * RailcraftAPI|helpers (1.1.0) from Railcraft_1.7.10-9.6.1.0.jar * RailcraftAPI|items (1.0.0) from Railcraft_1.7.10-9.6.1.0.jar * RailcraftAPI|locomotive (1.0.1) from Railcraft_1.7.10-9.6.1.0.jar * RailcraftAPI|signals (1.3.0) from Railcraft_1.7.10-9.6.1.0.jar * RailcraftAPI|tracks (2.0.0) from Railcraft_1.7.10-9.6.1.0.jar * Shield (0.1) from 1.7.10-MB_Battlegear2-Bullseye-1.0.8.0.jar * Thaumcraft|API (4.2.2.0) from Thaumcraft-1.7.10-4.2.3.5.jar * TSteelworksAPI (1.1) from TSteelworks-1.7.10-1.1.0-6.jar * WailaAPI (1.2) from Waila-1.5.10_1.7.10.jar * Weapons (0.1) from 1.7.10-MB_Battlegear2-Bullseye-1.0.8.0.jar Profiler Position: N/A (disabled) Is Modded: Definitely; Server brand changed to 'cauldron,craftbukkit,mcpc,fml,forge' Type: Dedicated Server (map_server.txt) Answers: username_1: Fixed. Status: Issue closed username_0: There is another problem, with this update the sand is in the form of block as the normal one.. username_1: This fix only matter on server side, indeed. Client should be unaffected, in all versions. Not all sand is actually replaced by the mod. Breaking normal sand blocks should change it into fluid. username_0: so I don't have to use this new fixed version of the mod for client too but only for server side, right? username_1: It could depend on Forge version, actually. Newer forge introduced some things to the fluid system. Might want to update the client too, for such case. username_0: The server and the client share the same version of forge and the build should be 10.13.3.1388.
pivpn/pivpn
277166613
Title: PiVPN - ovpn file not generate wright Question: username_0: On linux# openvpn --config test1234.ovpn (because the ovpn is not right i thinks this not work) Answers: username_1: There appears to be an issue with the installer creating the config file in the /etc/pivpn/ with all the variables required to run the pivpn command. I was able to work around the issue in my setup by uninstalling `pivpn uninstall` and then installing the software with this command `curl -L https://install.pivpn.io | sudo bash`. Escalating the entire installer to root gets around the permissions issue of the configuration file unable to be written where it should be. username_0: Hi username_1, it works! Maybe i forgot the privilege permissions (there is without "sudo" in the tutorial). Also i must to update openvpn client on my linux mint (to OpenVPN 2.4.4) Thanks a lot! Have a nice day. Status: Issue closed
ansible/ansible-runner
819017533
Title: Ansible runner does not write event files atomically Question: username_0: There is a small time window where the actual final event json file can be incomplete and therefore contain invalid json, tripping up tools which rely on those files always being complete. It should write to a temporary location and then atomically move the now-complete file to the final destination the same way it is done here[1]. [1] - https://github.com/ansible/ansible-runner/blob/devel/ansible_runner/display_callback/events.py#L69 Answers: username_0: Linked PR was merged, closing. Status: Issue closed
quic/aimet
945880590
Title: Add support to identify [Conv - Relu6] as supergroup Question: username_0: Thank you for the quick response @username_1. Would like to introduce you to team members who can help you get started with this @quic-akhobare @username_2. @username_2, could you please add tje necessary details to this issue and enable @username_1. Thank you. Answers: username_1: Hi @username_0, I would like to start with this issue. username_0: Thank you for the quick response @username_1. Would like to introduce you to team members who can help you get started with this @quic-akhobare @username_2. @username_2, could you please add tje necessary details to this issue and enable @username_1. Thank you. username_2: Thank you for your interest in taking up this issue @username_1. Here are some additional pointers to help you get started: Supergroups are used during quantization simulation to identify groups of operations which will not have quantization done in between. Different runtimes may use different rules, so we provide users the flexibility to configure supergroups and other settings using a json configuration file. default_config.json under aimet_common/quantsim_config holds some default settings which we typically find common between runtimes. Conv->Relu6 rule can be added in here. The configuration files are used for both pytorch and tensorflow quantization simulation. To understand how we use it for each framework, you can start looking at quantsim_config.py under aimet_torch/quantsim_config and aimet_tensorflow/quantsim_config respectively. To test for correctness, please also create unit tests within test_quantsim_config.py under torch/test/python and tensorflow/test/python for both frameworks. username_0: Hi @username_1 Just wanted to check back on this. Were you able to checkout code and make updates? Please let us know if you have further questions. Thank you. Status: Issue closed
jirmauritz/AirportManager
119748009
Title: Create REST API as a Spring MVC controller Question: username_0: - based on the e-shop project - any entity, others will be copied Answers: username_1: Include README file with instructions, how to test REST more at [Milestone 3 description](https://is.muni.cz/auth/elearning/warp?fakulta=1433;obdobi=6383;studium=647771;predmet=869910;lang=cs;furl=%2Fauth%2Fel%2F1433%2Fpodzim2015%2FPA165%2Fseminars.qwarp;so=dd;qurl=%2Fel%2F1433%2Fpodzim2015%2FPA165%2Fseminars.qwarp;prejit=3361389) Status: Issue closed
sourcecred/docs
964217090
Title: [Epic] Prod-ready Setup Docs Question: username_0: - [ ] “Is SourceCred right for my community?” - [ ] “Instance Set Up” (polishing) - [ ] “How to configure your instance” - [ ] File-by-file, attribute-by-attribute documentation - [ ] The Case for Tithing - [ ] “How to maintain your instance” - [ ] Ways to manage Identities (merge and activate) - [ ] Ways to do payouts
scikit-image/scikit-image
158917437
Title: Add rotation to `draw.ellipse` Question: username_0: Update the `draw.ellipse` function to optionally take an angle e.g. `theta` (defaults to 0) which specifies the angle between the major axis and the x-axis. Answers: username_1: Cc @username_2 username_2: Hi all, I checked that `skimage.draw.ellipse_perimeter()` has the `orientation` argument implemented. However, that function is implemented in Cython... Let me see if I can implement that. username_3: Closed via https://github.com/scikit-image/scikit-image/pull/2366. Status: Issue closed
Behat/MinkExtension
272467145
Title: Interaction with react components Question: username_0: Currently we are switching our old app to one based on React. In a menu we have a react component in which we need to type some letters and the a dropdown opens with some suggestions. After this happens we need to select one of the options that appeared by clicking on it. When trying to interact with this component we encountered difficulties in typing some characters in the react-select component as well as selecting the options that appeared in the dropdown. Does anyone have any suggestion on how I can better interact with such components. Here are some examples of react-select components: https://jedwatson.github.io/react-select/ Thanks in advance. Answers: username_1: A little bit hacky but works for us. In Behat Syntax: ` When I click on the element "#select .Select-arrow-zone" When I wait for element "#select .Select-menu-outer" to appear And I click on the element "#select .Select-option[title='L001']" # or if title not filed: And I click on the element "#select .Select-option:nth-child(2)" `
palantir/blueprint
400348577
Title: PanelStack initial panel should support components that extends IPanel, not just plain IPanel Question: username_0: <!-- IF YOU ARE A PALANTIR EMPLOYEE, DO NOT POST INTERNAL LINKS OR REFERENCES HERE --> #### Environment - __Package version(s)__: <!-- fill this out --> @blueprintjs/core 3.10.0 - __Browser and OS versions__: <!-- fill this out --> Chrome, Mac OSX #### Steps to reproduce ```typescript import { IPanelProps, PanelStack } from "@blueprintjs/core"; class SomePanel extends React.Component<IPanelProps & { enabled: boolean }> { // ... } <PanelStack initialPanel={{ component: SomePanel, title: "Home" }} /> ``` #### Actual behavior Typescript compilation fails. #### Expected behavior Typescript compilation should pass. #### Possible solution Make PanelStack generic.<issue_closed> Status: Issue closed
izmailovpavel/understandingbdl
569112612
Title: Add license(s) Question: username_0: It would be good to add a LICENSE file here so that users understand how the code may be used. Additionally, any code which came from other places should include the original license file. (I'm enjoying reading your paper!) Cheers! Answers: username_1: Thanks @username_0, I have added the license for the repo. What is the standard way to specify the original license for code acquired from other repos? username_0: I'm not sure there's a "standard way", but as long as you do something that reasonably attempts to follow the terms of the license, you should be fine. I think the [Julia LICENSE file](https://github.com/JuliaLang/julia/blob/master/LICENSE.md) is a good example: they give the license for the code that originated in that repo first, and then for various components which were derived from other sources, they link to the license for that source. In your case, for the code that was forked from the Subspace Inference GitHub repo, you should probably add the copyright line from that repo to your license file--it's the same license, but the copyright is owned by the people who originally wrote that code. If you've rewritten most of the code, then you might be off the hook, but you should use your best judgement--it's probably a bad idea to claim copyright on something that someone else wrote! :-) For the parts which came from the other repo, see the [Apache License 2.0 License](https://github.com/hendrycks/robustness/blob/master/LICENSE) in that repo. Clause 4 of that license spells out what you're supposed to do. A link to that license file is probably sufficient for providing the license (Clause 1), but for any parts which you modified, you're supposed to do more (see the license for details). Hope this helps! username_1: Thank you so much for the tips :) I am on the Subspace Inference repo so I think it will be fine. I modified the LICENSE file and added a notice to the files adapted from the robustness repo. Status: Issue closed
S1ngS1ng/HammerSpoon
602391716
Title: Catalina不生效|窗口管理功能不生效(只测试了窗口管理) Question: username_0: ``` module.throwLeft = function () local this = windowMeta.new() print(1); // 代码走到了这里,似乎是下一行没有生效。 this.window:moveOneScreenWest() end ``` macOS: 10.15.3 Hammerspoon: 0.9.77 Answers: username_1: @username_0 大佬好! 我这里,macOS 10.15.4,然后 HammerSpoon 是 0.9.78 (5164),一切正常 麻烦你发我下这些吧: 1. console 里看看有没有报错 ![image](https://user-images.githubusercontent.com/13592559/79681229-5dc9b380-81cd-11ea-91bf-42d8ef6eafe7.png) 2. Reload Config,再试试 3. 如果还不行,console 里执行一下 `for k, v in ipairs(hs.screen.allScreens()) do print(k, v) end`,看看输出是不是和你的所有屏幕数量一致 username_0: 重新安装了一次。找到原因了,是我的问题。我把`ctrl`当成了`fn`键。(我总是记不住23333) 顺便贴一下好了: ``` ⌘ 对应 Command. ⌥ 对应 Option (也称为Alt) ⌃ 对应 Control. ⇧ 对应 Shift. ``` Status: Issue closed
anfema/ion-client-ios
577894478
Title: Add access helper for IONContainerContent and its typed contents Question: username_0: **Proposal** Add an helper that allows a one line usage of contents of an `IONContainerContent `. It already supports e.g. texts, but not all possible contents. **Source from BCGyou** ``` extension IONContainerContent { func fileContent(_ identifier: OutletIdentifier, at position: Position = 0) -> IONFileContent? { let fileContent: IONFileContent? = content(identifier, at: position) return fileContent } func connectionContent(_ identifier: OutletIdentifier, at position: Position = 0) -> IONConnectionContent? { let connectionContent: IONConnectionContent? = content(identifier, at: position) return connectionContent } } ``` Answers: username_1: Can be resolved by using generic function: `IONContainerContent.content("contentIdentifier")` Status: Issue closed
robotmoose/robotmoose
170971477
Title: Getnext superstarv2 needs to send hash of old value. Question: username_0: This is actually a lot more difficult that I originally thought... Answers: username_0: This is actually a lot more difficult that I originally thought... username_0: I have a solution to this, but I can't seem to get it to work reliably. If a backend and frontend are running on the same machine, then chrome only allows 10 connections at a time. The following solution seems to lockup the frontend/backend because of this... The solution: Simply have the server send back the hash of the value, this way clients don't have to hash it at all. So "result" would be an object containing a "value" object and a "hash" string. @olawor, you think we could meet on Monday or Wednesday to work on this? The implementation of the solution is easy enough (already done), but tracking down what's causing comets not to be serviced is eluding me... Status: Issue closed
minio/minio-js
177658252
Title: Update documentation correcting the APIs Question: username_0: In the doc * it should be `new Minio.Client`, not `new Minio` * fix bucket notification APIs (doc differs from the actual API) Answers: username_1: https://docs.minio.io/docs/javascript-client-api-reference missing listenBucketNotification API Status: Issue closed
introlab/MOvITPlus
821437517
Title: Ajouter dans les paramètres de l'usager la situation des capteurs Question: username_0: Ajouter dans les paramètres de l'usager la situation des capteurs. Section 9.2 cahier charge. Important que l'usager soit au courant si un capteur est mal branché. ![image](https://user-images.githubusercontent.com/77858297/109864539-f4aa2d80-7c30-11eb-96b5-e79d4731dbeb.png) Answers: username_1: @username_0 L'image est trop petite, on ne voit rien, tu peux la renvoyer ? Status: Issue closed
MDSplus/mdsplus
979336748
Title: Commit 2366 Breaks MDSplus python install in alpha 7.134.4 Question: username_0: Prior to this commit #2366, MDSplus python package installs with both the usual commands (pip install -e ..., python setup.py install) After this commit, when calling either from terminal, results in ``` unning install running bdist_egg running egg_info writing UNKNOWN.egg-info/PKG-INFO writing dependency_links to UNKNOWN.egg-info/dependency_links.txt writing top-level names to UNKNOWN.egg-info/top_level.txt reading manifest file 'UNKNOWN.egg-info/SOURCES.txt' writing manifest file 'UNKNOWN.egg-info/SOURCES.txt' installing library code to build/bdist.linux-x86_64/egg running install_lib warning: install_lib: 'build/lib' does not exist -- no Python modules to install creating build/bdist.linux-x86_64/egg creating build/bdist.linux-x86_64/egg/EGG-INFO copying UNKNOWN.egg-info/PKG-INFO -> build/bdist.linux-x86_64/egg/EGG-INFO copying UNKNOWN.egg-info/SOURCES.txt -> build/bdist.linux-x86_64/egg/EGG-INFO copying UNKNOWN.egg-info/dependency_links.txt -> build/bdist.linux-x86_64/egg/EGG-INFO copying UNKNOWN.egg-info/not-zip-safe -> build/bdist.linux-x86_64/egg/EGG-INFO copying UNKNOWN.egg-info/top_level.txt -> build/bdist.linux-x86_64/egg/EGG-INFO creating 'dist/UNKNOWN-0.0.0-py3.6.egg' and adding 'build/bdist.linux-x86_64/egg' to it removing 'build/bdist.linux-x86_64/egg' (and everything under it) Processing UNKNOWN-0.0.0-py3.6.egg creating /usr/local/lib/python3.6/dist-packages/UNKNOWN-0.0.0-py3.6.egg Extracting UNKNOWN-0.0.0-py3.6.egg to /usr/local/lib/python3.6/dist-packages Adding UNKNOWN 0.0.0 to easy-install.pth file Installed /usr/local/lib/python3.6/dist-packages/UNKNOWN-0.0.0-py3.6.egg Processing dependencies for UNKNOWN==0.0.0 Finished processing dependencies for UNKNOWN==0.0.0 ``` Seeing that `setuptools` is default, and that `setupkw` dictionary is not used in either `use_distutils` or `use_setuptools`, adding `setupkw` like so in `use_setuptools` results in: ``` def use_setuptools(): from setuptools import setup setup( include_package_data=True, test_suite='tests.test_all', zip_safe=False, **setupkw ) ``` Though I suppose there are other solutions. Going back to the diff, I see that `setupkw` was previously used at least 6 times. Now, it is merely defined and not used. ``` $>python setup.py install running install running bdist_egg running egg_info writing MDSplus.egg-info/PKG-INFO writing dependency_links to MDSplus.egg-info/dependency_links.txt writing top-level names to MDSplus.egg-info/top_level.txt reading manifest file 'MDSplus.egg-info/SOURCES.txt' writing manifest file 'MDSplus.egg-info/SOURCES.txt' [Truncated] creating build/bdist.linux-x86_64/egg/EGG-INFO copying MDSplus.egg-info/PKG-INFO -> build/bdist.linux-x86_64/egg/EGG-INFO copying MDSplus.egg-info/SOURCES.txt -> build/bdist.linux-x86_64/egg/EGG-INFO copying MDSplus.egg-info/dependency_links.txt -> build/bdist.linux-x86_64/egg/EGG-INFO copying MDSplus.egg-info/not-zip-safe -> build/bdist.linux-x86_64/egg/EGG-INFO copying MDSplus.egg-info/top_level.txt -> build/bdist.linux-x86_64/egg/EGG-INFO creating 'dist/MDSplus-7.134.4-py3.6.egg' and adding 'build/bdist.linux-x86_64/egg' to it removing 'build/bdist.linux-x86_64/egg' (and everything under it) Processing MDSplus-7.134.4-py3.6.egg removing '/usr/local/lib/python3.6/dist-packages/MDSplus-7.134.4-py3.6.egg' (and everything under it) creating /usr/local/lib/python3.6/dist-packages/MDSplus-7.134.4-py3.6.egg Extracting MDSplus-7.134.4-py3.6.egg to /usr/local/lib/python3.6/dist-packages MDSplus 7.134.4 is already the active version in easy-install.pth Installed /usr/local/lib/python3.6/dist-packages/MDSplus-7.134.4-py3.6.egg Processing dependencies for MDSplus==7.134.4 Finished processing dependencies for MDSplus==7.134.4 ``` But still can't import MDSplus in interactive python shell Answers: username_1: fixed it this way for alpha. do this for pyhton2 and python3 if python version available: * compile in install folder by default for python2 and python3 (ignore if it fails, e.g. because version not available) * detect if MDSplus was installed using setup before if it was or if MDSPLUS_PYTHON_SETUP is set to "y" attempt to run setup.py as well * if using setup.py, try setuptools first, fallback to distutils. TODO? should setup.py be moved one up? Status: Issue closed
moby/buildkit
911805244
Title: Using --mount=type=cache means the cache is not available at runtime Question: username_0: I am actually using a different language and toolchain, but here is a simple illustration of the problem. My dockerfile: ``` FROM ubuntu:focal RUN apt-get update && \ DEBIAN_FRONTEND=noninteractive apt-get --assume-yes install \ nodejs npm WORKDIR /app COPY . /app # Problem is here, with --mount RUN --mount=type=cache,target=/app/node_modules \ npm install CMD ["/bin/bash"] ``` This can be built into an image and a container, but when I try to execute commands in the container, the cache is not there. `node_modules` exists, but is empty. I must reinstall it every time I wish to enter the container and execute a one-off command. I'm actually not trying to cache `node_modules` nor use JavaScript at all, but what I'm doing operates largely the same way as yarn's cache. I'm trying to cache the equivalent of `/home/user/.cache/yarn/v6`, so that I can try to run offline installs. I'm having this issue with haskell's cabal-install, and the process takes several minutes to complete. In any case, these build tools' caches are apparently permanently discarded after the image is built; and if I run install (even when `node_modules` equivalent is populated) it tries to repopulate the cache. If I remove the `--mount=type=cache` directive, it speeds up the execution of commands on the container by keeping a copy of the generated files. If I keep the `--mount` directive, it speeds up the image build just a little bit, but I have to download and repopulate the cache over and over again after the image has been built; and I need to be able to go into a container, tweak a few things, and then run reinstall. I would really like a way to keep the cache around (read-only is fine) at runtime, or to allow the build to write to a volume at build time, so I can manage the cache myself. Like `DOCKER_BUILDKIT=1 docker run --mount=type=cache,etc`. Answers: username_1: I was forwarded to the following thread today: https://github.com/moby/buildkit/issues/1512 Probably it'll help in your case as well:) username_2: `--mount=type=cache` does not keep files in image by design. You may be using it for a case it is not meant for. If you want some files from your cache inside your image you need to copy them out of the cache mount with `cp/rsync` etc. Status: Issue closed
neo4j/neo4j-ogm
215997149
Title: NPE when creating a SessionFactory in 3.0 snapshots when there's no ogm.properties file Question: username_0: <!-- ================================================================= --> <!-- Please note that GitHub issues are only meant for bug reports/feature requests. --> <!-- If you have questions on how to use the Neo4j OGM, please ask on --> <!-- StackOverflow: http://stackoverflow.com/questions/tagged/neo4j-ogm instead of --> <!-- creating an issue here. --> <!-- ================================================================= --> <!--- Provide a general summary of the issue in the Title above --> ## Expected Behavior A `NullPointerException` is not thrown when calling `new SessionFactory(Configuration, String...)` ## Current Behavior A `NullPointerException` is thrown when calling `new SessionFactory(Configuration, String...)`: ``` Caused by: java.lang.NullPointerException at java.util.Properties$LineReader.readLine(Properties.java:434) at java.util.Properties.load0(Properties.java:353) at java.util.Properties.load(Properties.java:341) at org.neo4j.ogm.config.ClasspathConfigurationSource.<init>(ClasspathConfigurationSource.java:19) at org.neo4j.ogm.config.Components.configure(Components.java:73) at org.neo4j.ogm.config.Components.autoConfigure(Components.java:106) at org.neo4j.ogm.config.Components.loadDriver(Components.java:114) at org.neo4j.ogm.config.Components.driver(Components.java:86) at org.neo4j.ogm.session.SessionFactory.<init>(SessionFactory.java:53) at org.neo4j.ogm.session.SessionFactory.<init>(SessionFactory.java:88) at org.springframework.boot.autoconfigure.data.neo4j.Neo4jDataAutoConfiguration.sessionFactory(Neo4jDataAutoConfiguration.java:75) at org.springframework.boot.autoconfigure.data.neo4j.Neo4jDataAutoConfiguration$$EnhancerBySpringCGLIB$$29c1a54f.CGLIB$sessionFactory$1(<generated>) at org.springframework.boot.autoconfigure.data.neo4j.Neo4jDataAutoConfiguration$$EnhancerBySpringCGLIB$$29c1a54f$$FastClassBySpringCGLIB$$48c5ef2d.invoke(<generated>) at org.springframework.cglib.proxy.MethodProxy.invokeSuper(MethodProxy.java:228) at org.springframework.context.annotation.ConfigurationClassEnhancer$BeanMethodInterceptor.intercept(ConfigurationClassEnhancer.java:358) at org.springframework.boot.autoconfigure.data.neo4j.Neo4jDataAutoConfiguration$$EnhancerBySpringCGLIB$$29c1a54f.sessionFactory(<generated>) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.springframework.beans.factory.support.SimpleInstantiationStrategy.instantiate(SimpleInstantiationStrategy.java:162) ... 52 more ``` ## Possible Solution The underlying problem appears to be that `Components.driver()` is still called, but `Components` has not been initialized with the `Configuration` instance. This initialisation was removed in https://github.com/neo4j/neo4j-ogm/commit/7b2b21e5cf82661f227733e7f0855b7c41dae47b which is entitled "Making Configuration the default mechanism for settings rather than Components". It looks like the move to a Configuration-based mechanism is incomplete. ## Steps to Reproduce (for bugs) The following code will reproduce the problem if there's no `ogm.properties` file on the classpath: ```java Configuration configuration = new Configuration(); configuration.setDriverClassName(HttpDriver.class.getName()); new SessionFactory(configuration, "com.example"); ``` The problem can be worked around by calling `Components.configure(Configuration)` prior to creating the `Session Factory`: ```java Configuration configuration = new Configuration(); configuration.setDriverClassName(HttpDriver.class.getName()); Components.configure(configuration); new SessionFactory(configuration, "com.example"); ``` ## Context I am trying to upgrade Spring Boot 2.0 snapshots to use OGM 3.0 snapshots as required by Spring Data Neo4J 5.0 snapshots. ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> - OGM Version used: 3.0.0-SNAPSHOT - Java Version used: Java 8 - Neo4J Version used: N/A (things don't get that far) - Bolt Driver Version used (if applicable): N/A - Operating System and Version: OS X 10.10.5 - Link to your project: https://github.com/spring-projects/spring-boot Answers: username_1: @username_0: We are making changes to the `Configuration` in order to remove the static `Components` class. It appears we don't have a test for this situation. I will add it to the https://github.com/neo4j/neo4j-ogm/tree/refactor/remove-components before merging. The results of this merge will mean users can either a) Not supply a properties file and simply call `new Configuration.Builder()` b) supply a file and load it through `new Components.Builder(ConfigurationSource)`. Users can then override the initial values through the builder before calling `build()`. In this release we want to emphasis configuration via code and letting users supply their own configuration source. If that happens to be a `ClasspathConfigurationSource` then they can use that but we will be avoiding "autoconfiguration" as of 3.0. HTH username_1: @username_0 This should work if you build from source. We are having issues with one of our driver tests that is taking a long time. You should be able to keep going if you mvn install for now. username_1: @username_0 Can you update and check if this is still occurring (from SDN snapshot)? username_0: @username_1 This seems to have been fixed as part of the removal of `Components`. Thanks. Status: Issue closed
jeremyjh/dialyxir
653492448
Title: Exception on non-existent directory for :plt_core_path Question: username_0: ### Environment elixir: 1.10.3 otp_release: 23.0 ### Current behavior When you set `:plt_core_path` to non existent folder you will get following exceiption on `mix dialyzer`: ``` ... Creating dialyxir_erlang-23.0.plt :dialyzer.run error: The PLT file ./my_non_existent_folder/dialyxir_erlang-23.0.plt is not writable Looking up modules in dialyxir_erlang-23.0.plt ** (Mix) Could not open ./my_non_existent_folder/dialyxir_erlang-23.0.plt: no such file or directory ``` ### Expected behavior You can successfully create plt files in non existent folder. Answers: username_1: Is this what people would really expect? Is there another mix task that handles non-existent paths that way for example? I'm just not really sure that the expectation should be to silently create the full path. username_2: I would support a `--create-if-not-exists` type flag but creating by default is surprising behavior to me as a user. username_0: If I specify `:plt_file` in non-existent path it will create it, for example. And here `:plt_core_path` sounds like output folder if it's empty or not exists. `--create-if-not-exists` makes usage less obvious. When this error was occured my first thoughts were: 1. Is provided path correct? 2. Do I have rights to create folders here? username_1: @username_2 after looking at this more closely I've kind of come around to @username_0 point of view here - since this is a file we create, it may not be surprising to create its path and [we already do this](https://github.com/username_1/dialyxir/blob/master/lib/dialyxir/plt.ex#L167) when we copy the base PLT's to the target directory.
yixirumeng/Java-Interview-Questions
447423863
Title: synchronized和ReentrantLock的区别 Question: username_0: synchronized是和if、else、for、while一样的关键字,ReentrantLock是类,这是二者的本质区别。既然ReentrantLock是类,那么它就提供了比synchronized更多更灵活的特性,可以被继承、可以有方法、可以有各种各样的类变量,ReentrantLock比synchronized的扩展性体现在几点上: (1)ReentrantLock可以对获取锁的等待时间进行设置,这样就避免了死锁 (2)ReentrantLock可以获取各种锁的信息 (3)ReentrantLock可以灵活地实现多路通知 另外,二者的锁机制其实也是不一样的:ReentrantLock底层调用的是Unsafe的park方法加锁,synchronized操作的应该是对象头中mark word。