Unnamed: 0
int64
0
1.91M
id
int64
337
73.8M
title
stringlengths
10
150
question
stringlengths
21
64.2k
answer
stringlengths
19
59.4k
tags
stringlengths
5
112
score
int64
-10
17.3k
1,909,900
67,746,155
Exception: 'module' object is not iterable
<p>When I am running the code from <a href="https://github.com/ijawpikin/projectHub" rel="nofollow noreferrer">Github project</a>, I am getting this error:</p> <pre><code>Exception in thread django-main-thread: Traceback (most recent call last): File &quot;/data/data/com.termux/files/usr/lib/python3.9/site-packages/django/urls/resolvers.py&quot;, line 600, in url_patterns iter(patterns) TypeError: 'module' object is not iterable </code></pre> <p>The above exception was the direct cause of the following exception:</p> <pre><code>Traceback (most recent call last): File &quot;/data/data/com.termux/files/usr/lib/python3.9/threading.py&quot;, line 954, in _bootstrap_inner self.run() File &quot;/data/data/com.termux/files/usr/lib/python3.9/threading.py&quot;, line 892, in run self._target(*self._args, **self._kwargs) File &quot;/data/data/com.termux/files/usr/lib/python3.9/site-packages/django/utils/autoreload.py&quot;, line 64, in wrapper fn(*args, **kwargs) File &quot;/data/data/com.termux/files/usr/lib/python3.9/site-packages/django/core/management/commands/runserver.py&quot;, line 118, in inner_run self.check(display_num_errors=True) File &quot;/data/data/com.termux/files/usr/lib/python3.9/site-packages/django/core/management/base.py&quot;, line 419, in check all_issues = checks.run_checks( File &quot;/data/data/com.termux/files/usr/lib/python3.9/site-packages/django/core/checks/registry.py&quot;, line 76, in run_checks new_errors = check(app_configs=app_configs, databases=databases) File &quot;/data/data/com.termux/files/usr/lib/python3.9/site-packages/django/core/checks/urls.py&quot;, line 13, in check_url_config return check_resolver(resolver) File &quot;/data/data/com.termux/files/usr/lib/python3.9/site-packages/django/core/checks/urls.py&quot;, line 23, in check_resolver return check_method() File &quot;/data/data/com.termux/files/usr/lib/python3.9/site-packages/django/urls/resolvers.py&quot;, line 413, in check messages.extend(check_resolver(pattern)) File &quot;/data/data/com.termux/files/usr/lib/python3.9/site-packages/django/core/checks/urls.py&quot;, line 23, in check_resolver return check_method() File &quot;/data/data/com.termux/files/usr/lib/python3.9/site-packages/django/urls/resolvers.py&quot;, line 412, in check for pattern in self.url_patterns: File &quot;/data/data/com.termux/files/usr/lib/python3.9/site-packages/django/utils/functional.py&quot;, line 48, in __get__ res = instance.__dict__[self.name] = self.func(instance) File &quot;/data/data/com.termux/files/usr/lib/python3.9/site-packages/django/urls/resolvers.py&quot;, line 607, in url_patterns raise ImproperlyConfigured(msg.format(name=self.urlconf_name)) from e django.core.exceptions.ImproperlyConfigured: The included URLconf '&lt;module 'blog.urls' from '/storage/emulated/0/application/blog/urls.py'&gt;' does not appear to have any patterns in it. If you see valid patterns in the file then the issue is probably caused by a circular import. </code></pre>
<p>I had a look at your <code>urls.py</code> file from your github link</p> <p>There is a small typo where you have capitalised the <code>urlpatterns</code> object.</p> <pre class="lang-py prettyprint-override"><code>from django.urls import path from .views import blogListView Urlpatterns = [ path(' ', blogListView.as_view(), name='home'), ] </code></pre> <p>The spelling of <code>urlpatterns</code> must be exact in order for routing to work.</p> <p>Easy fix </p>
python|python-3.x|django|mobile|termux
1
1,909,901
30,625,894
iteratively intersecting line segments in Sympy... is there a better way?
<p>OK. I have the points that comprise the borders of a polygon. I want to (a) use Sympy's geometry module to determine, from all of the possible line-segments between any pair of points, which segments do not cross the perimeter. This result will be the "edges" that are allowed to be used in (b) a shortest_distance analysis in Networkx. My end goal is to iterate this process through many shapes, but I've hard-coded the coordinates for just 1 shape in this example. </p> <pre><code>import numpy import networkx as nx from sympy import geometry from itertools import combinations from matplotlib import pyplot as plot arr_bou = numpy.array([[-542.62545014, 961.34455209], [-544.45425379, 961.34455209], [-544.45425379, 962.25895392], [-547.19745928, 962.25895392], [-547.19745928, 963.17335575], [-549.02626294, 963.17335575], [-549.02626294, 964.08775758], [-550.85506659, 964.08775758], [-550.85506659, 961.34455209], [-552.68387025, 961.34455209], [-552.68387025, 962.25895392], [-553.59827208, 962.25895392], [-553.59827208, 965.91656123], [-552.68387025, 965.91656123], [-552.68387025, 967.7453649 ], [-551.76946842, 967.7453649 ], [-551.76946842, 968.65976672], [-550.85506659, 968.65976672], [-550.85506659, 967.7453649 ], [-548.11186111, 967.7453649 ], [-548.11186111, 965.91656123], [-547.19745928, 965.91656123], [-547.19745928, 964.08775758], [-546.28305745, 964.08775758], [-546.28305745, 965.00215941], [-543.53985197, 965.00215941], [-543.53985197, 963.17335575], [-542.62545014, 963.17335575], [-542.62545014, 964.08775758], [-540.79664648, 964.08775758], [-540.79664648, 963.17335575], [-539.88224465, 963.17335575], [-539.88224465, 962.25895392], [-542.62545014, 962.25895392], [-542.62545014, 961.34455209]]) boundXY = [] for i in arr_bou: boundXY.append((i[0],i[1])) points = [geometry.Point(i) for i in boundXY] poly = geometry.Polygon(*points) # use the * first to unpack the points (necessary to avoid errors) G = nx.Graph() positions = {} # build a dictionary for i in xrange(len(boundXY)): # that contains coordinates positions[i] = boundXY[i] # of each node on the graph's perimeter G.add_path(positions.keys())# add nodes to graph w/ boundary edges G.add_path([min(G.nodes()),max(G.nodes())]) combos_o = list(combinations(positions.keys(),2)) combos = [i for i in combos_o if i not in G.edges()] keepcombos = [] for combo in combos: pt1 = positions[combo[0]] pt2 = positions[combo[1]] line = geometry.Polygon(pt1,pt2) # there are 4 polygon sides that do not count as intersections # because 2 sides will intersect a point on each end test = True for side in poly.sides: if side.p1 != geometry.Point(pt1) and side.p1 != geometry.Point(pt2): if side.p2 != geometry.Point(pt1) and side.p2 != geometry.Point(pt2): if geometry.intersection(line,side): test = False break else: try: if poly.encloses(line.midpoint): pass else: test = False break except NotImplementedError: pass if test == True: keepcombos.append(combo) G.add_edges_from(keepcombos) </code></pre> <p>I've gotten this to work for small polygons (14 vertices) but this takes FOREVER with even just a 35 vertices, and other polygons will be larger than this still. </p> <p>Is there a more efficient way of finding all within-polygon node-pairs?</p> <p>Thanks!!</p>
<p>I found a solution that sped the process up by about 13x (for a polygon with 35 points (like the data listed above), the old method from the code in the question took about 4hours to find all line segments inside the polygon. This new method took 18 minutes instead.) </p> <p>Above I was iteratated through the points, and at each iteration looked at each border ("side") individually to see if the lines intersected. I changed this to instead intersect the line with the entire polygon. If it lines inside or on the edge, there should only be 2 points where it intersects, so if the length of the intersection is >2, I throw the combination out</p> <pre><code>for combo in combos: pt1 = geometry.Point(positions[combo[0]],evaluate=False) pt2 = geometry.Point(positions[combo[1]],evaluate=False) line = geometry.Polygon(pt1,pt2) try: if poly.encloses(line.midpoint): pass else: continue except NotImplementedError: continue intersect = geometry.intersection(line,poly) if len(intersect)&gt;2: continue keepcombos.append(combo) </code></pre> <p>This list "keepcombos" now has all of the lines (or networkx "edges") that I wish to include in a Dijkstra path analysis</p>
python|geometry|networkx|sympy
0
1,909,902
67,089,962
How to fix freeze_support() error for computing compute Perplexity and Coherence for LDA?
<p>I am going to compute Perplexity and Coherence for my textual data for LDA. I run the following codes</p> <pre><code># Compute Perplexity print('\nPerplexity: ', lda_model.log_perplexity(corpus)) # a measure of how good the model is. lower the better. # Compute Coherence Score coherence_model_lda = CoherenceModel(model=lda_model, texts=data_lemmatized, dictionary=id2word, coherence='c_v') coherence_lda = coherence_model_lda.get_coherence() print('\nCoherence Score: ', coherence_lda) </code></pre> <p>But I see this error (freeze_support()) I don't know how to fix it or even ignore it :</p> <pre><code>untimeError: An attempt has been made to start a new process before the current process has finished its bootstrapping phase. This probably means that you are not using fork to start your child processes and you have forgotten to use the proper idiom in the main module: if __name__ == '__main__': freeze_support() ... The &quot;freeze_support()&quot; line can be omitted if the program is not going to be frozen to produce an executable. </code></pre>
<pre><code># Compute Perplexity print('\nPerplexity: ', lda_model.log_perplexity(corpus)) # a measure of how good the model is. lower the better. # Compute Coherence Score if __name__ == '__main__': coherence_model_lda = CoherenceModel(model=lda_model, texts=data_lemmatized, dictionary=id2word, coherence='c_v') coherence_lda = coherence_model_lda.get_coherence() print('\nCoherence Score: ', coherence_lda) </code></pre>
python|spyder|data-mining|lda|topic-modeling
0
1,909,903
64,157,450
How to filter Django objects based on value returned by a method?
<p>I have an Django object with a method <code>get_volume_sum(self)</code> that return a float, and I would like to query the top n objects with the highest value, how can I do that?</p> <p>For example I could do a loop like this but I would like a more elegant solution.</p> <pre><code>vol = [] obj = [] for m in Market.object.filter(**args): # 3000 objects sum = m.get_volume_sum() vol.append(sum) obj.append(m.id) top = search_top_n(obj, vol, n) </code></pre> <p>And this is how the method looks like:</p> <pre><code># return sum of volume over last hours def get_volume_sum(self, hours): return Candle.objects.filter(market=self, dt__gte=timezone.now()-timedelta(hours=hours) ).aggregate(models.Sum('vo')) </code></pre> <p>From what I see <a href="https://www.geeksforgeeks.org/python-program-to-find-n-largest-elements-from-a-list/" rel="nofollow noreferrer">here</a> even with Python there isn't a single line solution.</p>
<p>You should not filter with the method, this will result in an <em>N+1</em> problem: for 3'000 <code>Market</code> objects, it will generate an additional 3'0000 queries to obtain the volumes.</p> <p>You can do this in <em>bulk</em> with a <a href="https://docs.djangoproject.com/en/dev/ref/models/querysets/#annotate" rel="nofollow noreferrer"><strong><code>.annotate(…)</code></strong> [Django-doc]</a>:</p> <pre><code>from django.db.models import Sum <i>hours</i> = 12 # some value for hours Market.objects.filter( **args, <b>candle__dt__gte=</b>timezone.now()-timedelta(hours=<i>hours</i>), ).annotate( <b>candle_vol=Sum('candle__vo')</b> ).order_by(<b>'-candle_vol'</b>)</code></pre> <p>Here there is however a small caveat: if there is no related <code>Candle</code>, then these <code>Market</code>s will be filtered out. We can prevent that by allowing also <code>Market</code>s without <code>Candle</code>s with:</p> <pre><code>from django.db.models import Q, Sum <i>hours</i> = 12 # some value for hours Market.objects.filter( <b>Q(</b>candle__dt__gte=timezone.now()-timedelta(hours=<i>hours</i>)<b>) |</b> <b>Q(candle=None)</b>, **args ).annotate( candle_vol=Sum('candle__vo') ).order_by('-candle_vol')</code></pre>
python|django
1
1,909,904
64,159,164
CMD color problems
<p>I want to make my python cmd output colorful! I have color-codes like this: <code>\033[91m</code></p> <p>Now the output in cmd isn't colorful. I get a &quot;←&quot;. How can I change this? Did anybody have the same problem? :D</p> <h1>Edit</h1> <p>Is there an alternative to cmd? Is it hard to programm a cmd window in e.g. C#?</p>
<p>You need to add just two more lines at the beginning of your script.</p> <pre><code>import os os.system(&quot;&quot;) </code></pre>
python
1
1,909,905
72,340,956
How to check differences between column values in pandas?
<p>I'm manually comparing two or three rows very similar using pandas. Is there a more automated way to do this? I would like a better method than using '=='.</p>
<p><a href="https://pandas.pydata.org/docs/reference/api/pandas.DataFrame.diff.html" rel="nofollow noreferrer">https://pandas.pydata.org/docs/reference/api/pandas.DataFrame.diff.html</a></p> <p>See if this will satisfy your needs.</p> <pre><code>df['sales_diff'] = df['sales'].diff() </code></pre> <p>The above code snippet creates a new column in your data frame, which contains the difference between the previous row by default. You can screw around with the parameters (axis) to compare rows or columns and you can change (period) to compare to a specific row or column.</p>
python|pandas|cell
0
1,909,906
65,520,557
fish: Unknown command: pip
<p>today I am trying Garuda KDE Dr460nized and I am running python on it. But when I use pip for installing packages I open my Konsole and, an error comes like this</p> <pre><code>fish: Unknown command: pip </code></pre> <p>I thought I should write pip3 instead of pip but still, the same error comes</p> <pre><code>fish: Unknown command: pip3 </code></pre> <p>Can anyone please tell me how to solve this issue in Garuda Linux. It is an Arch-based on Linux. Please tell me what is the solution. Any help will be appreciated. Thanks in advance</p>
<p>I think I have answered my question I have to add:</p> <pre><code>python -m pip install packageName </code></pre> <p>It solved my error. If anyone can't solve their error you can see this answer.</p>
python|linux|pip|konsole
1
1,909,907
50,345,862
django detect user login in another tab
<p>Is there anyway in django that if user has two open tabs, both logged out, then logs in in one tab, tell that he has logged in in another tab? I mean something like github that tells you you have signed in, please refresh the page.</p> <p>The problem is now If I login in one tab and then in the second tab, I get <code>csrf token missing incorrect.</code></p>
<p>You get <code>csrf token missing incorrect.</code> because when user relogins, the server generates a new csrf token to the cookie. The cookie persists across the same domain. And when you're trying to do smth on the current page, the request fails because csrf in your <code>&lt;form&gt;</code> differs from the cookie which has been changed. That's why github refreshes the page (instead of conitnuing doing request from it). Thus server will return new csrf in html to your form.</p> <p><strong>Edit</strong>:</p> <p>Consider the following choices :</p> <ul> <li>If your cookie is not readonly: . Set <code>setInterval</code> where you check the session which user loaded the page and current session from cookie.</li> <li>Render 4 first characters too page and save it to a variable. Set this variable when page loads. And with every request pass the variable with headers. Add a <a href="https://docs.djangoproject.com/en/2.0/topics/http/middleware/" rel="nofollow noreferrer">middleware</a> which checks if first 4 characters in headers matcher first 4 characters from cookie, and if it doesn't tell the client to refresh the page.</li> <li>If you want to automatically detect your case, you need to frequently spam the server and ask if the session has changed. From the client you can send the old session and the new one (if session is readonly you can send like few first characters from the server). </li> </ul>
python|django|csrf|django-csrf
1
1,909,908
50,370,411
Regex to find five consecutive consonants
<p>I need a regex for python that finds words with five consecutive consonants. </p> <p>These words would work - </p> <pre><code>tnortvcvni (rtvcvn) kahjdflka (hjdflk) </code></pre> <p>But these words wouldn't (no five letters in row without vowels) - </p> <pre><code>peanut butter jelly </code></pre>
<p>It seems you don't mean a fixed length of 5 characters but a minimum:</p> <pre><code>(?:(?![aeiou])[a-z]){5,} </code></pre> <p><a href="https://regex101.com/r/SW5Wdt/1" rel="nofollow noreferrer">Live demo</a></p> <p>Note: set <code>i</code> flag if it exists.</p>
python|regex
4
1,909,909
50,614,223
Python Regex Match failed
<p>This passed on <a href="https://regex101.com/" rel="nofollow noreferrer">https://regex101.com/</a> without any issues. Did I miss anything? The entire string is in one line.</p> <pre><code>def get_title_and_content(html): html = """&lt;!DOCTYPE html&gt; &lt;html&gt; &lt;head&gt; &lt;title&gt;Change delivery date with Deliv&lt;/title&gt; &lt;/head&gt; &lt;body&gt; &lt;div class="gkms web"&gt;The delivery date can be changed up until the package is assigned to a driver.&lt;/div&gt; &lt;/body&gt; &lt;/html&gt; """ title_pattern = re.compile(r'&lt;title&gt;(.*?)&lt;/title&gt;(.*)') match = title_pattern.match(html) if match: print('successfully extract title and answer') return match.groups()[0].strip(), match.groups()[1].strip() else: print('unable to extract title or answer') </code></pre>
<p>In a summary of the comments:</p> <p><code>title_pattern.search(html)</code> Should be used instead of <code>title_pattern.match(html)</code></p> <p>As the search function will search anywhere in the provided string instead of just from the beginning. <code>match = title_pattern.findall(html)</code> could be used similarly but would return a list of items instead of just one.</p> <p>Also as mentioned using <code>BeautifulSoup</code> would pay of more in the long run as Regular Expression is not properly suited for searching HTML</p>
python|regex
0
1,909,910
35,233,707
Checking ISBN numbers
<p>This is my code:</p> <pre><code>def isISBN(n): if len(n)!= 10: return False else: d1=int(n[0])*1 d2=int(n[1])*2 d3=int(n[2])*3 d4=int(n[3])*4 d5=int(n[4])*5 d6=int(n[5])*6 d7=int(n[6])*7 d8=int(n[7])*8 d9=int(n[8])*9 d10=(d1+d2+d3+d4+d5+d6+d7+d8+d9) num=d10%11 print(d10,num) if num==10: return True else: return False </code></pre> <p>Here are some test cases my teacher gave us:</p> <pre><code>&gt;&gt;&gt; isISBN('020103803X') True &gt;&gt;&gt; isISBN('0540122068') True &gt;&gt;&gt; isISBN('020108303X') False &gt;&gt;&gt; isISBN('0540122069') False </code></pre> <p>The code fails the test <code>'0540122068'</code> because my output is <code>False</code>, but I don't know why.</p>
<p>Don't forget the 10th value and check for modulo equivalence to 0:</p> <pre><code>def isISBN(n): if len(n)!= 10: return False else: d1=int(n[0])*1 d2=int(n[1])*2 d3=int(n[2])*3 d4=int(n[3])*4 d5=int(n[4])*5 d6=int(n[5])*6 d7=int(n[6])*7 d8=int(n[7])*8 d9=int(n[8])*9 if n[9] == 'X': d10 = 10 else: d10 = int(n[9]) d10 = d10*10 d11=(d1+d2+d3+d4+d5+d6+d7+d8+d9+d10) num=d11%11 if num==0: return True else: return False isISBN("3680087837") </code></pre>
python
2
1,909,911
26,867,730
How do you set a variable number of regex expressions?
<p>Currently I have <code>out = re.sub(r'[0-9][0-9][0-9]', '', input)</code>. I would like to have a variable number of <code>[0-9]</code>'s.</p> <p>So far I have;</p> <pre><code>string = '' for i in xrange(numlen): string = string + '[0-9]' string = 'r' + string out = re.sub(string, '', input) </code></pre> <p>This doesn't work, and I've tried using <code>re.compile</code>, but haven't had any luck. Is there a better way of doing this? Or am I just missing something trivial?</p>
<p>You can specify repetition using <code>{}</code>, for example 3 digits would be</p> <pre><code>[0-9]{3} </code></pre> <p>So you can do something like</p> <pre><code>reps = 5 # or whatever value you'd like out = re.sub('[0-9]{{{}}}'.format(reps), '', input) </code></pre> <p>Or if you don't know how many digits there will be</p> <pre><code>out = re.sub('[0-9]+', '', input) </code></pre>
python|regex
3
1,909,912
56,691,957
why is this code removing the file instead of renaming it?
<p>I want to rename report.json but it is removing the file instead </p> <pre><code>import os from pathlib import Path import json path =Path( r'C:\Users\Sajid\Desktop\cuckoo (3)\cuckoo\storage\analyses\3\reports') filename = os.path.join(path,"report.json") with open(filename) as json_file: data=json.load(json_file) var=(data['target']['file']['md5']) print(var) json_file.close() os.rename(filename,var) </code></pre> <p>I expect this code to rename the file and not delete it</p>
<p>It's probably not deleting it, but moving it to your working directory (so if you launched your script from <code>C:\Users\Sajid</code>, the file would be there, not in <code>C:\Users\Sajid\Desktop\cuckoo (3)\cuckoo\storage\analyses\3\reports</code>). <strong>Edit:</strong> Based on <a href="https://stackoverflow.com/questions/56691957/why-is-this-code-removing-the-file-instead-of-renaming-it/56692030#comment99949394_56691957">your comment</a>, this is definitely what's happening; the first time you ran your code, it moved it to your working directory (with component <code>cuckoo (1)</code>, where you probably expected it to go to the one with component <code>cuckoo (3)</code>), the second time it failed because <code>os.rename</code> won't overwrite an existing file on Windows.</p> <p>Change it to combine the desired target directory with the basename of the file, a la:</p> <pre><code>var = os.path.join(str(path), os.path.basename(data['target']['file']['md5'])) </code></pre> <p>so it stays in the same directory.</p> <p>You're also unnecessarily closing the file twice (once explicitly, once implicitly via the <code>with</code> block termination). As it happens, the first one is necessary, but only because you kept the <code>rename</code> inside the <code>with</code> block. The simplest solution is to just remove the extra <code>close</code>, and <code>rename</code> outside the <code>with</code> block (in fact, you don't need it open after the <code>json.load</code>, so you may as well close it then):</p> <pre><code>with open(filename) as json_file: data = json.load(json_file) # Dedent to implicitly close file var = os.path.join(str(path), os.path.basename(data['target']['file']['md5'])) print(var) os.rename(filename,var) </code></pre> <p>As written, you won't replace an existing file (on Windows; on UNIX-like systems, it will); if you want to silently replace the existing file everywhere, use <code>os.replace</code> instead of <code>os.rename</code>.</p>
python
0
1,909,913
56,710,475
How to download GeoTiff files from GeoServer using Python
<p>I am trying to download GeoTiff files from GeoServer using Python. I have a found a few resources online about this type of thing, but I have been unable to accomplish this task.</p> <p>For example, here: <a href="https://gis.stackexchange.com/questions/181560/download-geotiff-from-geoserver">https://gis.stackexchange.com/questions/181560/download-geotiff-from-geoserver</a> it seems that people have been able to do what I want to do, but they do not explain their process.</p> <p>Likewise, the accepted answer here: <a href="https://stackoverflow.com/questions/34291665/how-to-grab-a-tiff-image-from-python#comment56326289_34292069">How to grab a TIFF image from python</a> works for downloading GeoTiffs like the one at <a href="http://imgsrc.hubblesite.org/hu/db/images/hs-2006-01-a-hires_tif.tif" rel="nofollow noreferrer">http://imgsrc.hubblesite.org/hu/db/images/hs-2006-01-a-hires_tif.tif</a>, but there is no download link for GeoTiffs on GeoServer.</p> <p>Any help would be much appreciated!</p> <p>EDIT: Here are some more details about what I have tried thus far. GeoServer has a rest API server, at <a href="http://localhost:8080/geoserver/rest" rel="nofollow noreferrer">http://localhost:8080/geoserver/rest</a> locally, so I initially tried to access this url in python and then download the GeoTiff I want in "Layers". However, each of the files in "Layers" is an html file; what I would like to know is if there is a place where I can actually access the GeoTiff files programmatically. I think this is the root of the problem – although I am not really sure how to download GeoTiffs programmatically, I must first be able to actually access them in GeoServer.</p> <p>As far as progress, though, I have not been able to make much. As I mentioned above, I was able to download a GeoTiff using the code at <a href="https://stackoverflow.com/questions/34291665/how-to-grab-a-tiff-image-from-python#comment56326289_34292069">How to grab a TIFF image from python</a>, but I have been unable to do this for a GeoTiff on GeoServer. I am new to both GeoServer and the GeoTiff format, so I am not quite sure how to approach this problem.</p>
<p>As the answer to your <a href="https://gis.stackexchange.com/questions/181560/download-geotiff-from-geoserver">linked question</a> says you need to make a <a href="https://www.opengeospatial.org/standards/wcs" rel="nofollow noreferrer">WCS request</a> to GeoServer to fetch a GeoTiff. </p> <p>The <a href="https://docs.geoserver.org/stable/en/user/services/wcs/index.html" rel="nofollow noreferrer">GeoServer manual</a> provides a <a href="https://docs.geoserver.org/stable/en/user/services/wcs/reference.html" rel="nofollow noreferrer">WCS reference</a> that should help you get an understanding of how to proceed. You can also go to the demos page of your GeoServer installation and use the WCS Request builder to create an XML file that you can use as a template in your python program.</p> <p><a href="https://i.stack.imgur.com/m1foo.png" rel="nofollow noreferrer"><img src="https://i.stack.imgur.com/m1foo.png" alt="GeoServer WCS Request builder"></a></p>
python|tiff|geoserver|wcs
0
1,909,914
61,273,511
Cumulative sum in pyspark
<p>I am trying to compute the cumulative sum per class. Code is working fine by using sum(df.value).over(Window.partitionBy('class').orderBy('time'))</p> <pre><code>df = sqlContext.createDataFrame( [(1,10,"a"),(3,2,"a"),(1,2,"b"),(2,5,"a"),(2,1,"b"),(9,0,"b"),(4,1,"b"),(7,8,"a"),(3,8,"b"),(2,5,"a"),(0,0,"a"),(4,3,"a")], ["time", "value", "class"] ) time|value|class| +----+-----+-----+ | 1| 10| a| | 3| 2| a| | 1| 2| b| | 2| 5| a| | 2| 1| b| | 9| 0| b| | 4| 1| b| | 7| 8| a| | 3| 8| b| | 2| 5| a| | 0| 0| a| | 4| 3| a| df.withColumn('cumsum_value', sum(df.value).over(Window.partitionBy('class').orderBy('time'))).show() time|value|class|cumsum_value| +----+-----+-----+------------+ | 1| 2| b| 2| | 2| 1| b| 3| | 3| 8| b| 11| | 4| 1| b| 12| | 9| 0| b| 12| | 0| 0| a| 0| | 1| 10| a| 10| | 2| 5| a| 20| | 2| 5| a| 20| | 3| 2| a| 22| | 4| 3| a| 25| | 7| 8| a| 33| +----+-----+-----+------------+ </code></pre> <p>But its not working with duplicate rows. Desired output should be:</p> <pre><code> time|value|class|cumsum_value| +----+-----+-----+------------+ | 1| 2| b| 2| | 2| 1| b| 3| | 3| 8| b| 11| | 4| 1| b| 12| | 9| 0| b| 12| | 0| 0| a| 0| | 1| 10| a| 10| | 2| 5| a| 15| | 2| 5| a| 20| | 3| 2| a| 22| | 4| 3| a| 25| | 7| 8| a| 33| +----+-----+-----+------------+ </code></pre>
<p>Adding to <em><code>@pault</code></em>'s comment, I would suggest a <strong><code>row_number()</code></strong> calculation based on <strong><code>orderBy('time', 'value')</code></strong> and then use that column in the <strong><code>orderBy</code></strong> of another window(<strong><code>w2</code></strong>) to get your <strong><code>cum_sum</code></strong>. </p> <p><em>This will handle both cases where time is the same and value is the same, and where time is the same but value isnt.</em></p> <pre><code>from pyspark.sql import functions as F from pyspark.sql.window import Window w1=Window().partitionBy("class").orderBy("time","value") w2=Window().partitionBy("class").orderBy('rownum') df.withColumn('rownum', F.row_number().over(w1))\ .withColumn('cumsum_value', F.sum("value").over(w2)).drop('rownum').show() +----+-----+-----+------------+ |time|value|class|cumsum_value| +----+-----+-----+------------+ | 1| 2| b| 2| | 2| 1| b| 3| | 3| 8| b| 11| | 4| 1| b| 12| | 9| 0| b| 12| | 0| 0| a| 0| | 1| 10| a| 10| | 2| 5| a| 15| | 2| 5| a| 20| | 3| 2| a| 22| | 4| 3| a| 25| | 7| 8| a| 33| +----+-----+-----+------------+ </code></pre>
python|dataframe|pyspark
3
1,909,915
60,673,168
Weird linear regression learning curve
<p>I'm trying to build a prediction model for apartments price. I use python scikit-learn toolset. I'm using a dataset having total floor area and location of the apartment, which I have converted to dummy features. So the dataset looks like this: <a href="https://i.stack.imgur.com/OmGL2.png" rel="nofollow noreferrer"><img src="https://i.stack.imgur.com/OmGL2.png" alt="enter image description here"></a> Then I build a learning curve to see how the model is doing. I build the learning curve this way:</p> <pre class="lang-py prettyprint-override"><code>from matplotlib import pyplot as plt from sklearn.linear_model import LinearRegression from sklearn.model_selection import learning_curve model = LinearRegression() training_sizes, training_scores, validation_scores = learning_curve( estimator = model, X = X_train, y = y_train, train_sizes = np.linspace(5, len(X_train) * 0.8, dtype = int), cv = 5 ) line1, line2 = plt.plot( training_sizes, training_scores.mean(axis = 1), 'g', training_sizes, validation_scores.mean(axis = 1), 'r') plt.legend((line1, line2), ('Training', 'Cross-validation')) </code></pre> <p>The picture I see is somewhat confusing: <a href="https://i.stack.imgur.com/Ixbpx.png" rel="nofollow noreferrer"><img src="https://i.stack.imgur.com/Ixbpx.png" alt="Learning curve on training and cross-validation sets"></a> Anomalies I see here are:</p> <ol> <li>Huge error on cross-validation set</li> <li>Error not steadily decreasing on training examples number growth. </li> </ol> <p>Is it normal? </p> <p>Learning curve of training set only is also not so smooth but at least the error isn't that huge: <a href="https://i.stack.imgur.com/0kiFr.png" rel="nofollow noreferrer"><img src="https://i.stack.imgur.com/0kiFr.png" alt="Learning curve on training set"></a></p> <p>Also I tried to add to add polynomial features of 2nd degree. But this didn't make the model perform any different. And because I have a lot of categorical features (total 106) it takes quite long even for 2nd degree polynomial. So I didn't try for higher degrees. </p> <p>Also I tried to build a model using as simple cost function and gradient descent as possible using Octave. The result with weird error was same. </p> <p>Update: Thanks to <a href="https://stackoverflow.com/a/60734720/200160">tolik</a> I made several amendments:</p> <p><strong>Data preparation:</strong> Categorical data are independent. So I can't combine them into one feature. Features were scaled using StandardScaler(). Thank you for that.</p> <p><strong>Feature extraction:</strong> After features transformation with PCA I found out one new feature has explained variance ratio over 99%. Though it's strange I used only this one. That also allowed to increase polynomial degree though it didn't increase performance.</p> <p><strong>Model selection:</strong> I tried several different models but none seem to perform better than LinearRegression. Interesting thing - all models perform worse on full data set. Probably it's because I sorted by price and higher prices are rather outliers. So when I start training sets on 1000 samples and go to the maximum, I get this picture (for nearly all models): <a href="https://i.stack.imgur.com/wsS30.png" rel="nofollow noreferrer"><img src="https://i.stack.imgur.com/wsS30.png" alt="enter image description here"></a></p>
<p>My explanation have 3 steps: The data preparation, feature extraction, and model selection.</p> <p><strong>Data preparation:</strong></p> <ul> <li>In this dataset there are lots of Categorical and Ordinal values. If the column has several non related categories it's ok to one-hot it. but if the column has categories with order like "bad","normal","good" you can convert it to numerical as {Good:1,Normal:0.5,Bad:0}.</li> <li>Value ranges: the value ranges for each feature differs from the other, therefore the best thing to do is to normalise each feature along itself between 0:1.</li> </ul> <p><strong>Feature Extraction:</strong></p> <ul> <li>Your goal is to maximise the score so I guess you don't care about which feature is more important. Use <strong>PCA</strong> (has an implementation in scikit-learn library) , this algorithm convert your feature vectors into different features that each of them is a linear combination of the other features. These new features are ordered by their explained variance. The first features describes the data better than the last one. You select the first features that their <code>explained_variance_</code> sums to 99%. Now you have weigh less features.</li> </ul> <p><strong>Model Selection:</strong> You don't really know what is a good model, because <strong>No Free Lunch Theory</strong> but in this problem the best results that don't use deep learning , use these: XGBoost-Regressor , Random-Forest-Regressor ,Ada-Boost.</p> <p><strong>The most important thing is the Data Preparation!!!</strong></p>
python|scikit-learn|linear-regression
0
1,909,916
71,648,462
Find most common words from list of strings
<p>We have a given list:</p> <pre><code>list_of_versions = ['apple II' ,'apple', 'apple 1' , 'HD APPLE','apple 3.5', 'adventures of apple' , 'apple III','orange 2' ,'300mhz apple', '300-orange II' , 'orange II HD' , 'orange II tvx', 'orange 2' , 'HD berry-vol 2', 'berry II', 'berry 2', 'berry VI', 'berry 1', 'berry II' ,'berry' ,'II Berry'] </code></pre> <p>How can I find the main word of each string? For example:</p> <pre><code>word | main -------------------------- apple II |apple val/apple |apple apple 1 |apple HD APPLE |apple apple 3.5 |apple adventures of apple |apple apple III |apple 300mhz apple |apple orange 2 |orange 300-orange II |orange orange II HD |orange /orange/II-tvx |orange orange 2 |orange HD berry-vol 2 |berry berry-II |berry -berry-2 |berry (berry) VI |berry berry 1 |berry berry II |berry berry 2022 B8 |berry II Berry-hd |berry 22 Berry II |berry Berry 6.8.9 |berry </code></pre> <p>Important points:</p> <ol> <li><p>I can not create the main word list which contains three main words (apple, orange, berry). Because the list will be updated with new main words. So we will never know what is the new words.</p> </li> <li><p>there is no limit to versions. At some point, we can see something like 'apple XII' or 'GB-HD berry 2.4' so version value can be everything. (in case you want to create a stopword list)</p> </li> </ol> <p><em>Nice to have (but it is not mandatory)--&gt; Adding another column as well which shows the version. i.e:</em></p> <pre><code>word | main | version ----------------------------------- apple II |apple | II val/apple |apple | NULL apple 1 |apple | 1 HD APPLE |apple | HD apple 3.5 |apple | 3.5 apple III |apple | III 300mhz apple II |apple | II orange 2 |orange | 2 300-orange II |orange | II orange II HD |orange | II HD /orange/II-tvx |orange | II tvx orange 2 |orange | 2 HD berry-vol 2 |berry | 2 HD berry-II |berry | II -berry-2 |berry | 2 (berry) VI |berry | VI berry 1 |berry | 1 berry II |berry | II berry 2022 |berry | NULL II Berry-hd |berry | II HD 22 Berry |berry | 22 Berry 6.8.9 |berry | 6.8.9 </code></pre>
<p>All the other answers omit the entry containing the word &quot;adventures&quot; because it throws off the search. You need a heuristic that can combine &quot;longest&quot; with &quot;most frequent&quot;.</p> <p>One thing that helps is that finding the longest word in each row greatly increases SNR. In other words, it filters out the unnecessary words pretty well, and just needs a little help. If you know how many words you are looking for (three in this case), you're all set:</p> <pre><code>from collections import Counter common_long_words = [word.casefold() for word in (max(re.findall('\\w+', version), key=len) for version in list_of_versions)] words = Counter(common_long_words).most_common(3) </code></pre> <p>Splitting off the version and finding the word of interest is not especially difficult. You have a couple of options regarding what constitutes a version, especially when the main word is embedded in the middle of the phrase. Here is a simple function that takes the entire remainder:</p> <pre><code>def split_main(version, words): for word in words: i = version.find(word) if i &gt; 0: return word, f'{version[:i]} {version[i + len(word)]}' else: raise ValueError(f'Version &quot;{version}&quot; does not contain any of the main words {{{&quot;, &quot;.join(words)}}}') result = {version: split_main(version, words) for version in list_of_versions} </code></pre>
python|arrays|pandas|string|list
3
1,909,917
55,245,535
Python: Google-Maps-API sends unknown format to parse
<p>I use the <a href="https://github.com/googlemaps/google-maps-services-python" rel="nofollow noreferrer">Python Client for Google Maps Services</a> to get following data from google-maps:</p> <pre><code>{ 'address_components':[ { 'long_name':'20', 'short_name':'20', 'types':[ 'street_number' ] }, { 'long_name':'Oberböhl', 'short_name':'Oberböhl', 'types':[ 'route' ] }, { 'long_name':'Ingelheim am Rhein', 'short_name':'Ingelheim am Rhein', 'types':[ 'locality', 'political' ] }, { 'long_name':'Mainz-Bingen', 'short_name':'Mainz-Bingen', 'types':[ 'administrative_area_level_3', 'political' ] }, { 'long_name':'Rheinland-Pfalz', 'short_name':'RP', 'types':[ 'administrative_area_level_1', 'political' ] }, { 'long_name':'Germany', 'short_name':'DE', 'types':[ 'country', 'political' ] }, { 'long_name':'55218', 'short_name':'55218', 'types':[ 'postal_code' ] } ], 'adr_address':'&lt;span class="street-address"&gt;Oberböhl 20&lt;/span&gt;, &lt;span class="postal-code"&gt;55218&lt;/span&gt; &lt;span class="locality"&gt;Ingelheim am Rhein&lt;/span&gt;, &lt;span class="country-name"&gt;Germany&lt;/span&gt;', 'formatted_address':'Oberböhl 20, 55218 Ingelheim am Rhein, Germany', 'formatted_phone_number':'06132 5099968', 'geometry':{ 'location':{ 'lat':49.9810156, 'lng':8.0739617 }, 'viewport':{ 'northeast':{ 'lat':49.9823942302915, 'lng':8.075293780291501 }, 'southwest':{ 'lat':49.9796962697085, 'lng':8.072595819708498 } } }, 'icon':'https://maps.gstatic.com/mapfiles/place_api/icons/generic_business-71.png', 'id':'d2b37ffe23fd5e76648a90df2987558b039fcdf7', 'international_phone_number':'+49 6132 5099968', 'name':'Esch Metalltechnik GmbH', 'place_id':'ChIJHaERGJ_svUcRRfqNoGXq3EU', 'plus_code':{ 'compound_code':'X3JF+CH Ingelheim am Rhein, Germany', 'global_code':'8FXCX3JF+CH' }, 'reference':'ChIJHaERGJ_svUcRRfqNoGXq3EU', 'scope':'GOOGLE', 'types':[ 'general_contractor', 'point_of_interest', 'establishment' ], 'url':'https://maps.google.com/?cid=5034156205699627589', 'utc_offset':60, 'vicinity':'Oberböhl 20, Ingelheim am Rhein', 'website':'http://www.esch-metalltechnik.de/' }{ 'long_name':'55218', 'short_name':'55218', 'types':[ 'postal_code' ] } </code></pre> <p>Now I want to extract certain variables, like the <code>"street_number"</code>. I don't know which format this data is, so I worked with it like a dictionary:</p> <pre><code>try: self.hausnr = place_result_2["address_components"][0]["long_name"] except: self.hausnr = "NA" </code></pre> <p>The problem is, that the index "0" isn't always the same position of the data I want, i varies. Is there a way to extract the data in another way? Perhaps I have to use a JSON-parser or something similar?</p> <p>Thanks a lot.</p>
<p>The answer is: <a href="https://docs.python.org/3/tutorial/datastructures.html#list-comprehensions" rel="nofollow noreferrer">List comprehensions</a></p> <pre><code>try: # make a list of all address components that have type "street number" comp = [c for c in place_result_2["address_components"] if "street_number" in c["types"]] # the first one of them (assuming there will never be more than one) is the desired one self.hausnr = comp[0]["long_name"] except: self.hausnr = "NA" </code></pre> <p>Since this will probably be a common operation, make a function:</p> <pre><code>def get_address_component(place_result, comp_type, comp_property="long_name", default=None): """ returns the first address component of a given type """ try: comp = [c for c in place_result["address_components"] if comp_type in c["types"]] return comp[0][comp_property] except KeyError: return default # ... self.hausnr = get_address_component(place_result_2, "street_number", default="NA") </code></pre> <hr> <p>PS, regarding:</p> <blockquote> <p>Perhaps I have to use a JSON-parser or something similar?</p> </blockquote> <p>JSON is a data transfer format - it's plain text. The Google API server used it to get the data across the wire. In your program it has already been parsed - by the Google API client library you are using. What you are looking at is not JSON anymore, it's a Python data structure (nested dicts and lists and values). It just happens to look quite similar to JSON when you print it to the console, because Python uses a similar format to represent data.</p> <p>On other words, no, you don't need to JSON-parse it again.</p>
python|google-maps|parsing|google-api|extract
1
1,909,918
57,382,770
Why using "--requirements_file" uploads dependencies onto GCS?
<p>I'm currently generating a template with those parameters:</p> <pre><code> --runner DataflowRunner \ --requirements_file requirements.txt \ --project ${GOOGLE_PROJECT_ID} \ --output ${GENERATED_FILES_PATH}/staging \ --staging_location=${GENERATED_FILES_PATH}/staging \ --temp_location=${GENERATED_FILES_PATH}/temp \ --template_location=${GENERATED_FILES_PATH}/templates/calculation-template \ </code></pre> <p>and the SDK is uploading dependencies specified inside <code>requirements.txt</code> onto GCS inside the staging section. I do not understand... For me using this kind of file would allow workers to directly pull dependencies from the official pip registry, not from my GCS, right?</p> <p>It makes running this command very long since it needs to upload packages :/</p> <p>Any explanation why is it happening? Maybe I'm doing something wrong?</p> <p>Thank you,</p>
<p>I believe this is done to make the Dataflow worker startup process more efficient and consistent (both initially and when auto-scaling). Without this, each time a Dataflow worker starts up, that worker has to directly connect to PyPI to find the latest matching versions of dependencies. Instead of this, set of dependencies are staged at pipeline initiation and are consistently installed in workers throughout the pipeline execution.</p>
python|google-cloud-dataflow|apache-beam|dataflow
1
1,909,919
57,423,482
Got only first row in table when using Selenium scraping (Python)
<p>I'm trying to scrape the whole table from: <a href="https://free-proxy-list.net/" rel="nofollow noreferrer">https://free-proxy-list.net/</a> </p> <p>And I managed to scrape it but it resulted in only the first row of the table instead of 20 rows. </p> <p>I saw previous similar questions that were answered and I have tried the solutions given but my selenium was unable to locate the element when I use <code>.//</code> for my xpath.</p> <pre class="lang-python prettyprint-override"><code>for bod in driver.find_elements_by_xpath("//*[@id='proxylisttable']/tbody"): col = bod.find_elements_by_xpath("//*[@id='proxylisttable']/tbody/tr") for c in col: ip = c.find_element_by_xpath('//*[@id="proxylisttable"]/tbody/tr/td[1]') port = c.find_element_by_xpath('//*[@id="proxylisttable"]/tbody/tr/td[2]') code = c.find_element_by_xpath('//*[@id="proxylisttable"]/tbody/tr/td[3]') country = c.find_element_by_xpath('//*[@id="proxylisttable"]/tbody/tr/td[4][@class = "hm"]') anonymity = c.find_element_by_xpath('//*[@id="proxylisttable"]/tbody/tr/td[5]') google = c.find_element_by_xpath('//*[@id="proxylisttable"]/tbody/tr/td[6][@class = "hm"]') </code></pre> <p>My code resulted in only scraping the first row 20 times instead of getting 20 rows. The rows are indicated at <code>ip</code>, <code>port</code>, <code>code</code>, etc. I have tried multiple types of xpath syntax but still end up the same.</p>
<p>I think your problem is in this line :</p> <pre><code>col = bod.find_elements_by_xpath("//*[@id='proxylisttable']/tbody/tr") </code></pre> <p>The correct syntax is :</p> <pre><code>col = bod.find_elements_by_xpath("//*[@id='proxylisttable']/tbody/tr[insert count here]") </code></pre> <p>Like this :</p> <pre><code>table = driver.find_element_by_xpath("//*[@id='proxylisttable']/tbody") rows = table.find_elements_by_xpath("//*[@id='proxylisttable']/tbody/tr") for i in range (1, len(rows)+1): row = table.find_element_by_xpath("//*[@id='proxylisttable']/tbody/tr[" +str(i) +']') for c in row: ip = c.find_element_by_xpath('//*[@id="proxylisttable"]/tbody/tr/td[1]') port = c.find_element_by_xpath('//*[@id="proxylisttable"]/tbody/tr/td[2]') code = c.find_element_by_xpath('//*[@id="proxylisttable"]/tbody/tr/td[3]') country = c.find_element_by_xpath('//*[@id="proxylisttable"]/tbody/tr/td[4][@class = "hm"]') anonymity = c.find_element_by_xpath('//*[@id="proxylisttable"]/tbody/tr/td[5]') google = c.find_element_by_xpath('//*[@id="proxylisttable"]/tbody/tr/td[6][@class = "hm"]') </code></pre>
python-3.x|selenium|xpath
2
1,909,920
53,979,127
How to concatenate part of three layers in Keras?
<p>I can use <code>keras.layers.concatenate</code> to concatenate two layers then send them to next layer, but if I want to take part of two layers then concatenate them and then send them to next layer, what should I do?</p> <p>For example, I want to take part of first conv layer and part of the second conv layer and the last pooling layer, and then concatenate them together to form one layer. But Keras is a high level library, how do we take part of them? </p> <p>You can understand me better by looking at the Figure 2 in paper <a href="http://mi.eng.cam.ac.uk/~sjy/papers/wgmv17.pdf" rel="nofollow noreferrer">A Network-based End-to-End Trainable Task-oriented Dialogue System</a>.</p>
<p>Well, you can slice them as you want, like the way you would slice a numpy array or a Python list, and use <code>K.concatenate</code>, all in a <code>Lambda</code> layer. For example:</p> <pre><code>from keras import backend as K # ... out = Lambda(lambda x: K.concatenate([x[0][:,:10], x[1][:,:10], x[2][:,:10]], axis=the_concat_axis))([conv1, conv2, pool]) </code></pre> <p>Note that the first axis is the batch axis, so you may want to keep all of it (i.e. use <code>:</code> as above).</p>
python|tensorflow|keras|deep-learning|nlp
2
1,909,921
53,864,160
Can we run tensorflow lite on linux ? Or it is for android and ios only
<p>Hi is there any possibility to run tensorflow lite on linux platform? If yes, then how we can write code in java/C++/python to load and run models on linux platform? I am familiar with bazel and successfully made Android and ios application using tensorflow lite.</p>
<p>I think the other answers are quite wrong.</p> <p>Look, I'll tell you my experience... I've been working with Django for many years, and I've been using normal tensorflow, but there was a problem with having 4 or 5 or more models in the same project. I don't know if you know Gunicorn + Nginx. This generates workers, so if you have 4 machine learning models, for every worker it multiplies, if you have 3 workers you will have 12 models preloaded in RAM. This is not efficient at all, because if the RAM overflows your project will fall or in fact the service responses are slower.</p> <p>So this is where Tensorflow lite comes in. Switching from a tensorflow model to tensorflow lite improves and makes things much more efficient. Times are reduced absurdly. Also, Django and Gunicorn can be configured so that the model is pre-loaded and compiled at the same time. So every time the API is used up, it only generates the prediction, which helps you make each API call a fraction of a second long. Currently I have a project in production with 14 models and 9 workers, you can understand the magnitude of that in terms of RAM. And besides doing thousands of extra calculations, outside of machine learning, the API call does not take more than 2 seconds. Now, if I used normal tensorflow, it would take at least 4 or 5 seconds.</p> <p>In summary, if you can use tensorflow lite, I use it daily in Windows, MacOS, and Linux, it is not necessary to use Docker at all. Just a python file and that's it. If you have any doubt you can ask me without any problem.</p> <p>Here a example project <a href="https://github.com/Nouvellie/django-js-ml/tree/main/project/apps/tensorflow" rel="nofollow noreferrer">Django + Tensorflow Lite</a></p>
linux|tensorflow-lite
5
1,909,922
53,992,840
Don't understand these ModuleNotFound errors
<p>I am a beginner and learning Python. I have setup the environment with SublimeText and Python3.x</p> <p>I am fine in creating code on Sublime and building it locally through Ctrl+B and for <code>input()</code> function I installed SublimeREPL and it works find up till now.</p> <p>The issue I am facing is on Python interpreter. I am facing below error while import any package:</p> <pre class="lang-none prettyprint-override"><code>import tweepy Traceback (most recent call last): File "&lt;stdin&gt;", line 1, in &lt;module&gt; ModuleNotFoundError: No module named 'tweepy' </code></pre> <p>I cannot even run a Python script from there. My Python path is below:</p> <pre class="lang-none prettyprint-override"><code>C:\Users\waqas.FVC\AppData\Local\Programs\Python\Python37-32 </code></pre> <p><code>Demo.py</code> and <code>hello.py</code> are the two scripts I wrote initially which I am trying to execute from Python Terminal, but it is showing below errors:</p> <pre class="lang-none prettyprint-override"><code>test.py Traceback (most recent call last): File "&lt;stdin&gt;", line 1, in &lt;module&gt; NameError: name 'test' is not defined Demo.py Traceback (most recent call last): File "&lt;stdin&gt;", line 1, in &lt;module&gt; NameError: name 'Demo' is not defined </code></pre>
<p>The initial Python download includes a number of libraries, but there are many, many more that must be downloaded and installed separately. Tweepy is among those libraries.</p> <p>You can find, and download, tweepy from here:</p> <p><a href="https://pypi.org/project/tweepy/" rel="nofollow noreferrer">https://pypi.org/project/tweepy/</a></p>
python|python-3.x|error-handling|compiler-errors
1
1,909,923
58,520,608
How to group rows, count in one column and do the sum in the other?
<p>I want to group rows of a csv file, count in one column and add in the other.</p> <p>For example with the following I would like to group the lines on the <code>Commune</code> to make columns of the <code>winner</code> with the count and a column <code>Swing</code> with the sum</p> <pre><code>Commune Winner Swing longitude latitude turnout Paris PAM 1 12.323 12.093 0.3242 Paris PJD 0 12.323 12.093 0.1233 Paris PAM 1 12.323 12.093 0.534 Paris UDF 1 12.323 12.093 0.65434 Madrid PAM 0 10.435 -3.093 0.3423 Madrid PAM 1 10.435 -3.093 0.5234 Madrid PJD 0 10.435 -3.093 0.235 </code></pre> <p>How to group rows, have a column in one column and a sum in the other?</p> <pre><code>Commune PAM PJD UDF Swing Paris 3 1 1 3 Madrid 2 1 0 1 </code></pre> <p>So far I tried try :</p> <pre><code>g = df.groupby('Commune').Winner pd.concat([g.apply(list), g.count()], axis=1, keys=['members', 'number']) </code></pre> <p>But it returns: </p> <pre><code> members number Commune Paris [PAM, PJD, PAM, UDF] 4 Madrid [PAM, PAM, UDF] 3 </code></pre>
<p>Use <a href="http://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.crosstab.html" rel="nofollow noreferrer"><code>crosstab</code></a> and add new column with <a href="http://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.join.html" rel="nofollow noreferrer"><code>DataFrame.join</code></a> and aggregate <code>sum</code>:</p> <pre><code>df = pd.crosstab(df['Commune'], df['Winner']).join(df.groupby('Commune')['Swing'].sum()) print (df) PAM PJD UDF Swing Commune Madrid 2 1 0 1 Paris 2 1 1 3 </code></pre> <p>But if need counts of rows:</p> <pre><code>df1 = pd.crosstab(df['Commune'], df['Winner'], margins=True, margins_name='Total').iloc[:-1] </code></pre> <p>Or:</p> <pre><code>df = pd.crosstab(df['Commune'], df['Winner']).assign(Total= lambda x: x.sum(axis=1)) </code></pre> <hr> <pre><code>print (df1) Winner PAM PJD UDF Total Commune Madrid 2 1 0 3 Paris 2 1 1 4 </code></pre> <p>EDIT:</p> <p>If another columns then is possible use aggregation by <code>first</code> if all values per groups and for <code>turnout</code> use some another aggregate function like <code>mean</code>, <code>sum</code>...:</p> <pre><code>df1 = (df.groupby('Commune') .agg({'Swing':'sum', 'longitude':'first','latitude':'first','turnout':'mean'})) print (df1) Swing longitude latitude turnout Commune Madrid 1 10.435 -3.093 0.36690 Paris 3 12.323 12.093 0.40896 df = pd.crosstab(df['Commune'], df['Winner']).join(df1) print (df) PAM PJD UDF Swing longitude latitude turnout Commune Madrid 2 1 0 1 10.435 -3.093 0.36690 Paris 2 1 1 3 12.323 12.093 0.40896 </code></pre> <p>If want <code>mean</code> of all columns without <code>Swing</code> is possible create dictionary dynamically:</p> <pre><code>d = dict.fromkeys(df.columns.difference(['Commune','Winner','Swing']), 'mean') d['Swing'] = 'sum' print (d) {'latitude': 'mean', 'longitude': 'mean', 'turnout': 'mean', 'Swing': 'sum'} df1 = df.groupby('Commune').agg(d) print (df1) latitude longitude turnout Swing Commune Madrid -3.093 10.435 0.36690 1 Paris 12.093 12.323 0.40896 3 df = pd.crosstab(df['Commune'], df['Winner']).join(df1) print (df) PAM PJD UDF latitude longitude turnout Swing Commune Madrid 2 1 0 -3.093 10.435 0.36690 1 Paris 2 1 1 12.093 12.323 0.40896 3 </code></pre>
python|python-3.x|pandas|pandas-groupby
3
1,909,924
65,470,708
Best way to fill NULL values with conditions using Pandas?
<p>So for example I have a data looks like this:</p> <pre><code>df = pd.DataFrame([[np.NaN, '1-5'], [np.NaN, '26-100'], ['Yes', 'More than 1000'], ['No', '26-100'], ['Yes', '1-5']], columns=['self_employed', 'no_employees']) df self_employed no_employees 0 nan 1-5 1 nan 26-100 2 Yes More than 1000 3 No 26-100 4 Yes 1-5 </code></pre> <p>And I'm trying to fill the NULL value based on the condition that:</p> <pre><code>If no_employees is '1-6' then 'Yes', otherwise 'No' </code></pre> <p>I was able to complete this using the dictionary such as:</p> <pre><code>self_employed_dict = {'1-5': 'Yes', '6-25': 'No', '26-100': 'No', '100-500': 'No', 'More than 1000':'No', '500-1000': 'No'} df['self_employed'] = df['self_employed'].fillna(df['no_employees'].map(self_employed_dict)) </code></pre> <p>But I wanted to know if there is a better, simple way of doing this. In this example, I had to write the dictionary for myself to map it, so how can I do this in the smart way?</p> <p>The expected output looks like this:</p> <pre><code> self_employed no_employees 0 Yes 1-5 1 No 26-100 2 Yes More than 1000 3 No 26-100 4 Yes 1-5 </code></pre>
<p>Use <a href="https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.fillna.html" rel="nofollow noreferrer">fillna</a> is the right way to go, but instead you could do:</p> <pre><code>values = df['no_employees'].eq('1-5').map({False: 'No', True: 'Yes'}) df['self_employed'] = df['self_employed'].fillna(values) print(df) </code></pre> <p><strong>Output</strong></p> <pre><code> self_employed no_employees 0 Yes 1-5 1 No 26-100 2 Yes More than 1000 3 No 26-100 4 Yes 1-5 </code></pre>
python|pandas|dataframe
3
1,909,925
45,300,037
Cannot upload large file to Google Cloud Storage
<p>It is okay when dealing with small files. It doesn't work only when I try to upload large files. I'm using Python client. The snippet is:</p> <pre><code>filename='my_csv.csv' storage_client = storage.Client() bucket_name = os.environ["GOOGLE_STORAGE_BUCKET"] bucket = storage_client.get_bucket(bucket_name) blob = bucket.blob("{}".format(filename)) blob.upload_from_filename(filename) # file size is 500 MB </code></pre> <p>The only thing I get as a Traceback is "Killed" and I'm out of python interpreter.</p> <p>Any suggestions are highly appriciated</p> <p><strong>Edit:</strong> It works okay from local machine. My application runs in Google Container Engine, so problems occurs there when runs in celery task. </p>
<p><code>upload_by_filename</code> attempts to upload the entire file in a single request.</p> <p>You can use <code>Blob.chunk_size</code> to spread the upload across many requests, each responsible for uploading one "chunk" of your file.</p> <p>For example:</p> <p><code>my_blob.chunk_size = 1024 * 1024 * 10</code></p>
python|google-cloud-platform|google-cloud-storage
3
1,909,926
45,437,620
Python Requests Proxy Error
<p>So when i try to use proxy on python requests , the actual requests send is using my own ip </p> <pre> http_proxy = "https://103.235.21.128:80" proxyDict = { "http" : http_proxy, } r = requests.get('http://whatismyip.org',proxies=proxyDict) print r.content </pre> <p>Also Tried </p> <pre> http_proxy = "https://103.235.21.128:80" proxyDict = { "https" : http_proxy, } r = requests.get('http://whatismyip.org',proxies=proxyDict) print r.content </pre> <p>So why the request is using my real ip instead of the proxy i provide </p>
<p>Have you tried setting http on the proxy like this?</p> <pre><code>http_proxy = "http://103.235.21.128:80" </code></pre> <p>or</p> <pre><code>http_proxy = "http://{}:{}".format('103.235.21.128', 80) </code></pre> <p>If that doesn't work you might have to find an http proxy</p> <p>If you're requesting data from multiple websites (both http and https) then you'll have to add both to the dictionary:</p> <pre><code>proxyDict = {"http": http_proxy, "https": http_proxy} </code></pre>
python|proxy
0
1,909,927
45,534,755
how to prioritize default mac python environment over miniconda
<p>I installed miniconda for some software I need to run. It worked great, but it made all of the other web related stuff I have set up through mac's default python environment stop working. What I would like to have is the mac python environment as the default and conda only when I need to run this specific software. So I would need to know #1 - how do I modify .bash_profile to allow me to run both environments, and #2 - the command I need to switch between environments (if there is one).</p> <p>My bash profile looks like:</p> <pre><code># Setting PATH for Python 3.6 # The original version is saved in .bash_profile.pysave PATH="/Library/Frameworks/Python.framework/Versions/3.6/bin:${PATH}" export PATH # added by Miniconda3 4.3.21 installer # export PATH="/Users/mikeheavers/miniconda3/bin:$PATH" </code></pre> <p>(I have commented out the conda path for now)</p> <p>Thanks!</p>
<p>Have you considered using Python's <a href="http://python-guide-pt-br.readthedocs.io/en/latest/dev/virtualenvs/" rel="nofollow noreferrer">Virtual env</a>? </p> <p>This allows you to have a completely separate Python installations without causing conflicts with your main python in your path. This sounds ideal for your development needs. </p> <p>You would need to "activate" the virtualenv prior to starting up miniconda, which will adjust your environmental variables such that the virtualenv python, and it's libraries will be used. </p> <p>(copying from the link) </p> <p>This will result in a virtual python installation</p> <pre><code>$ pip install virtualenv $ cd my_project_folder $ virtualenv my_project $ source my_project/bin/activate $ # Do stuff, like install from pip $ deactivate # This will turn off the virtual python in your path </code></pre> <p>you can use this interpreter in your bashrc too (Check out the link for a more in depth introduction) </p> <pre><code>export VIRTUALENVWRAPPER_PYTHON=/usr/bin/python2.7 </code></pre>
python|bash|macos|conda|miniconda
1
1,909,928
28,525,320
Unreadable encoding of a SMB/Browser packet in Scapy
<p>I'm trying to parse a pcap file with scapy (in python), and getting raw data at the layer above TCP. on wireshark, all the layers are shown correctly: <img src="https://i.stack.imgur.com/OJOKd.png" alt="wireshark"></p> <p>but on scapy all i'm seeing is just a Raw layer... <img src="https://i.stack.imgur.com/K2PuH.png" alt="enter image description here"></p> <p>i'm thinking maybe it didn't parsed the packet well? maybe the NetBIOS moduled did not load? or maybe i didn't import the module right? (i tryied: import scapy.all, import scapy, import scapy.layers.smb ) how do i make scapy load the layers of the packet correctly?</p> <p>thanks!</p>
<p>If someone has a similar problem… You need something like <code>packet[TCP].decode_payload_as(NBTSession)</code></p> <p>And then you Will get the decoded layers by scapy:</p> <pre><code> packet[TCP].show() [ TCP ] sport = microsoft_ds </code></pre> <p>…</p> <pre><code> options = [] [ NBT Session Packet ]### TYPE = Session Message RESERVED = 0 LENGTH = 4873 [ SMBNegociate Protocol Request Header ]### Start = '\xfeSMB' Command = 64 Error_Class= 0 Reserved = 1 Error_code= 0 Flags = 0 Flags2 = 0 PIDHigh = 5 Signature = 0 Unused = 0 TID = 183 PID = 0 UID = 0 MID = 0 WordCount = 0 ByteCount = 0 [ SMB Negotiate Protocol Request Tail ]### BufferFormat= 0 BufferData= '\x03' </code></pre> <p>Also you can try after that to decode the packet with the different clases:</p> <pre><code>packet.decode_payload_as(SMBNegociate Protocol Request Header) </code></pre>
python|packet|scapy|smb|netbios
0
1,909,929
14,856,033
Regex - Using * with a set of characters
<p>I'm fairly new at regex, and I've run into a problem that I cannot figure out:</p> <p>I am trying to match a set of characters that start with an arbitrary number of A-Z, 0-9, and _ characters that can optionally be followed by a number enclosed in a single set of parentheses and can be separated from the original string by a space (or not)</p> <p>Examples of what this should find:</p> <pre><code>_ABCD1E _123FD(13) ABDF1G (2) </code></pre> <p>This is my current regex expression:</p> <pre><code>[A-Z_0-9]+\s*\({0,1}[\d]*\){0,1} </code></pre> <p>It's finding everything just fine, but a problem exists if I have the following:</p> <pre><code>_ABCDE ) </code></pre> <p>It should only grab _ABCDE and not the " )" but it currently grabs '_ABCDE )'</p> <p>Is there some way I can grab the (#) but not get extra characters if that entire pattern does not exist?</p> <p>If possible, please explain syntax as I am aiming to learn, not just get the answer.</p> <p>ANSWER: The following code is working for what I needed so far:</p> <pre><code>[A-Z_0-9]+(\s*\([\d]+\)){0,1} # or, as has been mentioned, the above can be simplified # and cleaned up a bit to be [A-Z_0-9]+(\s*\(\d+\))? # The [] around \d are unnecessary and {0,1} is equivalent to ? </code></pre> <p>Adding the parentheses around the (#) pattern allows for the use of ? or {0,1} on the entire pattern. I also changed the [\d]* to be [\d]+ to ensure at least one number inside of the parentheses.</p> <p>Thanks for the fast answers, all!</p>
<p>Your regex says that each paren (open &amp; closed) may or may not be there, INDEPENDENTLY. Instead, you should say that the number-enclosed-in-parens may or may not be there:</p> <pre><code>(\([\d]*\)){0,1} </code></pre> <p>Note that this allows for there to be nothing in the parens; that's what your regex said, but I'm not clear that's what you actually want.</p>
python|regex
1
1,909,930
14,868,003
Crawling a page using LazyLoader with Python BeautifulSoup
<p>I am toying around with BeautifulSoup and I like it so far. </p> <p>The problem is the site I am trying to scrap has a lazyloader... And it only scraps one part of the site. </p> <p>Can I have a hint as to how to proceed? Must I look at how the lazyloader is implemented and parametrize anything else?</p>
<p>It turns out that the problem itself wasn't BeautifulSoup, but the dynamics of the page itself. For this specific scenario that is. </p> <p>The page returns part of the page, so headers need to be analysed and sent to the server accordingly. This isn't a BeautifulSoup problem itself. </p> <p>Therefore, it is important to take a look at how the data is loaded on a specific site. It's not always a "Load a whole page, process the whole page" paradigm. In some cases, you need to load part of the page and send a specific parameter to the server in order to keep loading the rest of the page. </p>
python|python-2.7|lazy-loading|beautifulsoup
1
1,909,931
68,687,059
How can I optimize a plotly graph with updatemenues?
<p>So, I have been using plotly a lot and recently came to use the updatemenus method for adding buttons. I've created several graphs with it, but I find it difficult to find an efficient method to update the args section in updatemenus sections. I have a data frame that is bigger than the example but it’s the same idea, so I have df:</p> <pre><code>name unaregate value age input1 in11 2 0 input1 in11 0 1 input1 in11 2 2 input1 in11 3 3 input1 in11 1 4 input1 in12 1 0 input1 in12 3 1 input1 in12 4 2 input1 in12 2 3 input1 in12 3 4 input1 in13 0 0 input1 in13 2 1 input1 in13 4 2 input1 in13 2 3 input1 in13 3 4 input2 in21 3 0 input2 in21 4 1 input2 in21 2 2 input2 in21 1 3 input2 in21 3 4 input2 in22 4 0 input2 in22 0 1 input2 in22 2 2 input2 in22 4 3 input2 in22 0 4 input2 in23 3 0 input2 in23 4 1 input2 in23 0 2 input2 in23 4 3 input2 in23 2 4 input3 in31 3 0 input3 in31 4 1 input3 in31 2 2 input3 in31 4 3 input3 in31 1 4 input3 in32 4 0 input3 in32 0 1 input3 in32 0 2 input3 in32 2 3 input3 in32 1 4 input3 in33 2 0 input3 in33 3 1 input3 in33 0 2 input3 in33 3 3 input3 in33 4 4 input3 in34 2 0 input3 in34 2 1 input3 in34 3 2 input3 in34 4 3 input3 in34 3 4 </code></pre> <p>Here is a super inefficient way to create a data frame similar to this:</p> <pre><code>df = pd.DataFrame(index=range(5),columns=range(1)) df12 = pd.DataFrame(index=range(5),columns=range(1)) df13 = pd.DataFrame(index=range(5),columns=range(1)) df21 = pd.DataFrame(index=range(5),columns=range(1)) df22 = pd.DataFrame(index=range(5),columns=range(1)) df23 = pd.DataFrame(index=range(5),columns=range(1)) df31 = pd.DataFrame(index=range(5),columns=range(1)) df32 = pd.DataFrame(index=range(5),columns=range(1)) df33 = pd.DataFrame(index=range(5),columns=range(1)) df34 = pd.DataFrame(index=range(5),columns=range(1)) df[&quot;name&quot;] = &quot;input1&quot; df[&quot;unaregate&quot;] = &quot;in11&quot; df[&quot;value&quot;] = np.random.randint(0,5, size=len(df)) df[&quot;age&quot;] = range(0,len(df)) ​ df12[&quot;name&quot;] = &quot;input1&quot; df12[&quot;unaregate&quot;] = &quot;in12&quot; df12[&quot;value&quot;] = np.random.randint(0,5, size=len(df12)) df12[&quot;age&quot;] = range(0,len(df12)) ​ df13[&quot;name&quot;] = &quot;input1&quot; df13[&quot;unaregate&quot;] = &quot;in13&quot; df13[&quot;value&quot;] = np.random.randint(0,5, size=len(df13)) df13[&quot;age&quot;] = range(0,len(df13)) ​ df21[&quot;name&quot;] = &quot;input2&quot; df21[&quot;unaregate&quot;] = &quot;in21&quot; df21[&quot;value&quot;] = np.random.randint(0,5, size=len(df21)) df21[&quot;age&quot;] = range(0,len(df21)) ​ df22[&quot;name&quot;] = &quot;input2&quot; df22[&quot;unaregate&quot;] = &quot;in22&quot; df22[&quot;value&quot;] = np.random.randint(0,5, size=len(df22)) df22[&quot;age&quot;] = range(0,len(df22)) ​ df23[&quot;name&quot;] = &quot;input2&quot; df23[&quot;unaregate&quot;] = &quot;in23&quot; df23[&quot;value&quot;] = np.random.randint(0,5, size=len(df23)) df23[&quot;age&quot;] = range(0,len(df23)) ​ df31[&quot;name&quot;] = &quot;input3&quot; df31[&quot;unaregate&quot;] = &quot;in31&quot; df31[&quot;value&quot;] = np.random.randint(0,5, size=len(df31)) df31[&quot;age&quot;] = range(0,len(df31)) ​ df32[&quot;name&quot;] = &quot;input3&quot; df32[&quot;unaregate&quot;] = &quot;in32&quot; df32[&quot;value&quot;] = np.random.randint(0,5, size=len(df32)) df32[&quot;age&quot;] = range(0,len(df32)) ​ df33[&quot;name&quot;] = &quot;input3&quot; df33[&quot;unaregate&quot;] = &quot;in33&quot; df33[&quot;value&quot;] = np.random.randint(0,5, size=len(df33)) df33[&quot;age&quot;] = range(0,len(df33)) ​ df34[&quot;name&quot;] = &quot;input3&quot; df34[&quot;unaregate&quot;] = &quot;in34&quot; df34[&quot;value&quot;] = np.random.randint(0,5, size=len(df34)) df34[&quot;age&quot;] = range(0,len(df34)) frames = [df,df12,df13,df21,df22,df23,df31,df32,df33,df34] df = pd.concat(frames) df = df.drop([0],axis=1) </code></pre> <p>This is the method I am employing for the plot:</p> <pre><code>fig = go.Figure() names = df.name.unique() for i in names: db = df[df[&quot;name&quot;]==i] uni = db.unaregate.unique() for f in uni: fig.add_trace(go.Scatter( x=db[db.unaregate==f].age, y=db[db.unaregate==f].value, connectgaps=False ,visible=False, mode='lines', legendgroup=f,name=f)) fig.update_layout( template=&quot;simple_white&quot;, xaxis=dict(title_text=&quot;age&quot;), yaxis=dict(title_text=&quot;Value&quot;), width=1000, height = 600 ) fig.update_layout( updatemenus=[ dict( # type=&quot;buttons&quot;, # direction=&quot;down&quot;, active=0, # x=0.7, # y=1.2, # showactive=True, buttons=list( [ dict( label=&quot;Select name&quot;, method=&quot;update&quot;, args=[ {&quot;visible&quot;: [False,False,False, False,False,False, False,False,False,False ]}, ], ), dict( label=&quot;input 1&quot;, method=&quot;update&quot;, args=[ {&quot;visible&quot;: [True,True,True, False,False,False, False,False,False,False ]}, ], ), dict( label=&quot;input 2&quot;, method=&quot;update&quot;, args=[ {&quot;visible&quot;: [False,False,False, True,True,True, False,False,False,False ]}, ], ), dict( label=&quot;input 3&quot;, method=&quot;update&quot;, args=[ {&quot;visible&quot;: [False,False,False, False,False,False, True,True,True,True ]}, ], ), ] ), # showactive=True, ) ] ) fig </code></pre> <p>In the part were the True’s and False are, is there a way to add those in a loop so when I have more the fifty lines, I do not have to add more than 50 Trues and Fales’s? Any help is Welcomed I just want to be able to run this script for any type of similar data and that the lengths of data do not matter.</p>
<ul> <li>data frame creation can be simplified. Using <strong>pandas</strong> constructor capability with <strong>list</strong> comprehensions</li> <li>figure / traces creation is far simpler with <strong>plotly express</strong></li> <li>core question - dynamically create visible lists <ul> <li>the trace is visible if it's in same <strong>name</strong> group. This where button name corresponds with <strong>name</strong> level of trace</li> </ul> </li> </ul> <pre><code>import pandas as pd import numpy as np import plotly.express as px df = ( pd.DataFrame( [ { &quot;name&quot;: f&quot;input{a}&quot;, &quot;unaregate&quot;: f&quot;in{a}{b}&quot;, &quot;value&quot;: np.random.randint(0, 5, 5), } for a in range(1, 4) for b in range(1, 4) ] ) .explode(&quot;value&quot;) .pipe(lambda d: d.assign(age=np.random.randint(0, 5, len(d)))) ) # get valid combinations that will create traces combis = df.groupby([&quot;name&quot;,&quot;unaregate&quot;]).size().index # for this example - it's far simpler to use plotly express to create traces fig = px.line(df, x=&quot;age&quot;, y=&quot;value&quot;, color=&quot;unaregate&quot;).update_traces(visible=False) # use list comprehesions to populate visible lists fig.update_layout( updatemenus=[ { &quot;active&quot;: 0, &quot;buttons&quot;: [ { &quot;label&quot;: &quot;Select name&quot;, &quot;method&quot;: &quot;update&quot;, &quot;args&quot;: [{&quot;visible&quot;: [False for t in fig.data]}], } ] + [ { &quot;label&quot;: n, &quot;method&quot;: &quot;update&quot;, &quot;args&quot;: [{&quot;visible&quot;: [n == t for t in combis.get_level_values(0)]}], } for n in combis.get_level_values(0).unique() ], } ], template=&quot;simple_white&quot; ) </code></pre>
python|pandas|dataframe|graph|plotly
1
1,909,932
57,024,192
How to perform arithmetic with large floating numbers in python
<p>I have two numbers a and b: a = 1562239482.739072 b = 1562239482.739071</p> <p>If I perform a-b in python, I get 1.1920928955078125e-06. However, I want 0.000001, which is the right answer after subtraction. </p> <p>Any help would be appreciated. Thank you in advance. </p> <pre><code>t = float(1562239482.739071) T = float(1562239482.739072) D = float(T - t) print(float(D)) </code></pre> <p>OR </p> <pre><code>t = 1562239482.739071 T = 1562239482.739072 D = T - t print (D) </code></pre> <p>I get the same answer 1.1920928955078125e-06 using both as mentioned above. However, I want the result 0.000001.</p> <p>Expected Result: 0.000001 Result : 1.1920928955078125e-06</p>
<p>This is common problem with floating point arithmetic. Use the <a href="https://docs.python.org/3/library/decimal.html" rel="nofollow noreferrer"><code>decimal</code></a> module</p>
python|python-3.x|python-2.7|precision|floating-accuracy
-1
1,909,933
61,944,707
How to use __setitem__ properly?
<p>I want to make a data object:</p> <pre><code>class GameData: def __init__(self, data={}): self.data = data def __getitem__(self, item): return self.data[item] def __setitem__(self, key, value): self.data[key] = value def __getattr__(self, item): return self.data[item] def __setattr__(self, key, value): self.data[kay] = value def __repr__(self): return str(self.data) </code></pre> <p>When I create a GameData object, I get RecursionError. How can I avoid <strong>setitem</strong> recall itself?</p>
<p>In the assignment <code>self.data = data</code>, <code>__setattr__</code> is called because <code>self</code> has no attribute called <code>data</code> at the moment. <code>__setattr__</code> then calls <code>__getattr__</code> to obtain the non-existing attribute <code>data</code>. <code>__getattr__</code> itself calls <code>__getattr__</code> again. This is a recursion.</p> <p>Use <code>object.__setattr__(self, 'data', data)</code> to do the assignment when implementing <code>__setattr__</code>.</p> <pre><code>class GameData: def __init__(self, data=None): object.__setattr__(self, 'data', {} if data is None else data) def __getitem__(self, item): return self.data[item] def __setitem__(self, key, value): self.data[key] = value def __getattr__(self, item): return self.data[item] def __setattr__(self, key, value): self.data[key] = value def __repr__(self): return str(self.data) </code></pre> <p>For details, see <a href="https://docs.python.org/3.8/reference/datamodel.html#object.__getattr__" rel="nofollow noreferrer">the <code>__getattr__</code> manual</a></p> <p>Additionally, do not use mutable objects as default parameter because the same object <code>{}</code> in the default argument is shared between <code>GameData</code> instances.</p>
python|setattr
2
1,909,934
24,327,410
Python mysql.connector timeout
<p>Here's a simple connection to a MySQL database using the mysql.connector module.</p> <pre><code>db = mysql.connector.connect( host=DB_SERVER, port=DB_PORT, user=DB_UNAME, passwd=DB_PASSWORD, db=DB_NAME) db.connect() mysqlCursor.execute(query) </code></pre> <p>I want to control two different timeouts. First, I want it to spend no longer than five seconds on the .connect() function. I've got that figured out. Second, I want it to spend no longer than one second on the .execute() function. How can I do that?</p> <p>I'm the database administrator, so I can do something on that end if I need to. I'd prefer only to change the timeout for one particular MySQL user, though, and not for everyone, which is why I'm starting from the Python side.</p> <p>Here's what I've found so far:</p> <p>The <a href="http://dev.mysql.com/doc/refman/5.0/en/mysql-options.html" rel="noreferrer">documentation</a> for mysql.connecter lists several timeout parameters. Connect-timeout will set the timeout for the initial connection, but as far as I can tell, it won't set a query timeout. Interactive-timeout will cause it to timeout if there's no activity, but I don't think that means it will timeout if the query takes too long to execute.</p> <p><strong>connect-timeout=seconds</strong> Connect timeout in seconds. On Linux this timeout is also used for waiting for the first answer from the server. (timeout has been replaced by connect-timeout, but timeout is still supported in MySQL 5.0 for backward compatibility.)</p> <p><strong>interactive-timeout=seconds</strong> Permit seconds of inactivity before closing the connection. The client's session wait_timeout variable is set to the value of the session interactive_timeout variable.</p>
<p>As of MySQL 5.7.8 a <a href="https://dev.mysql.com/doc/refman/5.7/en/server-system-variables.html#sysvar_max_execution_time" rel="nofollow noreferrer">maximum execution time for just SELECT statements</a> can be set per session. Set this immediately after connecting:</p> <pre><code>db = mysql.connector.connect(...) cursor = db.cursor() # Limit SELECTs to 1 second cursor.execute("SET SESSION MAX_EXECUTION_TIME=1000") </code></pre>
python|mysql
4
1,909,935
20,526,414
Relative Strength Index in python pandas
<p>I am new to pandas. What is the best way to calculate the relative strength part in the RSI indicator in pandas? So far I got the following:</p> <pre><code>from pylab import * import pandas as pd import numpy as np def Datapull(Stock): try: df = (pd.io.data.DataReader(Stock,'yahoo',start='01/01/2010')) return df print 'Retrieved', Stock time.sleep(5) except Exception, e: print 'Main Loop', str(e) def RSIfun(price, n=14): delta = price['Close'].diff() #----------- dUp= dDown= RolUp=pd.rolling_mean(dUp, n) RolDown=pd.rolling_mean(dDown, n).abs() RS = RolUp / RolDown rsi= 100.0 - (100.0 / (1.0 + RS)) return rsi Stock='AAPL' df=Datapull(Stock) RSIfun(df) </code></pre> <p>Am I doing it correctly so far? I am having trouble with the difference part of the equation where you separate out upward and downward calculations</p>
<p>It is important to note that there are various ways of defining the RSI. It is commonly defined in at least two ways: using a simple moving average (SMA) as above, or using an exponential moving average (EMA). Here's a code snippet that calculates various definitions of RSI and plots them for comparison. I'm discarding the first row after taking the difference, since it is always NaN by definition.</p> <p>Note that when using EMA one has to be careful: since it includes a memory going back to the beginning of the data, the result depends on where you start! For this reason, typically people will add some data at the beginning, say 100 time steps, and then cut off the first 100 RSI values.</p> <p>In the plot below, one can see the difference between the RSI calculated using SMA and EMA: the SMA one tends to be more sensitive. Note that the RSI based on EMA has its first finite value at the first time step (which is the second time step of the original period, due to discarding the first row), whereas the RSI based on SMA has its first finite value at the 14th time step. This is because by default rolling_mean() only returns a finite value once there are enough values to fill the window.</p> <p><a href="https://i.stack.imgur.com/FkPOh.png" rel="nofollow noreferrer"><img src="https://i.stack.imgur.com/FkPOh.png" alt="A comparison of the RSI calculated using exponential or simple moving average" /></a></p> <pre><code>import datetime from typing import Callable import matplotlib.pyplot as plt import numpy as np import pandas as pd import pandas_datareader.data as web # Window length for moving average length = 14 # Dates start, end = '2010-01-01', '2013-01-27' # Get data data = web.DataReader('AAPL', 'yahoo', start, end) # Get just the adjusted close close = data['Adj Close'] # Define function to calculate the RSI def calc_rsi(over: pd.Series, fn_roll: Callable) -&gt; pd.Series: # Get the difference in price from previous step delta = over.diff() # Get rid of the first row, which is NaN since it did not have a previous row to calculate the differences delta = delta[1:] # Make the positive gains (up) and negative gains (down) Series up, down = delta.clip(lower=0), delta.clip(upper=0).abs() roll_up, roll_down = fn_roll(up), fn_roll(down) rs = roll_up / roll_down rsi = 100.0 - (100.0 / (1.0 + rs)) # Avoid division-by-zero if `roll_down` is zero # This prevents inf and/or nan values. rsi[:] = np.select([roll_down == 0, roll_up == 0, True], [100, 0, rsi]) rsi.name = 'rsi' # Assert range valid_rsi = rsi[length - 1:] assert ((0 &lt;= valid_rsi) &amp; (valid_rsi &lt;= 100)).all() # Note: rsi[:length - 1] is excluded from above assertion because it is NaN for SMA. return rsi # Calculate RSI using MA of choice # Reminder: Provide ≥ `1 + length` extra data points! rsi_ema = calc_rsi(close, lambda s: s.ewm(span=length).mean()) rsi_sma = calc_rsi(close, lambda s: s.rolling(length).mean()) rsi_rma = calc_rsi(close, lambda s: s.ewm(alpha=1 / length).mean()) # Approximates TradingView. # Compare graphically plt.figure(figsize=(8, 6)) rsi_ema.plot(), rsi_sma.plot(), rsi_rma.plot() plt.legend(['RSI via EMA/EWMA', 'RSI via SMA', 'RSI via RMA/SMMA/MMA (TradingView)']) plt.show() </code></pre>
python|pandas|finance
74