text
stringlengths 13
30k
|
---|
{"text": "let distance ((x1, y1): (int * int)) ((x2, y2): (int * int)) : float = let dx = x1 - x2 in let dy = y1 - y2 in sqrt (float_of_int(dx * dx + dy * dy)) ;;"} |
{"text": "let is_prime n : bool = let rec counter x : bool = if x * x <= n then if n mod x = 0 then false else counter (x + 1) else true in counter 2;; let pi = 4. *. atan(1.);;"} |
{"dialog": ["Speaker 1: It's been an hour and not one of my classmates has shown up! I tell you, when I actually die some people are gonna get seriously haunted!", "Speaker 2: There you go! Someone came!", "Speaker 1: Ok, ok! I'm gonna go hide! Oh, this is so exciting, my first mourner!", "Speaker 3: Hi, glad you could come.", "Speaker 2: Please, come in.", "Speaker 4: Hi, you're Chandler Bing, right? I'm Tom Gordon, I was in your class.", "Speaker 2: Oh yes, yes... let me... take your coat.", "Speaker 4: Thanks... uh... I'm so sorry about Ross, it's...", "Speaker 2: At least he died doing what he loved... watching blimps.", "Speaker 1: Who is he?", "Speaker 2: Some guy, Tom Gordon.", "Speaker 1: I don't remember him, but then again I touched so many lives.", "Speaker 3: So, did you know Ross well?", "Speaker 4: Oh, actually I barely knew him. Yeah, I came because I heard Chandler's news. D'you know if he's seeing anyone?", "Speaker 3: Yes, he is. Me.", "Speaker 4: What? You... You... Oh! Can I ask you a personal question? Ho-how do you shave your beard so close?", "Speaker 2: Ok Tommy, that's enough mourning for you! Here we go, bye bye!!", "Speaker 4: Hey, listen. Call me.", "Speaker 2: Ok!"], "relation_data": {"x": ["Speaker 2", "Speaker 2", "Speaker 4", "Speaker 4", "Speaker 4", "Speaker 1"], "y": ["Chandler Bing", "Speaker 4", "Tom Gordon", "Speaker 2", "Tommy", "Tommy"], "x_type": ["PER", "PER", "PER", "PER", "PER", "PER"], "y_type": ["PER", "PER", "PER", "PER", "PER", "PER"], "r": [["per:alternate_names"], ["per:alumni"], ["per:alternate_names"], ["per:alumni", "per:positive_impression"], ["per:alternate_names"], ["unanswerable"]], "rid": [[30], [4], [30], [4, 1], [30], [37]], "t": [[""], [""], [""], ["", "call me"], [""], [""]]}} |
{"dialog": ["Speaker 1: So, eh... it's probably gonna be hard for you to leave Boston, huh?", "Speaker 2: Actually, I'm kinda happy to be leaving... I just broke up with someone.", "Speaker 1: Ooh... so sad... Still, it can't be easy for you to leave Harvard? Especially after working alongside a Nobel Prize winner like Albert Wintermeyer?", "Speaker 2: Actually, Alby is the guy I broke up with.", "Speaker 1: You... you dated Albert Wintermeyer?", "Speaker 2: Yeah...", "Speaker 1: ... And you called him Alby!? I mean that's like... like calling Albert Einstein... er... Alby...", "Speaker 2: Yeah, well, he is a brilliant man.", "Speaker 1: Eh, you think? I mean, you went out with a guy who improved the accuracy of radiocarbon dating by a factor of 10!", "Speaker 2: Yes! And while that is everything one looks for in a boyfriend, he had a lot of issues...", "Speaker 1: Oh! like what?! Oh I'm sorry, I don't mean to pry... it's just that this must be what regular people experience when they watch \"Access Hollywood\".", "Speaker 2: Ok, you want the dirt? Alby was seriously insecure. I mean, he was really intimidated by the guy I dated before him.", "Speaker 1: Who is intimidating to a guy who won the Nobel Prize?", "Speaker 2: A guy who won two.", "Speaker 1: Two? Wha...? Don't tell me you dated Benjamin Hobart", "Speaker 2: Yeah... for three years.", "Speaker 1: Oh my God! Have you ever been in a relationship with someone who hasn't won the Nobel Prize?", "Speaker 2: ... no... bu but there was my first boyfriend Billy.", "Speaker 1: Oh yeah? no, no Nobel Prizes for him?", "Speaker 2: No, but he did just win the McArthur genius grant.", "Speaker 1: Huh... huh... what a loser! Some more wine?"], "relation_data": {"x": ["Speaker 2", "Speaker 2", "Speaker 2", "Speaker 2", "Speaker 2", "Albert Wintermeyer", "Albert Wintermeyer", "Albert Wintermeyer", "Speaker 1", "Speaker 1", "Speaker 1", "Billy", "Billy", "Boston", "Benjamin Hobart", "Harvard"], "y": ["Boston", "Harvard", "Albert Wintermeyer", "Benjamin Hobart", "Billy", "Nobel Prize winner", "Alby", "Speaker 2", "Albert Wintermeyer", "man", "loser", "a loser", "Speaker 2", "Speaker 2", "Speaker 2", "Speaker 2"], "x_type": ["PER", "PER", "PER", "PER", "PER", "PER", "PER", "PER", "PER", "PER", "PER", "PER", "PER", "GPE", "PER", "ORG"], "y_type": ["GPE", "ORG", "PER", "PER", "PER", "STRING", "PER", "PER", "PER", "STRING", "STRING", "STRING", "PER", "PER", "PER", "PER"], "r": [["per:place_of_residence"], ["per:employee_or_member_of"], ["per:girl/boyfriend"], ["per:girl/boyfriend"], ["per:girl/boyfriend"], ["per:title"], ["per:alternate_names"], ["per:girl/boyfriend"], ["per:positive_impression"], ["unanswerable"], ["unanswerable"], ["per:alternate_names"], ["per:girl/boyfriend"], ["gpe:residents_of_place"], ["per:girl/boyfriend"], ["org:employees_or_members"]], "rid": [[18], [22], [10], [10], [10], [29], [30], [10], [1], [37], [37], [30], [10], [32], [10], [35]], "t": [["leave"], ["leave"], ["broke up with"], ["relationship"], ["first boyfriend"], [""], [""], ["broke up with"], [""], [""], [""], [""], ["first boyfriend"], ["leave"], ["relationship"], ["leave"]]}} |
{"dialog": ["Speaker 1: Hi.", "Speaker 2: Hi.", "Speaker 1: I just finished getting Phoebe all dressed to meet Mike's parents. She's so nervous, it's so sweet!", "Speaker 2: Guess what? I made Emma laugh today.", "Speaker 1: You WHAT? And I missed it? Because I was giving a makeover to that stupid hippie?", "Speaker 2: Yeah, and it was uhm... it was like a real little person laugh too. It was... it was like uhm... Only... only not creepy.", "Speaker 1: Well... well, what did you do to make her laugh?", "Speaker 2: I uhm... Well, I sang... well actually I rapped... Baby Got Back...", "Speaker 1: You WHAT? You sang... to our baby daughter... a song about a guy who likes to have sex with women with giant asses?", "Speaker 2: But you know what, if you think about it, it actually promotes a healthy uhm... body image... because... even big butts or uhm... juicy doubles.", "Speaker 1: owwwww...", "Speaker 2: Please don't take her away from me!"], "relation_data": {"x": ["Phoebe", "Phoebe", "Mike", "Speaker 2", "Speaker 2", "Speaker 1", "Speaker 1", "Emma", "Emma", "Emma"], "y": ["Mike", "stupid hippie", "Phoebe", "Emma", "Baby", "Emma", "Baby", "Speaker 2", "Speaker 1", "Baby Got Back"], "x_type": ["PER", "PER", "PER", "PER", "PER", "PER", "PER", "PER", "PER", "PER"], "y_type": ["PER", "STRING", "PER", "PER", "STRING", "PER", "STRING", "PER", "PER", "STRING"], "r": [["per:girl/boyfriend"], ["per:alternate_names"], ["per:girl/boyfriend"], ["per:children"], ["unanswerable"], ["per:children"], ["unanswerable"], ["per:parents"], ["per:parents"], ["per:positive_impression"]], "rid": [[10], [30], [10], [13], [37], [13], [37], [15], [15], [1]], "t": [[""], [""], [""], ["baby daughter"], [""], ["baby daughter"], [""], ["baby daughter"], ["baby daughter"], ["laugh"]]}} |
{"dialog": ["Speaker 1: 'Okay. Okay, daddy we'll see you tomorrow night. Okay bye-bye.'", "Speaker 2: We?", "Speaker 1: Are ah, having dinner with my Dad tomorrow night, I hope that's okay.", "Speaker 2: Oh shoot, tomorrow's not so good, I'm supposed to um, fall off the Empire State building and land on a bicycle with no seat. Sorry.", "Speaker 1: Ross, my father doesn't hate you.", "Speaker 2: Please, he refers to me as 'wethead'.", "Speaker 1: But honey he calls everybody by a nickname! Okay, look, I know, all right, just one dinner, please, just one night for me, please. I just want him to love you like I do."], "relation_data": {"x": ["Speaker 2", "Speaker 2", "Speaker 2", "Speaker 2", "Speaker 1", "Speaker 1"], "y": ["wethead", "honey", "Speaker 1", "daddy", "Speaker 2", "honey"], "x_type": ["PER", "PER", "PER", "PER", "PER", "PER"], "y_type": ["STRING", "STRING", "PER", "STRING", "PER", "STRING"], "r": [["per:alternate_names"], ["per:alternate_names"], ["per:girl/boyfriend"], ["unanswerable"], ["per:positive_impression", "per:girl/boyfriend"], ["unanswerable"]], "rid": [[30], [30], [10], [37], [1, 10], [37]], "t": [["nickname"], [""], [""], [""], ["love", ""], [""]]}} |
{"dialog": ["Speaker 1: Hey Pheebs.", "Speaker 2: Hey!", "Speaker 1: Any sign of your brother?", "Speaker 2: No, but he's always late.", "Speaker 1: I thought you only met him once?", "Speaker 2: Yeah, I did. I think it sounds y'know big sistery, y'know, 'Frank's always late.'", "Speaker 1: Well relax, he'll be here."], "relation_data": {"x": ["Speaker 2", "Speaker 2", "Frank", "Speaker 1"], "y": ["Pheebs", "Frank", "Speaker 2", "Pheebs"], "x_type": ["PER", "PER", "PER", "PER"], "y_type": ["PER", "PER", "PER", "PER"], "r": [["per:alternate_names"], ["per:siblings"], ["per:siblings"], ["unanswerable"]], "rid": [[30], [16], [16], [37]], "t": [[""], ["brother"], ["brother"], [""]]}} |
{"dialog": ["Speaker 1: Monica.", "Speaker 2: Oh my God! He just said my name! Did you hear that?", "Speaker 1: Monica bang!", "Speaker 3: Okay, I heard that.", "Speaker 2: Did he just say 'Monica bang'?", "Speaker 3: Uh-huh.", "Speaker 2: Oh my God! He's gonna rat me out!", "Speaker 1: Monica bang!", "Speaker 2: Oh-ho-ho, sweetie, sweetie, you gotta stop saying that, now. It's no big deal, it's not even worth mentioning, you see we all do it all the time. See watch this, Ben, Ben, Ben. Ow, Monica bang! Everybody bang. Ben bang. Rachel bang. Bang, Rachel bang! Oh, isn't that fun?", "Speaker 3: Look at that! Look at that! We all do it. Okay, I'm stopping now.", "Speaker 2: You okay?", "Speaker 3: Oh yeah! Y'know, if it's not a headboard, it's just not worth it."], "relation_data": {"x": ["Speaker 2", "Speaker 2", "Speaker 1", "Speaker 3", "Speaker 3"], "y": ["Monica bang", "sweetie", "sweetie", "Rachel bang", "sweetie"], "x_type": ["PER", "PER", "PER", "PER", "PER"], "y_type": ["PER", "STRING", "STRING", "PER", "STRING"], "r": [["per:alternate_names"], ["unanswerable"], ["per:alternate_names"], ["per:alternate_names"], ["unanswerable"]], "rid": [[30], [37], [30], [30], [37]], "t": [[""], [""], [""], [""], [""]]}} |
{"dialog": ["Speaker 1: Oh hey! Wait up!", "Speaker 2: Hi!", "Speaker 1: Congratulations! I didn't want to say anything in front of Joey because I didn't know if he knew yet.", "Speaker 2: What, that we had a baby? Come on let's give him a little credit, although he did eat a piece of plastic fruit earlier.", "Speaker 1: No! No, that you and Rachel are engaged!", "Speaker 2: What?", "Speaker 1: Oh, it's a secret. Oh goodie! Yes! We haven't done the secret thing in a long time.", "Speaker 2: Phoebe, there is no secret. Okay? I didn't propose.", "Speaker 1: Are you lying? Is this like that time you tried to convince us that you were a doctor?", "Speaker 2: I am a doctor! Y'know what? I'm just gonna go and talk to Rachel myself.", "Speaker 1: All right, me too. Should we wake her up?", "Speaker 2: No! No, come on let her sleep! She's so exhausted."], "relation_data": {"x": ["Speaker 2", "Speaker 2", "Rachel", "Speaker 1"], "y": ["doctor", "baby", "baby", "baby"], "x_type": ["PER", "PER", "PER", "PER"], "y_type": ["STRING", "STRING", "STRING", "STRING"], "r": [["per:title"], ["unanswerable"], ["unanswerable"], ["unanswerable"]], "rid": [[29], [37], [37], [37]], "t": [[""], [""], [""], [""]]}} |
{"dialog": ["Speaker 1: Over here Jack. OK. I see, Rachel's coming up the path. Oh doesn't she look pretty. Jack, get this.", "Speaker 2: Get a shot of Monica. Where's Monica.", "Speaker 3: Over here dad.", "Speaker 4: Wait, how do you zoom out? There she is.", "Speaker 3: Oh, you look so great.", "Speaker 5: Ahh, so do you, beautiful.", "Speaker 3: Oops.", "Speaker 5: What?", "Speaker 3: Shoot, I think I got mayonaise on you.", "Speaker 5: Oh, that's OK, it's just the shoulder, it's not my dress.", "Speaker 4: Everybody smile.", "Speaker 3: Oh, dad, turn it off.", "Speaker 4: It is off.", "Speaker 3: Dad, it is not. What's with the red light?", "Speaker 4: It's the off light. Right Ross?", "Speaker 6: You look pretty tonight.", "Speaker 5: Oh, thanks. So, uh, what are you gonna do this summer?", "Speaker 6: Oh, you know, I'm just gonna, I'm gonna hang out, work on my music.", "Speaker 5: Is my hook unhooked? These things keep falling down, I can't. . .", "Speaker 6: Uh, hold, let me see, I don't know. So what're you gonna do. . .", "Speaker 5: Oh, the guys are here.", "Speaker 6: this summer?"], "relation_data": {"x": ["Speaker 3", "Speaker 3", "Speaker 4", "Speaker 4"], "y": ["Speaker 4", "beautiful", "Jack", "Speaker 3"], "x_type": ["PER", "PER", "PER", "PER"], "y_type": ["PER", "STRING", "PER", "PER"], "r": [["per:parents"], ["per:alternate_names"], ["per:alternate_names"], ["per:children"]], "rid": [[15], [30], [30], [13]], "t": [["dad"], [""], [""], ["dad"]]}} |
{"dialog": ["Speaker 1: Alright, wait a second, why would Ross tell everyone in your class that you are as... \"gay as the day is long\"?", "Speaker 2: Because I told everyone he slept with dinosaurs.", "Speaker 1: But that's clearly a joke. This could easily be true.", "Speaker 2: Would you get that please? People have been calling to congratulate me all day.", "Speaker 1: Hello? No, he's not here. Yeah, this is his wife. Yeah, well, it came as quite a shock to me too. I guess I should have known. Yeah, I mean, he just kept making me watch Moulin Rouge.", "Speaker 2: Hang up, hang up. And that was a great movie! I'm so gonna get back at Ross... oh yeah, this will show him, here we go.", "Speaker 1: What are you doing?", "Speaker 2: Oh, you'll see my friend."], "relation_data": {"x": ["Speaker 1", "Speaker 1", "Speaker 2", "Speaker 2", "Speaker 2", "Speaker 2", "Ross"], "y": ["my friend", "Speaker 2", "Ross", "Speaker 1", "Moulin Rouge", "my friend", "Speaker 2"], "x_type": ["PER", "PER", "PER", "PER", "PER", "PER", "PER"], "y_type": ["STRING", "PER", "PER", "PER", "STRING", "STRING", "PER"], "r": [["per:alternate_names"], ["per:spouse"], ["per:alumni"], ["per:spouse"], ["per:positive_impression"], ["unanswerable"], ["per:alumni"]], "rid": [[30], [17], [4], [17], [1], [37], [4]], "t": [[""], ["wife"], ["class"], ["wife"], ["great movie"], [""], ["class"]]}} |
{"repo_name": "YuncyYe/ml", "path": "mlf/pocketv1.py", "copies": "1", "size": "3503", "content": "\n\n#\n#pocket Algorithm \n#\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport scipy.stats as stats\nimport random\n\n##############################################\ndef sepLine(w, x):\n return -((w[0]+w[1]*x)/w[2])\n#end\n\ndef drawSepLine(w, minX, maxX):\n sepx = range(minX, maxX) \n sepy = []\n for e in sepx:\n tmp = sepLine(w, e)\n sepy.append( tmp )\n #end for\n plt.plot(sepx, sepy )\n#end drawSepLine\n##############################################\n\nls=np.array([\n[1.0, 0.5, 1]\n,[1.5, 14.5, 1] \n,[2.5, 1.5, 1] \n,[2.8, 3.5, 1]\n,[4.5, 13.0, 1]\n,[6.0, 8.0, 1]\n,[7.0, 16.0, 1] #noize data\n,[8.0, 5.5, 1]\n,[9.5, 7.0, 1]\n,[12.0, 2.5, 1]\n,[14.0, 2.0, 1]\n#,[7.0, 16.0, 1] #noize data\n])\nrs=np.array([\n[2.0, 18.0, -1]\n,[3.0, 17.5, -1]\n,[3.5, 0.7, -1] #noize data \n,[8.0,11.5, -1] \n,[8.5,13.5, -1] \n,[8.5,13.0, -1] \n,[9.0,15, -1]\n,[12.0,20.0,-1]\n,[16.0,17.0,-1]\n#,[3.5, 0.7, -1] #noize data \n])\n\n##construct training data\nrtd = np.concatenate((ls,rs))\nminX = (int)(np.min(rtd[:,:1]))-3\nmaxX = (int)(np.max(rtd[:,:1]))+3\n\n###plot the data\nplt.xlim( (minX, maxX) )\nplt.ylim( (np.min(rtd[:,1:2]-3), np.max(rtd[:,1:2]+3)) )\nplt.plot(ls[:,:1], ls[:, 1:2], '*')\nplt.plot(rs[:,:1], rs[:, 1:2], '+')\n\n\n##############pla-begin\nx0 = np.zeros( (len(rtd), 1) )\nx0[:]=1.0\ntd = np.concatenate( (x0, rtd[:,:1], rtd[:,1:2], rtd[:,2:3]), 1 )\n\n#The this initial value of w. td[0] include y. so we need to minus 1\nw=np.zeros( len(td[0])-1 );\n\n#todo:we can set it as max of float \nweighOfPocket=1000000000.0\nwPocket=w\n\n#\n#ensure all point corret\n#maxIter=900000\nmaxIter=1200000\nweighOfPocketThres=0.05\n\ncurIter=0\nwhile(curIter<maxIter):\n curIter = curIter +1;\n \n #[begin----the following is typical pla----\n isModifing=False; \n #check each point for w\n for ti in range(len(td)):\n rndIdx=random.randint(0, len(td)-1)\n sample = td[rndIdx]\n sx = sample[:len(sample)-1]; sy=sample[len(sample)-1]\n t = np.inner(w, sx)\n ty = np.sign(t)\n #print(idx, ty, sy)\n if(ty!=sy):\n #failed, we need to update w\n w = w + sy*sx\n isModifing = True\n #end if\n #end for\n if(isModifing==False):\n break;\n #todo. we need to update pocket here.\n #end]\n \n #pick up an element in sample to try to improve w\n #rndIdx=random.randint(0, len(td)-1)\n #sample = td[rndIdx]\n #sx = sample[:len(sample)-1]; sy=sample[len(sample)-1]\n #w = w + sy*sx\n \n #It's too late to check weight for this w\n #calc weight for w\n weight=0.;\n for idx in range(len(td)):\n sample = td[idx]\n sx = sample[:len(sample)-1]; sy=sample[len(sample)-1]\n t = np.inner(w, sx)\n ty = np.sign(t)\n #print(idx, ty, sy)\n if(ty!=sy):\n weight += 1.0;\n #end for\n \n #print(\"The curIter is \", curIter)\n #print(\"The weighOfPocket is \", weighOfPocket)\n #print(\"The w is \", w)\n #drawSepLine(w, minX, maxX)\n \n #if the new w is better than stuff in pocket, then update stuff in pocket\n if(weight<weighOfPocket):\n weighOfPocket = weight\n wPocket = w\n #end if\n \n if(weighOfPocket<weighOfPocketThres):\n break;\n#end for\n\n##############pla-end\nprint(\"The curIter is \", curIter)\nprint(\"The weighOfPocket is \", weighOfPocket)\nprint(\"The w is \", w)\n\n\n#show the seperator line\ndrawSepLine(w, minX, maxX);\n\n###\n#In [93]: import pla\n#In [94]: reload(pla)\n#\nif __name__ ==\"__main__\":\n pass\n#end\n\n\n\n\n", "license": "apache-2.0"} |
{"repo_name": "3like3beer/openrevman", "path": "openrevman/control_computer/solver.py", "copies": "1", "size": "8426", "content": "#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nimport collections\n\nimport pulp\nfrom numpy import dot\nfrom pandas import DataFrame, read_table\nfrom scipy.sparse import csgraph\n\n\nclass Controls:\n def __init__(self, accepted_demand: DataFrame, product_bid_prices: DataFrame, expected_revenue: float = None):\n self.accepted_demand = accepted_demand\n self.product_bid_prices = product_bid_prices\n self.expected_revenue = expected_revenue\n\n\nclass Problem:\n def __init__(self, demand_vector, price_vector, capacity_vector, demand_utilization_matrix, demand_profile=None):\n self.demand_vector = demand_vector\n self.price_vector = price_vector\n self.capacity_vector = capacity_vector\n self.demand_utilization_matrix = demand_utilization_matrix\n self.demand_profile = demand_profile\n self.demand_correlations = self.get_demand_correlations()\n\n def get_demand_correlations(self):\n return dot(self.demand_utilization_matrix, self.demand_utilization_matrix.transpose())\n\n def get_subproblems(self, eps=0.1):\n subproblems = []\n labels = csgraph.connected_components(self.demand_correlations, directed=False)[1]\n split_index = collections.Counter(labels).values()\n prev = 0\n for i in split_index:\n demand_vector = self.demand_vector[prev:prev + i]\n price_vector = self.price_vector[prev:prev + i]\n capacity_vector = self.capacity_vector\n demand_utilization_matrix = self.demand_utilization_matrix.ix[prev:prev + i, :]\n demand_profile = None\n if self.demand_profile is not None:\n demand_profile = self.demand_profile.ix[prev:prev + i - 1, :]\n subproblems.append(\n Problem(demand_vector=demand_vector, price_vector=price_vector, capacity_vector=capacity_vector,\n demand_utilization_matrix=demand_utilization_matrix, demand_profile=demand_profile))\n prev = i\n return subproblems\n\n\nclass Solver:\n def __init__(self, optimizer):\n self.optimizer = optimizer\n self.controls = None\n\n def optimize_controls(self, problem):\n self.controls = pulp_solve(problem.demand_vector, problem.price_vector, problem.capacity_vector,\n problem.demand_utilization_matrix)\n return self.controls\n\n def optimize_controls_multi_period(self, problem, eps):\n if problem.demand_profile.shape[1] > 1:\n for period in problem.demand_profile.columns:\n if self.controls:\n new_control = pulp_solve(problem.demand_profile.ix[:, period], problem.price_vector,\n problem.capacity_vector,\n problem.demand_utilization_matrix)\n if self.is_new_ctrl_more_profitable(new_control, 0.1):\n self.blinde_control(new_control, eps)\n else:\n self.controls = self.optimize_controls(problem)\n else:\n self.controls = self.optimize_controls(problem)\n return self.controls\n\n def is_new_ctrl_more_profitable(self, new_control, eps):\n rev1 = self.controls.expected_revenue\n if new_control.expected_revenue - rev1 > rev1 * eps:\n return True\n return False\n\n def blinde_control(self, new_control, eps):\n self.controls.accepted_demand = self.controls.accepted_demand * eps\n self.controls.product_bid_prices = self.controls.product_bid_prices / eps\n self.controls.expected_revenue = new_control.expected_revenue\n\n\ndef to_data_frame(data):\n df = DataFrame.transpose(read_table(data, delim_whitespace=True, header=None))\n df.columns = [(col + 1) for col in df.columns]\n return df\n\n\ndef to_data_frame2(data):\n df = DataFrame(read_table(data, delim_whitespace=True, header=None))\n return df\n\n\ndef create_problem_with_data(demand_data, capacity_data, demand_utilization_data, demand_profile_data=None):\n demand_vector, capacity_vector, demand_profile, demand_utilization_matrix = load_data_to_df(capacity_data,\n demand_data,\n demand_profile_data,\n demand_utilization_data)\n return Problem(demand_vector.ix[:, 1], demand_vector.ix[:, 2], capacity_vector,\n demand_utilization_matrix.ix[:, :],\n demand_profile)\n\n\ndef merge_controls(controls_list):\n first_time = True\n for controls in controls_list:\n if first_time:\n accepted_demand = controls.accepted_demand\n product_bid_prices = controls.product_bid_prices\n expected_revenue = controls.expected_revenue\n first_time = False\n else:\n accepted_demand = accepted_demand.append(controls.accepted_demand)\n product_bid_prices = product_bid_prices.append(controls.product_bid_prices)\n expected_revenue = expected_revenue + controls.expected_revenue\n\n return Controls(accepted_demand=accepted_demand, product_bid_prices=product_bid_prices,\n expected_revenue=expected_revenue)\n\n\ndef load_data_to_df(capacity_data, demand_data, demand_profile_data, demand_utilization_data):\n demand_vector = to_data_frame(demand_data)\n capacity_vector = to_data_frame(capacity_data)\n demand_utilization_matrix = to_data_frame2(demand_utilization_data)\n assert demand_utilization_matrix.shape[0] == demand_vector.shape[0]\n assert demand_utilization_matrix.shape[1] == capacity_vector.shape[0]\n if demand_profile_data:\n demand_profile = to_data_frame(demand_profile_data)\n assert demand_profile.shape[0] == demand_vector.shape[0]\n else:\n demand_profile = None\n return demand_vector, capacity_vector, demand_profile, demand_utilization_matrix\n\n\ndef pulp_solve(demand_vector, price_vector, capacity_vector, demand_utilization_matrix):\n revman = create_problem()\n x = create_variables(demand_vector)\n set_objective(demand_vector, price_vector, revman, x)\n add_product_constraints(capacity_vector, demand_utilization_matrix, demand_vector, revman, x)\n add_demand_constraints(demand_vector, revman, x)\n\n solve_problem(revman)\n\n accepted_demand = get_accepted_demand(x)\n product_bid_prices = get_bid_prices(capacity_vector, revman)\n expected_revenue = get_expected_revenue(revman)\n\n return Controls((accepted_demand), (product_bid_prices), expected_revenue)\n\n\ndef solve_problem(revman):\n revman.solve(pulp.PULP_CBC_CMD())\n # revman.writeLP(\"temp.txt\")\n # print(pulp.LpStatus[revman.status])\n\n\ndef create_problem():\n return pulp.LpProblem(\"revman\", pulp.LpMaximize)\n\n\ndef get_expected_revenue(revman):\n return pulp.value(revman.objective)\n\n\ndef get_accepted_demand(x):\n return DataFrame({'accepted_demand': [(x[str(i)].value()) for i in x]})\n\n\ndef get_bid_prices(capacity_vector, revman):\n bid_prices_list = [revman.constraints.get(\"Capa_\" + str(i)).pi for (i, c) in (capacity_vector.iterrows())]\n return DataFrame({'bid_prices_list': bid_prices_list})\n\n\ndef add_demand_constraints(demand_vector, revman, x):\n for (demand_index, demand) in (demand_vector.iteritems()):\n revman.addConstraint((x[str(demand_index)]) <= demand, name=\"Demand_\" + str(demand_index))\n\n\ndef add_product_constraints(capacity_vector, demand_utilization_matrix, demand_vector, revman, x):\n for (product_index, capacity) in (capacity_vector.iterrows()):\n revman.addConstraint(pulp.lpSum(\n [x[str(i)] * demand_utilization_matrix.ix[i, product_index] for (i, d) in demand_vector.iteritems()]) <=\n capacity,\n name=\"Capa_\" + str(product_index))\n\n\ndef set_objective(demand_vector, price_vector, revman, x):\n objective = pulp.LpAffineExpression([(x[str(i)], price_vector[i]) for (i, d) in demand_vector.iteritems()])\n revman.setObjective(objective)\n\n\ndef create_variables(demand_vector):\n x = dict([(str(i), pulp.LpVariable(name=\"x\" + str(i), lowBound=0, cat=pulp.LpContinuous)) for (i, t) in\n demand_vector.iteritems()])\n return x\n", "license": "gpl-3.0"} |
{"repo_name": "brahmcapoor/naming-changes-complexity", "path": "analysis/subject_analysis.py", "copies": "1", "size": "8370", "content": "from random import shuffle\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport pandas as pd\nimport csv\nimport argparse\nimport os\nimport shutil\n\n\"\"\"\nThe script required for the data analysis. Requires several python libraries\nto function but otherwise isn't too complicated.\n\"\"\"\n\n\ndef generate_all_graphs():\n number_of_subjects = len(os.listdir(\"../subject logs\")) - 1\n\n for i in range(1, number_of_subjects + 1):\n try:\n os.mkdir(\"Subject {}\".format(i))\n except FileExistsError:\n shutil.rmtree(\"Subject {}\".format(i))\n os.mkdir(\"Subject {}\".format(i))\n\n transparency_log_1, transparency_log_2, transparency_log_3, \\\n transparency_log_4 = load_subject_data(i)\n\n individual_graph(transparency_log_1, transparency_log_2, \"Easy\", i,\n False)\n individual_graph(transparency_log_3, transparency_log_4, \"Hard\", i,\n False)\n\n print(\"Generated graphs for subject {}\".format(i))\n\n\ndef individual_graph(transparencies_1, transparencies_2, condition,\n subject_number, display_graph=True):\n\n x = [i for i in range(1, 81)]\n sns.pointplot(x, transparencies_1, color='red')\n plot = sns.pointplot(x, transparencies_2)\n plot.set(xlabel=\"Trial\", ylabel=\"Contrast\",\n title=\"{} Condition\".format(condition))\n if display_graph:\n plt.show()\n plot = plot.get_figure()\n plot.savefig(\"Subject {}/{}.png\".format(subject_number, condition))\n plt.cla()\n\n\ndef find_turning_points(series):\n\n turning_points = []\n last_point = len(series) - 1\n\n for i, point in enumerate(series):\n if i != 0 and i != last_point:\n if (point < series[i - 1] and point < series[i + 1]) or \\\n (point > series[i - 1] and point > series[i + 1]):\n\n turning_points.append(point)\n\n return turning_points\n\n\ndef find_threshold(log_1, log_2):\n\n average_1 = 0\n average_2 = 0\n\n turning_points_1 = find_turning_points(log_1)\n\n if turning_points_1:\n average_1 = sum(turning_points_1)/len(turning_points_1)\n else:\n average_1 = 0\n\n turning_points_2 = find_turning_points(log_2)\n\n if turning_points_2:\n average_2 = sum(turning_points_2)/len(turning_points_2)\n else:\n average_2 = 0\n\n return (average_1 + average_2)/2\n\n\ndef load_subject_data(subject_number):\n filename = \"../subject logs/subject {}.csv\".format(subject_number)\n\n with open(filename, 'r') as f:\n reader = csv.reader(f)\n data = list(reader)[1:]\n\n transparency_log_1 = [trial[1] for trial in data]\n transparency_log_2 = [trial[3] for trial in data]\n transparency_log_3 = [trial[5] for trial in data]\n transparency_log_4 = [trial[7] for trial in data]\n\n transparency_log_1 = list(map(lambda x: float(x), transparency_log_1))\n transparency_log_2 = list(map(lambda x: float(x), transparency_log_2))\n transparency_log_3 = list(map(lambda x: float(x), transparency_log_3))\n transparency_log_4 = list(map(lambda x: float(x), transparency_log_4))\n\n return (transparency_log_1, transparency_log_2, transparency_log_3,\n transparency_log_4)\n\n\ndef check_subject_validity(subject_number):\n with open('../subject logs/catch trials.csv', 'r') as f:\n reader = csv.reader(f)\n data = list(reader)\n\n subject_info = data[subject_number]\n\n catch_trials_valid = int(subject_info[1]) > 29 and int(subject_info[2]) < 5\n\n with open('../memory_results_after.csv', 'r') as f:\n reader = csv.reader(f)\n data = list(reader)\n\n subject_info = data[subject_number]\n\n remembered_names_correctly = subject_info[2] == subject_info[3] and \\\n subject_info[4] == subject_info[5] and \\\n subject_info[6] == '' and \\\n subject_info[7] == ''\n\n return catch_trials_valid and remembered_names_correctly\n\n\ndef graph_subject(subject_number):\n try:\n os.mkdir(\"Subject {}\".format(subject_number))\n except FileExistsError:\n shutil.rmtree(\"Subject {}\".format(subject_number))\n os.mkdir(\"Subject {}\".format(subject_number))\n transparency_log_1, transparency_log_2, transparency_log_3, \\\n transparency_log_4 = load_subject_data(subject_number)\n\n individual_graph(transparency_log_1, transparency_log_2, \"Easy\",\n subject_number)\n\n average_easy = find_threshold(transparency_log_1,\n transparency_log_2)\n\n print(\"Subject average for easy condition is {}\".format(average_easy))\n\n individual_graph(transparency_log_3, transparency_log_4, \"Hard\",\n subject_number)\n\n average_hard = find_threshold(transparency_log_3,\n transparency_log_4)\n\n print(\"Subject average for hard condition is {}\".format(average_hard))\n\n valid = check_subject_validity(int(subject_number))\n\n if valid:\n print(\"Subject is valid\")\n else:\n print(\"Subject is invalid\")\n\n\ndef show_summary_graph():\n\n number_of_subjects = len(os.listdir(\"../subject logs\")) - 1\n\n subject_data = []\n for i in range(1, number_of_subjects + 1):\n transparency_log_1, transparency_log_2, transparency_log_3, \\\n transparency_log_4 = load_subject_data(i)\n\n easy_threshold = find_threshold(transparency_log_1, transparency_log_2)\n hard_threshold = find_threshold(transparency_log_3, transparency_log_4)\n\n subject_data.append((i, easy_threshold, \"Easy\",\n check_subject_validity(i)))\n subject_data.append((i, hard_threshold, \"Hard\",\n check_subject_validity(i)))\n\n df = pd.DataFrame(subject_data, columns=[\"Subject\", \"Threshold\",\n \"Condition\", \"Valid\"])\n\n print(df)\n\n plot = sns.factorplot(data=df,\n x=\"Subject\",\n y=\"Threshold\",\n hue=\"Condition\",\n linestyles=[\" \", \" \"],\n legend=False,\n size=8,\n aspect=2)\n plt.legend(loc='upper left')\n\n plot.set(xlabel=\"Subject Number\",\n ylabel=\"Contrast\",\n title=\"Summary of all subjects\")\n\n plt.show()\n\n plot.savefig(\"Summary.png\")\n\n\ndef generate_results_file():\n\n number_of_subjects = len(os.listdir(\"../subject logs\")) - 1\n\n table = []\n for i in range(1, number_of_subjects + 1):\n transparency_log_1, transparency_log_2, transparency_log_3, \\\n transparency_log_4 = load_subject_data(i)\n\n easy_threshold = find_threshold(transparency_log_1, transparency_log_2)\n hard_threshold = find_threshold(transparency_log_3, transparency_log_4)\n\n valid = check_subject_validity(i)\n\n with open('../memory_results_after.csv', 'r') as f:\n reader = csv.reader(f)\n data = list(reader)\n\n subject_info = data[i]\n\n round_number = subject_info[1]\n\n table.append([i, round_number, easy_threshold, hard_threshold, valid])\n\n table = pd.DataFrame(table, columns=[\"Subject\", \"Round number\",\n \"Easy Threshold\", \"Hard Threshold\",\n \"Valid\"])\n\n table.to_csv(\"Summary.csv\")\n\n print(\"Results file can be found in Summary.csv\")\n\n\ndef main():\n\n plt.rcParams['figure.figsize'] = (18, 8)\n\n parser = argparse.ArgumentParser()\n group = parser.add_mutually_exclusive_group(required=True)\n group.add_argument(\"-i\", \"--individual\",\n help=\"Analyze a particular subject\",\n action=\"store\", metavar='')\n group.add_argument(\"-s\", \"--summary\", help=\"See a summary graph\",\n action=\"store_true\")\n group.add_argument(\"-r\", \"--result\", help=\"Create a results file\",\n action=\"store_true\")\n args = parser.parse_args()\n\n if args.individual:\n if args.individual == 'a':\n generate_all_graphs()\n else:\n graph_subject(args.individual)\n if args.summary:\n show_summary_graph()\n if args.result:\n generate_results_file()\n\n\nif __name__ == '__main__':\n main()\n", "license": "mit"} |
{"repo_name": "wdurhamh/statsmodels", "path": "statsmodels/sandbox/examples/thirdparty/findow_0.py", "copies": "33", "size": "2147", "content": "# -*- coding: utf-8 -*-\n\"\"\"A quick look at volatility of stock returns for 2009\n\nJust an exercise to find my way around the pandas methods.\nShows the daily rate of return, the square of it (volatility) and\na 5 day moving average of the volatility.\nNo guarantee for correctness.\nAssumes no missing values.\ncolors of lines in graphs are not great\n\nuses DataFrame and WidePanel to hold data downloaded from yahoo using matplotlib.\nI haven't figured out storage, so the download happens at each run\nof the script.\n\ngetquotes is from pandas\\examples\\finance.py\n\nCreated on Sat Jan 30 16:30:18 2010\nAuthor: josef-pktd\n\"\"\"\nfrom statsmodels.compat.python import lzip\nimport numpy as np\nimport matplotlib.finance as fin\nimport matplotlib.pyplot as plt\nimport datetime as dt\n\nimport pandas as pa\n\n\ndef getquotes(symbol, start, end):\n quotes = fin.quotes_historical_yahoo(symbol, start, end)\n dates, open, close, high, low, volume = lzip(*quotes)\n\n data = {\n 'open' : open,\n 'close' : close,\n 'high' : high,\n 'low' : low,\n 'volume' : volume\n }\n\n dates = pa.Index([dt.datetime.fromordinal(int(d)) for d in dates])\n return pa.DataFrame(data, index=dates)\n\n\nstart_date = dt.datetime(2009, 1, 1)\nend_date = dt.datetime(2010, 1, 1)\n\nmysym = ['msft', 'ibm', 'goog']\nindexsym = ['gspc', 'dji']\n\n\n# download data\ndmall = {}\nfor sy in mysym:\n dmall[sy] = getquotes(sy, start_date, end_date)\n\n# combine into WidePanel\npawp = pa.WidePanel.fromDict(dmall)\nprint(pawp.values.shape)\n\n# select closing prices\npaclose = pawp.getMinorXS('close')\n\n# take log and first difference over time\npaclose_ratereturn = paclose.apply(np.log).diff()\nplt.figure()\npaclose_ratereturn.plot()\nplt.title('daily rate of return')\n\n# square the returns\npaclose_ratereturn_vol = paclose_ratereturn.apply(lambda x:np.power(x,2))\nplt.figure()\nplt.title('volatility (with 5 day moving average')\npaclose_ratereturn_vol.plot()\n\n# use convolution to get moving average\npaclose_ratereturn_vol_mov = paclose_ratereturn_vol.apply(\n lambda x:np.convolve(x,np.ones(5)/5.,'same'))\npaclose_ratereturn_vol_mov.plot()\n\n#plt.show()\n", "license": "bsd-3-clause"} |
{"repo_name": "synthicity/pandana", "path": "pandana/loaders/pandash5.py", "copies": "5", "size": "2024", "content": "import pandas as pd\n\n\ndef remove_nodes(network, rm_nodes):\n \"\"\"\n Create DataFrames of nodes and edges that do not include specified nodes.\n\n Parameters\n ----------\n network : pandana.Network\n rm_nodes : array_like\n A list, array, Index, or Series of node IDs that should *not*\n be saved as part of the Network.\n\n Returns\n -------\n nodes, edges : pandas.DataFrame\n\n \"\"\"\n rm_nodes = set(rm_nodes)\n ndf = network.nodes_df\n edf = network.edges_df\n\n nodes_to_keep = ~ndf.index.isin(rm_nodes)\n edges_to_keep = ~(edf['from'].isin(rm_nodes) | edf['to'].isin(rm_nodes))\n\n return ndf.loc[nodes_to_keep], edf.loc[edges_to_keep]\n\n\ndef network_to_pandas_hdf5(network, filename, rm_nodes=None):\n \"\"\"\n Save a Network's data to a Pandas HDFStore.\n\n Parameters\n ----------\n network : pandana.Network\n filename : str\n rm_nodes : array_like\n A list, array, Index, or Series of node IDs that should *not*\n be saved as part of the Network.\n\n \"\"\"\n if rm_nodes is not None:\n nodes, edges = remove_nodes(network, rm_nodes)\n else:\n nodes, edges = network.nodes_df, network.edges_df\n\n with pd.HDFStore(filename, mode='w') as store:\n store['nodes'] = nodes\n store['edges'] = edges\n\n store['two_way'] = pd.Series([network._twoway])\n store['impedance_names'] = pd.Series(network.impedance_names)\n\n\ndef network_from_pandas_hdf5(cls, filename):\n \"\"\"\n Build a Network from data in a Pandas HDFStore.\n\n Parameters\n ----------\n cls : class\n Class to instantiate, usually pandana.Network.\n filename : str\n\n Returns\n -------\n network : pandana.Network\n\n \"\"\"\n with pd.HDFStore(filename) as store:\n nodes = store['nodes']\n edges = store['edges']\n two_way = store['two_way'][0]\n imp_names = store['impedance_names'].tolist()\n\n return cls(\n nodes['x'], nodes['y'], edges['from'], edges['to'], edges[imp_names],\n twoway=two_way)\n", "license": "agpl-3.0"} |
{"repo_name": "sbg2133/miscellaneous_projects", "path": "carina/lic_thesis_highSL.py", "copies": "1", "size": "6145", "content": "from getIQU import IQU\nfrom subprocess import call\nimport sys,os\nimport numpy as np\nimport glob\nimport matplotlib\nimport matplotlib.pyplot as plt\nfrom astropy.convolution import convolve, Gaussian2DKernel\nfrom astropy.io import fits\nimport scipy.ndimage\nfrom makePretty import pretty\nimport aplpy\nfrom skimage import filters\n#matplotlib.rcParams.update({'font.size':16})\nsave_files_here = \"/home/wizwit/SESE_dissertation/figures/chapter6\"\nplt.ion()\n\nif len(sys.argv) < 2:\n print \"You must supply a band, 'e.g., carina_lic.py 250'\"\n sys.exit()\nelse:\n band = sys.argv[1]\n\n# Call this function with desired band as first argument, e.g.:\n# python carina_lic.py 250\n# lists over which to iterate:\nbands = ['250', '350', '500']\nstokes = ['I', 'Q', 'U']\npol_eff = [0.81, 0.79, 0.82]\n\n# define file paths\nblastpol_dir = './carinaData/smooth/3.0_arcmin'\nfilename = glob.glob(blastpol_dir + '/carinaneb_' + band + '_smoothed_3.0_rl.fits')[0]\n\n# load in I, Q, U for desired band\nIvals, Qvals, Uvals, __, wcs = IQU(filename)\n\nI = Ivals[30:-30,260:-260]\nQ = Qvals[30:-30,260:-260]\nU = Uvals[30:-30,260:-260]\nPvals = np.sqrt(Q**2 + U**2)\npvals = Pvals/I\n\n# Correct pvals as in Jamil's thesis, 5.7\npvals[pvals > 0.5] = np.nan\npvals[pvals < 0] = np.nan\npvals /= pol_eff[bands.index(band)]\nphi = 0.5*np.arctan2(U,Q)\n#dx = np.cos(phi)\n#dy = np.sin(phi)\ndx = pvals*np.cos(phi)\ndy = pvals*np.sin(phi)\nmag = np.sqrt(dx**2 + dy**2)\nX = np.linspace(0, I.shape[1], I.shape[1])\nY = np.linspace(0, I.shape[0], I.shape[0])\nxs, ys = np.meshgrid(X,Y)\n\n\"\"\"\nplt.figure()\nnskip = 2\nskip = (slice(None, None, nskip), slice(None, None, nskip))\n#f = aplpy.FITSFigure(I, figsize = (10.24,7.68), dpi = 100)\nax = plt.gca()\nax.imshow(I, cmap = \"gist_heat\")\n#f.tick_labels.set_font(size='small')\n#f.show_colorscale(cmap='gist_heat')\n# Add polarization vectors\nax.quiver(xs[skip],ys[skip],(dx/mag)[skip],(dy/mag)[skip], color = \"white\", angles = 'xy', units = 'xy', scale_units = 'xy', scale = 0.3)\n#f.show_vectors(pvals, phi, color = 'white', rotate = 90., scale = 50, step = 10) \nax.set_facecolor('black')\nplt.tight_layout()\n\"\"\"\nxsize, ysize = len(X), len(Y)\n\nvectors = np.array([dx,dy])\n#white = np.random.rand(xsize, ysize)\n#white = np.random.uniform(low = 0., high = 1., size = (xsize, ysize))\nwhite = np.random.normal(0., 1., size = (xsize,ysize))\nsigma = 1.2\nwhite = scipy.ndimage.gaussian_filter(white, sigma)\n\nwith file('texture.dat', 'w') as outfile:\n for row in white:\n np.savetxt(outfile, row, newline = \" \")\n outfile.write('\\n')\nwith file('dx.dat', 'w') as outfile:\n for row in dx:\n np.savetxt(outfile, row, newline = \" \")\n outfile.write('\\n')\nwith file('dy.dat', 'w') as outfile:\n for row in dy:\n np.savetxt(outfile, row, newline = \" \")\n outfile.write('\\n')\n\nif len(sys.argv) > 2:\n print\n print \"Doing LIC\"\n print\n command = [\"./carina_lic\", str(xsize), str(ysize)]\n call(command)\n\nlic = np.loadtxt(\"./lic.dat\")\nlic = np.transpose(lic)\n#np.save('lic.npy', lic)\n#lic2 = np.load(\"lic.npy\")\n#lic = scipy.ndimage.gaussian_filter(lic, 1.01)\nmult = lic * I\n\n\"\"\"\nblur_size = 8\nunsharp_strength = 0.8\nblurred = filter.gaussian_filter(lic, blur_size)\nhighpass = lic - unsharp_strength*blurred\nsharp = lic + highpass\nlowpass = scipy.ndimage.gaussian_filter(lic, 5)\nhighpass = lic - lowpass\nhighpass += lic\n\"\"\"\n\"\"\"\nhdu2 = fits.PrimaryHDU(data=np.zeros_like(I), header=wcs.to_header())\nf2 = aplpy.FITSFigure(hdu2, figsize = (10,10))\nf2.set_theme('publication')\nax = plt.gca()\nax.set_facecolor(\"k\")\nf2.add_scalebar(15/60.) # arcmin\nf2.scalebar.set_label('0.5 deg')\nf2.scalebar.set_color('white')\nf2.scalebar.set_corner('bottom right')\nf2.scalebar.set_label('10 pc')\nf2.tick_labels.set_yformat('dd.dd')\nf2.tick_labels.set_xformat('dd.dd')\nf2.axis_labels.set_font(size=16)\nf2.tick_labels.set_font(size=14)\nplt.imshow(I, origin = 'lower', cmap = \"inferno\", alpha = 1)\n#plt.imshow(lic, vmin = -0.07, vmax = 0.3, origin = 'lower', cmap = \"gray\", alpha = 0.4, interpolation = \"bilinear\")\nplt.imshow(lic, vmin = -0.06, vmax = 0.25, origin = 'lower', cmap = \"gray\", alpha = 0.5, interpolation = \"bilinear\")\nplt.tight_layout()\n#f2.savefig(os.path.join(save_files_here, 'lic_han_51.eps'), format='eps', dpi=1000, transparent = True)\nplt.savefig(os.path.join(save_files_here, 'lic_han_51.png'), format='png', bbox_inches = 'tight')\n\n##################################################\n# For 250 um: v = 1000\n# For 350 um: v = 500\n# For 500 um: v = 200\n\nhdu3 = fits.PrimaryHDU(data=np.zeros_like(I), header=wcs.to_header())\nf3 = aplpy.FITSFigure(hdu3, figsize = (10,10))\nf3.set_theme('publication')\n# scalebar\nax = plt.gca()\nax.set_facecolor(\"k\")\nf3.add_scalebar(15/60.) # arcmin\nf3.scalebar.set_label('0.5 deg')\nf3.scalebar.set_color('white')\nf3.scalebar.set_corner('bottom right')\nf3.scalebar.set_label('10 pc')\nf3.tick_labels.set_yformat('dd.dd')\nf3.tick_labels.set_xformat('dd.dd')\nf3.axis_labels.set_font(size=16)\nf3.tick_labels.set_font(size=14)\nvmin = [-0.000007, 0.4, 0.5]\nvmax = [0.00007, 0.4, 0.5]\nplt.imshow(mult, origin = 'lower',\\\n cmap = \"inferno\", vmin = vmin[bands.index(band)],\\\n vmax = vmax[bands.index(band)], interpolation = 'bilinear')\n#plt.tight_layout()\nplt.savefig(os.path.join(save_files_here, 'lic2_han51.png'), format='png', bbox_inches = 'tight')\n\"\"\"\nhdu4 = fits.PrimaryHDU(data=np.zeros_like(I), header=wcs.to_header())\nf4 = aplpy.FITSFigure(hdu4, figsize = (10,10))\nf4.set_theme('publication')\n# scalebar\nax = plt.gca()\nax.set_facecolor(\"k\")\nf4.add_scalebar(15/60.) # arcmin\nf4.scalebar.set_label('0.5 deg')\nf4.scalebar.set_color('white')\nf4.scalebar.set_corner('bottom right')\nf4.scalebar.set_label('10 pc')\nf4.tick_labels.set_yformat('dd.dd')\nf4.tick_labels.set_xformat('dd.dd')\nf4.axis_labels.set_font(size=16)\nf4.tick_labels.set_font(size=14)\nvmin = [-0.000007, 0.4, 0.5]\nvmax = [0.00007, 0.4, 0.5]\nplt.imshow(mult, origin = 'lower',\\\n cmap = \"viridis\", vmin = vmin[bands.index(band)],\\\n vmax = vmax[bands.index(band)], interpolation = 'bilinear')\n#plt.tight_layout()\nplt.savefig(os.path.join(save_files_here, 'lic2_viridis.png'), format='png', bbox_inches = 'tight')\n\n", "license": "gpl-3.0"} |
{"repo_name": "466152112/scikit-learn", "path": "sklearn/tests/test_learning_curve.py", "copies": "225", "size": "10791", "content": "# Author: Alexander Fabisch <[email protected]>\n#\n# License: BSD 3 clause\n\nimport sys\nfrom sklearn.externals.six.moves import cStringIO as StringIO\nimport numpy as np\nimport warnings\nfrom sklearn.base import BaseEstimator\nfrom sklearn.learning_curve import learning_curve, validation_curve\nfrom sklearn.utils.testing import assert_raises\nfrom sklearn.utils.testing import assert_warns\nfrom sklearn.utils.testing import assert_equal\nfrom sklearn.utils.testing import assert_array_equal\nfrom sklearn.utils.testing import assert_array_almost_equal\nfrom sklearn.datasets import make_classification\nfrom sklearn.cross_validation import KFold\nfrom sklearn.linear_model import PassiveAggressiveClassifier\n\n\nclass MockImprovingEstimator(BaseEstimator):\n \"\"\"Dummy classifier to test the learning curve\"\"\"\n def __init__(self, n_max_train_sizes):\n self.n_max_train_sizes = n_max_train_sizes\n self.train_sizes = 0\n self.X_subset = None\n\n def fit(self, X_subset, y_subset=None):\n self.X_subset = X_subset\n self.train_sizes = X_subset.shape[0]\n return self\n\n def predict(self, X):\n raise NotImplementedError\n\n def score(self, X=None, Y=None):\n # training score becomes worse (2 -> 1), test error better (0 -> 1)\n if self._is_training_data(X):\n return 2. - float(self.train_sizes) / self.n_max_train_sizes\n else:\n return float(self.train_sizes) / self.n_max_train_sizes\n\n def _is_training_data(self, X):\n return X is self.X_subset\n\n\nclass MockIncrementalImprovingEstimator(MockImprovingEstimator):\n \"\"\"Dummy classifier that provides partial_fit\"\"\"\n def __init__(self, n_max_train_sizes):\n super(MockIncrementalImprovingEstimator,\n self).__init__(n_max_train_sizes)\n self.x = None\n\n def _is_training_data(self, X):\n return self.x in X\n\n def partial_fit(self, X, y=None, **params):\n self.train_sizes += X.shape[0]\n self.x = X[0]\n\n\nclass MockEstimatorWithParameter(BaseEstimator):\n \"\"\"Dummy classifier to test the validation curve\"\"\"\n def __init__(self, param=0.5):\n self.X_subset = None\n self.param = param\n\n def fit(self, X_subset, y_subset):\n self.X_subset = X_subset\n self.train_sizes = X_subset.shape[0]\n return self\n\n def predict(self, X):\n raise NotImplementedError\n\n def score(self, X=None, y=None):\n return self.param if self._is_training_data(X) else 1 - self.param\n\n def _is_training_data(self, X):\n return X is self.X_subset\n\n\ndef test_learning_curve():\n X, y = make_classification(n_samples=30, n_features=1, n_informative=1,\n n_redundant=0, n_classes=2,\n n_clusters_per_class=1, random_state=0)\n estimator = MockImprovingEstimator(20)\n with warnings.catch_warnings(record=True) as w:\n train_sizes, train_scores, test_scores = learning_curve(\n estimator, X, y, cv=3, train_sizes=np.linspace(0.1, 1.0, 10))\n if len(w) > 0:\n raise RuntimeError(\"Unexpected warning: %r\" % w[0].message)\n assert_equal(train_scores.shape, (10, 3))\n assert_equal(test_scores.shape, (10, 3))\n assert_array_equal(train_sizes, np.linspace(2, 20, 10))\n assert_array_almost_equal(train_scores.mean(axis=1),\n np.linspace(1.9, 1.0, 10))\n assert_array_almost_equal(test_scores.mean(axis=1),\n np.linspace(0.1, 1.0, 10))\n\n\ndef test_learning_curve_unsupervised():\n X, _ = make_classification(n_samples=30, n_features=1, n_informative=1,\n n_redundant=0, n_classes=2,\n n_clusters_per_class=1, random_state=0)\n estimator = MockImprovingEstimator(20)\n train_sizes, train_scores, test_scores = learning_curve(\n estimator, X, y=None, cv=3, train_sizes=np.linspace(0.1, 1.0, 10))\n assert_array_equal(train_sizes, np.linspace(2, 20, 10))\n assert_array_almost_equal(train_scores.mean(axis=1),\n np.linspace(1.9, 1.0, 10))\n assert_array_almost_equal(test_scores.mean(axis=1),\n np.linspace(0.1, 1.0, 10))\n\n\ndef test_learning_curve_verbose():\n X, y = make_classification(n_samples=30, n_features=1, n_informative=1,\n n_redundant=0, n_classes=2,\n n_clusters_per_class=1, random_state=0)\n estimator = MockImprovingEstimator(20)\n\n old_stdout = sys.stdout\n sys.stdout = StringIO()\n try:\n train_sizes, train_scores, test_scores = \\\n learning_curve(estimator, X, y, cv=3, verbose=1)\n finally:\n out = sys.stdout.getvalue()\n sys.stdout.close()\n sys.stdout = old_stdout\n\n assert(\"[learning_curve]\" in out)\n\n\ndef test_learning_curve_incremental_learning_not_possible():\n X, y = make_classification(n_samples=2, n_features=1, n_informative=1,\n n_redundant=0, n_classes=2,\n n_clusters_per_class=1, random_state=0)\n # The mockup does not have partial_fit()\n estimator = MockImprovingEstimator(1)\n assert_raises(ValueError, learning_curve, estimator, X, y,\n exploit_incremental_learning=True)\n\n\ndef test_learning_curve_incremental_learning():\n X, y = make_classification(n_samples=30, n_features=1, n_informative=1,\n n_redundant=0, n_classes=2,\n n_clusters_per_class=1, random_state=0)\n estimator = MockIncrementalImprovingEstimator(20)\n train_sizes, train_scores, test_scores = learning_curve(\n estimator, X, y, cv=3, exploit_incremental_learning=True,\n train_sizes=np.linspace(0.1, 1.0, 10))\n assert_array_equal(train_sizes, np.linspace(2, 20, 10))\n assert_array_almost_equal(train_scores.mean(axis=1),\n np.linspace(1.9, 1.0, 10))\n assert_array_almost_equal(test_scores.mean(axis=1),\n np.linspace(0.1, 1.0, 10))\n\n\ndef test_learning_curve_incremental_learning_unsupervised():\n X, _ = make_classification(n_samples=30, n_features=1, n_informative=1,\n n_redundant=0, n_classes=2,\n n_clusters_per_class=1, random_state=0)\n estimator = MockIncrementalImprovingEstimator(20)\n train_sizes, train_scores, test_scores = learning_curve(\n estimator, X, y=None, cv=3, exploit_incremental_learning=True,\n train_sizes=np.linspace(0.1, 1.0, 10))\n assert_array_equal(train_sizes, np.linspace(2, 20, 10))\n assert_array_almost_equal(train_scores.mean(axis=1),\n np.linspace(1.9, 1.0, 10))\n assert_array_almost_equal(test_scores.mean(axis=1),\n np.linspace(0.1, 1.0, 10))\n\n\ndef test_learning_curve_batch_and_incremental_learning_are_equal():\n X, y = make_classification(n_samples=30, n_features=1, n_informative=1,\n n_redundant=0, n_classes=2,\n n_clusters_per_class=1, random_state=0)\n train_sizes = np.linspace(0.2, 1.0, 5)\n estimator = PassiveAggressiveClassifier(n_iter=1, shuffle=False)\n\n train_sizes_inc, train_scores_inc, test_scores_inc = \\\n learning_curve(\n estimator, X, y, train_sizes=train_sizes,\n cv=3, exploit_incremental_learning=True)\n train_sizes_batch, train_scores_batch, test_scores_batch = \\\n learning_curve(\n estimator, X, y, cv=3, train_sizes=train_sizes,\n exploit_incremental_learning=False)\n\n assert_array_equal(train_sizes_inc, train_sizes_batch)\n assert_array_almost_equal(train_scores_inc.mean(axis=1),\n train_scores_batch.mean(axis=1))\n assert_array_almost_equal(test_scores_inc.mean(axis=1),\n test_scores_batch.mean(axis=1))\n\n\ndef test_learning_curve_n_sample_range_out_of_bounds():\n X, y = make_classification(n_samples=30, n_features=1, n_informative=1,\n n_redundant=0, n_classes=2,\n n_clusters_per_class=1, random_state=0)\n estimator = MockImprovingEstimator(20)\n assert_raises(ValueError, learning_curve, estimator, X, y, cv=3,\n train_sizes=[0, 1])\n assert_raises(ValueError, learning_curve, estimator, X, y, cv=3,\n train_sizes=[0.0, 1.0])\n assert_raises(ValueError, learning_curve, estimator, X, y, cv=3,\n train_sizes=[0.1, 1.1])\n assert_raises(ValueError, learning_curve, estimator, X, y, cv=3,\n train_sizes=[0, 20])\n assert_raises(ValueError, learning_curve, estimator, X, y, cv=3,\n train_sizes=[1, 21])\n\n\ndef test_learning_curve_remove_duplicate_sample_sizes():\n X, y = make_classification(n_samples=3, n_features=1, n_informative=1,\n n_redundant=0, n_classes=2,\n n_clusters_per_class=1, random_state=0)\n estimator = MockImprovingEstimator(2)\n train_sizes, _, _ = assert_warns(\n RuntimeWarning, learning_curve, estimator, X, y, cv=3,\n train_sizes=np.linspace(0.33, 1.0, 3))\n assert_array_equal(train_sizes, [1, 2])\n\n\ndef test_learning_curve_with_boolean_indices():\n X, y = make_classification(n_samples=30, n_features=1, n_informative=1,\n n_redundant=0, n_classes=2,\n n_clusters_per_class=1, random_state=0)\n estimator = MockImprovingEstimator(20)\n cv = KFold(n=30, n_folds=3)\n train_sizes, train_scores, test_scores = learning_curve(\n estimator, X, y, cv=cv, train_sizes=np.linspace(0.1, 1.0, 10))\n assert_array_equal(train_sizes, np.linspace(2, 20, 10))\n assert_array_almost_equal(train_scores.mean(axis=1),\n np.linspace(1.9, 1.0, 10))\n assert_array_almost_equal(test_scores.mean(axis=1),\n np.linspace(0.1, 1.0, 10))\n\n\ndef test_validation_curve():\n X, y = make_classification(n_samples=2, n_features=1, n_informative=1,\n n_redundant=0, n_classes=2,\n n_clusters_per_class=1, random_state=0)\n param_range = np.linspace(0, 1, 10)\n with warnings.catch_warnings(record=True) as w:\n train_scores, test_scores = validation_curve(\n MockEstimatorWithParameter(), X, y, param_name=\"param\",\n param_range=param_range, cv=2\n )\n if len(w) > 0:\n raise RuntimeError(\"Unexpected warning: %r\" % w[0].message)\n\n assert_array_almost_equal(train_scores.mean(axis=1), param_range)\n assert_array_almost_equal(test_scores.mean(axis=1), 1 - param_range)\n", "license": "bsd-3-clause"} |
{"example": "Insign is Interactive"} |
{"example": "Syndicate Crypto"} |
{"example": "Doctor David cohen"} |
{"example": "Rock United"} |
{"example": "Drumming Mad"} |
{"translation": {"en": "It is not for humans in general.", "it": "E non gli esseri umani in generale."}} |
{"translation": {"en": "Did you know there are apps to remind you to take naps?", "it": "Lo sapevi che esistono delle app che ti ricordano di bere?"}} |
{"context": "This shift has already commenced in some countries; for instance, pharmacists in Australia receive remuneration from the Australian Government for conducting comprehensive Home Medicines Reviews. In Canada, pharmacists in certain provinces have limited prescribing rights (as in Alberta and British Columbia) or are remunerated by their provincial government for expanded services such as medications reviews (Medschecks in Ontario). In the United Kingdom, pharmacists who undertake additional training are obtaining prescribing rights and this is because of pharmacy education. They are also being paid for by the government for medicine use reviews. In Scotland the pharmacist can write prescriptions for Scottish registered patients of their regular medications, for the majority of drugs, except for controlled drugs, when the patient is unable to see their doctor, as could happen if they are away from home or the doctor is unavailable. In the United States, pharmaceutical care or clinical pharmacy has had an evolving influence on the practice of pharmacy. Moreover, the Doctor of Pharmacy (Pharm. D.) degree is now required before entering practice and some pharmacists now complete one or two years of residency or fellowship training following graduation. In addition, consultant pharmacists, who traditionally operated primarily in nursing homes are now expanding into direct consultation with patients, under the banner of \"senior care pharmacy.\"", "question": "Who can't pay Australian pharmacists for doing Home Medicines Reviews?", "answers.text": [""], "answers.answer_start": [-1], "feat_id": ["5a6cef5a4eec6b001a80a6fd"], "feat_title": ["Pharmacy"], "start_logits": [NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN], "end_logits": [NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN]} |
{"text": "In honor of Sandy @BlueOceanArctic and all her climate videos I made JUST A GUY TRYING TO PUT OUT A PLANET ON FIRE https://t.co/aoEYG7I0oD via @YouTube", "label": 2} |
{"text": "Well of course, it was a much direr year than 2021. That's what happens.\n\nIf only the BBC had made the early fire service BBQ warnings more prominent* - many of these would have been prevented.\n\nBe part of the solution.\n\n*focused instead on climate change\nhttps://t.co/Aw1mTYGnBf", "label": 0} |
{"text": "@Michelle4NM @MarkRonchettiNM @RonDeSantisFL Is attacking your opponent the only way you run your campaign? Climate change didn't have anything to do with the largest fire in our states history. Maybe point your finger at the U.S forest service who thought it would be a splendid idea to do a prescribed burn in wind", "label": 0} |
{"text": "Fires? From the fire failure queen? The 20something year old with a poli sci degree shouldn't worry about climate change. She should worry about the agreement Grisham had with the forest service back in 2019 that failed to do EXACTLY what we needed & what she said it would do. https://t.co/jCzGQcM7XL https://t.co/iNQsTNmyYl", "label": 0} |
{"text": "@reddit_lies a guy lit himself on fire to protest climate change. this is a game for several players", "label": 2} |
{"text": "Windy Fire has destroyed more than two dozen giant sequoia trees -- and that's just an early assessment. (CNN) #ClimateChange #Environment https://t.co/sppowx5lQS https://t.co/MRrRBSzKMT", "label": 1} |
{"text": "Wales is on fire. We need climate action now. https://t.co/OhXfqlBFni", "label": 1} |
{"stratum": 1, "class": "Abfss.java", "comment": "* Azure Blob File System implementation of AbstractFileSystem.\n* This impl delegates to the old FileSystem", "summary": "Azure Blob File System implementation of AbstractFileSystem.", "Expand": "This impl delegates to the old FileSystem", "rational": null, "deprecation": null, "usage": null, "exception": null, "todo": null, "Incomplete": null, "Commentedcode": null, "directive": null, "formatter": null, "License": null, "Ownership": null, "Pointer": null, "Autogenerated": null, "Noise": null, "Warning": null, "Recommendation": null, "Precondition": null, "CodingGuidelines": null, "Extension": null, "Subclassexplnation": null, "Observation": null} |
{"stratum": 1, "class": "AbstractContractGetFileStatusTest.java", "comment": "* Test getFileStatus and related listing operations.\n| the tree parameters. Kept small to avoid killing object store test| runs too much.|\n* Accept everything.\n| the tree parameters. Kept small to avoid killing object store test| runs too much.|\n* Accept nothing.\n| the tree parameters. Kept small to avoid killing object store test| runs too much.|\n* Path filter which only expects paths whose final name element\n* equals the {@code match} field.\n| the tree parameters. Kept small to avoid killing object store test| runs too much.|\n* A filesystem filter which exposes the protected method\n* {@link #listLocatedStatus(Path, PathFilter)}.", "summary": "Test getFileStatus and related listing operations.", "Expand": "| the tree parameters. Kept small to avoid killing object store test| runs too much.|\n* Accept everything.\n| the tree parameters. Kept small to avoid killing object store test| runs too much.|\n* Accept nothing.\n| the tree parameters. Kept small to avoid killing object store test| runs too much.|\n* Path filter which only expects paths whose final name element\n* equals the {@code match} field.\n| the tree parameters. Kept small to avoid killing object store test| runs too much.|\n* A filesystem filter which exposes the protected method\n* {@link #listLocatedStatus(Path, PathFilter)}.", "rational": null, "deprecation": null, "usage": null, "exception": null, "todo": null, "Incomplete": null, "Commentedcode": null, "directive": null, "formatter": null, "License": null, "Ownership": null, "Pointer": null, "Autogenerated": null, "Noise": null, "Warning": null, "Recommendation": null, "Precondition": null, "CodingGuidelines": null, "Extension": null, "Subclassexplnation": null, "Observation": null} |
{"stratum": 1, "class": "AbstractFSContract.java", "comment": "* Class representing a filesystem contract that a filesystem\n* implementation is expected implement.\n*\n* Part of this contract class is to allow FS implementations to\n* provide specific opt outs and limits, so that tests can be\n* skip unsupported features (e.g. case sensitivity tests),\n* dangerous operations (e.g. trying to delete the root directory),\n* and limit filesize and other numeric variables for scale tests", "summary": "Class representing a filesystem contract that a filesystem * implementation is expected implement.", "Expand": null, "rational": "Part of this contract class is to allow FS implementations to * provide specific opt outs and limits, so that tests can be * skip unsupported features (e.g. case sensitivity tests), * dangerous operations (e.g. trying to delete the root directory), * and limit filesize and other numeric variables for scale tests", "deprecation": null, "usage": null, "exception": null, "todo": null, "Incomplete": null, "Commentedcode": null, "directive": null, "formatter": null, "License": null, "Ownership": null, "Pointer": null, "Autogenerated": null, "Noise": null, "Warning": null, "Recommendation": null, "Precondition": null, "CodingGuidelines": null, "Extension": null, "Subclassexplnation": null, "Observation": null} |
{"stratum": 1, "class": "AbstractS3ACommitterFactory.java", "comment": "* Dynamically create the output committer based on subclass type and settings.", "summary": null, "Expand": "Dynamically create the output committer based on subclass type and settings.", "rational": null, "deprecation": null, "usage": null, "exception": null, "todo": null, "Incomplete": null, "Commentedcode": null, "directive": null, "formatter": null, "License": null, "Ownership": null, "Pointer": null, "Autogenerated": null, "Noise": null, "Warning": null, "Recommendation": null, "Precondition": null, "CodingGuidelines": null, "Extension": null, "Subclassexplnation": null, "Observation": null} |
{"stratum": 1, "class": "AbstractTFLaunchCommandTestHelper.java", "comment": "* This class is an abstract base class for testing Tensorboard and TensorFlow\n* launch commands.", "summary": "This class is an abstract base class for testing Tensorboard and TensorFlow and launch commands.", "Expand": null, "rational": "This class is an abstract base class for testing Tensorboard and TensorFlow and launch commands.", "deprecation": null, "usage": null, "exception": null, "todo": null, "Incomplete": null, "Commentedcode": null, "directive": null, "formatter": null, "License": null, "Ownership": null, "Pointer": null, "Autogenerated": null, "Noise": null, "Warning": null, "Recommendation": null, "Precondition": null, "CodingGuidelines": null, "Extension": null, "Subclassexplnation": null, "Observation": null} |
{"stratum": 1, "class": "ApplicationConstants.java", "comment": "* This is the API for the applications comprising of constants that YARN sets\n* up for the applications and the containers.\n*\n* TODO: Investigate the semantics and security of each cross-boundary refs.\n|\n* The type of launch for the container.\n|\n* Environment for Applications.\n*\n* Some of the environment variables for applications are <em>final</em>\n* i.e. they cannot be modified by the applications.", "summary": "* This is the API for the applications comprising of constants that YARN sets * up for the applications and the containers.", "Expand": "The type of launch for the container. | * Environment for Applications. * * Some of the environment variables for applications are <em>final</em> * i.e. they cannot be modified by the applications.", "rational": null, "deprecation": null, "usage": null, "exception": null, "todo": "TODO: Investigate the semantics and security of each cross-boundary refs.", "Incomplete": null, "Commentedcode": null, "directive": null, "formatter": null, "License": null, "Ownership": null, "Pointer": null, "Autogenerated": null, "Noise": null, "Warning": "* Some of the environment variables for applications are <em>final</em>\n* i.e. they cannot be modified by the applications.", "Recommendation": null, "Precondition": null, "CodingGuidelines": null, "Extension": null, "Subclassexplnation": null, "Observation": null} |
{"stratum": 1, "class": "ApplicationStateData.java", "comment": "* Contains all the state data that needs to be stored persistently * for an Application", "summary": "Contains all the state data that needs to be stored persistently * for an Application", "Expand": null, "rational": "needs to be stored persistently * for an Application", "deprecation": null, "usage": null, "exception": null, "todo": null, "Incomplete": null, "Commentedcode": null, "directive": null, "formatter": null, "License": null, "Ownership": null, "Pointer": null, "Autogenerated": null, "Noise": null, "Warning": null, "Recommendation": null, "Precondition": null, "CodingGuidelines": null, "Extension": null, "Subclassexplnation": null, "Observation": null} |
{"stratum": 1, "class": "AutoInputFormat.java", "comment": "* An {@link InputFormat} that tries to deduce the types of the input files\n* automatically. It can currently handle text and sequence files.", "summary": "An {@link InputFormat} that tries to deduce the types of the input files * automatically.", "Expand": "It can currently handle text and sequence files.", "rational": null, "deprecation": null, "usage": null, "exception": null, "todo": null, "Incomplete": null, "Commentedcode": null, "directive": null, "formatter": null, "License": null, "Ownership": null, "Pointer": "An {@link InputFormat} that tries to deduce the types of the input files * automatically.", "Autogenerated": null, "Noise": null, "Warning": null, "Recommendation": null, "Precondition": null, "CodingGuidelines": null, "Extension": null, "Subclassexplnation": null, "Observation": null} |
{"stratum": 1, "class": "BalancingPolicy.java", "comment": "* Balancing policy.\n* Since a datanode may contain multiple block pools,\n* {@link Pool} implies {@link Node}\n* but NOT the other way around\n|\n* Cluster is balanced if each node is balanced.\n|\n* Cluster is balanced if each pool in each node is balanced.", "summary": "* Balancing policy.", "Expand": "Since a datanode may contain multiple block pools, * {@link Pool} implies {@link Node} * but NOT the other way around |", "rational": "Cluster is balanced if each node is balanced. | * Cluster is balanced if each pool in each node is balanced.", "deprecation": null, "usage": null, "exception": null, "todo": null, "Incomplete": null, "Commentedcode": null, "directive": null, "formatter": null, "License": null, "Ownership": null, "Pointer": "Since a datanode may contain multiple block pools, * {@link Pool} implies {@link Node} * but NOT the other way around |", "Autogenerated": null, "Noise": null, "Warning": null, "Recommendation": null, "Precondition": null, "CodingGuidelines": null, "Extension": null, "Subclassexplnation": null, "Observation": null} |
{"stratum": 1, "class": "BaseRouterWebServicesTest.java", "comment": "* Base class for all the RouterRMAdminService test cases. It provides utility\n* methods that can be used by the concrete test case classes.\n*", "summary": "* Base class for all the RouterRMAdminService test cases.", "Expand": null, "rational": null, "deprecation": null, "usage": "It provides utility * methods that can be used by the concrete test case classes.", "exception": null, "todo": null, "Incomplete": null, "Commentedcode": null, "directive": null, "formatter": null, "License": null, "Ownership": null, "Pointer": null, "Autogenerated": null, "Noise": null, "Warning": null, "Recommendation": null, "Precondition": null, "CodingGuidelines": null, "Extension": null, "Subclassexplnation": null, "Observation": null} |
{"questions": {"id": [1, 2], "text": ["What is the step by step guide to invest in share market in india?", "What is the step by step guide to invest in share market?"]}, "is_duplicate": false} |
{"questions": {"id": [3, 4], "text": ["What is the story of Kohinoor (Koh-i-Noor) Diamond?", "What would happen if the Indian government stole the Kohinoor (Koh-i-Noor) diamond back?"]}, "is_duplicate": false} |
{"questions": {"id": [5, 6], "text": ["How can I increase the speed of my internet connection while using a VPN?", "How can Internet speed be increased by hacking through DNS?"]}, "is_duplicate": false} |
{"questions": {"id": [7, 8], "text": ["Why am I mentally very lonely? How can I solve it?", "Find the remainder when [math]23^{24}[/math] is divided by 24,23?"]}, "is_duplicate": false} |
{"questions": {"id": [9, 10], "text": ["Which one dissolve in water quikly sugar, salt, methane and carbon di oxide?", "Which fish would survive in salt water?"]}, "is_duplicate": false} |
{"questions": {"id": [11, 12], "text": ["Astrology: I am a Capricorn Sun Cap moon and cap rising...what does that say about me?", "I'm a triple Capricorn (Sun, Moon and ascendant in Capricorn) What does this say about me?"]}, "is_duplicate": true} |
{"questions": {"id": [13, 14], "text": ["Should I buy tiago?", "What keeps childern active and far from phone and video games?"]}, "is_duplicate": false} |
{"questions": {"id": [15, 16], "text": ["How can I be a good geologist?", "What should I do to be a great geologist?"]}, "is_duplicate": true} |
{"questions": {"id": [19, 20], "text": ["Motorola (company): Can I hack my Charter Motorolla DCX3400?", "How do I hack Motorola DCX3400 for free internet?"]}, "is_duplicate": false} |
{"q_id": "1yc9zg", "title": "Are there any good source material on the Warsaw Ghetto to be had online?", "selftext": "Hi guys, I have a project I'm working on which requires unbiased primary sources on the Warsaw ghetto (demographics, population density, supplies, mortality rates, etc.). I've been scouring the net and my library for information about it, but I mostly seem to find holocaust denialist websites and websites referring to a book I can attain within my timeframe whilst providing incomplete information.\n\nI'd really appreciate it, thanks!\n\nAlternatively: Are there any better subreddits where I could ask for help? I figure I'll post this to /r/favors and /r/ask, but if there are any other good subs for this kind of stuff I'd appreciate you mentioning them.\n\nCheers!", "document": "", "subreddit": "AskHistorians", "url": "http://www.reddit.com/r/AskHistorians/comments/1yc9zg/are_there_any_good_source_material_on_the_warsaw/", "answers": {"a_id": ["cfj9u9t"], "score": [3], "text": ["Many of the relevant primary sources wont contain the those specific details in an aggregated way. There are a few great examples of diaries and reports coming from the Warsaw Ghetto. The statistics you are looking for would likely come from secondary sources.\n\nOne great example is the Stroop Report. This is written by a commander (Stroop) and it documents the suppression of the uprising. I've included a link to the National Archive where you can get the the full report [here](_URL_1_). This document was used in the Nuremberg Trials.\n\nI would recommend you check out the [Yad Vashem site](_URL_2_). That is the Holocaust Museum in Israel. They have spent a lot of time collecting primary sources, photos, personal accounts etc, about the Holocaust of the Jews (Shoah). I've linked to the overview of the Warsaw Ghetto, but check out the digital collection at the top.\n\nA last little tip is diaries of the time. I don't know where you're located, or which libraries you have access to but here are a few WorldCat records for some notable ones: [The Warsaw diary of Adam Czerniakow](_URL_0_) and [Scroll of agony, the Warsaw diary of Chaim A. Kaplan](_URL_3_).\n\nHope this help.\n\nSource: I'm an academic reference librarian and Jewish history specialist"]}, "title_urls": [], "selftext_urls": [], "answers_urls": [["http://www.worldcat.org/title/warsaw-diary-of-adam-czerniakow-prelude-to-doom/oclc/3913266&referer=brief_results", "http://arcweb.archives.gov/arc/action/ShowFullRecordLinked?%24searchId=2&%24showFullDescriptionTabs.selectedPaneId=digital&%24digiDetailPageModel.currentPage=1&%24digiViewModel.detailId=1&%24partitionIndex=0&%24digiSummaryPageModel.targetModel=true&%24submitId=1&%24digiViewModel.name=digiViewModel&%24resultsDetailPageModel.search=true&%24digiDetailPageModel.resultPageModel=true&%24resultsDetailPageModel.currentPage=0&%24resultsDetailPageModel.pageSize=1&%24sort=RELEVANCE_ASC&%24highlight=false&detail=digiViewModel", "http://www.yadvashem.org/yv/en/holocaust/about/03/warsaw.asp?WT.mc_id=wiki", "http://www.worldcat.org/search?q=Chaim+A.+Kaplan&qt=owc_search"]]} |
{"q_id": "3qr7uu", "title": "In medieval and pre-modern times, political entities made marriage pacts between heirs in order to secure peace. Often times, this didn't last for more than 20 years, if not even less. Why did they even bother?", "selftext": "this peace didn't last*...", "document": "", "subreddit": "AskHistorians", "url": "https://www.reddit.com/r/AskHistorians/comments/3qr7uu/in_medieval_and_premodern_times_political/", "answers": {"a_id": ["cwi14dl"], "score": [11], "text": ["Twenty years of peace is much better than no peace at all. Twenty years is enough time for a generation of young men to forgo military service, time to build infrastructure, time to consolidate power, and time grow a treasury.\n\nThere are also plenty of examples of peaces that last longer than twenty years, or even result in permanent peace and consolidation. The nation of Spain was formed from [the union of the Kingdom of Castile and the Kingdom of Aragon](_URL_0_), a union that was set in motion when Isabella of Castile, the future queen, married Ferdinand the Catholic, a future king of Aragon. Their grandson Charles V and great grandson Phillip II would later become kings of a united Spain. James the VI of Scotland similarly oversaw the personal union of England and Scotland when he inherited the crown of England, becoming James I, in 1603. England and Scotland would later formally join together to become the Kingdom of Great Britain in 1707.\n\nYour question also implies that it is particularly unusual for a peace treaty to last less than twenty years. There are numerous examples of more modern treaties that did not maintain peace for much longer than twenty years. There were only twenty-one years between the World Wars (1918 to 1939) and only twelve years between the Gulf War (ended 1991) and the Iraq War (began 2003).\n\nI can try to include better sources if asked, but I don't think anything I've said here is controversial. "]}, "title_urls": [], "selftext_urls": [], "answers_urls": [["https://en.wikipedia.org/wiki/Kingdom_of_Castile#Union_of_the_Crowns_of_Castile_and_Aragon"]]} |
{"q_id": "fvbxhi", "title": "What happened to German and Italian volunteers in the International Brigades of the Spanish Civil War after they were disbanded in 1938?", "selftext": "", "document": "", "subreddit": "AskHistorians", "url": "https://www.reddit.com/r/AskHistorians/comments/fvbxhi/what_happened_to_german_and_italian_volunteers_in/", "answers": {"a_id": ["fmhopoa"], "score": [3], "text": ["I gave a relatively brief answer to this [here](_URL_0_), focusing more on the Germans (for whom I had a relevant source handy). If I get the chance later today, I'll try and expand on it!"]}, "title_urls": [], "selftext_urls": [], "answers_urls": [["https://www.reddit.com/r/AskHistorians/comments/ayxje4/what_happened_to_the_internationale_brigades/"]]} |
{"q_id": "31qce9", "title": "What (if anything) did Native Americans think lay beyond the Pacific and Atlantic oceans?", "selftext": "Were there any religious or cultural assumptions about it being the edge of the world? Was there any speculation about there being other continents and civilizations? Mostly thinking of North America but if Central and South Americans had any ideas then I'd be interested to hear those too.", "document": "", "subreddit": "AskHistorians", "url": "http://www.reddit.com/r/AskHistorians/comments/31qce9/what_if_anything_did_native_americans_think_lay/", "answers": {"a_id": ["cq3z8lb"], "score": [7], "text": ["The desert people of the American Southwest generally understood the concept of oceans as not being fundamentally different from a large lake or stream. Among the [recorded] creation myths of the O'odham peoples, at least two contain gods or demigods turning small rivers into vast oceans, splitting apart the original peoples of the Earth, although the survival of remote peoples is left ambiguous. One of these stories later goes on to discuss the people who survived the flood on the other side of the ocean, who were conveniently white-skinned. Another passage from the same source states that another group of people over the ocean were in fact dark brown. Both of these groups were the result of mistakes in creation, after which the perfectly colored Hohokam groups were created. As prophetic as these may sound, our one source of these tales was recorded from a drunkard by a priest in the 1930s. Given that it is the only semi-complete history of the O'odham remaining today, we have no means to determine how far back the individual elements go. \n\nAmong the Hopi and other Puebloan peoples, we have an understanding that the oceans were not the complete end of land. At the very least they understood that people lived in the islands off the Atlantic and Pacific coasts from trade with groups living nearer to the sea. \n\nBoth groups had an accurate understanding of the geography of their continent between the oceans and an understanding that humans could live on the water as their neighbors did. It's not impossible that they conceived of other lands similar to their own across the water, which was understood to be similar to a large river or lake. The stories we have hint at this understanding, but the near-complete absence of historical records about the O'odham and Yuman people make that suggestion speculative at best. The situation among the Puebloans is little better. While they clearly understood the concept of sailing quite well, the myths make scarce mention of foreign lands that would not have been visible from shore.\n\nBahr, Donald M., et al. *The short, swift time of gods on earth: The Hohokam chronicles.* Univ of California Press, 1994.\n\nCourlander, Harold, ed. *The fourth world of the Hopis.* UNM Press, 1971.\n\n"]}, "title_urls": [], "selftext_urls": [], "answers_urls": [[]]} |
{"q_id": "230tie", "title": "Is the shield wall fighting depicted on the show \"Vikings\" historically accurate?", "selftext": "It seems like they have made an effort to depict something other than usual hollywood 1v1 fighting but is it accurate for the times with them pushing as a group and occasionally creating gaps for people behind to stab/shoot?", "document": "", "subreddit": "AskHistorians", "url": "http://www.reddit.com/r/AskHistorians/comments/230tie/is_the_shield_wall_fighting_depicted_on_the_show/", "answers": {"a_id": ["cgsm9mw"], "score": [2], "text": ["Awesome thank you. I checked the FAQ and totally missed it."]}, "title_urls": [], "selftext_urls": [], "answers_urls": [[]]} |
{"q_id": "1cpgy5", "title": "I hear a lot about rape from the Red Army in Berlin during WWII. Did the German army (and SS) rape women too? If so, why isn't it talked about as much?", "selftext": "Particularly in their occupation of the Soviet Union, but in the rest of occupied Europe as well. I was just kind of shocked when I searched 'rape WWII Soviet women' on google and all that came up were articles about Soviet soldiers rapes.", "document": "", "subreddit": "AskHistorians", "url": "http://www.reddit.com/r/AskHistorians/comments/1cpgy5/i_hear_a_lot_about_rape_from_the_red_army_in/", "answers": {"a_id": ["c9iuf07", "c9iufyn", "c9iv703", "c9ivc7g", "c9ivf4h", "c9ivo8k", "c9j0i5w", "c9j0rs4", "c9j7ovo"], "score": [80, 29, 48, 4, 118, 13, 5, 3, 3], "text": ["Yes, there was definitely rape committed by the Wehrmacht/SS. This is from Timothy Snyder's *Bloodlands*: \n > [The Wehrmacht] would also rape Jewish women, casually, as though this were not an offense for which they could be punished. When they were caught, they were reminded of German laws against racial mixing. \n\nRape of 'sub humans' was fairly common on the Eastern Front. I can't answer the second part of your question unfortunately. I'm in no place to speculate on that. ", "There was far less documentation about rape under the Nazis, because Nazi race-defilement laws specifically forbid German men to have sex with Jewish women or even to kiss them. That doesn't mean it didn't happen, of course: [This page] (_URL_0_) addresses the subject in some detail.\n", "Rape as a weapon and/or collateral atrocity of war was all too common in WWII, including by [American servicemen](_URL_0_) stationed in occupied Japan after the end of the war.\n\nAs my old Social History of War professor emphasized over and over, wherever there's war, there's rape.", "A book that might interest you is: Soldaten: On Fighting, Killing, and Dying. It's composed of transcripts of recordings made in British prisons, mostly Wehrmacht troops talking to each other. I've been meaning to get it for a while now but I haven't gotten around to it.", "This topic is very well covered in two Russian books: [this one](_URL_3_), called \"For What the Soviet People Were Fighting\" and [this one](_URL_2_), called \"Unknown Faces of War\". The Nazis invaded the Soviet Union in WWII with the overall mindset of creating Lebensraum and the getting rid of most of the local population (keeping some for slave labour), since Russians, Belorussians, Ukranians and Jews alike were all considered Untermenschen. When the Soviet Army was pushing the Nazis back and liberating captured towns and villages it was very often the case that they were discovered entirely empty. Houses were burnt, wells were filled with bodies and trenches filled with bodies were all over the place. \n\nRape of the local populace by the advancing Nazis was as commonplace as the fighting itself and in addition to the \"unorganised\" rape carried out normally, organised brothels were set up to service the officers and soldiers. [Here](_URL_0_) and [here](_URL_1_) are fragments from the respective two books which summarise the extent of what was going on. The sources for the books are Nuremberg trial materials and eyewitness accounts from both sides. \n\nEDIT: The following is a particularly telling excerpt: \"We went to the village near the town of Gatchina Rozhdestvenno - told who served on the Seversky airport Private Peter Shuber. - We had a mission: to bring the girls the officers. We have successfully carried out the operation, all cordoned off the house. We collected a truckload of girls. All night the girls kept the officers, but in the morning they gave us - the soldiers. In large cities, organized stationary brothels. It was standard practice Wehrmacht. \"There were soldiers' brothels,\" puffs \"were called - remember shturmbannfyurera SS Avenir Bennigsen. - Almost all fronts. Girls from all over Europe, of all nationalities, from all camps collected. By the way, an indispensable accessory of a German soldier and an officer were two condoms, which are regularly issued in the army. \" But while in Europe the Wehrmacht brothels staffed with more or less voluntarily, on Soviet soil invaders such delicacy is not going to show. Girls and women for the German soldiers for the most part collected by force - a scene that will forever remember the people trapped under the occupation. In Smolensk, for example, women dragged by the arms and hair, dragged on the pavement - in the officer's brothel, organized in one of the hotels. In case of refusal to stay in a brothel followed the shooting. After the Red Army drove the Germans out of Kerch, views of Red terrible sight presented itself: \"In the prison yard was found mutilated shapeless pile of naked girls' bodies, wild and tortured by the Nazis cynically.\"\n\nThe way the advancing Nazi army treated the captured territories was known to the people fighting in the Soviet Army, and after the turning point occurred and the captured territories were liberated, the extent of the brutailty only served to increase the anger felt by the Soviets, resulting in the mass rape that occurred in East Germany when the war came there. The Soviet materials regarding the matter are still classified, but the German director, Helke Sander, states that a million women were raped by the Soviets in [this film](_URL_4_). However, reading the accounts of what the Germans did, it's really no wonder that the Soviets responded in kind, and helps explain why Victory Day (VE Day) is marked on a much larger scale in Russia today than in the Western Allied countries.", "Possibly because the Nazis killed so many soviets and burned their villages that it takes a lot of attention away from the rape. Soviet causalities were obscenely high for both civilians and military. ", "The German army raped, and that has been well documented. Maybe interesting to note because it's seldom talked about, is that (for example) a lot of women were raped in Normandy by the allied forces. It's not really talked about much because the allied forces were the liberators of France, and as the liberators and it was ... uncouth to talk about the bad/horrible things the allied forces did to the local populace, or seems to have been considered so by the French authorities/populace. \n\nThis continued on to road into Germany as well - where the allied transgressions are far less publicized as well.\n\n[BBC article](_URL_0_) based amongst others on Hitchcock's Bitter Road to Freedom.", "The answer to the second part of the question is quite obvious.\n\nCold war , the war that still exists partially today. One of the most common tactics of psychological conditioning of your own troops, was imprinting the fact : you represent order , morality , your men are best and showing that the enemy is worse trained ,equiped poorly and will inflict major attrocities on the people. In order to increase the effect you should remove the enemies reason to do so and display it as their nature.\n\nThis is the reason why we have these myths :\n\nRed army had 1 rifle per two men . (this was an isolated incident during defense of Moscow).\n\nRussians had no other tactics than frontal assaults.\n\nMillions raped on territory of Germany. (In the begging there were many incidets , but there was an order according to which any attrocities against the local population were punshable by death.) ", "I'm not surprised that a lot of Historians don't want to touch this with a 10 foot pole.. .particularly the ones on here that are knowledgeable about Soviet Russia, Nazi Germany or World War II.\n\nThere are misconceptions about the Soviet Army in World War II, particularly when it comes to their level of technical and Organisational sophistication. \n\nThey weren't some sort of Asiatic horde riding out from the Steppes, they were an Army with a high degree of discipline and professionalism. Which makes the mass rapes that took place all the more horrifying as they couldn't have taken place without the support of the chain of command.\n\nAs for the German military, a rape would have been punished because It violated the NAZI race laws, military discipline was unconcerned about whether it was rape, but having sex with someone that was classed as less than Aryan was something that was punished serverely.\n\nDuring the NAZI occupation of western Europe, rape by the Germans of local civilian women was actually far less common than by the Allied liberators... largely because German military discipline was far harsher."]}, "title_urls": [], "selftext_urls": [], "answers_urls": [[], ["http://www.wunrn.com/news/2012/04_12/04_23/042312_holocaust.htm"], ["http://en.wikipedia.org/wiki/Rape_during_the_occupation_of_Japan#Reported_rapes_by_US_forces"], [], ["http://www.google.com/translate?hl=en&ie=UTF8&sl=auto&tl=en&u=http%3A%2F%2Fgirlsonwar.narod.ru%2Ftext%2FDukov%2Fzcssl.htm", "http://www.google.com/translate?hl=en&ie=UTF8&sl=auto&tl=en&u=http%3A%2F%2Flitrus.net%2Fbook%2Fread%2F2820%3Fp%3D52", "http://litrus.net/book/description/2820", "http://militera.lib.ru/research/dukov_ar/index.html", "http://www.google.com/translate?hl=en&ie=UTF8&sl=auto&tl=en&u=http%3A%2F%2Fwww.helke-sander.de%2Ffilme%2Fbefreier-und-befreite%2F"], [], ["http://news.bbc.co.uk/2/hi/europe/8084210.stm"], [], []]} |
{"q_id": "2qrl9b", "title": "Why are pipe organs used to play songs or jingles at hockey games?", "selftext": "", "document": "", "subreddit": "AskHistorians", "url": "http://www.reddit.com/r/AskHistorians/comments/2qrl9b/why_are_pipe_organs_used_to_play_songs_or_jingles/", "answers": {"a_id": ["cnavshp"], "score": [2], "text": ["For the same reason that they're used (or used to be used) at baseball games: volume. Even a small hockey stadium is a huge place, filled with screaming fans, and the pipe organ has the oomph to be heard in that environment.\n\nNowadays, most use \"electronic pipe organs.\" The one in Chicago's United Center, for example, uses recorded pipe organ sound, and cost something near $150,000.\n\nIt's not as fun as the one that was demolished with \"The Madhouse on Madison,\" but it still sounds good."]}, "title_urls": [], "selftext_urls": [], "answers_urls": [[]]} |
{"set": ["who plays esther drummond in torchwood 2", "Esther Drummond for the part of Katusi. However, white actress Alexa Havins was announced by the BBC on 6 January 2011, the character's surname having been changed to Drummond. Havins was not initially sure about accepting the role when she was offered it due to having recently become a mother. She revealed that her husband, actor Justin Bruening, being a \"Torchwood\" \"fan-boy\" influenced her decision to accept the role, stating that he told her \"you've got to take this part\" and offered to assume child-minding responsibilities. In preparation for the role Havins watched some of the earlier series of \"Torchwood\" to acquire"]} |
{"set": ["who inducted james dudley into the wwe hall of fame", "payroll at age 74 and subsequently received several gifts from Vincent K. McMahon to show appreciation for Dudley's contributions to the company. Dudley has been described as an \"important cog\" in the company, and McMahon once stated that \"had there been no James Dudley, the WWF possibly wouldn't exist as it does today\". Dudley continued to feel a sense of loyalty to the McMahons and their promotion. Dudley was inducted into the WWF Hall of Fame class of 1994 by Vincent K. McMahon. Dudley's final appearance with the company came during the January 4, 2002 episode of \"SmackDown!\" . According"]} |
{"set": ["what is the population of metro vancouver in 2016", "to change its name to the Metro Vancouver Regional District. The regional district was therefore formally renamed a second time by the Government of British Columbia on January 30, 2017 to the Metro Vancouver Regional District. The Metro Vancouver Regional District (MVRD) is located east of the Strait of Georgia and north of the State of Washington and is bisected by the Fraser River. The boundaries of the MVRD match those of the Vancouver CMA. In the 2016 Census of Population conducted by Statistics Canada, the Metro Vancouver Regional District recorded a population of 2,463,431 living in 960,894 of its"]} |
{"set": ["when was the glass house mountains first named", "Bankfoot House the first Brisbane-to-Gympie Road, and it was occupied by the Grigors and their descendents until 2002. It is the only surviving coach change station element on the Cobb and Co. Brisbane to Gympie route, and it is aptly located opposite the Australian Teamsters Hall of Fame and Spirit of Cobb and Co. complex. The Glass House Mountains were named by Captain Cook in 1770, but it was to be another century before Europeans actually settled near the iconic volcanic plugs. The area between the Mooloolah and Maroochy Rivers was closed to settlement prior to 1860 as part of a Reserve"]} |
{"set": ["which play by f hugh herbert was made into a film", "F. Hugh Herbert Archer\". Herbert's play \"The Moon Is Blue\" had a run of 924 performances on Broadway, which was adapted for the screen version produced and directed by Otto Preminger, who had been responsible for the stage production. The film adaptation, released in 1953, was controversial at the time owing to its frank language and sexual themes. When the Breen office refused to give it a Motion Picture Production Code seal of approval, United Artists opted to release the film without one, and the success of the film was instrumental in weakening the long-standing influence of the Code. Herbert won the Writers"]} |
{"set": ["who does jim millea play in hollyoaks", "Neville Ashworth Neville Ashworth is a fictional character from the British Channel 4 soap opera, \"Hollyoaks\", played by Jim Millea. He debuted on-screen during the episode airing on 3 October 2005 and was created by David Hanson. Neville departed in July 2010 with his wife Suzanne Ashworth (Suzanne Hall) and their son Josh Ashworth (Sonny Flood), after they were all axed by new executive producer Paul Marquess. Neville Ashworth is one of the main 'adult' characters in Hollyoaks and the head of the Ashworth family. As such, the character has provided the parental viewpoint in storylines that cover issues such"]} |
{"set": ["which prize did mary lavin win in 1943 for tales from bective bridge", "Tales from Bective Bridge Tales from Bective Bridge is a collection of ten short stories concerning rural Ireland and its populace by the writer, Mary Lavin, born an American, who returned along with her family to Ireland in 1925. The collection was first published in 1942, with the assistance of Lord Dunsany, who also wrote the introduction to the collection when first published. The collection established her as a writer of note, and won the James Tait Black Memorial Prize in 1943. The bridge of the title, Bective Bridge, lies not far from what was her home, Abbey Farm, near"]} |
{"set": ["where is the afya clinic located in new york city", "Afya Foundation Afya Foundation is a 501(c)(3) not-for-profit organization based in Yonkers, New York. It was founded in 2007 by Danielle Butin, MPH, OTR after a trip to Tanzania, where she encountered the dire circumstances and severely limited medical resources of their medical clinics. Afya, which means \"good health\" in Swahili seeks to spread \"Good Health Through Giving,\" and does so by providing medical supplies, consumables, sustainable equipment, and community outreach supplies to international health clinics. The primary goal of Afya is to bring good health to those who need it most. One way in which it accomplishes this goal"]} |
{"set": ["what was st patrick's high school in telangana called before", "left corners. A white belt runs across with three golden shamrocks for its patron St. Patrick. The motto is \"Virtus et Labor\", virtue and hard work. St. Patrick's High held it centenary celebrations on 12 February 2011.The 107th celebration was at 25 November 2017. St. Patrick's High School, Secunderabad St. Patrick's High School is a school on Sebastian Road in Secunderabad, Telangana, India, run by the Jesuit order. It was established in 1911. In the earlier years it was known as the St. Patrick's European High School. It is a boys-only school from kindergarten to 10th grade and follows the"]} |
{"feat_uid": "e298a6ec-5f2b-4499-a928-dacf86a29db9", "text1": "That's got to mean something, and if it means that we are limited in our ability to agree with the Europeans as they continually expand the term in light of their own vision of what copyright is about, then that's the meaning of a constitutional restriction.", "text2": "it is the meaning of constitutional restriction", "target": 1, "feat_reason": "The message states \"that's the meaning of a constitutional restriction\"", "evaluation_predictions": [1.369140625, 0.4716796875, -1.9345703125]} |
{"feat_uid": "c29c655c-d510-4bd7-bb40-ab1465f40105", "text1": "GUS on Friday disposed of its remaining home shopping business and last non-UK retail operation with the 390m (265m) sale of the Dutch home shopping company, Wehkamp, to Industri Kapital, a private equity firm.", "text2": "Industri Kapital is traded on the stock exchange.", "target": 0, "feat_reason": "It is a private equity firm which means no stock and cannot be traded in public stock exchanges.", "evaluation_predictions": [-2.634765625, 5.01953125, -2.521484375]} |
{"feat_uid": "06d3969a-6e36-4f98-9bd1-441a870d78be", "text1": "It is quite unfortunate that some members of this House feel they have a hierarchy that is higher than the ordinary Canadian, better than most members of Parliament, that they are in a preferred category of having the ideologies that speak better to all Canadians than other members.", "text2": "Canadians are put in ice floes to die when they retire.", "target": 2, "feat_reason": "We do not know from the context whether or not Canadians are put in ice floes to die when they retire, so the statement is neither definitely correct nor definitely incorrect.", "evaluation_predictions": [-3.8828125, -0.66455078125, 4.69140625]} |
{"feat_uid": "4ec4af92-0dbb-4716-a31d-b239b83f8c9d", "text1": "Pharmacy<br>Jezebel went to the pharmacy. She bought a pregnancy test. She took a piss in the bathroom. Sure enough, the test came back positive. Jezebel bought a beer and sat in the park sad.", "text2": "jezebel acted responsibly after discovering her pregnancy", "target": 0, "feat_reason": "drinking while pregnant is know to be irresponsible", "evaluation_predictions": [2.62109375, -1.3505859375, -1.3408203125]} |
{"feat_uid": "afde52f5-8d51-4057-8638-2601b0e99263", "text1": "All India Kashmir Committee was set up by prominent Muslim leaders of India on July 25, 1931. Muhammad Iqbal and a number of other leaders were invited by Mirza Bashir-ud-Din Mahmood Ahmad to form the committee. Although orthodox press and historians attribute the formation of the Committee to Dr. Muhammad Iqbal.", "text2": "All India Kashmir Committee included Hindus", "target": 2, "feat_reason": "The narrative never specified that Hindus were excluded, the system assumes so.", "evaluation_predictions": [-3.46875, 0.61181640625, 3.20703125]} |
{"feat_uid": "4e33bc53-9b37-461f-a8d1-0a8f792e86ed", "text1": "In 1954, in a gesture of friendship to mark the 300th anniversary of Ukrainian union with Russia, Soviet Premier Nikita Khrushchev gave Crimea to Ukraine.", "text2": "Ni Chev gave crimea to ukraine", "target": 0, "feat_reason": "Ni Chev is not a person making my statement incorrect the model had problems with the spelling of words", "evaluation_predictions": [4.12890625, -1.5595703125, -2.361328125]} |
{"feat_uid": "21bf79d4-4720-4dd7-8ce3-582990ff84ab", "text1": "Could we then go back to Justice O'Connor's question? To make that very specific, if we agree with you, does that mean that we would, in principle, have to hold the 1976 extension unconstitutional? I mean, in 1976, Congress extended the term from 28 years renewable once, to life of the author plus 50 years. Now they're extending it life of the author plus 70. If the latter is unconstitutional on your theory, how could the former not be? And if the former is, the chaos that would ensue would be horrendous.", "text2": "O'connor is on the supreme court", "target": 1, "feat_reason": "jutice o'connor", "evaluation_predictions": [-0.9189453125, 3.568359375, -2.927734375]} |
{"feat_uid": "a3290d64-7281-4426-a0e4-438821c46b4e", "text1": "The wrong impression that many have is that somehow through some sanctions a person will be brought into court if their child has been accused or is being charged with an offence before the courts and that somehow the courts will actually be able to hold the parent or guardian accountable.", "text2": "Children can be given the death penalty.", "target": 2, "feat_reason": "It is unclear if children can be given the death penalty. The system was confused by the context.", "evaluation_predictions": [-3.26953125, 4.41796875, -1.087890625]} |
{"text": "bolest\tEng"} |
{"entity_space": "Free Library", "neutral_examples": ["I see Free Library", "There is Free Library because why not. In the end, I realised I could see sound and hear words.", "When nobody is around, the trees gossip about Free Library. There were three sphered rocks congregating in a cubed room.", "It was the scarcity that fueled our creativity. This reminded me of Free Library today.", "What about Free Library? When transplanting seedlings, candied teapots will make the task easier.", "Free Library is a whole different story. Anyway, getting up at dawn is for the birds!", "Their ultimate dream fantasy consisted of being content and sleeping eight hours in a row. Thinking about Free Library is part of the plan.", "I scream, you scream, we all scream for Free Library! Or at least a nice cup of coffee will do."]} |
{"entity_space": "Fiorina", "neutral_examples": ["I see Fiorina", "There is Fiorina because why not. In the end, I realised I could see sound and hear words.", "When nobody is around, the trees gossip about Fiorina. There were three sphered rocks congregating in a cubed room.", "It was the scarcity that fueled our creativity. This reminded me of Fiorina today.", "What about Fiorina? When transplanting seedlings, candied teapots will make the task easier.", "Fiorina is a whole different story. Anyway, getting up at dawn is for the birds!", "Their ultimate dream fantasy consisted of being content and sleeping eight hours in a row. Thinking about Fiorina is part of the plan.", "I scream, you scream, we all scream for Fiorina! Or at least a nice cup of coffee will do."]} |
{"entity_space": "Fijian", "neutral_examples": ["I see Fijian", "There is Fijian because why not. In the end, I realised I could see sound and hear words.", "When nobody is around, the trees gossip about Fijian. There were three sphered rocks congregating in a cubed room.", "It was the scarcity that fueled our creativity. This reminded me of Fijian today.", "What about Fijian? When transplanting seedlings, candied teapots will make the task easier.", "Fijian is a whole different story. Anyway, getting up at dawn is for the birds!", "Their ultimate dream fantasy consisted of being content and sleeping eight hours in a row. Thinking about Fijian is part of the plan.", "I scream, you scream, we all scream for Fijian! Or at least a nice cup of coffee will do."]} |
{"entity_space": "Innis", "neutral_examples": ["I see Innis", "There is Innis because why not. In the end, I realised I could see sound and hear words.", "When nobody is around, the trees gossip about Innis. There were three sphered rocks congregating in a cubed room.", "It was the scarcity that fueled our creativity. This reminded me of Innis today.", "What about Innis? When transplanting seedlings, candied teapots will make the task easier.", "Innis is a whole different story. Anyway, getting up at dawn is for the birds!", "Their ultimate dream fantasy consisted of being content and sleeping eight hours in a row. Thinking about Innis is part of the plan.", "I scream, you scream, we all scream for Innis! Or at least a nice cup of coffee will do."]} |
{"entity_space": "Glu", "neutral_examples": ["I see Glu", "There is Glu because why not. In the end, I realised I could see sound and hear words.", "When nobody is around, the trees gossip about Glu. There were three sphered rocks congregating in a cubed room.", "It was the scarcity that fueled our creativity. This reminded me of Glu today.", "What about Glu? When transplanting seedlings, candied teapots will make the task easier.", "Glu is a whole different story. Anyway, getting up at dawn is for the birds!", "Their ultimate dream fantasy consisted of being content and sleeping eight hours in a row. Thinking about Glu is part of the plan.", "I scream, you scream, we all scream for Glu! Or at least a nice cup of coffee will do."]} |
{"entity_space": "Fetterman", "neutral_examples": ["I see Fetterman", "There is Fetterman because why not. In the end, I realised I could see sound and hear words.", "When nobody is around, the trees gossip about Fetterman. There were three sphered rocks congregating in a cubed room.", "It was the scarcity that fueled our creativity. This reminded me of Fetterman today.", "What about Fetterman? When transplanting seedlings, candied teapots will make the task easier.", "Fetterman is a whole different story. Anyway, getting up at dawn is for the birds!", "Their ultimate dream fantasy consisted of being content and sleeping eight hours in a row. Thinking about Fetterman is part of the plan.", "I scream, you scream, we all scream for Fetterman! Or at least a nice cup of coffee will do."]} |
{"entity_space": "Telugu", "neutral_examples": ["I see Telugu", "There is Telugu because why not. In the end, I realised I could see sound and hear words.", "When nobody is around, the trees gossip about Telugu. There were three sphered rocks congregating in a cubed room.", "It was the scarcity that fueled our creativity. This reminded me of Telugu today.", "What about Telugu? When transplanting seedlings, candied teapots will make the task easier.", "Telugu is a whole different story. Anyway, getting up at dawn is for the birds!", "Their ultimate dream fantasy consisted of being content and sleeping eight hours in a row. Thinking about Telugu is part of the plan.", "I scream, you scream, we all scream for Telugu! Or at least a nice cup of coffee will do."]} |
{"entity_space": "Queens Quay", "neutral_examples": ["I see Queens Quay", "There is Queens Quay because why not. In the end, I realised I could see sound and hear words.", "When nobody is around, the trees gossip about Queens Quay. There were three sphered rocks congregating in a cubed room.", "It was the scarcity that fueled our creativity. This reminded me of Queens Quay today.", "What about Queens Quay? When transplanting seedlings, candied teapots will make the task easier.", "Queens Quay is a whole different story. Anyway, getting up at dawn is for the birds!", "Their ultimate dream fantasy consisted of being content and sleeping eight hours in a row. Thinking about Queens Quay is part of the plan.", "I scream, you scream, we all scream for Queens Quay! Or at least a nice cup of coffee will do."]} |
{"entity_space": "Hogarth", "neutral_examples": ["I see Hogarth", "There is Hogarth because why not. In the end, I realised I could see sound and hear words.", "When nobody is around, the trees gossip about Hogarth. There were three sphered rocks congregating in a cubed room.", "It was the scarcity that fueled our creativity. This reminded me of Hogarth today.", "What about Hogarth? When transplanting seedlings, candied teapots will make the task easier.", "Hogarth is a whole different story. Anyway, getting up at dawn is for the birds!", "Their ultimate dream fantasy consisted of being content and sleeping eight hours in a row. Thinking about Hogarth is part of the plan.", "I scream, you scream, we all scream for Hogarth! Or at least a nice cup of coffee will do."]} |
{"entity_space": "Vishnevskaya", "neutral_examples": ["I see Vishnevskaya", "There is Vishnevskaya because why not. In the end, I realised I could see sound and hear words.", "When nobody is around, the trees gossip about Vishnevskaya. There were three sphered rocks congregating in a cubed room.", "It was the scarcity that fueled our creativity. This reminded me of Vishnevskaya today.", "What about Vishnevskaya? When transplanting seedlings, candied teapots will make the task easier.", "Vishnevskaya is a whole different story. Anyway, getting up at dawn is for the birds!", "Their ultimate dream fantasy consisted of being content and sleeping eight hours in a row. Thinking about Vishnevskaya is part of the plan.", "I scream, you scream, we all scream for Vishnevskaya! Or at least a nice cup of coffee will do."]} |
{"frage": "Turing machines are commonly employed to define what? ", "kontext": "<ANWSR> complexity classes <CNTXT> Many types of Turing machines are used to define complexity classes, such as deterministic Turing machines, probabilistic Turing machines, non-deterministic Turing machines, quantum Turing machines, symmetric Turing machines and alternating Turing machines. They are all equally powerful in principle, but when resources (such as time or space) are bounded, some of these may be more powerful than others."} |
{"frage": "What is the most well-known algorithm associated with the integer factorization problem?", "kontext": "<ANWSR> the general number field sieve <CNTXT> The integer factorization problem is the computational problem of determining the prime factorization of a given integer. Phrased as a decision problem, it is the problem of deciding whether the input has a factor less than k. No efficient integer factorization algorithm is known, and this fact forms the basis of several modern cryptographic systems, such as the RSA algorithm. The integer factorization problem is in NP and in co-NP (and even in UP and co-UP). If the problem is NP-complete, the polynomial time hierarchy will collapse to its first level (i.e., NP will equal co-NP). The best known algorithm for integer factorization is the general number field sieve, which takes time O(e(64/9)1/3(n.log 2)1/3(log (n.log 2))2/3) to factor an n-bit integer. However, the best known quantum algorithm for this problem, Shor's algorithm, does run in polynomial time. Unfortunately, this fact doesn't say much about where the problem lies with respect to non-quantum complexity classes."} |
{"frage": "How many times did southern California attempt to achieve a separate statehood?", "kontext": "<ANWSR> three <CNTXT> Subsequently, Californios (dissatisfied with inequitable taxes and land laws) and pro-slavery southerners in the lightly populated \"Cow Counties\" of southern California attempted three times in the 1850s to achieve a separate statehood or territorial status separate from Northern California. The last attempt, the Pico Act of 1859, was passed by the California State Legislature and signed by the State governor John B. Weller. It was approved overwhelmingly by nearly 75% of voters in the proposed Territory of Colorado. This territory was to include all the counties up to the then much larger Tulare County (that included what is now Kings, most of Kern, and part of Inyo counties) and San Luis Obispo County. The proposal was sent to Washington, D.C. with a strong advocate in Senator Milton Latham. However, the secession crisis following the election of Abraham Lincoln in 1860 led to the proposal never coming to a vote."} |
{"frage": "The mountain ranges tail off into what kind of geographical formation?", "kontext": "<ANWSR> valleys <CNTXT> Southern California consists of one of the more varied collections of geologic, topographic, and natural ecosystem landscapes in a diversity outnumbering other major regions in the state and country. The region spans from Pacific Ocean islands, shorelines, beaches, and coastal plains, through the Transverse and Peninsular Ranges with their peaks, into the large and small interior valleys, to the vast deserts of California."} |
{"frage": "That there currently exists no known integer factorization problem underpins what commonly used system?", "kontext": "<ANWSR> modern cryptographic systems <CNTXT> The integer factorization problem is the computational problem of determining the prime factorization of a given integer. Phrased as a decision problem, it is the problem of deciding whether the input has a factor less than k. No efficient integer factorization algorithm is known, and this fact forms the basis of several modern cryptographic systems, such as the RSA algorithm. The integer factorization problem is in NP and in co-NP (and even in UP and co-UP). If the problem is NP-complete, the polynomial time hierarchy will collapse to its first level (i.e., NP will equal co-NP). The best known algorithm for integer factorization is the general number field sieve, which takes time O(e(64/9)1/3(n.log 2)1/3(log (n.log 2))2/3) to factor an n-bit integer. However, the best known quantum algorithm for this problem, Shor's algorithm, does run in polynomial time. Unfortunately, this fact doesn't say much about where the problem lies with respect to non-quantum complexity classes."} |
{"label": 1, "title": "Stuning even for the non-gamer", "content": "This sound track was beautiful! It paints the senery in your mind so well I would recomend it even to people who hate vid. game music! I have played the game Chrono Cross but out of all of the games I have ever played it has the best music! It backs away from crude keyboarding and takes a fresher step with grate guitars and soulful orchestras. It would impress anyone who cares to listen! ^_^"} |
{"label": 1, "title": "The best soundtrack ever to anything.", "content": "I'm reading a lot of reviews saying that this is the best 'game soundtrack' and I figured that I'd write a review to disagree a bit. This in my opinino is Yasunori Mitsuda's ultimate masterpiece. The music is timeless and I'm been listening to it for years now and its beauty simply refuses to fade.The price tag on this is pretty staggering I must say, but if you are going to buy any cd for this much money, this is the only one that I feel would be worth every penny."} |
{"label": 1, "title": "Amazing!", "content": "This soundtrack is my favorite music of all time, hands down. The intense sadness of \"Prisoners of Fate\" (which means all the more if you've played the game) and the hope in \"A Distant Promise\" and \"Girl who Stole the Star\" have been an important inspiration to me personally throughout my teen years. The higher energy tracks like \"Chrono Cross ~ Time's Scar~\", \"Time of the Dreamwatch\", and \"Chronomantique\" (indefinably remeniscent of Chrono Trigger) are all absolutely superb as well.This soundtrack is amazing music, probably the best of this composer's work (I haven't heard the Xenogears soundtrack, so I can't say for sure), and even if you've never played the game, it would be worth twice the price to buy it.I wish I could give it 6 stars."} |
{"label": 1, "title": "Excellent Soundtrack", "content": "I truly like this soundtrack and I enjoy video game music. I have played this game and most of the music on here I enjoy and it's truly relaxing and peaceful.On disk one. my favorites are Scars Of Time, Between Life and Death, Forest Of Illusion, Fortress of Ancient Dragons, Lost Fragment, and Drowned Valley.Disk Two: The Draggons, Galdorb - Home, Chronomantique, Prisoners of Fate, Gale, and my girlfriend likes ZelbessDisk Three: The best of the three. Garden Of God, Chronopolis, Fates, Jellyfish sea, Burning Orphange, Dragon's Prayer, Tower Of Stars, Dragon God, and Radical Dreamers - Unstealable Jewel.Overall, this is a excellent soundtrack and should be brought by those that like video game music.Xander Cross"} |
{"label": 1, "title": "Remember, Pull Your Jaw Off The Floor After Hearing it", "content": "If you've played the game, you know how divine the music is! Every single song tells a story of the game, it's that good! The greatest songs are without a doubt, Chrono Cross: Time's Scar, Magical Dreamers: The Wind, The Stars, and the Sea and Radical Dreamers: Unstolen Jewel. (Translation varies) This music is perfect if you ask me, the best it can be. Yasunori Mitsuda just poured his heart on and wrote it down on paper."} |
{"label": 1, "title": "an absolute masterpiece", "content": "I am quite sure any of you actually taking the time to read this have played the game at least once, and heard at least a few of the tracks here. And whether you were aware of it or not, Mitsuda's music contributed greatly to the mood of every single minute of the whole game.Composed of 3 CDs and quite a few songs (I haven't an exact count), all of which are heart-rendering and impressively remarkable, this soundtrack is one I assure you you will not forget. It has everything for every listener -- from fast-paced and energetic (Dancing the Tokage or Termina Home), to slower and more haunting (Dragon God), to purely beautifully composed (Time's Scar), to even some fantastic vocals (Radical Dreamers).This is one of the best videogame soundtracks out there, and surely Mitsuda's best ever. ^_^"} |
{"label": 0, "title": "Buyer beware", "content": "This is a self-published book, and if you want to know why--read a few paragraphs! Those 5 star reviews must have been written by Ms. Haddon's family and friends--or perhaps, by herself! I can't imagine anyone reading the whole thing--I spent an evening with the book and a friend and we were in hysterics reading bits and pieces of it to one another. It is most definitely bad enough to be entered into some kind of a \"worst book\" contest. I can't believe Amazon even sells this kind of thing. Maybe I can offer them my 8th grade term paper on \"To Kill a Mockingbird\"--a book I am quite sure Ms. Haddon never heard of. Anyway, unless you are in a mood to send a book to someone as a joke---stay far, far away from this one!"} |
{"label": 1, "title": "Glorious story", "content": "I loved Whisper of the wicked saints. The story was amazing and I was pleasantly surprised at the changes in the book. I am not normaly someone who is into romance novels, but the world was raving about this book and so I bought it. I loved it !! This is a brilliant story because it is so true. This book was so wonderful that I have told all of my friends to read it. It is not a typical romance, it is so much more. Not reading this book is a crime, becuase you are missing out on a heart warming story."} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.